Try changing your user agent to a Chrome one (e.g. Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/120.0.0.0 Safari/537.36). Works a treat!
Feels like we’re back to 2007 again when spoofing firefox user agent to IE would fix websites not working properly, only now we spoof it to chrome instead.
Not really. The example listed above is perfectly readable.
Knowing the versions of webkit, browser version, etc. is important due to inconsistencies, new features, mossing features, and deprecated features. Sure it can be faked, but that is on the end user.
There is more information in there that isn’t actually true and only supposed to trick some old web servers into treating it a certain way than there is actually correct information,
It mentions three different browsers, only one of which is actually true, and three different rendering engines, none of which is actually what’s used.
Chrome doesn’t use Webkit, and the referenced Webkit version is probably 10 years old at this point. The user agent is full of stuff for backwards compatibility. That’s why it’s being deprecated in favour of a better API (client hints)
There’s an API called “client hits” that’s replacing user-agent. Some of the hints will require the user to provide permission for the site to use them, since they could be used for fingerprinting.
Major browsers (Chrome and I thibk Firefox) are freezing the user-agent. The only thing that’ll be changing in user agents is the major browser version. Other parts including platform will be static. Chrome on Windows will always report itself as Windows 10 for example. https://www.chromium.org/updates/ua-reduction/
Try changing your user agent to a Chrome one (e.g.
Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/120.0.0.0 Safari/537.36
). Works a treat!Feels like we’re back to 2007 again when spoofing firefox user agent to IE would fix websites not working properly, only now we spoof it to chrome instead.
Sidenote:
HTTP user agents have become absolutely bonkers over the years.
deleted by creator
Not really. The example listed above is perfectly readable.
Knowing the versions of webkit, browser version, etc. is important due to inconsistencies, new features, mossing features, and deprecated features. Sure it can be faked, but that is on the end user.
There is more information in there that isn’t actually true and only supposed to trick some old web servers into treating it a certain way than there is actually correct information,
It mentions three different browsers, only one of which is actually true, and three different rendering engines, none of which is actually what’s used.
Chrome doesn’t use Webkit, and the referenced Webkit version is probably 10 years old at this point. The user agent is full of stuff for backwards compatibility. That’s why it’s being deprecated in favour of a better API (client hints)
I like how this guy explains the history of browser user agents and why they have this strange configuration today:
https://webaim.org/blog/user-agent-string-history/
And that’s why you shouldn’t parse them and use feature detection instead.
There’s an API called “client hits” that’s replacing user-agent. Some of the hints will require the user to provide permission for the site to use them, since they could be used for fingerprinting.
Major browsers (Chrome and I thibk Firefox) are freezing the user-agent. The only thing that’ll be changing in user agents is the major browser version. Other parts including platform will be static. Chrome on Windows will always report itself as Windows 10 for example. https://www.chromium.org/updates/ua-reduction/
Oh so like how other browsers reported Windows NT for decades… cool.
Useragent parsing is still a thing?
It is similiar in nature to greping html.