I'm trying to parse one site. I'm sending a naked http request, just by setting a User-Agent.
It's surprising how a request using "requests" returns a 200 response. But the same query using httpx returns 403. I have tried making the request with both the 1.1 version of the http protocol and the 2.0 version. It didn't give any result.
I know that this site uses some protection against bots.
But why does the primitive "requests" do the job, as opposed to httpx?
I noticed that httpx adds a 'Host' header in its default request, which 'requests' does not. But I don't know how to get rid of it.
30 sec video: https://i.imgur.com/tWOe0sZ.mp4