python中使用不同代理的url请求
我正在尝试检索一些谷歌搜索结果并缓存的页面。实际上我现在有两个问题。我通常可以下载前十个结果,但无法让它转到“下一个”10 页。我尝试使用正则表达式来查找页面中的下一个控件并从那里开始,但我无法让它工作。 但我的另一个问题,更重要的是,我的 IP 地址被 google 屏蔽了,我无法再查看缓存的结果。因此,我想到的一个解决方案是使用代理来发出 url 请求,并在某些时间间隔更新此代理。但我不知道如何完成这项工作。我尝试在网上搜索解决方案,但找不到好的解决方案。 如果您能给我一些建议,我将非常高兴。
I am trying to retrive some pages which are google search results and cached. Actually i have two problems for now. I can normally download the first ten results, but cant get it to work to go to "next" 10 pages. I have tried to use regular expressions to find the next control in the page and go from there, i couldn't get it to work.
But my other problem, which is much more important, is that my ip address is blocked by google that i can no longer view the cached results. So one solution coming to my mind is to use a proxy for making the url requests, and renewing this proxy in some intervals. But i do not have an idea of how to make this job. I tried to search the web for solutions, but i could not come up with a good one.
I would be very happy if you can give me some advice.
如果你对这篇内容有疑问,欢迎到本站社区发帖提问 参与讨论,获取更多帮助,或者扫码二维码加入 Web 技术交流群。
绑定邮箱获取回复消息
由于您还没有绑定你的真实邮箱,如果其他用户或者作者回复了您的评论,将不能在第一时间通知您!
发布评论
评论(1)
第五个示例?或者也许是第三个示例?
5th example? Or maybe 3rd example?