Is there a google API to read cached content? [closed]

心已入冬 提交于 2020-01-13 08:36:28

问题


I know you can go to http://webcache.googleusercontent.com/search?q=cache:http://example.com/ to view Google's cache of any URL, but do they provide an API to hit thousands of these and pay for access?

I don't want to just make HTTP GETs to these URLs too fast and get my IP addresses banned or upset Google.

Just wondering if they offer a way to pay and do this through official channels like they do with their search API.


回答1:


Google doesn't seem to have an API to access the cached results:

There are some attempts to scrape it and wrap it in APIs, such as this perl module

Other than that the Wayback Machine has an API, of cached versions of sites. Perhaps that will do?




回答2:


Currently there's no tool that I've found that does it. You'd have to create your own script to individually cache a certain number of pages. To avoid Google blocking you, I suggest capping the number of urls scraped. Not ideal, but running a script 10 times is better than looking at 1000 cached urls individually. :/

If you want to see if anything you edit on your site would effect your potential rankings in Google, I'd check out SEORadar.com, they'll do that for you.



来源:https://stackoverflow.com/questions/19010131/is-there-a-google-api-to-read-cached-content

易学教程内所有资源均来自网络或用户发布的内容,如有违反法律规定的内容欢迎反馈
该文章没有解决你所遇到的问题?点击提问,说说你的问题,让更多的人一起探讨吧!