How to request Google to re-crawl my website? [closed]

丶灬走出姿态 提交于 2019-11-28 14:56:39
kevinmicke

There are two options. The first (and better) one is using the Fetch as Google option in Webmaster Tools that Mike Flynn commented about. Here are detailed instructions:

  1. Go to: https://www.google.com/webmasters/tools/ and log in
  2. If you haven't already, add and verify the site with the "Add a Site" button
  3. Click on the site name for the one you want to manage
  4. Click Crawl -> Fetch as Google
  5. Optional: if you want to do a specific page only, type in the URL
  6. Click Fetch
  7. Click Submit to Index
  8. Select either "URL" or "URL and its direct links"
  9. Click OK and you're done.

With the option above, as long as every page can be reached from some link on the initial page or a page that it links to, Google should recrawl the whole thing. If you want to explicitly tell it a list of pages to crawl on the domain, you can follow the directions to submit a sitemap.

Your second (and generally slower) option is, as seanbreeden pointed out, submitting here: http://www.google.com/addurl/

The usual way is to either resubmit your site in your Google Webmaster Tools or submit it here: http://www.google.com/addurl/

Google says that it is unable to control when your site is re-crawled. Regardless, you could also check this post on "forcing rewcrawls", I haven't tried it myself but it's worth a shot if you're desperate.

On another note, I might add that you make sure you have a sitemap.xml up as this will also help with SEO.

As far I know, if you resubmit a sitemap it will trigger and crawler of your site.

Jérôme Verstrynge

Nowadays, the revisiting of a website pretty much depends on its popularity, authority and how often its content changes. Having a sitemap.xml containing all URLs is always better. You can also set the lastmod tag of each URL entries. If you don't abuse it, crawlers will take it into account.

标签
易学教程内所有资源均来自网络或用户发布的内容,如有违反法律规定的内容欢迎反馈
该文章没有解决你所遇到的问题?点击提问,说说你的问题,让更多的人一起探讨吧!