diTii.com Digital News Hub

Sign up with your email address to be the first to know about latest news and more.

I agree to have my personal information transfered to MailChimp (more information)


Guidelines to Move Website to Help Search Engines; Google Cleans Webmaster Tools’ ‘Subscriber stats, Robots.txt tool, and Site performance’ Features

Planning to move your contents, read this Google Webmaster guide to help search engines understand your new site structure better and make your site more user-friendly.

“While maintaining a website, webmasters may decide to move the whole website or parts of it to a new location. For example, you might move content from a subdirectory to a subdomain, or to a completely new domain. Changing the location of your content can involve a bit of effort, but it’s worth doing it properly,” writes Fili Wiese (Ad Traffic Quality) & Kaspar Szymanski (Search Quality) in a “how to move you content to a new location” blog post.

Wiese and Szymanski suggests in the case of moving contents – make sure to follow these guidelines — per Google post:

  • It’s important to redirect all users and bots that visit your old content location to the new content location using 301 redirects. To highlight the relationship between the two locations, make sure that each old URL points to the new URL that hosts similar content. If you’re unable to use 301 redirects, you may want to consider using cross domain canonicals for search engines instead.
  • Check that you have both the new and the old location verified in the same Google Webmaster Tools account.
  • Make sure to check if the new location is crawlable by Googlebot using the Fetch as Googlebot feature. It’s important to make sure Google can actually access your content in the new location. Also make sure that the old URLs are not blocked by a robots.txt disallow directive, so that the redirect or rel=canonical can be found.
    If you’re moving your content to an entirely new domain, use the Change of address option under Site configuration in Google Webmaster Tools to let us know about the change.
  • If you’ve also changed your site’s URL structure, make sure that it’s possible to navigate it without running into 404 error pages. Google Webmaster Tools may prove useful in investigating potentially broken links. Just look for Diagnostics > Crawl errors for your new site.
  • Check your Sitemap and verify that it’s up to date.
  • Once you’ve set up your 301 redirects, you can keep an eye on users to your 404 error pages to check that users are being redirected to new pages, and not accidentally ending up on broken URLs. When a user comes to a 404 error page on your site, try to identify which URL they were trying to access, why this user was not redirected to the new location of your content, and then make changes to your 301 redirect rules as appropriate.
  • Have a look at the Links to your site in Google Webmaster Tools and inform the important sites that link to your content about your new location.
  • If your site’s content is specific to a particular region you may want to double check the geotargeting preferences for your new site structure in Google Webmaster Tools.
  • As a general rule of thumb, try to avoid running two crawlable sites with completely or largely identical content without a 301 redirection or specifying a rel=”canonical”.
  • Lastly, we recommend not implementing other major changes when you’re moving your content to a new location, like large-scale content, URL structure, or navigational updates. Changing too much at once may confuse users and search engines.

Google Webmaster Tools: Spring Cleaning Subscriber stats, Create robots.txt tool, Site performance feature

Also, in another blog post, Google announced that they’ll be removing the “Subscriber stats, the Create robots.txt tool, and the Site performance” features in the next two weeks in the spring cleaning of Webmaster Tools.

“Subscriber stats reports the number of subscribers to a site’s RSS or Atom feeds. This functionality is currently provided in Google owned Feedburner.

Due to low usage, the Create robots.txt tool, which provides a way to generate robots.txt files for the purpose of blocking specific parts of a site from being crawled by Googlebot Will be removed from Webmaster Tools.

Site performance, that provides information about the average load time of your site’s pages, is also shutting down due to low usage,” posted Jonathan Simon, Webmaster Trends Analyst.

Share This Story, Choose Your Platform!

Do NOT follow this link or you will be banned from the site!