Wednesday, July 2, 2014

Difference between Crawling and Indexing?

Difference between Crawling and Indexing?

Crawling means if Google bot (softwares) visits a page in your site and reads the content inside it is called crawling.
Indexing refers to when Google saves your site address in its index (collection of webpages, content, anchor text).

What is NoIndex Meta-Tags:

The ‘Noindex’ allows search engines to read the pages but instructs them to remove it from memory that it was ever indexed. That means when the search engine comes to a page with noindex meta tag, it will continue to read the content inside it including the links (so link juice is passed) but will forget it after reading and will not index it.

What is NoFollow?

If a search engine finds a “no follow” backlink it will just ignore the link and will not get deeper into crawling and no link juice is passed.

What is DoFollow?
When Google bots are crawling a website they look for external links point to other websites. If this links is “do follow” then search engines will follows the link and so link juice gets passed.


How different search engines treat Nofollow?
Google: Google does not follow the nofollow link.
Yahoo: Yahoo follows that link for finding new content but link juice is not passed.
Ask: It does not follow nofollow tag and treats it as normal dofollow link.


Keep reading for more informative posts...




Interesting facts regarding Robots.txt file

Interesting facts regarding Robots.txt file

June 30, 2014 is celebrated as the 20th Anniversay for Robots.txt. Robots.txt was created by Martijn Koster in 1994. He created this file when he was working at Nexor to overcome the issues with crawlers hitting his sites too hard. All major Search Engines including WebCrawler, Lycos and AltaVista, quickly adopted it, even 20 years later continue to support it and obey it.

Most common robots.txt mistakes ever noticed:

To disallowing URLs it is always not important to implement Robots.txt file. When you are thinking to implement Robots.txt file to disallow URL, take it as a last option. You first think about the 410 HTTP response, noindex meta tags and rel=canonical.


Keep Reading...

Thursday, June 5, 2014

What is the meaning of Webspam and how Google’s Webspam team work?



What is the meaning of Webspam and how Google’s Webspam team work? – Chapter 5th

Webspam or Search Spam is a phrase used to describe webpages that are designed to "spam Google search results" using SEO tactics that are against Google publishers guidelines.  Google’s webspam team keeps organic search results free from non-quality sites which use Black-hat techniques to get high ranking and penalize these sites to work against the Google’s guidelines.



Matt Cutts is the head of Google’s Webspam team. All the updates are announced by Matt Cutts. In year 2000 Matt Cutts joined Google as a Software Engineer. From the University of North Carolina at Chapel Hill Matt Cutts done his Ph.D.  and started his career in Search.  

Before working at the Search Quality group at Google, Cutts worked at the ads engineering group and SafeSearch, Google's family filter. There he earned the nickname "porn cookie guy" by giving his wife's homemade cookies to any Googler who provided an example of unwanted pornography in the search results.




Keep reading this blog for such interesting posts regarding Google and SEO…