Around the Net

Patent Detects Duplicate Content

Bill Slawski tells us about a patent granted this week to Google that describes why Google might look for duplicate content during the crawling of Web pages, how the engine might handle crawling and anchor text, and describes how some URLs for Web pages are crawled in a round-robin format over days, weeks or longer.

Slawski writes that the patent also describes different methods for identifying duplicate content, such as taking fingerprints of content found on pages to match content from another page.

Read the whole story at SEO by the Sea »

Next story loading loading..