![]() |
Google's similarity algorithm
Does anyone have any information on this.
I have multiple category pages under one domain. What does Google look for in similar attributes on your pages? Similar to the main domain page or all the sub pages together? If it just body text, then it is easy to keep from getting a lowered PR. I think it is much more than that. Page size (71K for all lower pages), scripts, meta language, word patterns in text, who knows this. GoogleGuy - is she/he for real? :helpme |
do you change titles, meta tags, text on the page? Or is all the content the same except for the keywords?
|
not quite sure what you mean with the "similar" thing but..
If you mean when it finds 86 of 300 and the rest are the same it just means it already has a page in its cache that is quite similar from your domain. if you want to keep them seperate make them subdomains |
Sosa,
the content on each page is totally different. |
Smokeythebear
thanks, you just told me what I need to know. :thumbsup :pimp :winkwink: |
It has to do with content. Otherwise dynamic pages would be considered too similiar
|
I've been observing the similarity filters and dup content thresholds on G and other SE's over several generations of algos.
G's algos for flagging pages as being similar are quite sophisticated. For example, pages made up of small snippets from other pages, however random, at some point start being seen as similar. This applies to links pages which use the same titles and descriptions (as everyone else). Those titles and descriptions get 'burned' and they are not considered unique any more (when they are all over the web). The most assured way to minimize tripping dup/similar filters is to write each page's text as a new piece of work. If you have multiple pages which link to a common page and your links include descriptions, use original, new descriptions for each link on every page. Even so, when you start cranking out hundreds and thousands of pages, statistically, you will start repeating and while you may not see it from within your site, an algo can trip a filter across large numbers of pages. In summary, a web page is considered (by SEs) as: a) content page - when most of the text content on the page is original (not repeated anywhere else on your site or others) and with minimal off-site links. b) hub page - when there is a high ratio of off-site links to actual original text content. Link lists are examples of hub pages. c) both - when there is a mix of off-site links and original content. From my experience, it seems to be best to create either content or hub pages. They are more effective for what they are. In any case, when linking to internal and offsite pages, it is important that you link to pages which relate to the theme of the page you are linking from. -Dino |
Debbie - in my experiences - and I have many "category" pages that rank #1 for their niche or at least in the top 10 for highly competitive phrases - the difference should be around 15-25%, not counting the outgoing links - if you have around 200 or so links out going and a good amount coming in the percentage can be a little lower.
As far as Googleguy - yes he is real - he is a self described engineer that works for Google and most think he is Matt Cutts - he has a blog here: http://www.mattcutts.com/blog/ where he describes his responses to different types of schemes and SEO tricks |
| All times are GMT -7. The time now is 09:05 PM. |
Powered by vBulletin® Version 3.8.8
Copyright ©2000 - 2026, vBulletin Solutions, Inc.
©2000-, AI Media Network Inc123