Block Segmentation, Duplicate Content, SEO and IR

Often you here about the need to prevent duplicate content within your own website as it applies to SEO, but why? This is our take on why creating uniquely distinct nodes on each page is imperative for your content to produce not only context, but rankings as opposed to tripping search engine filters.

Block segment analysis dictates the value of weight / ranking  factors a web page is given for a set portion of the page. For example, text in the header and footer of a document are treated differently than say text in the body section of a page.

The header is given more weight than the footer and typically the body text is given more value than both (since that is where the page has the ability to distinguish itself from borrowed elements, i.e. navigation, etc.).

The reason for this is simple, page composition dictates that proximity and distinction (the more unique one document is from another) by distinguishing your content, the better. Search engines can easily discern duplication (through shingle analysis, singular value decomposition, etc.) in areas such as navigation, sidebars/blogrolls with the same links and footers all using the same text. This has the tendency of diffuse a pages ranking potential.

In order to leverage a page for SEO, it must (1) get indexed and (2) stay indexed in order to pass its ideal ranking factor, be used to bridge ranking factor as a hub by building internal links to and from that page. If search engines determine that your pages are redundant or ubiquitous the dampener kicks in algorithmically and the ranking factor can be suppressed via de-indexing or filters (like the supplemental index) as a basis of sequencing leaving the same repetitive hash mark (from document to document).

This aspect of the ranking algorithm is based on the c-index whereby term weight and similarity are used to assign relevance through singular value decomposition as a base vector across web pages. If enough of the singular values are identified across the global collection of documents then you have co-occurrence (which may in this instance work against your ranking objectives).

Duplication via a template or though tag pages that lack unique content can go supplemental in a secondary search index, if they are not nurtured with enough internal link flow. The gist here is, the larger your website grows, the higher the probability of diffusing or diluting your global nodes / top ranking keywords (a.k.a keyword cannibalization) like an over-optimization penalty.

Just like a base that shifts to accommodate scalability, the foundation (like a triangle) must continue to broaden its base if the vertical threshold continues to rise. Similarly, the more pages you add about (1) topic A shift the focus from topic B. The common thread in most websites is the navigation (which if left to a simplistic ontology is not sufficient to feed an entire website).

You can stem a websites topic into multiple topics, however each has a threshold to cross and moving on many fronts requires more inbound links and internal stability though secondary navigation to sculpt the on page factors that identify each segments (to pass value).

This is why virtual theming (linking from relevant keywords in one document to another document) is so important. It allows the ranking factor to transfer link weight to the champion page through term frequency which essentially raises the bar in how search engines will interpret each respective keyword when considering it for retrieval.

Deep linking (acquiring inbound links to a specific page vs. just the homepage) is equally as beneficial for eliminating the tendency for pages to go supplemental (or lose ranking factor as a unique asset). When you spread a site too thin (expand the content through automation) such as using a shopping cart that pulls similar values from a database. You must ensure that you have the ability to customize data (titles, h1, and content) to make each page distinct enough to add additional leverage to your primary, secondary and tertiary keywords that define your website.

Or Stick Around and Read More Posts

13 Trackbacks

  1. By Variables that Affect SEO | SEO Design Solutions on April 25, 2009 at 10:54 am

    [...] of content or context to reach or exceed the needed “term frequency” to impact [...]

  2. [...] can find term weights or relevance thresholds by currently examining a website or web page specifically and then seeing [...]

  3. [...] Aside from the topical ranking a website may exhibit, search engines are known for parsing entire sites to determine the amount of relevance a given website or web page has for a specific keyword or term. [...]

  4. [...] their collective ranking power. Search engines already do this when they parse the index using singular value decompression(SVD) like the card game concentration “and match queries and content” with the [...]

  5. [...] that have a tendency for duplicity or duplicate content (like a re-occurring sidebar, main navigation, etc.) do add significant ranking factor, but they [...]

  6. [...] is an aspect of global term weights, the relevance model /weighting scheme and page level thresholds, but inverse document frequency [...]

  7. [...] sources from outside a website to override the search engines natural tendency to extract term frequency from a website to deem a page [...]

  8. [...] approach, you may have excessive or redundant elements that require more distinction to increase term weight or relevance score. Pages sparse in content are often overlooked and if they go supplemental or are [...]

  9. By How to Avoid SEO Over Optimization on February 24, 2010 at 11:26 pm

    [...] engines are smart enough to distinguish navigation blocks within your website template from the body area or contextual links vs. links in the sidebar or [...]

  10. By Each Page Should Pack an SEO Punch on March 29, 2010 at 8:53 pm

    [...] engines parse pages as a vector, meaning that the entire body of documents is seen as one master document acting as one unified [...]

  11. [...] context of overlapping shingles or keywords which have closely knit semantic relevance “as vectors” based on a node of context/relevance all increase the depth of word relatedness and increase [...]

  12. [...] last thing you would want is for every instance of that keyword (shingle) to link to the target URL and trip a filter for over optimization. Hence, we created two methods [...]

  13. By WordPress SEO Tips to Avoid Search Engine Penalties on February 20, 2013 at 12:46 am

    [...] By eliminating digital stragglers (thin pages), you’ll bolster your websites relevancy by having a higher percentage of unique indexed content. [...]

Post a Comment

Your email is never shared. Required fields are marked *

*
*

Web Analytics