Skip to main content

replicate, replicate, replicate

Aside from the usual reasons why it's silly to duplicate static content from Web page to Web page, here is yet another:

If your Web site has content copied and pasted from one page to another, it is very likely that Google is filtering out some or all of the involved Web pages from search results. The reason: The Google search engine does its best to optimize user experience by returning unique content. Because no one wants search results listing page after page of the same stuff.


always crashing in the same car: recurring mistakes and misuses of the web


Higher Education and other organizational Web sites tend to needlessly replicate content. This happens perhaps most frequently when information is repurposed as marketing material. The same content ends up appearing at its original source as well as at one or two marketing pages. Unfortunately, if this marketing material resides above the core content in the site hierarchy - and it usually does - it can end up replacing the core content in Google rankings.

The result is that your users, who once were able to sail directly to the information they need in its entirety and in the best context, now have to wade through marketing fodder to find it.

This can be made worse, believe it or not, if the marketing page helpfully adds a "for more information" link back to the core content. The reason: this is exactly what content spammers do. Their raison d'etre is to dump content in multiple locations with links from one content dump to the next. Fortunately Google has gotten quite good at recognizing this pattern.

What can you do to prevent replicated information from killing search placement of your critical material? There's always the option of adding a robots.txt file to the directory containing the page you do not wish to have indexed. But to my mind, this is a skewed solution. More than 50 percent of users arrive at their destination via Web searches. So why go to the trouble to develop a Web page and then purposely block it from searches?

Better, if at all possible, is to keep redundant content to a minimum on your site. And this means redundant page titles, meta-information, and the like too - this repetition also can cause filtering. If you are still running a "text only" version of your site, LOSE IT. Switch to using Web Standards instead. Write fresh content when developing marketing pieces and similar information, keep it brief, and link to core content rather than redistributing it.

The added benefit is that you are no longer confusing your users with seemingly duplicate pages. Or, for that matter, thudding them over the heads with repetitive verbiage.

More Information:

Popular posts from this blog

facebook, time to grow up

Originally published on August 28, 2006 I appreciate how Facebook has enabled me to connect with colleagues, and (younger) family members in a manner that is both informative and expressly cordial. It attracts students like Nutella attracts chocolate lovers, and because of that, I see interesting potential here. In fact, one of our faculty members at Penn State plans to try running his human-computer interaction course through Facebook this fall . Definitely worth pursuing.

Announcing the 45th Eastern Primitive Rendezvous

The 45th Eastern Primitive Rendezvous takes place September 23-October 1, 2022.   For more information, visit the official EPR website and Facebook group . We will be hosting the 45th Eastern Primitive Rendezvous on our family farm, near East Smithfield, PA. The dates are September 23 - October 1, 2022.  This is a living-history event depicting 18th-century activities. Visitors can tour the camp each day from 10 a.m. to 4 p.m.  Tuesday, September 27 is School Tours Day, during which we welcome classes from all of the area schools. Campers need to preregister ( nrlhf.org/pdf/pre-reg.pdf ), and period-correct clothing and gear are required.  For the exact location and more details, visit the  official EPR website  and Facebook group . For those of you who attended the 2017 EPR, this is the same location.

the case for incremental redesign: part ii

If you are in any way responsible for a Web site, you should have some understanding of the principles of Extreme Programming . Cultivated as a discipline of software development, it is a combination of ensuring that designs remain uncomplicated, centering changes around user requirements, and employing the concept of the "Whole Team." The result is that small changes are released as they are needed - and endorsed - by the client. Not surprisingly, Extreme Programming speaks well to Web management. Consider its core values: simplicity, communication, feedback, and courage. These are the bedrock incremental redesign. Simplicity - Integrate all site changes in small doses. Avoid tectonic disruption of the entire Web presence. Document faithfully, but do not get bogged down in over-documenting. Or overplanning. Leverage reusable objects. Better yet, get the site into a content management system - one that is scaled to its requirements. Eliminate unmanageable code morass by fol...