freshpromo Toronto SEO

Predicting Search Engine Algorithm Changes

By John Metzler

With moderate search engine optimization knowledge, some common sense, and a resourceful and imaginative mind, one can be able to keep his or her web site in good standing with search engines even through the most significant algorithm changes. The recent Google update of October/November 2005, dubbed “Jagger”, is what inspired me to write this, as I saw some web sites that previously ranked in the top 20 results for extremely competitive keywords suddenly drop down to the 70th page. Yes, the ebb and flow of search engine rankings is nothing to write home about, but when a web site doesn’t regain many ranking spots after such a drop it can tell us that the SEO done on the site may have had some long-term flaws. In this case, the SEO team has not done a good job predicting the direction a search engine would take with its algorithm.

Impossible to predict, you say? Not quite. The ideas behind Google’s algorithm come from the minds of fellow humans, not supercomputers. I’m not suggesting that it’s easy to “crack the code” so to speak because the actual math behind it is extremely complicated. However, it is possible to understand the general direction that a search engine algorithm will take by keeping in mind that any component of SEO which is possible to manipulate to an abnormal extent will eventually be weighted less and finally rendered obsolete.

One of the first such areas of a web site that started to get abused by webmasters trying to raise their rankings was the keywords meta tag. The tag allows a webmaster to list the web site’s most important keywords so the search engine knows when to display that site as a result for a matching search. It was only a matter of time until people started stuffing the tag with irrelevant words that were searched for more frequently than relevant words in an attempt to fool the algorithm. And they did fool it, but not for long. The keywords meta tag was identified as an area that was too susceptible to misuse and was subsequently de-valued to the point where the Google algorithm today doesn’t even recognize it when scanning a web page.

Another early tactic which is all but obsolete is repeating keywords at the bottom of a web page and hiding them by changing the color of the text to match the background color. Search engines noticed that this text was not relevant to the visitor and red- flagged sites that employed this method of SEO.

This information is quite basic, but the idea behind the aforementioned algorithm shifts several years ago is still relevant today. With the Jagger update in full swing, people in the SEO world are taking notice that reciprocal links may very well be going the way of the keywords meta tag. (I.e. extinct) Webmasters across the world have long been obsessed with link exchanges and many profitable web sites have existed offering services that help webmasters swap links with ease. But with a little foresight, one could see that link trading had its days numbered, as web sites would obtain thousands of incoming links from webmasters who may have never even viewed the web site they were trading with. In other words, a web site’s popularity was manipulated by excessively and unnaturally using an SEO method.

So with keyword meta tags, keyword stuffing within content, and now link exchanges simply a part of SEO history, what will be targeted in the future? Well, let’s start with what search engines currently look at when ranking a web site and go from there:

On-page Textual Content.
In the future, look for search engines to utilize ontological analysis of text. In other words, not only your main keywords will play a factor in your rankings, but also words that relate to them. For example, someone trying to sell NFL jerseys online would naturally mention the names of teams and star players. In the past, algorithms might have skipped over those names, deemed them irrelevant to a search for “NFL jerseys.” But in the future, search engines will reward those web sites with a higher ranking than those that excessively repeat just “NFL jerseys.” With ontological analysis, web sites that speak of not only the main keywords but other relevant words can expect higher rankings.

The conclusion: Write your web site content for your visitors, not search engines. The more naturally written sites can expect to see the better results in the future.

Offering Large Amounts of Content.
This can frequently take the form of dynamic pages. Even now, search engines can have a difficult time with dynamic content on web sites. These page usually have lengthy URLs consisting of numbers and characters such as &, =, and ? The common problem is that the content changes so frequently on these dynamic pages and the page becomes “old” in the search engine’s database, thus leaving the search users seeing results that contain old information. Since many dynamic pages are created by web sites displaying hundreds or thousands of products they sell, and the number of people selling items on the Internet will obviously increase in the coming years, you can expect that search engines will improve their technology and do a better job indexing dynamic content in the future.

The conclusion: Put yourself ahead of the game if you are selling products online and invest in database and shopping cart software that is SEO-friendly.

Incoming Links.
Once thought to be a very difficult thing to manipulate, incoming links to one’s web site have been abused by crafty SEOs and webmasters the world over. It is finally at a point where Google is doing a revamp of what constitutes a “vote from [one site to another]” as they explain it in their webmaster resources section. Link exchanges are worth significantly less now than ever to the point where the only real value in obtaining them is to make sure a new web site gets crawled by search engine spiders.

Over the years, many web sites reached top spot for competitive keywords by flexing their financial muscle and buying thousands of text links pointing to their site with keywords in the anchor text. Usually these links would appear like advertisements along sidebars or navigation areas of web sites. Essentially this was an indirect way of paying for high Google rankings, something which Google is no doubt trying to combat with each passing algorithm update. One idea of thought is that different areas of a web page from a visual point of view will be weighted differently. For example, if a web site adds a link to your site within the middle of their page text, that link should count for more than one at the bottom of the site near the copyright information.

This brings up the value of content distribution. By writing articles, giving away free resources, or offering something else of value to people, you can create a significant amount of content on other web sites that will include a link back to your own.

The conclusion: It all starts with useful content. If you are providing your web site visitors with useful information, chances are many other sites will want to do the same. SEO doesn’t start with trying to cheat the algorithm; it starts with an understanding of what search engines look for in a quality web site.