With
moderate search engine optimization knowledge, some common sense,
and a resourceful and imaginative mind, one can be able to keep
his or her web site in good standing with search engines even
through the most significant algorithm changes. The recent Google
update of October/November 2005, dubbed "Jagger", is what inspired
me to write this, as I saw some web sites that previously ranked
in the top 20 results for extremely competitive keywords suddenly
drop down to the 70th page.
|
Yes, the ebb and flow of search engine rankings is nothing to write
home about, but when a web site doesn't regain many ranking spots
after such a drop it can tell us that the SEO done on the site may
have had some long-term flaws. In this case, the SEO team has not
done a good job predicting the direction a search engine would take
with its algorithm.
Impossible to predict, you say? Not quite. The ideas behind Google's
algorithm come from the minds of fellow humans, not supercomputers.
I'm not suggesting that it's easy to "crack the code" so to speak
because the actual math behind it is extremely complicated. However,
it is possible to understand the general direction that a search
engine algorithm will take by keeping in mind that any component
of SEO which is possible to manipulate to an abnormal extent will
eventually be weighted less and finally rendered obsolete.
One of the first such areas of a web site that started to get
abused by webmasters trying to raise their rankings was the keywords
meta tag. The tag allows a webmaster to list the web site's most
important keywords so the search engine knows when to display
that site as a result for a matching search. It was only a matter
of time until people started stuffing the tag with irrelevant
words that were searched for more frequently than relevant words
in an attempt to fool the algorithm. And they did fool it, but
not for long. The keywords meta tag was identified as an area
that was too susceptible to misuse and was subsequently de-valued
to the point where the Google algorithm today doesn't even recognize
it when scanning a web page.
Another early tactic which is all but obsolete is repeating keywords
at the bottom of a web page and hiding them by changing the color
of the text to match the background color. Search engines noticed
that this text was not relevant to the visitor and red- flagged
sites that employed this method of SEO.
This information is quite basic, but the idea behind the aforementioned
algorithm shifts several years ago is still relevant today. With
the Jagger update in full swing, people in the SEO world are taking
notice that reciprocal links may very well be going the way of
the keywords meta tag. (I.e. extinct) Webmasters across the world
have long been obsessed with link exchanges and many profitable
web sites have existed offering services that help webmasters
swap links with ease. But with a little foresight, one could see
that link trading had its days numbered, as web sites would obtain
thousands of incoming links from webmasters who may have never
even viewed the web site they were trading with. In other words,
a web site's popularity was manipulated by excessively and unnaturally
using an SEO method.
So with keyword meta tags, keyword stuffing within content, and
now link exchanges simply a part of SEO history, what will be
targeted in the future? Well, let's start with what search engines
currently look at when ranking a web site and go from there:
On-page Textual Content. In the future, look for search engines
to utilize ontological analysis of text. In other words, not only
your main keywords will play a factor in your rankings, but also
words that relate to them. For example, someone trying to sell
NFL jerseys online would naturally mention the names of teams
and star players. In the past, algorithms might have skipped over
those names, deemed them irrelevant to a search for "NFL jerseys."
But in the future, search engines will reward those web sites
with a higher ranking than those that excessively repeat just
"NFL jerseys." With ontological analysis, web sites that speak
of not only the main keywords but other relevant words can expect
higher rankings.
The conclusion: Write your web site content for your visitors,
not search engines. The more naturally written sites can expect
to see the better results in the future.
Offering Large Amounts of Content. This can frequently take the
form of dynamic pages. Even now, search engines can have a difficult
time with dynamic content on web sites. These page usually have
lengthy URLs consisting of numbers and characters such as &, =,
and ? The common problem is that the content changes so frequently
on these dynamic pages and the page becomes "old" in the search
engine's database, thus leaving the search users seeing results
that contain old information. Since many dynamic pages are created
by web sites displaying hundreds or thousands of products they
sell, and the number of people selling items on the Internet will
obviously increase in the coming years, you can expect that search
engines will improve their technology and do a better job indexing
dynamic content in the future.
The conclusion: Put yourself ahead of the game if you are selling
products online and invest in database and shopping cart software
that is SEO-friendly.
Incoming Links. Once thought to be a very difficult thing to
manipulate, incoming links to one's web site have been abused
by crafty SEOs and webmasters the world over. It is finally at
a point where Google is doing a revamp of what constitutes a "vote
from [one site to another]" as they explain it in their webmaster
resources section. Link exchanges are worth significantly less
now than ever to the point where the only real value in obtaining
them is to make sure a new web site gets crawled by search engine
spiders.
Over the years, many web sites reached top spot for competitive
keywords by flexing their financial muscle and buying thousands
of text links pointing to their site with keywords in the anchor
text. Usually these links would appear like advertisements along
sidebars or navigation areas of web sites. Essentially this was
an indirect way of paying for high Google rankings, something
which Google is no doubt trying to combat with each passing algorithm
update. One idea of thought is that different areas of a web page
from a visual point of view will be weighted differently. For
example, if a web site adds a link to your site within the middle
of their page text, that link should count for more than one at
the bottom of the site near the copyright information.
This brings up the value of content distribution. By writing
articles, giving away free resources, or offering something else
of value to people, you can create a significant amount of content
on other web sites that will include a link back to your own.
The conclusion: It all starts with useful content. If you are
providing your web site visitors with useful information, chances
are many other sites will want to do the same. SEO doesn't start
with trying to cheat the algorithm; it starts with an understanding
of what search engines look for in a quality web site.