Anatomy Of An Internet Search Engine

0
46

For some unfortunate souls SEO is simply the learning of tricks and techniques that, according to their understanding, should propel their site into the top rankings on the major search engines.

This understanding of the way SEO works can be effective for a time however it contains one basic flaw the rules change. Search engines are in a constant state of evolution in order to keep up with the SEO’s in much the same way that Norton, McAfee, AVG or any of the other anti-virus software companies are constantly trying to keep up with the virus writers.

Basing your entire websites future on one simple set of rules (read: tricks) about how the search engines will rank your site contains an additional flaw, there are more factors being considered than any SEO is aware of and can confirm. That’s right, I will freely admit that there are factors at work that I may not be aware of and even those that I am aware of I cannot with 100% accuracy give you the exact weight they are given in the overall algorithm. Even if I could, the algorithm would change a few weeks later and what’s more, hold your hats for this one; there is more than one search engine.

So if we cannot base our optimization on a set of hard-and-fast rules what can we do? The key my friends, is not to understand the tricks but rather what they accomplish. Reflecting back on my high school math teach Mr. Barry Nicholl I recall a silly story that had a great impact. One weekend he had the entire class watch Dumbo The Flying Elephant (there was actually going to be a question about it on our test). Why? The lesson we were to get from it is that formulas (like tricks) are the feather in the story. They are unnecessary and yet we hold on to them in the false belief that it is the feather that works and not the logic. Indeed, the tricks and techniques are not what works but rather the logic they follow and that is their shortcoming.

And So What Is Necessary?

To rank a website highly and keep it ranking over time one must optimize it with one primary understanding, that a search engine is a living thing. Obviously this is not to say that search engines have brains, I will leave those tales to Orson Scott Card and other science fiction writers, however their very nature results in a lifelike being with far more storage capacity.

If we consider for a moment how a search engine functions; it goes out into the world, follows the road signs and paths to get where it’s going, and collects all of the information in its path. From this point, the information is sent back to a group of servers where algorithms are applied in order to determine the importance of specific documents. How are these algorithms generated? They are created by human beings who have a great deal of experience in understanding the fundamentals of the Internet and the documents it contains and who also have the capacity to learn from their mistakes, and update the algorithms accordingly. Essentially we have an entity that collects data, stores it, and then sorts through it to determine what’s important which it’s happy to share with others and what’s unimportant which it keeps tucked away.

So Let’s Break It Down

To gain a true understanding of what a search engine is, it’s simple enough to compare it to the human anatomy as, though not breathing, it contains many of the same core functions required for life. And these are:

The Lungs & Other Vital Organs – The lungs of a search engine and indeed the vast majority of vital organs are contained within the datacenters in which they are housed. Be it in the form of power, Internet connectivity, etc. As with the human body, we do not generally consider these important in defining who we are, however we’re certainly grateful to have them and need them all to function properly.

The Arms & Legs – Think of the links from the engine itself as the arms and legs. These are the vehicles by which we get where we need to go and retrieve what needs to be accessed. While we don’t commonly think of these as functions when we’re considering SEO these are the purpose of the entire thing. Much as the human body is designed primarily to keep you mobile and able to access other things, so too is the entire search engine designed primarily to access the outside world.

The Eyes – The eyes of the search engine are the spiders (AKA robots or crawlers). These are the 1s and 0s that the search engines send out over the Internet to retrieve documents. In the case of all the major search engines the spiders crawl from one page to another following the links, as you would look down various paths along your way. Fortunately for the spiders they are traveling mainly over fiber optic connections and so their ability to travel at light speed enables them to visit all the paths they come across whereas we as mere humans have to be a bit more selective.

The Brain – The brain of a search engine, like the human brain, is the most complex of its functions and components. The brain must have instinct, must know, and must learn in order to function properly. A search engine (and by search engine we mean the natural listings of the major engines) must also include these critical three components in order to survive.

The Instinct – The instinct of a search engines is defined in it’s core functions, that is the crawling of sites and either the inability to read specific types of data, or the programmed response to ignore files meeting a specific criteria. Even the programmed responses become automated by the engines and thus fall under the category of instinct much the same as the westernized human instinct to jump from a large spider is learned. An infant would probably watch the spider or even eat it meaning this is not an automatic human reaction.

The instinct of a search engines is important to understand however once one understands what can and cannot be read and how the spiders will crawl a site this will become instinct for you too and can then safely be stored in the “autopilot” part of your brain.

The Knowing – Search engines know by crawling. What they know goes far beyond what is commonly perceived by most users, webmasters and SEOs. While the vast storehouse we call the Internet provides billions upon billions of pages of data for the search engines to know they also pick up more than that. Search engines know a number of different methods for storing data, presenting data, prioritizing data and of course, way of tricking the engines themselves.

While the search engine spiders are crawling the web they are grabbing the stores of data that exist and sending it back to the datacenters, where that information is processed through existing algorithms and sp@m filters where it will attain a ranking based on the engine’s current understanding of the way the Internet and the documents contained within it work.

Similar to the way we process an article from a newspaper based on our current understanding of the world, the search engines process and rank documents based on what they understand to be true in the way documents are organized on the Internet.

The Learning – Once it is understood that search engines rank documents based on a specific understanding of the way the Internet functions, it then follows that in order to insure that new document types and technologies are able to be read and that the algorithm be changed as new understandings of the functionality of the Internet are uncovered a search engine must have the ability to “learn”.

Aside from a search engine needing the ability to properly spider documents stored in newer technologies, search engines must also have the ability to detect and accurately penalize sp@m and as well as accurately rank websites based on new understandings of the way documents are organized and links arranged. Examples of areas where search engines must learn in an ongoing basis include but are most certainly not limited to:

Understanding the relevancy of the content between sites where a link is found

Attaining the ability to view the content on documents contained within new technologies such as database types, Flash, etc.

Understanding the various methods used to hide text, links, etc. in order to penalize sites engaging in these tactics

Learning from current results and any shortcoming in them, what tweaks to current algorithms or what additional considerations must be taken into account to improve the relevancy of the results in the future.

The learning of a search engine generally comes from the uber-geeks hired by and the users of the search engines. Once a factor is taken into account and programmed into the algorithm it them moves into the “knowing” category until the next round of updates.

How This Helps in SEO

This is the point at which you may be asking yourself, “This is all well-and-good but exactly how does this help ME?” An understanding of how search engines function, how they learn, and how they live is one of the most important understandings you can have in optimizing a website. This understanding will insure that you don’t simply apply random tricks in hopes that you’ve listened to the right person in the forums that day but rather that you consider what is the search engine trying to do and does this tactic fit with the long term goals of the engine.

For a while keyword density sp@mming was all the rage among the less ethical SEOs as was building networks of websites to link together in order to boost link popularity. Neither of these tactics work today and why? They do not fit with the long-term goals of the search engine. Search engines, like humans, want to survive. If the results they provide are poor then the engine will die a slow but steady death and so they evolve.

When considering any tactic you must consider, does this fit with the long-term goals of the engine? Does this tactic in general serve to provide better results for the largest number of searches? If the answer is yes then the tactic is sound.

For example, the overall relevancy of your website (i.e. does the majority of your content focus on a single subject) has become more important over the past year or so. Does this help the searcher? The searcher will find more content on the subject they have searched on larger sites with larger amounts of related content and thus this shift does help the searcher overall. A tactic that includes the addition of more content to your site is thus a solid one as it helps build the overall relevancy of your website and gives the visitor more and updated information at their disposal once they get there.

Another example would be in link building. Reciprocal links are becoming less relevant and reciprocal-links between unrelated sites are virtually irrelevant. If you are engaging in reciprocal link building insure that the sites you link to are related to your site’s content. As a search engine I would want to know that a site in my results also provided links to other related sites thus increasing the chance that the searcher was going to find the information that they are looking for one way or another without having to switch to a different search engine.

In Short

In short, think ahead. Understand that search engines are organic beings that will continue to evolve. Help feed them when they visit your site and they will return often and reward your efforts. Use unethical tactics and you may hold a good position for a while but in the end, if you do not use tactics that provide for good overall results, you will not hold your position for long. They will learn.

Dave Davies is the CEO of Beanstalk Search Engine Positioning, Inc. One of the few SEO firms to offer guaranteed optimization services. Visit their site for a free SEO & web design review and their blog to keep updated on the latest going on in the search engine world.

LEAVE A REPLY

Please enter your comment!
Please enter your name here