Making Sense of Search Engines
tpromocom (image)

Making Sense of Search Engines

It's likely that it'll be a search engine that brings your website to the attention of prospective customers and other interested parties on the Internet more so than anything else. That's why it's important for you to know how search engines work and how they present information to someone doing research on the Net.

There are basically two types of search engines. The first one is performed by a robot called a crawler or spider. Search Engines use spiders to index websites. When you submit your website pages to a search engine by completing their required submission page, the search engine spider will index your entire site.

A 'spider' is actually an automated program that's run by the search engine system. Spiders visit a web site, read the content, examine the site's Meta tags, and they follow links contained within it. The spider then returns all that information to a central depository where the data is indexed. This or another spider will then return to the site periodically to check for new information. The frequency with which this occurs is determined by the moderators of the search engine.

A search engine will visit each link you have on your website and it will index them for future reference. Some spiders will only index a certain number of pages, so don't create a site with 500 pages and expect every one of them to appear in a search engine's search results.

A spider is almost like a book where it contains the table of contents, the content, and the links and references. It does this for all the websites it finds during its search of the Internet. Many of them will index up to a million pages a day!

Examples of popular search engines include DuckDuckGo, Start Page, Excite, Lycos, AltaVista and Google. The first two in this short list do not save your search data while the others are known to do so, especially Google.

When you ask a search engine to locate information, it will actually search through its index which it created without actually searching the Web. Different search engines produce different rankings because not every one of them uses the same algorithms.

One of the things that a search engine algorithm scans for is the frequency and location of keywords on a web page. It can also detect artificial keyword stuffing or spamdexing. Then the algorithms analyze the way a page links to other pages on the Web. By checking how pages link to one another, a search engine can determine what a page is about and whether the keywords of the linked pages are similar to the keywords on the original page.

For an extensive list of search engines, use our ToganX.Info Search page at There are approximately 50 search engines on ToganX to choose from. For additional information about Tpromo.Com, go to For more information on Al Colombo, go to

General Questions, Comments, and Suggestions:




Newsletter: Click Here

For a free consultation with Allan Colombo, call 330-956-9003 or email: Copywriter@Tpromo.Com.

Click here for our Micro Blog menu.
Click here for our main website.