Chapter 8. Tools of the Trade

Driving traffic to your site is macro-level SEO strategy. The other important tools of the trade have to do with crafting your web sites—and web pages—so that they “put their best faces forward” in search engine ranking.

Working with the Bot

To state the obvious, before your site can be indexed by a search engine, it has to be found by the search engine. Search engines find web sites and web pages using software that follows links to crawl the web. This kind of software is variously called a crawler, a spider, a search bot, or simply a bot (bot is a diminutive for “robot”).

To be found quickly by search engine bot, it helps to have inbound links to your site. More important, the links within your site should work properly. If a bot encounters a broken link, it cannot reach—or index—the page pointed to by the broken link.

Note

As I explained earlier, you don’t have to wait for the bot to find your site on its own if you list your site—manually or using a submission tool—with the important search engines. However, it’s still important, even if you list your site, that it be “bot friendly.”

Images

Pictures don’t mean anything to a search bot. The only information a bot can gather about pictures come from the alt attribute used within a picture’s <img> tag, and from text surrounding the picture. Therefore, always take care to provide description information via the <img> tag’s alt attribute along with your images, and to provide at least one text-only (e.g., outside of an image map) link to all pages on your site.

Links

Certain kinds of links to pages (and sites) simply cannot be traversed by a search engine bot. The most significant issue is that a bot cannot login to your site. (This is probably a very good thing, or we’d all be in big trouble!)

So if a site or page requires a user name and a password for access, then it probably will not be included in a search index.

Don’t be fooled by seamless page navigation using such techniques as cookies or session identifiers. If an initial login was required, then these pages can probably not be accessed by a bot.

Complex URLs that involve a script can also confuse the bot (although only the most complex dynamic URLs are absolutely non-navigable). You can generally recognize this kind of URL because a ? is included following the script name. Pages reached with this kind of URL are dynamic, meaning that the content of the page varies depending upon the values of the parameters passed to the page generating the script (the name of the script, often code written in PHP, comes before the ? in the URL).

You can try this example by comparing the two URLs to see for yourself the difference a changed parameter makes!

Dynamic pages opened using scripts that are passed values are too useful to avoid. Most search engine bots can traverse dynamic URLs provided they are not too complicated. But you should be aware of dynamic URLs as a potential issue with some search engine bots, and try to keep these URLs as simple—using as few parameters—as possible.

File Formats

Most search engines, and search engine bots, are capable of parsing and indexing many different kinds of file formats. For example, Google indexes file types including: pdf, asp, jsp, html, shtml, xml, cfm, php, doc, xls, ppt, rtf, wks, lwp, wri, and swf.

However, simple is often better. To get the best search engine placement, you are well advised to keep your web pages, as they are actually opened in a browser, to straight HTML.

Note

Even though a file opens in straight HTML in a browser, it can be generated using a server-side script. How a file was created is essentially irrelevant to the search engine, which cares only about the actual file that is browsed.

So check the source file as shown in a browser rather than the script file used to generate a dynamic page to see what the search engine will index.

Google puts the “simple is best” idea this way: “If fancy features such as JavaScript, cookies, session IDs, frames, DHTML, or Flash keep you from seeing all of your site in a text browser, then search engine spiders may have trouble crawling your site.” The only way to know for sure whether a bot will be unable to crawl your site is to check your site using an all-text browser.

Viewing Your Site with an All-Text Browser

Improvement implies a feedback loop: you can’t know how well you are doing without a mechanism for examining your current status. The feedback mechanism that helps you improve your site from an SEO perspective is to view your site as the bot sees it. This means viewing the site using a text-only browser. A text-only browser, just like the search engine bot, will ignore images and graphics, and only process the text on a page.

The best-known text-only web browser is Lynx. You can find more information about Lynx at http://lynx.isc.org/. Generally, the process of installing Lynx involves downloading source code and compiling it.

The Lynx site also provides links to a variety of pre-compiled Lynx builds you can download.

Don’t want to get into compiled source code, or figuring out which idiosyncratic Lynx build to download? There is a simple Lynx Viewer available on the Web at http://www.delorie.com/web/lynxview.html. You’ll need to follow directions carefully to use it. Essentially, these instructions involve adding a file to your web site to prove you own the site. The host of the Lynx Viewer is offering a free service, and doesn’t want to be deluged, but it is not hard to comply

Using Lynx Viewer, it’s easy to see the text that the search bot sees when you are not distracted by the “eye candy” of the full image version.

Tip

Users, web designers, and advertisers are often very fond of fancy graphics and graphic effects on the Web. But bear in mind that search engines are almost exclusively concerned with words (meaning plain text rather than pictures). Therefore, your SEO efforts should be focused on text, not the “look and feel” of a site. When practicing SEO stick to word-craft, and don’t get sidetracked by irrelevant visual issues.

Meta Information

Meta information—sometimes called meta tags for short—is a mechanism you can use to provide information about a web page.

The term derives from the Greek word “meta,” which means “behind” or “hidden”. “Meta” refers to the aspect of something that is not immediately visible, perhaps because it is in the background, but which is there nonetheless and has an impact.

The most common meta tags provide a description and keywords for telling a search engine what your web site and pages are all about. Each meta tag begins with a name attribute that indicates what the meta tag represents. The meta tag

<meta name="description" ...></meta>

means that this tag will provide descriptive information. The meta tag

<meta name="keywords" ...></meta>

means that the tag will provide keywords.

The description and keywords go within a content attribute in the meta tag. For example, here’s a meta description tag (often simply called the meta description):

  <meta name="description" content="Quality information, articles about a variety of
topics ranging from Photoshop, programming to business, and investing."></meta>

Keywords are provided in a comma-delimited list. For example:

  <meta name="keywords" content="Photoshop, Wi-Fi, wireless networking, programming, C#,
  business, investing, writing, digital photography, eBay, pregnancy,
information"></meta>

Meta Information Tactics

In the early days of SEO, intelligent use of meta information was a crucial element of SEO. But since anyone can add any meta keyword and description information they’d like to a site or page, the feature has been widely abused—and search engines discount meta data in favor of automated page content analysis using PageRank and other variables.

That said, the price of adding meta information—it’s free—is right, so you should add it to each page as a matter of standard SEO practice. Try to provide targeted

meta keyword lists and descriptions (for example, the samples I provided just above are way too general to be helpful as SEO).

Meta keywords should be limited to a dozen or so terms. Don’t load up the proverbial kitchen sink. Think hard about the keywords that you’d like to lead to your site when visitors search.

Note

Meta tags can contain a lot more than just descriptions and keywords, including (but not limited to) a technical description of the kind of content on a page, and even the character encoding used:

<meta http-equiv="Content-Type" content="text/html; charset=iso-8859-1" />

One kind of page that really does need a meta description and keyword list—even today—is the page that primarily consists of images. Recall that I’ve suggested that as a matter of SEO you avoid such pages. But if you can’t help yourself, or have bowed to unspeakable pressure applied by your web designer, you should know that Google and other search engines won’t have a clue what is on your page—unless you describe it for the search engine using meta information.

Tip

If you want to include a phrase containing more than one term in your keyword list, quote it. For example: “digital photography”. However, there is not much point in including a compound term if the words in the phrase (“digital” and “photography”) are already included as keywords.

For the keywords that are really significant to your site, you should include both single and plural forms, as well as any variants. For example, a site about photography might well want to include both “photograph” and “photography” as meta tags.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
18.118.12.222