Now that we have your site up and we have taken some baby steps toward getting some visitors to your site, it’s time to manage and analyze the site. As luck would have it, there are plenty of tools for just these things and better yet, these tools are free.
There is an entire toolbox at your disposal at http://www.google.com/webmasters/
The first of these useful items are the webmaster tools which can do site diagnostics and basically give you an indication of how the site is working, which search terms are bringing people to your site and even show you if malware has been found on your site. Overall site performance (page load times) are also indexed and graphed.
The other tool is Google Analytics, which does require you to add a snippet of code to your pages, but the statistics gathered are very useful. Analytics gives you tremendous insight into your visitors, how they are getting to your site, how long they stay and what they view. With this data in hand you can hone your site, tailor it to suit your visitors needs and maximize it’s potential. You can also set “goals” - pages you want visited or actions you want performed (like filling out a request for quotation) and tune your site to funnel visitors toward those goals more effectively.
This is all very Google-centric and there now appears to be similar tools for Bing - but it is probably a better idea to tackle the Google side first due to it’s dominance in the search market. If you feel ambitious, look into doing something similar with Bing and Yahoo!
This is not a set and forget kind of thing, you should expect to check back frequently to monitor how your site is doing. You can also have Analytics send you scheduled emails to save you from visiting the site. It also not a bad idea to do the malware check regularly - it’s not always readily apparent to the site owner when your been compromised. All in all these free tools can help you or your webmaster really tune-up your site and give you far more information that just page hits.
Bringing in the Traffic
In part one of the series, I talked about getting a site and doing it right. Now what? Well if you have a site and you are content with the form and function, it is likely a good time to find ways to become noticed and drive traffic to your site. While I often suggest using other media to announce your web presence - print, radio, etc. - today I'm sticking just to using the web and the ever-present search engine. One of the main goals for your site is to have a good page ranking - this means being listed in the top results when a potential client does a search using keywords that relate to your business. This general process is known as Search Engine Optimization - SEO, which is equal parts science and art. There is also no lack of people claiming to be SEO experts and will gladly take your money and deliver questionable results - take due care.
First your site needs a little prep work, you need to take care of three things right off the bat - your Meta Tags, SiteMap and robots.txt. All these might be something your web developer will take care of (and they should); but it does not hurt to know what they are all about.
Meta Tags are pieces of HTML code - hidden to the casual observer - that hint to the search engines what your site is all about. These hints help your site get properly indexed, so when someone does a search for “antique watch insurance southern Ohio” your site gets listed, preferably on the first page of results. The two key meta tags for this are called description and keywords - the first is a brief description of your site or that particular page, the latter is a collection of keywords that relate to the content, essentially these are the search words that should bring a potential client to your doorstep. Here is what they look like in the page source code: http://www.w3schools.com/tags/tag_meta.asp
A sitemap is an xml file - sitemap.xml, that resides at the document root on the webserver - it is a file that search engines can use to understand the layout and content of your site - it is a guide or map that crawlers refer to. Since many websites have dynamic content, a good sitemap helps to reveal that content and the sites true hierarchy which would otherwise be hidden from the crawler and therefore not indexed. Sitemaps do not have to be handcrafted, there are free generators available to simplify the task - http://www.xml-sitemaps.com/
Finally, robots.txt; now that you want to entice the various web crawlers, spiders and bots to come for a visit, you may also want them to play nice and index that which you want indexed and possibly ignore other content. The robots.txt file gives them basic instruction on what part of the site is okay to crawl and what is off limits. Keep in mind - not all crawlers are respectable and they can ignore robots.txt - dealing with them is a whole other story though. Again this file goes into your web document root directory. Using robots.txt is not a security measure and does not ensure the privacy of content and should not be used with that goal in mind. More info - http://www.robotstxt.org/robotstxt.html
With those items taken care of, now it’s time to submit your site to some major search engines. This is basically the process of alerting the various search engines to your site’s existence by submitting the URL to their various services:
Do not expect instant results, but your site will be crawled and indexed, and good formatting and content along with proper meta tags, sitemaps and a robots.txt should help your page ranking. That about wraps things up, I hope you can put some of this to good use on your site.