Celtnet: how to make money online
Internet Marketing Make Money Online

Sunday, July 30, 2006

Is Product the Key to Success?



After working at this whole 'internet marketing' thing for quite a while now I've finally come to the conclusion that the only real way to make money from Internet marketing is to have your own products. I'm certain that it's possible to make a good living from affiliate marketing and Google AdSense click-throughs. However, to make real money with these programs you need content and lots of it.

Admittedly, I'm going to persevere with these schemes, especially on pages such my spice guide page that includes both Google AdSense ads and Amazon links to relevant books. Then there are my auctions pages that mix AdSense advertising with eBay searches, Amazon comparisons, ClickBank searches and eBay misspelling searches. Yielding as many sources of income as possible in one place. Again, such pages can bring-in quite reasonable income.

Then I thought to myself. I'd gone through quite a lot of effort to write the pages that afforded me with the functionality on these pages. Wouldn't I gain more income if I actually marketed and sold the software itself?

In putting-up my website I'd also learnt a lot about how to create websites, get them indexed, improve their website rankings and market them. This should all be grist to producing eBooks that could again be marketed and sold. In additon there were the other things I was interested in, such as my recipes site which could also be converted into eBooks. Creating real products.

Obviously, once you have tangible products (even electronic ones) you need a way to sell them. The obvious way is to create a web-store based on PayPal as a search methodology. I've almost completed my first eBook and I've been gathering information all over the web about creating a PayPal-based store which I'll start implementing next week.

Once it's up and running I'll post back on this blog as to how it was created (and how you can create an equivalent store for your selves). Then I'll begin updating on how this new marketing strategy is going.

Wednesday, July 26, 2006

Images with AdSense Ads, the Next Big Thing?



Google's AdSense program has become and almost ubiquitous source of income for webmasters both large and small. As a result may people have been playing with the format and placemet of AdSense ad units to maximize click-through rates.

The greatest amount of information about optimizing adsense unit placement and format comes from Google (hardly surprising, as they make money from these ad units themselves) and their data clearly indicate that wider units tend to out-perform taller ones with the formats: 336x280, 300x250 and 160x600 resulting in the highest overall number of click-throughs.

It has long been known that plain text ads out-perfrom image-based ads (my page on internet advertising effectivenesss explains why). Also, ads that look as if they are part of the text (same background, no border and with blue links) out-perform ads that stand out as separate blocks. It's even better if you can incorporate the ads into your text. Indeed, a recent study by Tim Carter (at AskTheBuilder.com) showed a 20% increase in revenues when he palced a large rectangle (336x280) ad unit positioned within his articles and sited in the upper left-hand corner.

If you generate pages with lots of text then using multiple ad units can help with click through rates (just note that if you have multiple ad units then the ads which lie lower down your page will have lower bid prices). But it's important that the ads with the highest click-through rates must appear first in your HTML code (you can use CSS to locate it anywhere you want on your page).

The latest trend, however, is the use of images to make ads stand out on a page (see Draw a Pig for an example. Now, you can't use things like flashing GIF arrows pointing at your ad block as this would violate the AdSense terms and conditions by enticing visitors to click. However, you can have a lerge letterboard (say 728x90) and if your site is about flowers, say, then you can legitimately place a row of flowers above the ad block as this simply fits-in with the overall theme and content of your site.

Indeed, if recent reports are to be believed then this is strategy that even Google themselves are experimenting with (see the Digital Inspiration blog). This seems only to be a beta test for Google at the moment, but a number of people have now reported seeing versions of the Google five-ad tower where the first ad in the block is an image rather than the expected text ad. I've just started using this image enhancement technique on my own websites (you can see and example here on my history of the spice trade page) where the images are all ancient means of transporting spices and thus relate directly to the content of the article itself.

Over the next few days I'll begin converting more of my high-traffic pages to similar image-based formats and I'll report back on my progress in abut three weeks. It certainly seems to be a sensible way froward and anything that improves click-through rates can only be a good thing (especially if Google are experimenting with this themselves).

Monday, July 24, 2006

Blogging as a Resource



Your Blog is potentially a very valuable resource, are you making the most of it? If you have an active blog then you potentially have a very valuable resource that you should be making the most of. The first thing is to monetize your Blog. This could be as simple as adding some advertising banners on your blog (one across the top and one on your navigation pane). If you have some affiliateships or some of your own products your can also provide links to these on your blog.

If you also have your won website then your Blog should be integrated into this. You should have a link from your website to your blog and links from your Blog to valuable pages or resources on your website. Ideally these should be links to deep pages within your site as your Blog can help you overcome the problem of the search engines not indexing deeply enough into your website. The reason is this: If you have a Blog then all your older posts are archived, but these have the same website border as your current blogs. As a result you have a large number of copies of any links your've added to your blogs on multiple pages. This looks good to the web indexers and as a results they will index these pages (often very quickly).

The only hurdle to being indexed in this manner is that of actually getting your Blog crawled and indexed. If you're posting frequently then you can use 'ping' services such as pingomatic or pingoat to perform this task for you. Your new content will be posted to a number of Blog indexes and will be seen by the Search Engines and included in their indexes. However, as I've previously posted this may only be a transient inclusion. If you really want your Blog indexed on a premanent basis (along with all those pages the Blog points to) then you will need to treat your Blog exactly like any other web page.

What does this mean? Essentially you need to play the 'indexing' game. This means that you need to get sufficient in-bound links pointing to yor website that the serach engines can't help but index the pages and also have to maintanin them in their indexes. This means that you have to get your Blog included in as many Blog directories as possible (a good starting point for doing this is the RSSTop55 directory. This lists almost 185 blog directories and RSS feed aggregators.).

The other thing you need to do is to get your Blog published as an RSS xml feed as well (most blog generators do this for you) so that you can also add the content of your Blog to the various news aggregators out there.

This is only a brief outline of the advantages that a good Blog can bring you. For more information see my article on Blogging.

Sunday, July 23, 2006

The Importance of <meta> tags



I've been performing some tidying-up work on my website and a quick comparison of my <meta> tags with what Google reported for some of my web pages indicated that Google was using a the <META NAME="description" CONTENT=""> as well as the content tag (sometimes instead of) to add descriptions of some of my pages.

Now, I'd read somewhere the the <meta> tags were declining in terms of their usage, so I'd taken my eye off the boll as far as these components of my pages were concerned. Basically I'd missed a major trick — and now I was paying for it.

'Why is this important?' you may ask. Well, some of my pages, though they had different content had the same <meta> tags as others and Google was considering them as 'similar' and dumping the second page as a 'supplemental' result. Essentially, because of a stupid oversight I was losing pages of content from Google's main index. More importantly I was losing free advertising.

'How so?' you may well ask. For each page of your site that's indexed Google has a title of up to 66 characters and a description of up to 160 characters. This is your advertisement, your pitch to get visitors to come to various sections of your site. I know that the holy grail of SEO at the moment is to get your web page to be the first in Google's (or whatever search system you're targetting) list of results. If you're striving for this, then take a step back and consider how you're actually using the search results yourself. When you type in a search term or string and Google returns the results then you'll see about eight results of the ten that Google returns. Do you immediately go to the result on the first position and click, or do your eyes gaze over the results and if there's a title or strapline that interests you, do you then click on that one?

If you're perfectly honest with yourself (and studies of how people actually use web pages confirms this) then you're probably in the majority of those who glance at the descriptions and straplines to see something that interests them. Those who have bold headings and can write good content (within the 160-word limit of the strapline) will get the most traffic. OK, so it's important to get on the first page of results if you can, but it's far better for you to write good copy that will draw visitors in rather than trying to squeeeze the final few placings for your page.

This means that you will need to take a close look at the <title> tag of your web pages as this is what provides the title in the search engines. Beyond this you need to take a close look at anything that might be used to provide the description for your web page. This includes the <META NAME="description" CONTENT=""> tag and your first >h1> heading as well as the first sentence of your main text. If take care of all this then you should improve the chances of your web page being spotted from amongst the crowd. However, you won't really know if this effort has been worth it until your page does get indexed and you can see what text's being reported. Once you see this text then you may well want to edit it to make the information initially reported abut your web page be 'punchier' and stand out more.

Like all advertising making the most of how people perceive your page in that first search engine report is a question of editing, refining and testing. This is your first contact with a potential customer and your should make the most of it.

Thursday, July 20, 2006

Thoughts and Considerations



I must admit that I've broken my own rule of posting twice a week for the past week now. Unortunately there are only eighteen useful hours in a day and unfortunately I can't write everywhere at once.

The main reason for the quiet on this medium being that I've been rather busy on researching and writing something else. To bolster my recipes section on my site I've been researching some of the main spices used by various cultures and how these spices have been traded. As a result I now have a web page on the spice trade which details the spice trade and the influence it's had on the world from 5000 years ago to the present. Allied to this is a page giving detailing the spices themselves and what they are. These pages have taken a while to research but today they went 'live' for the first time.

What, however, I hear you ask, has this to do with this blog. Well, the reaserch is actually going to be used to create an eBook on spices, giving a history of the spice trad, a description of the plant origins for each spice and a number of recipes that show the spice off to it's best abilities.

You'll also notice that the pages are PHP scripts that also allow me to link these pages into my Amazon affiliate store so that I can recommend book dealing with the subject matter of the page. In addition, using my localization system I can poinit users to either the US or UK amazon stores, as befits their country of origin. I'm therefore providing an useful information source on the website, linked to Google AdSense and Amazon. The information from there will also provide a product that I can sell from my own website and which I can also supply to various eBook stores to sell for me.

As a result a little thought, some research and a fair amount of writing provides me with three independent sources of income. Not bad for a week's work!

Friday, July 14, 2006

Using Blogs to get your Site Indexed


Or not...



It's often said that writing a Blog, pinging this to a number of blog sites and getting it listed in Blog indexes can help with the indexing of your site and that a frequency of about two posts every week is about right for this. On learning this I thought it would be a great way of getting more of my site indexed in the main search engines.

As a result I started increasing the post frequency on this blog and began using pingoat to ping the main blog directories when my blog had been updated. The frequency increased to twice a week and then every other day (sometimes more frequently). Once I had sufficient content I also began adding the name of this Blog to some blog directories, spreading the word about the blog's existence as wide as possible.

I also attacked the 'look and feel' of the blog, adding links to important pages and areas of my site. By June 29th the Blog had been crawled and about six pages from it entered Googles Index on July 3rd. Then, on July 4th those pages vanished from Google's index. The Blog was spidered again on July 9th and on Jul 11th all the Blog pages as well as ten additional pages from my main site that were referenced by the blog were in Google's index. Of course on July 12th I lost my site (see the story below). On Jult 13th, wen my site came back up the pages were still in Google. Then, today on July 14th all the 30+ pages relating to this Blog had vanished again!

Is this another example of Google Losing the Plot? Basically a page should either be good enough to be included in Google's index or it should not. This 'is it there, or isn't it?' is a complete madness, however you look at it. It either suggests that Google's indexing system is broken, or it indicates that they're out of space on their servers and are dumping certain indexed pages based on some kind of quality algorithm that seems to make little or no sense to me.

It's no surprise that many sites are becoming completely frustrated with Google. Especially as Google's taking their usual attitude of 'everything's fine'. Despite trying to do everything right I've almost given up on getting Google to index the vast majority of my site and I'm relying on advertising to get traffic. Which is maybe Google's whole strategy here: trying to get as many people to used their AdWords system as possible. After all, if your pages are never going to enter their index what else can you do?

Admittedly I've had a very frustrating week and despite working very hard to improve and extend my website I'm seeing very little advance for all this work, but from what I see out there on the 'net I don't think that this problem is entirely limited to me.

Wednesday, July 12, 2006

Your Hosting Company can ruin your day



Today something happened to me that made me realize just how dependent I (and anyone else who makes their living on the internet) am on the whims of others. It just so happens that over the past few days my Hosting Company's been migrating their unix hosting servers over to new IP numbers. Most of thesse migrations seem to have been fairly painless. However, today at 07:00 this morning my website vanished from the internet (which was fine and totally expected).

By 09:00, though, when I entered my site's URL (www.celtnet.org.uk) int a browser to check if it was back up I was instantly re-directed to another site entirely. My heart began to beat faster as I feared the worst — had my site been hijacked? Half an hour later and my site's URL seemed to be pointing to another site entirely. Obviously something was completely amiss.

Immediately I dashed-off an email to the people at LCN (my hosting company) enquiring as to what was going on and why my site seemed to be pointing at other people. About an hour after my email was sent, at 10:30 they updated their information page to day that, unlike the previous day's IP address moves today's migration was progressing much slower and it might take 24 hours for the problems to resolve!

Soon after I received a response to my email, with an apology and a link to the updated information page (which I'd already seen). Not very helpful, admittedly — but probably fair enough. By now, in addition, a request for www.celtnet.org.uk was pointing at the name of the machine whose IP address was being moved: oldraq8.lcn.com which simply displayed a 'Server not Found' message. Well, at least my URL wasn't directing people to someone else's site any more. Of course, by now, the next consquence of all this was striking me. Fore each hour that my site remained off-line I was actively losing money!

Indeed, Wednesday is one of my stronger days and on a typical Wednesday I can expect to take-in more than the ~$220 that my hosting plan with LCN costs for an entire year. Essentially in this one day's outage (assuming that it's only going to be a day) I'll lose enough money to have otherwise paid for a year's hosting. Then there's also the general loss of credibility that the site's unavailability causes. Especially as the URL was directing people to a random website and a 'Server not Found' error.

Basically today's looking to me like being a complete disaster. It's now 14:40 — over seven hours since the site first went down and I've taken pen and paper to go and write this missive. I've also realized that though I'll be able to write this up in my Blog no-one will be able to read it as the blog's hosted on my main site — which is down.

It's now 16:20 and my URL's finally pointing at my side, though the site iself is still not available and I'm getting a 'connection timed out' message. It's an improvement, but still not as good as actually having my site available. Indeed, it's 22:30 now and still no sign of the site being back up. I'll set-up a system to ping the site so tiat if it comes back sometime tonight I'll know when. It now seems likely that this is going to impact Google's crawl tonight — a really bad thing after all the hard work I've been doing in terms of getting my site into Google's index.

Finally! It's 10:15 the following day and my site's back up. I can now upload this blog and get on with some work on the site.

Please don't read this as me getting at LCN. Whilst I've been them the've been a great Hosting Company. The tale is here to show just how dependent websites are on other people and how something very small can have a severe effect on a site-owner's bottom line.

Tuesday, July 11, 2006

Has Google Lost the Plot?


Part 2



In yesterday's first part of this article you learnt a little about Google's business model and how the business analysts view the company. Today's article will look at Google from the viewpoint of website owners and internet marketers.

Google (as well as all the other search engines) are in a very peculiar position as the're delivering value on top of other peoples' content. The value being the delivery of that content to the user in a way that's useful to the user (which usually means the ranking of the results returned). To remain in top position the search engines must convince their users that they're delivering content that is 'the best of the web'.

Admittedly, Google comes for a fair amount of stick in articles such as this one; mostly this is because Google is the dominant player in the web indexing business and because of this a small change in the way that Google indexes or caches websites can make a web-based company succeed or fail. Admittedly, Google, along with all the other search engines, have a problem. The web is expanding month on month and though some of the new sites are providing good and useful content many of the others are spam sites, whose sole function is to promote other peoples' products. Others are 'scraper' sites, stealing content and presenting it as their own. Yet others are 'clone' sites. Sites that have Amazon or eBay marketing areas that are duplicates of other sites and use the same code and/or pages as them. There are also those websites that take pre-existing articles from article websites and publish these as new content.

Including multiple copies of such websites would both degrade the results of any search engine and would simply swamp any good information-rich website in a sea of junk. Over the past eight months Google have been busy tweaking their algorithms and delivering a system called Bigdaddy. What few messages that have been coming out of Google have been very 'up' about this, but the news from webmasters and a number of recent disasters indicate that Google may well have dropped the ball over these updates.

Part of the problem with all this is the difficulty in actually getting any information out of Google. All their systems are automated and there are pages where you can send questions to Google, but all you ever get back from these is a standard automated response. Google can also black-list websites because of content or practices that Google deem to be bad. But they don't tell you that this has happened. The only way to find out if you have one of these Google penalties is to use a form to tell Google that you think you've had a penalty, that you've corrected the problem and that you'll never, ever, do it again. There's an implicit assumption of guilt and that all webmasters are somehow 'bad'. The only other indication of a Google penalty is that the number of pages indexed in a site (as well as the site's overall rankings) drop dramatically.

Google also change their algorithms and indexing practices at will without making much of a general announcement about it. And it's this implicit environment of secrecy that annoys many webmasters. Especially as any updates to Google's indexing strategies may make a site fall in terms of rankings or may dramatically change the number of pages from a site that are indexed. A change in Google's algorithms may have the same effect as gaining a Google penalty, which obviously dramatically increases the stress levels for may web authors.

Despite their slogan of ''Don't be evil' Google come across as being a cold and uncaring company who have an almost religious belief in their indexing algorithm. Tweaking this algorithm is what they do and it sometimes appears to outsiders as if Google firmly believes that their algorithm can never be wrong.

I can understand why Google have done this. If they didn't then they'd be inundated by umpteen million e-mails every day. However, because of this they've painted themselves as being cold and aloof: uncaring of the needs of the very people who create the content from which they make their money.

The current upheaval started about seven months ago when Google began to roll-out their new Bigdaddy datacentres. These were, according to Matt Cutts 'the human voice of Google' were new data centre infrastructures rather than a major algorithm change in indexing. Bigdaddy is, in effect, a brand-new data center that Google uses to perform core search engine tasks like cataloging Web sites or serving up localized features. Google operates thousands of data centers across the globe.

However, as the Bigdaddy datacentres were rolled out during January and February many website designers and authors reported very dramatic reductions in the number of pages from their sites that were indexed. What seemed to have happened was that as Google rolled out some new technology to fix a few things this resulted in their breaking other things, But it wasn't until May 10th that Google's co-founder, Larry Page, addressed this issue saying the problems took the company by surprise and it was now investigating what is going on. "We have a team studying it now," he told an audience of journalists at Google's annual Press Day event. It looked as if in upgrading to Bigdaddy Google was erasing its existing data caches and replacing them completely with new indexes. During this process many web pages that were previously indexed simply disappeared from Google.

At about the same time Google CEO Eric Schmodt told the New York Times about a 'server crisis' where the machines holding their indexes were completely clogged.

Another problem was that real web pages were being replaced with 'supplemental results' often from six months or so previously. These supplemental results are pretty useless in terms of searching though they do artificially inflate the number of pages indexed from a given site.

Then on May 16th Matt Cutts offered a very long post on the timeline of Google's new indexing and Bigdaddy roll-out. This said that many of the supplemental results would be refreshed. Then came the scary stuff: If a site didn't have enough incoming links then it might be considered as having 'low confidence' and wouldn't get crawled as deeply. Even worse, depending on the number of links to your site Google may be crawling your site every day to see if there's interesting content but none of that content may be indexed (this is happening to me now). So Google is basically stealing my bandwidth without giving me anything for it.

Maybe Google doesn't want the good content of the web indexed. After all, if you're not getting any visitors for the search engines you're going to have to resort to advertising and one of the big sources of advertising out there is Google's AdWords program. The truth is that Google is no longer reflecting the internet as it is, but they're reflecting the Internet as Google thinks it should be. Moreover, other search engines are catching-up with Google and are indexing more of the web than Google are.

In their recent update it seems as if Google re only really counting the number of uni-directional (ie non-reciprocal) links coming into your site from related sites as a measure of a site's worth. However, if you do a search for your site's URL on Google and then perform a links search with the links:www.mysite.com you'll see that the links: search only returns a third of the total number of links coming into your site. So Google may well be under-estimating the number of sites linking to you and so your ranking and the number of pages indexed in your site will suffer as a result. Also if you're in a specialized niche then you may not have many inbound links, but those you do come from natural links built in your area of interest. However, as you have few incoming links most of your pages won't be indexed (as will be the case for many of the sites that link to you) so even if you have good links the non-indexing of these linking pages mean that both you and the sites linking to you aren't given the credit that you should.

The upshot of this is that a perfectly good, clean, website cannot now get all or its pages indexed if it doesn't have a sufficiently good score in terms of both in-bound and out-bound links. Too few inbound links and the site won't be fully indexed. As a result Google is effectively editorializing a site's content because Google doesn't think it's 'good enough'. In reality Google's users are being deprived of many good resources because of this. Users are only being exposed to a portion of an internet — a portion that Google defines for them. Does this really mean that Google is short of indexing space and that they can only keep going by 'pruning' the web? It's certainly looking more and more as if this is the case.

Added to this Google have had other debacles of late. One of the most notable were the reports in a number of SEO fora that Google had suddenly and very mysteriously added a few billion pages to their index. This might be a cause for celebration: had Google finally relented and indexed the pages that they'd dropped or excluded from their index? Analysis of my own site indicated no for Google still only indexed 85 pages (MSN was indexing 1100+) [as an aside, since I wrote this Google have added a further 40 pages to my site's index]. Rather, all those new pages were blatant spam and contained only pay-per-click (PPC) advertising as well as scraped conttent. Even worse, many of these pages were displaying well in the search indexes. A Google representative responded via forums to the issue by calling it a "bad data push," something that met with various groans throughout the SEO community.

What seems to have happened is that a Moldovan had the enterprising idea of using the way that Google handles subdomains (in reality these are sub-components of a website but Google handles them as if they were totally independent sites). This allowed he enterprising individual to craft scripts that would generate an endless number of subdomains as soon as the googlebot dropped by his site. As Google indexed these subdomians on an essentially 'no questions asked' basis he was able to deliver single pages containing keyword-rich scraped content, keyworded links, and PPC ads for those keywords. Spambots were then sent out to put GoogleBot on the scent via referral and comment spam to tens of thousands of blogs around the world. The spambots provided the broad setup, and it didn't take much to get the dominos to fall.

As a result this individual found himself with 5+ million indexed pages on Google. Word of this spread like wildfire in the SEO communities, though the general public seems essentially ignorant of Google's latest snafu. Google responded saying that only 'millions' of pages had been indexed and that they were working on tweaking their algorithms to remove the possibility of this happening again. Though close examination strongly indicates that the domains were actually being removed manually from Google's index.

What stuck in many webmasters' craw is not the fact that Google were caught out, but the fact that this scamster managed to get millions of pages into Google whilst Google's managing to ignore many thousands of websites with perfectly good content and is not including them in its index.

Then came the story at the end of June that Google had dropped 'amazon.com' from its entire search index (and its not the only big player to go missing from Google). If Google can't do the big players, then what hope have we, the minnows in the Internet pool?

Will I be sticking with Google? The answer is that for the moment I will. Google is generating some of my best income due to the AdSense program. However, bizarrely, most of the traffic that creates my AdSense income comes from non-Google search engines!

If you want to see the entire article then you can find the full article here

Monday, July 10, 2006

Has Google Lost the Plot?


Part 1



As a web developer and webmaster of the www.celtnet.org.uk website I have a vested interest in the way that search engines work and operate. A recent article in the Economist Newspaper indicates that the market share of the various search engines is as follows:

Google: 50%
Yahoo!: 28%
MSN: 13%
Ask: 6%

Google has more market penetration that this, however, as it tends to be used by more 'web savvy' users who are far more likely to buy products on-line. Indeed, the popularity of Google grew out of this community because Google at the time was seen as 'best in classs'. The results returned by this search engine were better than those of its rivals both in terms of the accuracy of the results delivered and the total number of sites and pages indexed. With its search algorithm and its page ranking facility Google delivered both the most and the best results. Even after its stock market floatation Google was basking in the rosy glow of how it was perceived by its user base. But does this view of Google still hold true?

Since its stockmarket debut Google (in common with many other large companies cf Microsoft) has been adding new and often quite different products to 'supplement' its core search engine business. Some have criticized this as an example of a large company straying from its core business. However, Brin and Page, Google's founders, have calculated that Google's engineers should spend apout 70% of their time on core products, 20% on relevant but tangential products and 10% of their time on the wildly fun things that might (but probably won't) lead to a product.

As a result a large number of tiny teams within Google have been working on all kinds of different products. However, many webmasters see this as Google not caring about their core business any more. In fact this is a view that many in the investment sphere now share and Google's share price has been rather volatile of late. Indeed, the view is held that at floatation Google was a fairly simple story: in effect they were a marketing company deploying online ads on top of the most popular search engine on the internet. Now, however, Google looks rather like a mess and comparisons are being made with Microsoft during the mid 1990s where they had a single core strength (Windows OS) and were churning-out a string of rather poor me-to applications.

For example Google Video has been trumped by YouTube and Google News lags far behind Yahoo! News. Google's instant messaging offering lags far behind AOL's Yahoo!'s and MSN's offerings. Even Google's slogan of 'Don't be evil' is looking faintly ridiculous as they use their status as a 500lb gorilla to stifle the growth of web-based developers and entrepreneurs.

Over the past few years Google has almost certainly offered the best search engine experience on the internet. This is significantly due to Google's search and indexing algorithm. The problem now is that Google actively regard themselves as the best and this has coloured their attitudes at all levels and represents one of the reason that they don't seem to have a clear strategy for the future and are currently dabbling aimlessly at a variety of non-core projects. This over-confidence has also led Google to have what might be viewed as an almost-religious faith in their mathematical algorithms which means that there is almost no human intervention as to what gets indexed and why. For Google the important thing to know about a website is how many incoming links are pointing to it and a website's overall rank (its Page Rank) is directly linked to the number of incoming links. This is great for old websites on popular topic, but if you're a new website in a specialist niche your indexing and rankings will suffer as a result.

This is why Ask (formerly Ask Jeeves) is employing a different track. They also use incoming links to rank web pages but they also cluster sites and pages by theme. Thus rank isn't absolute as it's measured with respect to a site's or a page's comparison to its peers ߞ an approach that often returns better results than Google's. A cursory comparison certainly suggests that Ask's general web search is at least as good as Google's and it's searches for maps and images are currently superior. As a result Ask is gaining a small but dedicated following.

The potentially really big player in this arena, however, is Microsoft who are weighing-in with their MSN search engine. This is not as good for searching as Google, but has indexed more of the web (for example Google has indexed 86 pages from my site but MSN has 1100!). As Microsoft tweaks MSN's search abilities this product may well offer a very strong challenge to Google. After all, Microsoft has a tradition of entering a marketplace rather late with a fairly mediocre 'me too' product that gradually improves until it prevails. Moreover, MSN will be including an on-screen search box in its next operating system Vista that will be pre-configured to search MSN as a default.

The search engine wars are definitely not over and in tomorrow's article I'll be looking at Google's relationship (or rather lack of it) with their main content providers, the webmasters. Return here tomorrow for the second part of this article.

Sunday, July 09, 2006

The Importance of Website Search



The paradigm of the search engine is the glue by which the modern internet works. If you need to find any information about a website then you will go to a search engine (probably Google, judging by the search statistics) and you will enter your search terms and find a list of sites that you might be interested in. This is the kind of functionality that visitors expect to find whether they are performing a general web search or whether they are looking for something specific on an individual website.

If you're serious about your website then you will need to provide some kind of search functionality for your site. This both provides your visitors with a sense of professionalism about your site. It also gives them a way to find things that are present on your site but which may not be immediately obvious. You are also providing your visitors with a means of navigating your site that's completely and entirely familiar to them. After all it's how the web works in general.

Google provides this kind of search functionality which you can use on your site via their Adsense program. This is free to join and you can use the Google Adsense button on the left-hand panel of this Blog to join. However, there is one major disadvantage to using Google as your site's search mechanism. Though the system is free (and you don't have to store potentially large indexes on your site) it does rely on Google having indexed your entire site. If you're a new site or you have very few back-links to your site then it's becoming increasingly difficult to enter google's indexes and very few of your pages may be indexed by Google. As a result Google's search system may be of little or no use to you.

Personally I've eschewed Google's site index search system for the very reasons detailed above (though I do include Google web searching tools on my site). Instead I use the Fluid Dynamics Search Engine to power searches on my site. This is a Perl-based system that you can either install for free (or you can pay $40 for a license that gives you slightly more functionality).

This is an impressive system which you can easily download and install (as long as you can execute CGI scripts on your server). It allows you to define your search results page as you want as well as allowing you to decide which pages you want to index and which ones you want to omit. This way you can index anything you want from your site (or your can omit 90% of your site from the index if you want). Moreover, once you've installed the scripts on your site you can use a standard web browser to call the script and execute it, making running the indexing of your site very easy and flexible. The system is very quick as well, both in terms of indexing your site and in terms of the searches you can perform. Even better, you can also specify ads that you want to place on your search results.

The basic search functionality is provided by a search box of the type that you can find on my home page. However, if you want a very simple search box that you can place anywhere then all you need to do is modify the form below to use on your own site:

<form method="get" action="http://my_site_url/cgi-bin/search/search.pl" style="margin:1px;">
<input type="hidden" name="Match" value="1" />
<input type="hidden" name="Realm" value="All" />
<input type="text" name="Terms" value="search site" onfocus="clearform(this);" style="width:100px;" />
<br style="line-height:2px;" />
<input type="submit" value="Site Search" />
</form>

As the Perl script is usually in the cgi-bin directory simply change my_site_url to the URL of your own website. (Depending on your configuration you may have to name search.pl to search.cgi. IF you want to see how this simple site search box looks and behaves then have a look at my recipe site search function.

Adding search functionality to your site is something easy to do and you can do it with a tool that's completely free and which is under your own control. Don't neglect this important functionality as it's an essential part of making your site both look and feel professional.