Category Archives: search

A smile at Google Developer Day London

I’m at Wembley Stadium for Google Developer Day. No, Google is not filling the entire stadium (yet); but it is a decent venue. On arrival we were given the usual USB stick though this one is distinctive:

Smiley Google USB guy

Made me smile after reading nothing but gloomy economic news on the journey over.

I’m not expecting news announcements today; just a bit more detail on initiatives like App Engine and Gears. The agenda was put together before the Chrome announcement; but I’m hoping Google’s new browser will make an appearance as well.

If you are here and would like to chat please come and say hello – tell me what you think of Google’s developer activities.

Will anyone untangle the nofollow mess?

I get frequent emails asking to place advertising on itwriting.com. My reply is that I’m happy to take appropriate ads, provided that they are clearly marked as such, non-intrusive, and that links are either inserted by script or tagged “nofollow” in accordance with Google’s guidelines. Usually I don’t hear back from the advertiser.

The problem is that these advertisers are not primarily interested in having readers of this site click their ad or see their brand. Their real concern is search engine optimization – improving the rank of a web site in Google searches by adding to the number of incoming links. Counting incoming links to determine PageRank was Google’s original secret sauce, by which it determines which pages are of the highest quality, though it is probably less important now precisely because the principle is so well known. Advertising like this is often called paid links.

Although it upsets Google, there is nothing illicit about accepting paid links. Advertising is the commercial engine that drives much of the web. It is risky for the host site, since Google might downgrade its ranking because of the paid links. In other words, the advertiser is proposing a trade – buying more PageRank for its own sites, at the expense of some of yours.

The other risk, which is the reality that Google’s algorithms are attempting to simulate, is that a site filled with spammy links is less attractive to readers and therefore likely to become less popular. After a while the advertising will dry up too.

Balancing the demands of advertisers and editorial is nothing new. My main complaint about these advertising offers is that they are not transparent about their intentions.

Even so, the whole SEO/Paid links/Google dance is a horrible mess. One of the problems is that the Web has de-facto authority. People search the web for information, and believe the top-ranking results as the answer to their question. Another real-world issue is that countless businesses now depend on Google for their success and survival. I often hear from organizations which are desperate to improve their search ranking and vulnerable to SEO consultants whose services are often not good value.

All this makes the nofollow idea good, in that it is tackling the right problem; but unfortunately it is not working. We are seeing the nofollow tag frequently applied to links that are “good”, and frequently not applied to links that are “bad”. For example, links in comments to a WordPress blog are automatically tagged nofollow; yet some of the best-informed advice on the web is found in comments to WordPress blogs. Links in Wikipedia are tagged nofollow; yet an external link that survives in an important Wikipedia article is a high recommendation of quality.

At the same time, both the spammers and the SEO folk are having few problems placing paid links sans nofollow all over the web.

It would not surprise me if the secret search ranking algorithms used by Google and others actually accord more significance to links tagged nofollow than is acknowledged. It would actually make sense for the search engines to try and distinguish between good and bad links regardless of nofollow. In other words, the nofollow tag has become almost pointless. That said, I have no idea whether, if Google declared the tag obsolete, the net spamminess of the web would improve or deteriorate.

I don’t know how to fix the problem; though I suspect that Google as the Web’s current gatekeeper is the entity with most power to improve it. Another thing that would help: more focus on content and less focus on SEO.

Technorati tags: , , , , ,

The trouble with Knol

Is that that it’s going to be full of rubbish. Wikipedia, which arguably has less authority because contributions can be anonymous, will likely have more authority, since it is more-or-less restricted by its community to one entry per topic.

Another way of looking at this is that on Wikipedia, if you want to contribute to a topic that already has an article, you have no choice but to (try and) improve the existing one. On Google Knol, there’s every incentive to start a new one, never mind if it duplicates existing content, or is worse than an existing one.

Take Search Engine Optimization, for example. Wikipedia has an article that looks decent. Knol has thirty articles; or if you search for SEO, more like seventy. And that’s after just a few days.

Google is good at ranking, and users can rate pages, but Knol is still a mess. You can be sure that many articles will be written primarily to drive traffic to the author’s web site, or to attract Adsense clicks. Wikipedia is not immune to spam; but at least contributors can delete it. All you can do on Knol is to give a spammy article a low rating.

Thinking aloud, what might work is some kind of Slashdot-style filtering. For example, you could have it so that by default Knol searches only show articles which have:

  • More than 10 ratings
  • An average rating of at least 4
  • or are the only article on the subject

or some such; vary the constants according to taste.

Then again, you could have a team of editors (and become Britannica); or enforce one article per topic and become more like Wikipedia.

Technorati tags: , , , ,

Web stats: do you believe Google, or the web site owner?

Escherman’s Andrew Smith, in technology PR, asks whose site traffic figures do you trust – Google’s (via Ad Planner), or the site owner?

I don’t have Ad Planner, but because I run AdSense I can see Google’s stats on Adsense views on this site. I also have web logs, analyzed via awstats.

I took a look at my own figures for June. My stats show about 6.5 times more page views than AdSense reports.

This isn’t hits vs pages, incidentally. “Hits” record every request, so a page with several images requires several hits. Hits is therefore always the biggest number, but pages is in theory more meaningful.

It is a huge discrepancy. What’s the reason? I can think of several:

  • Google only counts page views that run its AdSense script. Bots like web crawlers are not likely to run these.
  • Not all my pages have AdSense on them, though most do.
  • Every time a request is made for my RSS feed, awstats will count that as a page view, but Google (rightly) will not.
  • Google will try to eliminate the rubbish, like spam bots posting comments that end up in the Akismet junk box.

Still, 6.5 times is a huge difference, more than I would expect. The page view discrepancy on the site Smith chose to look at is a mere 4.2 times – though we don’t know how that particular web site calculates its figures.

I don’t have any firm conclusions, though my own figures suggest that any web site which simply quotes figures from its logs will come up with something much larger than Google’s filtered stats.

I’d have thought the answer for advertisers would be to use tracking images and the like in ads so they can get their own statistics.

Finally, this prompts another question. Just how much Web traffic is bot-driven? We know that somewhere between 65% up to, by some estimates, 90%+ of email is spam. Web crawlers and RSS feeds are not bad things, but they are not human visitors either. Add that to the spam bots, and what proportion does it form?

Technorati tags: , , ,

Why I can’t use Microsoft Live Search for real work

I’ve complained before about the quality of Microsoft’s Live Search vs Google; but today’s example seemed too good an illustration not to mention.

I needed to update Windows XP to SP3. In particular, I wanted what Microsoft calls the “network” download; that is, the entire service pack, not the launcher app that initiates a piecemeal download tailored to the specific machine.

I pulled up Live Search and entered “windows xp sp3 download”.

To my surprise, Live Search offered me only third party download sites in its first page of results. Actually, that’s not strictly true. At number 8 is the download for “Windows XP SP3 RC2 Refresh” (obsolete); and at number 10 the general home page for XP downloads:

Find popular Windows XP downloads, including PowerToys, trial software, tools and utilities

I tried Google. Same search. The official network download is in first place. The official piecemeal download is second.

I know: you can argue that this is just an isolated search, and that some other search might show Google in an equally bad light. However, I find this constantly: Google gets me much better results. Further, this case is particularly telling, since a third-party download site is not what you want when patching Windows. Quite likely those other sites do point you to the correct official download eventually; but getting Microsoft downloads from Microsoft’s site is safer.

I am not surprised Microsoft has a tiny share of the search market; and I don’t believe this is simply because of Google’s clever marketing.

Update PS: The above screen grab still matches what I get today. However, users in different countries may get different results; from the comments below I suspect that US users get better results in this instance. Maybe Live Search is worse in the UK than in the US; I’d be interested to know.

Technorati tags: , ,

Web 2.0 for the rest of us?

We all know what Web 2.0 means. Google, Flickr, Facebook, Yahoo, mash-ups usually with Google Maps or Flickr, Salesforce.com, and anything but Microsoft. But what does it mean for everyday businesses, like some of the small businesses I talk to from time to time? Some are sceptical. One I can think of sells a successful software application but does not even run a support forum – why make it easy for others to discuss and publicise flaws and problems in your product?

I was interested therefore in a recent book by Amy Shuen, called Web 2.0: A strategy Guide. A foreword by Tim O’Reilly says, "it is the first book that really does justice to my ideas". It was O’Reilly who popularized the Web 2.0 concept – and yes, it is anO’Reilly book.

Shuen writes enthusiastically about network effects, using Flickr, Netflix, Google, Facebook, LinkedIn, Amazon and Apple (iPod/iTunes/iPhone) as case studies. I enjoyed it, but the problem with this kind of book is the chasm between these few web giants and everyone else. Another problem is the tendency to ignore the Web 2.0 graveyard – thousands of start-ups that fail, or moribund and/or spam-infested blogs and forums. Since there are more failures than successes, it would be sobering to investigate these rather than riding a wave of Web 2.0 hype. Nevertheless, it is a thought-provoking book with an extensive bibliography, and not a bad starting point for investigating Web 2.0 concepts. I liked the “five steps to Web 2.0”, which begin with finding collective value and end with perhaps the most important, which is what Shuen calls “recombining innovations”:

New-style click-and-mortar, online-offline network partnerships focus on bridging and building new networks rather than replacing or disrupting the infrastructures of offline companies.

I’ve also received a short Web 2.0 book by Marco Cantù, called The Social Web. It is a brisk tour of the sites and concepts that form today’s online communities. Typical readers of this blog probably won’t find anything new here; but I liked the common-sense tips on things like blogging and creating interactive web sites.

I would argue that almost all businesses either are, or should be, “click-and-mortar” entities. Whatever business you are in, a useful question is: what proportion of purchases in your sector begin with or include a Google search? If the answer is significant, you are in the Web 2.0 business.

That does not mean SEO (Search Engine Optimization) is the answer to everything. I am an SEO sceptic. All too often, SEO is lipstick on a pig. Optimise your web site for users, not robots. Further, it is no good trying to get users to interact with you, if you are not willing to interact with them. Surprisingly, I see this all the time. I suggest spending less time worrying about high Google ranking, and more time worrying about what users find when they do land on your site.

The case studies that interest me most are where old-style businesses have found ways to engage successfully with Web 2.0 innovations. For example, I’ve written about kitchons.com, which services domestic appliances and tunes its business via Google ads. I came across another example today: a financial company which lets you put an image from Flickr on your credit card. Clever.

Web 2.0: A Strategy Guide by Amy Shuen (ISBN: 978-0-596-52996-3). O’Reilly $24.99.

The Social Web written and published by Marco Cantù. $17.39 print or $8.70 electronic, from Lulu.

Technorati tags: ,

Why you can’t trust a Google ad

An interesting facet of the recent problems with UK non-supplier Zavvi Direct is that all the purchasers I spoke to found the fake web site via a Google ad. Put another way, without the ease of advertising through Google and eBay, it is likely that far fewer people would have found the site and potentially lost their money.

That raises the question: does Google do anything to verify that its advertisers are genuine? Here’s the answer, from a Google spokesperson:

Google, along with other online and offline advertising platforms are not able to proactively check the legitimacy of each and every advertiser. Consumers should always check the validity of what is being sold to them and how they are asked to pay for items. If Google is alerted to a potential fraud then we will work with the relevant legal authorities to help them resolve such matters.

This was clarified to me as follows. Google will assume ads are OK unless it receives complaints. If it receives a few complaints it might pass them on to the merchant. If it receives numerous complaints it might warn the advertiser and eventually disable the account.

I guess it is unreasonable to expect Google to conduct checks on every advertiser. Still, there is a related point: does Google do enough to highlight the difference between advertisements, and links identified by its famous search ranking algorithms? Here is a snapshot of a search I just made:

I’ve sized the browser small to get everything in; there are more search results than I’ve shown. However, it shows three panels of results. The top left is tinted and marked in unobtrusive gray type “Sponsored links”. The top right is narrow, not tinted, and also marked in gray type “Sponsored links”. The bottom left is what most tech-savvy folk think of as the main results area.

Judging by my interviews, some people are not really aware of the distinction between a “sponsored link” and a search result. In some cases, the buyer could not tell me what kind of link they clicked. To them it was just “Google”.

It would be easy to make the ads more distinct. Google could use the plain English “Advertisements” rather than the “sponsored links” circumlocution. It could use something bolder than gray text to identify them. It could use a different font and colour for the links in the right-hand column. It is good that the top left links are in a tinted panel; yet some may perceive this simply as best-match links, rather than links in an entirely different category than those that follow.

Overall, it seems to me that Google deliberately makes its ads look the same as search results. Which is good for advertisers, but can be bad news for buyers.

Technorati tags: , ,

Now it’s YahGoog

Yahoo has signed up for Adsense:

By offering Google’s industry-leading technology to Yahoo!, the whole system becomes more efficient, and everyone benefits.

This is efficient in the same way that having everyone run Windows is efficient. Hmmm.

Google observes that the deal is non-exclusive; Yahoo can still sell its own ads, etc etc. I tend to agree with Om Malik, who says:

In my opinion, with this deal, Yahoo has publicly acknowledged that Google is superior to them when it comes to search & contextual advertising.

Yes. But how much does that matter? Outsourcing what you are less good at, in order to concentrate on core competencies, can be a smart business move.

The snag here: advertising is Yahoo’s primary business activity. Here are its revenue figures for the first quarter 2008:

  • Marketing services: $1,818 million
  • Fees: $245,milion

Outsourcing the core of your business is bad PR.

Technorati tags: , ,

Microsoft: forget the Live Search Cashback, just improve the engine

Microsoft is paying users to use its search engine with a new search cashback scheme. Looks like an affiliate scheme where the commission is paid back to the customer. US only.

I think Microsoft should focus on improving its search engine. This morning, I needed to call a local electrician and figured that search would be quicker than using a phone book. I entered the name of the retailer and the town. For some reason, this stymied Live Search: the result I was looking for was not on the first 10 pages. Identical search on Google: the first four results matched, and the address and telephone number were at the top of the page with a little map.

In a poll last year 51% thought Google delivered the best results for an example search, while 35% preferred Live Search and 31% Yahoo. That’s an inconclusive result, and this is not an exact science; but personally I find Google almost always delivers better results, sometimes (as in the case this morning) dramatically so.

If Microsoft managed to reverse this I would switch to Live Search in a heartbeat.

Technorati tags: , , , ,