Tag Archives: Google

Are you new to SEO? Do you have a basic understanding of how web pages are built, and want to tweak yours for better rankings? I’ve wanted to collect a few authoritative, practical, and non-technical resources on SEO for my clients for some time. The good news is that it’s easier than you think to build a page that does a number of things that the search engines like. The guides below should help explain what to look for when evaluating whether a site has been developed with an eye for SEO.

Google has created a compact Search Engine Optimization Starter Guide of best practices that developers can follow to improve their sites’ crawlability and indexing. This guide is available as a PDF.

The SEOmoz Beginner’s Guide to SEO provides comprehensive information you need to get on the road to professional quality SEO. Read it online or download it as a PDF.

SEOmoz’s CEO and co-founder wrote a very nice article on building the perfectly optimized page that explains the most important on-page factors and how they affect rankings in today’s environment.

Sometimes, a client will ask for SEO as though it’s something to be tacked on after the site has been built. Happily, good development practices benefit human visitors as well as SEO, and they go into every site I build, as I build it, from the very first line of code. I’ve written up a list of on-page factors and other considerations within a developer’s direct control that are important not only for SEO, but for good web development in general, at Aleph Studios | SEO Consulting.

I’d like to note that, as of today, Google Webmaster Tools is reporting over one million inbound links to pages on ardamis.com.

one million inbound links

One million inbound links!

I’ve been spending quite a bit of time on ardamis.com lately, giving it a new look, working at improving the site’s navigation, cultivating some inbound links, and posting more regularly. It’s rewarding to see that the effort is paying off.

Over the last few days, I’ve been concentrating on reducing page load times by sending the proper headers and compressing files.

I’ll give it some time and see how performance improves.

According to Google’s Webmaster Tools’ performance overview, with Super Cache running, a single minified CSS, a single minified JavaScript, etc. but no compression or header tweaks:
On average, pages in your site take 2.8 seconds to load (updated on Jun 28, 2010). This is faster than 53% of sites.

The chart illustrating page load times is pretty much all over the place, but at no time has the site dipped into the 20th percentile, indicating a ‘fast’ site. I’m trying to change that.

I was tired of seeing the majority of my posts’ comments feeds show up in Google’s Supplemental Index, so I changed all the individual posts’ comments RSS links to rel=”nofollow”. This should at least cause Googlebot to stop passing PageRank through those links, but what I really want is for Googlebot to stop spidering the individual posts’ comment feeds, in hopes that they’ll eventually be removed from the index. To see only those pages of a site that are in the Supplemental Index, use this neat little search feature: site:DOMAIN.com *** -view. For example, to see which pages of Ardamis.com are in the SI, I’d search for: site:ardamis.com *** -view. This is much easier than the old way of scanning all of the indexed pages and picking them out by hand.

To change all the individual posts’ comments feed links to rel=”nofollow”, open ‘wp-includesfeed-functions.php’ and add rel=”nofollow” to line 84 (in WordPress version 2.0.6), as so:

echo "<a href="$url" rel="nofollow">$link_text</a>";

One could use the robots.txt file to disallow Googlebot from all /feed/ directories, but this would also restrict it from the general site’s feed and the all-inclusive /comments/feed/, and I’d like the both of these feeds to continue to be spidered. Another, minor consequence of using robots.txt to restrict Googlebot is that Google Sitemaps will warn you of “URLs restricted by robots.txt”.

To deny all spiders from any /feed/ directory, add the following to your robots.txt file:

User-agent:*
Disallow: /feed/

To deny just Googlebot from any /feed/ directory, use:

User-agent: Googlebot
Disallow: /feed/

For whatever reason, the whole-site comments feed at //ardamis.com/comments/feed/ does not appear among my indexed pages, while the nearly empty individual post feeds are indexed. Also, the general site feed at //ardamis.com/feed/ is in the Supplemental Index. It’s a mystery to me why.