It’s a Ultimate SEO Guide
for those who want to make their website search engine friendly.
Picking an SEO Friendly
CMS
When looking to build a website, determining which CMS (content
management system) you are going to use is a critical first step. Sure, you
could have someone code your site from scratch, but why waste the effort when
you don’t have to (not to mention that’s probably a terrible idea, for lots of
reasons)?
There are literally hundreds of CMS options, and it can be hard
to determine which one is the best. Unfortunately, not all content management
systems are created equal, and some can actually do more harm than good (as an
SEO rule of thumb, avoid anything .NET/ASP, and anything made by Adobe).
As an SEO company, one of the problems we see most often is
clients who’ve built their websites on a CMS that does not support SEO best
practices, thus making it difficult (if not impossible) to rank for their
target keywords. Without further ado, here is a list of features you should
insist on in an SEO friendly CMS:
11 Must Have Features
1. Static, customizable URLs – You should
be able to define both the page name and the directory structure for the entire
website, page-by-page (not database driven, unless it can be manually
overridden if necessary). Keyword rich and search engine friendly URLs are an
SEO must have. If this feature is missing it should be a deal killer.
2. Support for custom URL
redirects (301s, 302s, etc). – At some point you will change
a page name, move it, or change its structure in your site hierarchy. In order
to keep the trust and value of any inbound links to that page and to avoid
creating a poor user experience, you MUST properly redirect it when you move it
using a 301 permanent redirect. Your CMS should allow you to do this one page
at a time if needed, or for blocks of pages. You should also be able to directly
edit your .htaccess file, if needed.
3. Customizable Title tags and
Meta tags on Every Page – The Title tag and Meta description tags are
important SEO elements, and should be unique for each page. They should always
be carefully written and never duplicated from page to page. Meta keywords tags have no value, and should be left blank. These
tags should not be database driven, but
written manually and thoughtfully for each page.
4. Custom Image File names and
Image ALT Tags for Each Image – Search engines look at image
file names and ALT tags for keyword usage to help define the topic of a page
and the site as a whole. You need to be able to choose the image file name, and
you should have the ability to define an ALT tag and if needed an image Title
tag.
5. Support for Rel=Canonical tags
and Meta Robots tags on Each Page – These elements help to
prevent duplicate content penalties, especially with eCommerce websites. You
should be able to add this tag on a page-by-page basis as needed.
6. The Ability to Directly Edit the
HTML on ANY Page – This is important for customizing content, link anchor
text, NoFollow tags, heading tags and other HTML elements.
7. Automatically Generated and
Updated Sitemap.xml File – Search engines use this file to find and index all
of the pages on your site. It can be a real pain to maintain manually, so the
automatic feature is extremely handy. If you have videos on your site, you
should also have a separate video sitemap.
8. Support for Blog Integration – Every
website needs a blog, since that is the best way to ensure ongoing content
growth (which search engines love).
9. Clean, Lean Coding – Your CMS
should make use of HTML best practices, using clean, current code (validates
100% with W3C), and that loads quickly. Valid, fast loading code is appealing
both to visitors and search engines. Avoid code filled with useless or
excessive JavaScript, unnecessary non-breaking spaces and antiquated tables
instead of clean CSS. If possible, find a CMS that supports HTML5/CSS3. For
images, use compression, and for layout elements consider using CSS sprites.
Make sure to set page caching quite a ways out.
10. Rich Snippet Support – Schema and RDFa,
for example. Authorship and Publisher markup in particular should be possible,
but other rich text markup such as reviews, products, apps, locations and
services are also a good idea if relevant.
11. Mobile Support – This could be
responsive design, where the site auto-adjusts to any screen size or device
type, or it could be a dedicated mobile version of a site. Whatever option you
go with, at this point in time not having a mobile friendly version of your
site is a major negative factor.
4 Very Nice to Have Features
1. Web 2.0 Integration – If you want
people to share what they read on Facebook or Twitter or any other social site,
you need the ability to integrate buttons to make that sharing easy. You might
also want things like open graph tags, Twitter cards, Facebook insights, etc.
2. Custom 404 Page – Hopefully
your site will never have any 404 errors, but if it does it’s nice to have
something show up other than “404 error – page not found”. Create a custom 404
page that lists your most popular pages and perhaps a search box. Keep them on
the site! Consider making your 404 page funny, like these:http://fab404.com/
3. Automatic Link Management – Whenever
you change the name of a page, any links to that page break. While implementing
a 301 redirect for the changed page will fix this, you would ideally want to go
through the site and change all links to reflect the new page name. Doing this
manually sucks, so find a CMS that supports automatic link updating.
4. User Generated Content – Make it possible
for visitors/customers to rate products or services, leave reviews, etc. For
some industries this is a must have, not a nice to have.
This isn’t a comprehensive list of features by any means, but
any CMS with these elements will be about as SEO friendly as it can get. Beyond
this list, look for additional features as dictated by your company needs.
Hosting
Believe it or not, server speed, server
up-time, server location, IP address and type of hosting are all elements that
can impact your rankings, some directly and some indirectly. Since Google wants
its users to have an awesome experience on the sites that Google delivers in
search results, Google tries to analyze sites from a UX (user experience)
perspective.
A site that loads slowly, that is down often, or that shares IP
addresses and/or hosting servers with less than savory sites (shared hosting)
can negatively impact that user experience. Beyond that, being on a shared
server can pose major security risks for your website, depending on how well
your hosting provider has things configured.
To avoid any such problems, it is recommended that you host your
website on a dedicated server that has the capacity to handle your site
features and traffic (with capacity to spare) to avoid any downtime. Having
secondary and tertiary servers as redundancies is also highly recommended
(cloud platforms can help with this). Though certainly not required, you may
also find some benefit in using a server located in the country that matches
your TLD (i.e. USA for .com, .net, UK for .co.uk, etc.).
It is also highly recommended that you have a dedicated C class
IP address for your website, and that you don’t share that IP address with ANY
other websites. This avoids any potential cross-linking devaluation issues, as
well as preventing bad neighborhood issues. A bad neighborhood occurs when a
number of sites on a shared IP address are identified by Google as spammy, thus
hindering the rank-ability of any other site on the same IP address.
Make sure to schedule regular backups
of your website and all databases at non-peak traffic times (2am is usually a
great time). Also, make sure to regularly test the page load speed on your
site. If you’re running WordPress, the best option available is probably WP Engine.
Page Load Speed
On April 9th, 2010 (almost 3 years ago, come on
already!), page load speed officially became a
part of the Google search ranking algorithm.
For usability reasons, best practices dictate that a web page should load
within 1-2 seconds on a typical DSL connection. However, according to Google a
load time of 1.4 seconds is the threshold between a fast page and a slow page.
That means, ideally, that every page on your website should load in 1.4 seconds
or less, to receive the maximum SEO and usability benefit for fast loading
pages.
Google gathers page load time data through actual user
experience data collected with the Google search toolbar, and may also be
combining that with data collected as Google crawls a website and from other
sources. As such, page load speed in terms of the ranking algorithm is being
measured using the total load time for a page, exactly as a user would
experience it, and not just user perceived load time (visual load time).
There are a few key things you can do
to increase that page load speed to reach Google’s recommended page load speed
of 1.4 seconds or less. Three things that would be most impactful would be
leveraging browser caching, CSS Sprites for images where possible, and reducing
the image file sizes as much as possible for images that can’t be sprited
(different file types, removing unnecessary color channels, etc.). You might
also see benefits by using a content delivery network
(CDN) for your images.
We would also recommend reducing the total number of CSS and
JavaScript files by combining them into fewer files, and minimizing the file
sizes by using compression and code minification where feasible.
W3 Total Cache is
an excellent WordPress plug-in that can help with page load speed issues, and a
simple CDN can be set-up via Amazon AWS for very little money. You can learn
how to do this here.
Code Validation
Because there are so many programming languages and so many ways
to accomplish any one thing using each language search engines rely on certain
rules in which they read the content of the website.
Having code that adheres to these rules removes and helps to
minimize errors when parsing or separating the code from the content of any one
page.
Search engines such as Google have openly stated that W3C
standards are what they suggest when making the code easy to understand for
them. We typically only test the home page of the website, because many issues
can be easily fixed a crossed the entire website using just its page templates.
Site Hierarchy and Navigation
Picture your website as a series of pyramids,
5-10 sitting side by side. It should look something like this:
The tip of each pyramid represents a top level page, such as
Home Page, About Us, Products, Services, Contact, etc. Each of these top level
pages should be optimized for a root keyword with significant search volume.
The middle and base of the pyramids represent sub-pages, pages under the
top-level pages that are topically related. These should be optimized for
variations of the top-level or root keyword.
For a website about, say, Bicycle Repair, you might optimize the
pages as follows:
·
Home Page – “Bicycle Repair”
·
About Us – “Bicycle Repair Shop”
·
Services – “Bicycle Repair Services”
The root keyword, the main topic of the entire site, would be
Bicycle Repair. As such, the home page is optimized for Bicycle Repair, and all
other main pages for variations of that keyword.
Then, you go one more level down. Since Services is optimized
for Bicycle Repair Services, pages under that page should be optimized for
things like Cheap Bicycle Repair Services, San Francisco Bicycle Repair
Services, Bicycle Repair Services in Utah, etc. The higher up a page is within
the site hierarchy, the more search volume the keyword assigned to that page
should have. The lower the page is in the hierarchy, the lower the search
volume of the keyword phrase.
This is called creating keyword silos, and it is a critical
component of SEO and user friendly site design. Not only does this make your
site topically clear to search engines, but it makes site navigation super
simple for users as well. For optimal search engine and user usability, don’t
go more than 3-4 levels deep (i.e. Top Level Page, Sub-Page, Sub-Sub-Page). A
user or search engine should be able to get to any page on your site in 4
clicks or less. This is called a flat architecture.
Your site’s navigational elements will follow this exact same
structure. Speaking of navigational elements, make sure that all site
navigation consists of HTML and CSS only. Don’t ever use JavaScript or Flash
for your navigation, as that can cause both search engine indexing problems and
user experience problems. The details are complicated, so just don’t.
Never underestimate the value of
long-tail traffic. More and more queries each year are long-tail, and by
anticipating questions people might type into search engines and creating
matching content, you can get far ahead of your competitors. You can also mine
data from your analytics organic keyword data, your Google and Bing Webmaster
tools search query data, and other sources like Google Trends, Google’s Keyword Tool, SpyFu, Quora, and Yahoo Answers.
URL Structure
A site’s URL structure is extremely important
to both users and search engines. Poor URL structure can not only hurt
rankings, but can prevent pages from being indexed and lower the
click-through-rate (CTR) in the SERPs (Search Engine Results Pages). Though
they have dialed back some on that value, Google is still biased towards exact
keyword match non-hyphenated domains, so get one if you can.
It is extremely important that URLs be readable, user friendly,
and that they contain the keyword of the page. It should never be longer than
256 characters (though ideally under 100 characters), and should contain no
query parameters, strange number sequences, spaces or symbols. If a number
sequence is needed for eCommerce reasons (like a product ID), append the number
to the end of the search engine friendly URL. (i.e.
http://www.awesomesite.com/killer-keyword-url-1234/)
If relevant, a geo-qualifier (such as Seattle WA) should also be
included in the URL. It should use a file ending that identifies the page type
(.html, .php, etc).
A proper URL will consist of
lowercase words separated by dashes/hyphens only (no
underscores, since Google combines words separated by underscores, and no
uppercase letters).
The URL structure should never go more than 3 sub-directories deep,
just as the site navigation should never go more than 3 directories deep. The
more important the page, the higher up in the directory structure that page
should be. The URL structure should function as a bread crumb, telling visitors
exactly where they are within the site.
An ideal URL structure is as follows:
http://www.keywordrichdomain.com/keyword-category/keyword-rich-phrase.html
This URL structure is written perfectly for both SEO and search
engine indexing. It is short (74 characters), descriptive, keyword rich, and
contains no query parameters. The use of a category can help with usability,
but it is not necessary for SEO purposes. In fact, it can be beneficial to have
all pages in the root directory.
Internal Linking and Anchor Text
It is very important to cross link within the
pages and content of one’s site using keyword rich anchor text and/or from
within keyword rich content that matches the linked-to page. Pages of similar
topic should cross link to each other using relevant keywords in the anchor
text or surrounding body text. There should ideally be 5-10 top-level pages on
your site that effectively cross-link to each other and that have more inbound
links than any other pages on your site. These pages will become your site links.
When linking internally, keep in mind
that all internal links should use absolute URLs (i.e.
http://www.domain.com/page-name.php), not relative URLs (i.e.
/page-name.php).
In addition to linking from within
the text of a page, keyword rich anchor text should be used in the main
navigation elements. Where space prevents the use of the keyword for the page
being linked to in the navigation, it is important to include the title element in the navigation anchor tag, as
follows:
<a
href=”http://www.domain.com/” title=”Home of the awesome
keyword”>Home</a>
The same goes for links outside of your site. When you get a
link on a blog, forum or press release, some of those links should include the
keyword of the page being linked to in the anchor text. At present, the ratio
of anchor text from inbound links that is considered safe is 15-30%.
Also, the keyword used in the anchor text going to a page should
be topically consistent. One should not use “Keyword Topic A” in the anchor
text for one link, and then “Keyword Topic B” in the anchor text of another
link to the same page (unless Keyword Topic B is a very close variation of
Keyword Topic A). A page should have one core topic, so be consistent.
While Google is more than capable of
crawling and executing basic JavaScript, they don’t always get things right. To
be on the safe side, all links that you want crawled should be in plain old
HTML (this includes navigation elements). If you’re considering using something
like AJAX on your site, be aware of the risks and how to minimize them.
Broken Links
Because Google and other search engines crawl
the web link-to-link, broken links can cause SEO problems for a website. When
Google is crawling a site and hits a broken link, the crawler immediately
leaves the site. If Google encounters too many broken links on a site it may
deem that site a poor user experience, which can cause a reduced crawl
rate/depth, and both indexing and ranking problems.
Unfortunately, broken links can also happen due to someone
outside of your site linking in incorrectly. While these types of broken links
can’t be avoided, they can be easily fixed with a 301 redirect.
If broken links are found, you need to implement a 301 redirect
per the guidelines in the URL Redirect section. You can also use your Google or
Bing Webmaster Tools account to check for broken links that have been found on
your site.
URL Redirects
Unless a redirect is truly temporary, such as
for a time sensitive promotion, 302 redirects should never be used. 302
redirects don’t pass any link value, and are essentially a dead end for SEO. In
almost every scenario where a redirect is needed, a 301 (permanent) redirect
should be used.
Any page that changes URLs or is
deleted needs a 301 redirect to tell search engines and users that the page has
moved/is gone. There should never be more than one URL path to a page. You can
learn more about redirects here: http://moz.com/learn/seo/redirection
On an Apache server, redirects will be configured via the
mod_rewrite module and your .htaccess file. Making these sorts of changes is a
very technical task, and can break your website if done incorrectly. Unless
you’re very technical, it’s best to leave this one to your web developer.
That said, here are a couple of common issues, and the correct
code to use in your .htaccess file to fix them:
## Always include this at the start of your htaccess file ##
Options +FollowSymlinks
RewriteEngine On
## Redirect HTTPS URLs to HTTP URLs ##
RewriteCond %{SERVER_PORT} ^443$
RewriteRule ^(.*)$ http://www.%{HTTP_HOST}/$1 [R=301,L]
## Redirect Non-WWW to WWW ##
RewriteCond %{HTTP_HOST} !^www\.
RewriteRule ^(.*)$ http://www.%{HTTP_HOST}/$1 [R=301,L]
## Rewrite all MiXed Case uRLs to lower case URL’s and 301
redirect ##
RewriteMap lc int:tolower
RewriteCond %{REQUEST_URI} [A-Z]
RewriteRule (.*) ${lc:$1} [R=301,L]
## Redirect index.htm, index.html or index.php to the trailing
slash recursively for the entire site ##
RewriteCond %{THE_REQUEST} /index\.html? [NC]
RewriteRule ^(.*/)?index\.html?$ /$1 [R=301,L]
RewriteCond %{THE_REQUEST} /index\.php? [NC]
RewriteRule ^(.*/)?index\.php?$ /$1 [R=301,L]
## Ensure all URLs have a trailing slash ##
RewriteBase /
RewriteCond %{REQUEST_FILENAME} !-f
RewriteCond %{REQUEST_URI} !(.*)/$
RewriteRule ^(.*)$ http://www.%{HTTP_HOST}/$1/ [L,R=301]
## Remove Spammy Query Strings ##
<ifModule mod_rewrite.c>
RewriteCond %{QUERY_STRING} enter|query|strings|here [NC]
RewriteRule .* http://www.%{HTTP_HOST}/$1? [R=301,L]
</ifModule>
## End of htaccess file ##
Obviously, if any of these don’t apply to your site, don’t use
them. Test these carefully before you roll them live on a site! (I take no
responsibility for broken sites, you’ve been warned).
If your site has HTTPS pages for a good reason, you definitely
don’t want to go redirecting them to HTTP pages! For the spammy query strings
(say, someone pointed a ?pill query string at your site), just replace the fake
text with the real query strings, separated by pipes if there are multiple
query strings.
Robots.txt Configuration
XML Sitemap Configuration
In a nutshell, make sure your sitemap.xml file exists immediately after
your root directory (www.domain.com/sitemap.xml), and that it contains a
properly formatted list of every page on your site that you want Google and
other search engines to list in their indices. If a page wouldn’t make a good
search engine entry point, it’s probably best not to list it. You can learn the
nuts and bolts of creating a sitemap here, and specialized sitemaps here.
Title Tags
After the URL, the second most important place
on a page to have the keyword is the Title tag. A proper Title tag will:
·
Be less than 55 characters in length (and never shorter than 15
characters).*
·
Be unique to and descriptive of that page (never use identical
or mostly identical title tags on multiple pages on the same site).
·
Use the keyword of that page twice if space permits (once at the
start, followed by a separator such as a colon and a space, and then once again
in a call to action). If the character limit prevents the use of the keyword
twice, use it once in a good call to action, with the keyword as close to the
beginning of the title tag as possible.
·
If relevant, include a geo-qualifier (such as Seattle WA).
·
Typically not be used for branding purposes, unless you have a
major brand whose name would increase SERP click-through-rates. If you must
include your brand name, use it at the end of the Title tag, not at the
beginning.
* – While you may hear some SEOs
say to limit Titles to 70 characters, it’s a misconception. Google is actually
using a pixel width limit, not a character limit. Title tags appear in 13.5pt
Arial font by default (18px), with searched for keywords bolded, and Google has
a pixel width limit of 512 pixels.
You can see if a Title will truncate
by doing the following: simply use Excel, set column width to 512px, set
columns to wrap text, and font to Arial 13.5pt. Type in your Title, and bold
the main keyword. If the line breaks, your Title tag will truncate. (You can
also use this tool to check: Title Length Tool).
Using 55 characters is now considered the safe upper limit, as
this Title character limit will avoid truncation 95% of the time.
Proper title tag structure is as follows:
<title>Keyword Phrase:
Call to Action Using Keyword Phrase</title>
A colon and a space are used as the separator because it uses
the least amount of characters. You can also use – or | (a dash or a pipe).
The Title tag is the first description of the page that search
engine users will read, and it is extremely important to both users and search
engines that it contain the keyword they are searching for. This will not only
help to improve rankings, but can significantly improve the click-through-rate
on search engine results pages (SERPs).
Meta Description Tags
Meta description tags are used as the
description searchers will see in the search engine results pages (SERPs).
Having the keyword used properly in the Meta description tag is not a part of
the ranking algorithm, but will increase the likelihood that users will click
on the link to the page. It should adhere to the following guidelines:
·
Be unique and relevant to that page, be written as ad text, and
contain a call to action.
·
Be no more than 160 characters in length, including spaces and
punctuation (and no less than 50 characters).
·
Contain 1-2 complete sentences, correct grammar and punctuation.
·
Use the keyword once or twice (once per sentence, as close to
the start of each sentence as possible).
·
Include a geo-qualifier (City and/or State), such as “Seattle
WA”, only if relevant.
A proper Meta description tag would be:
<meta name=”description” content=”Keyword Phrase used in a question? Keyword Phrase used in a good
click-inducing call to action.” />
To further encourage the search engines to use the description
you provide, add this tag to every page:
<meta name=”robots”
content=”noodp, noydir” />
This tells search engines not to use snippets from DMOZ or the
Yahoo Directory. Some people think this tag is no longer needed. We disagree,
and have seen some pretty wonky descriptions come from not blocking this.
Meta Keywords Tags
These no longer have any SEO value on
Google, Yahoo or Bing. The current best practice is to leave these off of your
site entirely. If your CMS forces you to have this tag present, either leave it
blank or make sure that there are no more than 1-2 keywords in the tag, and
that those keywords are used at least twice each on the page in question.
While these tags have no SEO value, misusing them can still have
a negative impact on your rank-ability. Keyword stuffed meta keywords tags are
still a negative signal, and could negatively impact your ability to rank
organically.
Heading Tags
Search engines weight text for SEO value based
on text size and position on the page. Heading tags are supposed to be larger
than the other text on the page, and should typically be the first thing in the
content section of the page; thus the added benefit of having the keyword in
the heading tags.
Every page should have an H1 tag, as search engines look to the
H1 to help determine the topic of a page. It should be the first thing in the
body text of the page, and should appear prominently.
The keyword of a page needs to be
used in the H1 tag, and in at least half of the total heading tags on a page.
There should never be more than one H1 on a
page. H1 tags should never wrap images or logos, only text.
From a usability perspective, paragraphs should never be longer
than 5 lines of text, and it is wise to break up a page every 2-3 paragraphs
with a sub-heading in the form of an H tag (H2 or H3). Testing has shown that
when users are faced with a large block of unbroken text, most either skim over
the text or skip it altogether.
We recommend no more than 1 heading tag per 150 words on the
page. It is VERY important that the keyword of a page be used in the H1 tag, as
close to the beginning of the H1 as possible. Ideally, if a page has at least
300 words of content, there should be at least one additional H tag on each
page that contains the keyword, for added SEO value.
Image Names and ALT Tags
For the benefit of search engines, code
compliance, and visually impaired users, every image MUST have an ALT tag. The
ALT tag should accurately describe the image, and should contain a keyword
relevant to that page only if the keyword is relevant to the image as well. The
ALT tag should be the same for every instance of that image on the entire site.
<img
src=”keyword-rich-image-name.jpg” alt=”Describe Image” />
Image file names should be comprised of words separated by
dashes, should be descriptive to both users and search engines, and should
accurately describe the image. If relevant, it should also use a keyword
relevant to the page/domain.
NoFollow Tags
All links pass a bit of trust, PageRank and
anchor text. This 3rd party endorsement system is at the heart of the Google
search algorithm. It measures how different pages link together, and weights
those links based on traffic, popularity, relevance, age, size, content, and
hundreds of other components.
When pages that Google deems relevant link to other pages, some
of that trust and authority flows through that link to the site being linked
to. A “followed” link is essentially endorsing the page being linked to. Anchor
text also passes through links.
Enter the rel=”nofollow” tag. Google introduced this tag to help
prevent ranking manipulation through comment spam. While the use of this tag
used to prevent ANY trust or anchor text from being passed on, Google has
recently made some changes.
Now, when the tag rel=”nofollow” is
used in an anchor tag (link), Google will usually just pass less trust to the page being linked to (~50% less,
from what we have heard). Using this tag is like saying, this page is nice, but
we don’t want to endorse it. Of course, Google can ignore you and pass whatever
they choose…NoFollow is little better than a suggestion these days, especially
when it comes to NoFollow’d social links (Twitter, Facebook, etc.).
With that in mind, from an SEO perspective, you would probably
only choose to use the rel=”nofollow” tag in the following instances:
1. On any user created
link within blog or forum comments, author profiles, etc. If it’s likely to be
spammed, use NoFollow.
2. On any internal
link that links to a page with no SEO value (i.e. login pages, RSS feed pages,
etc) – If the page has no user value, or if it’s a page you don’t really want
showing up in search engines as an entry point, feel free to use the
rel=”nofollow” tag on links to that page. Just make sure that every page on the
site is linked to from at least one other page on the site, so that all pages
can be indexed (a user sitemap is the best place for this).
Remember, NoFollow may still pass link juice to the linked to
page, but usually less of it.
A proper NoFollow tag would be used like this:
<a
href=”https://www.domain.com” rel=”nofollow”>Page Name</a>
Keep in mind though, it is perfectly normal to link out to other
sites on occasion, and linking out to at least some other sites with followed
links is a part of appearing “normal” in Google’s eyes. Don’t link out to
external domains from every page, but definitely do so with followed links from
at least a few pages, particularly blog posts as relevant.
Content
Last but far from least, one of the most
important on-site SEO elements is text content. Search engines want to see
roughly 200-300 words or more of unique text on
each page (there’s some debate on the minimum, but this is a safe number).
We recommend having roughly 400-600 words of unique text per
page, and never more than 800 words on a single page (this is for usability
reasons, not SEO reasons, and blog posts are an exception to this rule;
1000-3000+ word blog posts rank amazingly well).
This content on a page should contain the exact target keyword
about 3-4 times, and perhaps a few more times on long pages. Having unique,
keyword rich, topic focused text on a page can help to improve search engine
rankings significantly.
It is also highly beneficial to use
variations of the keyword. For example, for “chocolate pudding”, you would want
to use “chocolate” and “pudding” somewhere on the same page. It is also highly
valuable to use related keywords. In the case of “chocolate pudding”, you might
also want to use “creamy”, “Jell-O”, “dessert” and “tasty”. This technique is
referred to as LDA (Latent Dirichlet Allocation) and helps to establish
excellent topical relevancy, and increases your chance of ranking
significantly.
You should also use the keyword at least once in a
<strong> tag and/or an <em> tag on every page.
That said; at the end of the day make sure you are writing
content for your human users, and not just for search engines. If it’s not
worthy of commenting on or sharing, it’s probably not worth writing.
To rank and stay ranking in a competitive space and to appeal to
Google’s QDF algorithm (query deserves freshness), regular content growth is
very important. We recommend that every site have a blog, and that they write
in that blog at least once per week (once per day if possible).
It is also beneficial to update the content of existing pages
from time to time. Google can tell how often a site is updated, and takes that
into account in rankings. The more often changes are made to a site, and the
more often unique content is added, the greater the value of that site to users
in the eyes of the search engines.
In addition to content growth, search engines and users love to
see content diversity. Text is great, but images, videos, polls, PDFs, and
other interactive resources have both user experience and SEO benefits. Videos
and images, in addition to helping with SEO, can also drive additional traffic
to your site via image search and video search results.
Duplicate content is viewed as
a big negative by every major search engine, Google in particular. It can not
only hurt rankings, but can prevent a page from ranking, and sometimes result
in de-indexing of an entire domain. Search engines want to see unique content
on a site, and in their search results. A search results page with 10 different
websites that contain the same content would be a poor results page indeed.
There are two main types of duplicate content:
1. Duplicate content
within your own domain (such as is often caused by CMS issues, like WordPress
tag and author pages). This is a very common problem with e-commerce websites.
2. Cross-domain
duplicate content, where your site is hosting content that is identical to
content used on other websites. This is also a common problem with e-commerce
websites, as well as article sites, news sites, and less than scrupulous
scraper sites.
Duplicate content is generally defined as any page that is 30%
or more the same as content elsewhere on the web. Duplicate content is found
using N-grams, which look for identical sequence of 10 words or more, excluding
common usages and stop words. Too many blocks of identical content on a page =
duplicate.
If content must be duplicated within your own site, for
whatever reason, you have three options:
1. make use of the
rel=”canonical” tag on duplicate pages
2. use the <meta
name=”robots” content=”noindex, follow” />tag on duplicate pages
3. block indexing of
duplicate pages in the Robots.txt file
When it comes to cross-domain duplicate content, you can use the
above options, but the best practice is to simply not have cross-domain
duplicate content anywhere on your site. It is better to be safe than sorry in
this regard.
Because Google is putting more emphasis on user experience, page
load times are also now a factor in search algorithms. You should ensure that
every page on the site will load in 1-2 seconds on a typical DSL connection.
When a user exits the site via the back button, it hurts rankings, and few
things elicits the use of the back button more than a slow loading page. With
this in mind, make sure the file size of images or video in your content are as
small as possible.
Using flash or images in place of
text is considered a big negative. While search engines like a variety of
media, it must not take the place of text.
Flash looks good, but it is not SEO friendly. In addition, requiring that users
have Flash to view a website has been shown to increase load times and bounce
rates, and decrease usability. It also means that your site would appear broken
to someone using an Apple mobile device such as the iPhone or iPad. If you need
Flash-like visual elements, there are alternatives.
However, if you absolutely must use
Flash, there is an SEO friendly technique for websites using Flash known as SIFR, where the flash is replaced with text if the user does
not have flash installed, and that makes the HTML text always available to the
spiders. This way, the flash content is presented to user’s who have flash, but
is always readable by the search engines, as well as users who don’t have flash
installed.
Avoid all Black Hat SEO
techniques. When trying to rank for a competitive keyword, you may be
tempted to try some less than kosher SEO tactics…DON’T! Black Hat SEO is a very
big negative, and if discovered could result in your site being removed
completely from Google’s search index.
This means:
·
No keyword stuffing. This means NONE, not in any way. Don’t hide
keywords in DIVs, strange made up Meta tags, or anywhere else on the page or in
the code. It is not worth the risk of being de-indexed.
·
No disabling of the back button, and no pop-ups/pop-unders
stopping you from exiting a site
·
No sneaky redirects
·
No pulling large blocks of content or entire sites with iFrames
(no SEO value)
·
No Meta refreshes (unless the refresh time is 0)
·
No hidden text (unless you’re using a user accessible collapsing
DIV, which is fine)
·
No hidden links
·
No text of the same or very similar color as the background
·
No displaying different content based on user agent (except for
URL names, which is OK)
Though advertising can be a great way to make money, it
seriously detracts from both the user experience and the SEO-ability of both a
website as a whole, as well as individual pages. Ad calls slow down page load
times, often create a poor user experience, and in some instances can trigger ranking
penalties tied to the Google Panda updates.
You should carefully weigh the pros and cons of using ads, based
on the purpose and goals of your website. For some sites, the tradeoff will be
worth it.
Here is a visual example of a
perfectly optimized page (courtesy of Moz):
|
Perfectly Optimized Page |
Optimal Content Sample:
A sample page optimized for the keyword “Seattle SEO”, ideally
written, is as follows (HTML in Orange):
<h1>Seattle SEO</h1>
<p>If you are a
business in Seattle with an online presence, then your business could likely
benefit from professional Seattle SEO services. What is SEO you say? Quite
simply, it is the art of optimizing a website or webpage to be more search
engine and user friendly, ideally resulting in improved search engine rankings.</p>
<p>The vast majority
of internet users make use of search engines to find what they are looking for.
Regardless of whether you own an online business or a brick-and-mortar
operation, a website is a must-have to maximize your revenue potential.
Unfortunately, just any old website isn’t going to cut it. To get the most
benefit from your online presence, making effective use of <a
href=”http://www.vudumarketing.com/”>Search Engine Optimization</a> is a
necessity.</p>
<p>So, your business
is in Seattle, and you want a local company to build you a website and/or do
some SEO work…now what? How do you find such a company?</p>
<h2>Finding the Right
Seattle SEO Firm</h2>
<p>With dozens of
companies offering SEO services in the Seattle area, it can be a real challenge
to find the right one for your business. Of course everyone will claim to be
the best, but how do you really know if what they are selling is what you
should be buying? To that end, we offer the following advice:</p>
<ul><li><strong>Know What Questions
to Ask</strong> – It is important to always negotiate from a
position of strength. Learn the basics of <a href=”http://www.vudumarketing.com/”>SEO</a>, such as the value
of keyword rich URLs and title tags, before you go searching. Even if you just
sound like you know what you are talking about, you are much more likely to
effectively weed out imposters.</li>
<li><strong>Ask for Examples of
Past SEO Work</strong> – From testimonials to actual optimized sites and
rankings, ask to see some work they’ve done. Make a note of the examples, and
contact those companies if possible. Ask them if they had a positive experience
with the SEO company in question, and if they would recommend them.</li>
<li><strong>Determine How They
Measure Success</strong> – If their only measure of success is improved
rankings, they are probably not the right company for you. SEO, like any form
of marketing, is about making more money. To that end, tracking traffic and
conversions in addition to rankings is the ideal scenario. If they don’t
measure rankings plus some additional metrics, you should probably look
elsewhere for a better company.</li></ul>
<h2>Learn More About
Seattle SEO</h2>
<p>While there are
many other tips that could be offered, the three tips above are all that you
should need to find a high quality SEO company, one that can add real value to
your bottom line. <a href=”http://www.vudumarketing.com/contact-us/”>Learn More About
SEO</a> today!</p>
In summary of the above content:
·
It’s about 400 words in length, not counting code, putting it in
the ideal 400-600 word range.
·
The keyword is used in the H1 tag, and in at least 50% of the
other H tags, and the page is effectively broken up by sub-headings (H2 or H3
Tags). There is approximately one H tag per 150 words of content.
·
The keyword is used 5 times throughout the page, and also
appears in a strong tag (bold) and in a list item.
·
Variations of the keyword and other related keywords are used
throughout the text, helping to page to conform to the rules of LDA (Latent
Dirichlet Allocation).
·
The text of the page contains cross-links to other related pages
using keyword rich anchor text relevant to the pages being linked to. Every
page should cross-link to 1 or 2 other relevant pages on your site.
The page closes with a call to action, which is a critical
element of content writing. All content should provide real benefit to the
reader, and should encourage the reader to take some sort of further action to
secure a conversion.
Rich Snippets (microdata, microformats, RDFa)
If you’ve done a search on Google recently and
seen a search result with an image next to it, or star ratings, a video next to
the SERP, or any one of a dozen other advanced search results, you’ve seen the
impact of Schema, RDFa, etc. You can learn more from Google here,
and on Schema.org.
One of the most common implementation
out there has to do with rel=”author” and rel=”publisher”. This is where the
image of the person or business who wrote an article or page shows up next to
their search result. These are a great way to increase your traffic, and are
very simple to set-up. There’s a great slide deck here that walks you through this
particular implementation.
If your site provides products, services, recipes, apps, music,
videos, or any of dozens of other options, the chances are good that you could
use Schema to upgrade your search results!
These page-level code markups allow your search results to stand
out from the crowd, and can significantly increase your organic
click-through-rates. I’ve seen instances where the #3 or #4 search result
actually had a higher CTR than the #1 ranked result, simple because of Schema
markup.
Localization
If you’re a local business, local search
traffic is becoming more important than ever. More and more people are
searching using localized search phrases, or at least have local intent while
searching. If you want to rank well for these sorts of local searches, there
are a few key things you should do/have in place:
·
Create and/or Claim Your Local Profiles – You can
certainly do this manually on Google, Bing, Yelp and other sites, but the
easiest way is to use a tool like Yextto quickly and easily do this across all of
the most common sites at once.
·
Claim Your Social Profiles – With Local SEO, citations are
key (any place online that has your name, address and phone number exactly as
it appears in your local listings). The key with this is to make sure all of
your online business listings use the exact same format for your name, address,
phone number and website URL (and we mean EXACT…1-800-798-2430 is different
from (800) 798-2430…be EXACT). One of the quickest and easiest ways to get a
bunch of citations (and to prevent brand squatters) is to claim all of your
profiles via a service like KnowEm.
·
Create a KML File – A KML file can
help to make sure your business shows up accurately on maps.
·
List Your Business Info in Schema – Using Schema to
identify your business name, location, business hours and other key elements
can not only help with local rankings, but can help with getting rich snippets
in your search results.
·
Use Your NAP (Name, Address, Phone) and Local Keywords Throughout
Your Site – Throughout your site, in the footer, on your contact
page, and in Titles, Metas and page content as relevant. It should be crystal
clear to Google and visitors that you service a particular area.
Social Sharing Elements