For more basic but also in-depth advice, our Periodic Table Of SEO Success Factors, shown below, introduces you to all the key concepts you need to know:
- On The Page SEO
- Off The Page SEO
- Violations
Within each group are subgroups, as each chapter of this guide will explain. These subgroups contain one or more individual SEO factors with a specific weight or importance.
Violations, while a group unto themselves, are displayed under the group and subgroup to which they’re associated.
Those two letter acronyms you see on the chart? That’s our play on the periodic table of elements and the letter representations, or symbol, of each element. You may have had to remember that the symbol for gold was Au or that iron’s was Fe.
We’ve tried to make it slightly more intuitive. The first letter of each “SEO element” comes from the subgroup that it’s in and the second letter stands for the individual factor.
Factors Work In Combination
No single SEO factor will guarantee search engine rankings. Having a great HTML title won’t help if a page has low quality content. Having many links won’t help if they are all low in quality. Having several positive factors can increase the odds of success while the presence of negative factors can worsen those odds.
On The Page Factors
On The Page search ranking factors are those that are entirely within the publisher’s own control. What type of content do you publish? Are you providing important HTML clues that help search engines (and users) determine relevancy? How does your site architecture help or hinder search engines?
Off The Page Factors
Off The Page ranking factors are those that publishers do not directly control. Search engines use these because they learned early on that relying on publisher controlled signals alone didn’t always yield the best results. For instance, some publishers may try to make themselves seem more relevant than they are in reality.
With billions of web pages to sort through, looking only at ‘on the page’ clues isn’t enough. More signals are needed to return the best pages for any particular search.
Violations
Make no mistake, search engines want people to perform SEO because it can help improve their search results. Search engines provide help in the form of guidelines, blog posts and videos to encourage specific SEO techniques.
However, there are some techniques that search engines deem “spam” or “black hat”, which could result in your pages receiving a ranking penalty or, worse, being banned from the search engines entirely.
Violations are generally tactics meant to deceive or manipulate a search engine’s understanding of a site’s true relevancy and authority.
Weighting
All the factors we show are weighted on a scale of one to three, as shown in the top right corner of each factor as well as reflected in the hue of that factor. A weighting of three is most important and is something you should pay special attention to because it has a bigger impact than other factors.
That doesn’t mean that factors weighted two or one aren’t important. They are. It’s just that they are of less importance, relatively speaking, in terms of the other factors on the chart. Violations are also weighted, but in negative numbers, with negative three being the worst and potentially most harmful to your SEO success.
The weighting is based on a combination of what search engines have said, surveys of the SEO community as well as our own expertise and experience in watching the space over time. We don’t expect them to be perfect. Not everyone will agree. Your mileage may vary.
But we’re confident it is a useful general guide.
“Missing” Factors & The Guide’s Philosophy
Experienced SEOs may be wondering why some factors aren’t shown. How come ALT text or bolding words aren’t included as HTML factors, for example?
The answer? We don’t think those things are as important, relatively speaking. We’re not trying to boil the ocean and encompass every possible signal (Google has over 200 of them) and sub-signals (Google has over 10,000 of those).
Instead, the goal of the Periodic Table Of SEO Success Factors and this online companion guide is to help those new to SEO focus on the big picture and perhaps allow experienced SEOs to hit the “reset” button if they’ve gotten lost staring at specific trees in the SEO forest.
That’s why this guide doesn’t address having your most important keywords be at the beginning or end of an HTML title tag. Nor are we trying to assess how much more weight an H1 header tag carries than an H2 tag.
We’re purposely avoiding being ultra specific because such things often distract and pull us down the rabbit hole. Instead, we hope you gain an understanding that pages should have descriptive titles, that indicating page structure with header tags may help, and topping things off with structured data is a good idea.
Do these things well and you’ve probably addressed 90% of the most important HTML factors.
Similarly, it’s not whether a good reputation on Twitter is worth more than on Facebook. Instead, we’re trying to help people understand that having social accounts that are reputable in general, which attract a good following and generate social shares, may ultimately help you achieve search success.
Chapter 2: Content & Search Engine Success Factors
Content is king. You’ll hear that phrase over and over again when it comes SEO success. Indeed, that’s why the Periodic Table Of SEO Success Factors begins with the content “elements,” with the very first element being about content quality.
Get your content right, and you’ve created a solid foundation to support all of your other SEO efforts.
Cq: Content Quality
More than anything else, are you producing quality content? If you’re selling something, do you go beyond being a simple brochure with the same information that can be found on hundreds of other sites?
Do you provide a reason for people to spend more than a few seconds reading your pages?
Do you offer real value, something of substance to visitors, that is unique, different, useful and that they won’t find elsewhere?
These are just some of the questions to ask yourself in assessing whether you’re providing quality content. This is not the place to skimp since it is the cornerstone upon which nearly all other factors depend.
Cr: Content Research / Keyword Research
Perhaps the most important SEO factor after creating good content is good keyword research. There are a variety of tools that allow you to discover the specific ways that people may be searching for your content.
You want to create content using those keywords, the actual search terms people are using, so you can produce content that effectively “answers” that query.
For example, a page about “Avoiding Melanoma” might use technical jargon to describe ways to prevent skin cancer. But a search engine might skip or not rank that page highly if people are instead searching for “skin cancer prevention tips”. Your content needs to be written in the right ‘language’ – the language your customer or user is using when searching.
Cw: Content Words / Use Of Keywords
Having done your keyword research (you did that, right?), have you actually used those words in your content? Or if you’ve already created some quality content before doing research, perhaps it’s time to revisit that material and do some editing.
Bottom line, if you want your pages to be found for particular words, it’s a good idea to actually use those words in your copy.
How often? Repeat each word you want to be found for at least five times or seek out a keyword density of 2.45%, for best results.
No no no, that was a joke! There’s no precise number of times. Even if “keyword density” sounds scientific, even if you hit some vaunted “ideal” percentage, that would guarantee absolutely nothing.
Just use common sense. Think about the words you want a page to be found for, the words you feel are relevant from your keyword research. Then use them naturally on the page. If you commonly shift to pronouns on a second and further references, maybe use the actual noun again here and there, rather than a pronoun.
Ce: Content Engagement
Quality content should produce meaningful interactions with users. Search engines may try to measure this interaction – engagement – in a variety of ways.
For example, how long do users stay on your page? Did they search, clickthrough to your listing but then immediately “bounce” back to the results to try something else? That “pogosticking” behavior can be measured by search engines and could be a sign that your content isn’t engaging.
Conversely, are people sending a relatively long time reviewing your content, in relation to similar content on other sites? That “time on site” metric or “long click” is another type of engagement that search engines can measure and use to assess the relative value of content.
Social gestures such as comments, shares and “likes” represent another way that engagement might be measured. We’ll cover these in greater detail in the Social section of this guide.
Search engines are typically cagey about the use of engagement metrics, much less the specifics of those metrics. However, we do believe engagement is measured and used to inform search results.
Cf: Content Freshness
Search engines love new content. That’s usually what we mean when we say ‘fresh’.
So you can’t update your pages (or the publish date) every day thinking that will make them ‘fresh’ and more likely to rank. Nor can you just add new pages constantly, just for the sake of having new pages, and think that gives you a freshness boost.
However, Google does have something it calls “Query Deserved Freshness (QDF)”. If there’s a search that is suddenly very popular versus its normal activity, Google will apply QDF to that term and look to see if there’s any fresh content on that topic. If there is, that new or fresh content is given a boost in search results.
Chapter 3: HTML Code & Search Engine Success Factors
HTML is the underlying code used to create web pages. Search engines can pick up ranking signals from specific HTML elements. Below are some of the most important HTML elements to achieve SEO success.
Ht: HTML Title Tag
Imagine that you wrote 100 different books but gave them all the same exact title. How would anyone understand that they are all about different topics?
Imagine that you wrote 100 different books, and while they did have different titles, they weren’t very descriptive — maybe just a single word or two. Again, how would anyone know, at a glance, what the books are about?
HTML titles have always been and remain the most important HTML signal that search engines use to understand what a page is about. Bad titles on your pages are like having bad book titles in the examples above. In fact, if your HTML titles are deemed bad, Google changes them.
So think about what you hope each page will be found for, relying on the keyword research you’ve already performed. Then craft unique, descriptive titles for each of your pages.
Hd: The Meta Description Tag
The meta description tag, one of the oldest supported HTML elements, allows you to suggest how you’d like your pages to be described in search listings. If the HTML title is the equivalent to a book title, the meta description is like the blurb on the back describing the book.
SEO purists will argue that the meta description tag isn’t a “ranking factor” and that it doesn’t actually help your pages rank higher. Rather, it’s a “display factor,” something that helps how you look if you appear in the top results due to other factors.
Technically, that’s correct. And it’s one of the reasons we decided to call these “success” factors instead of ranking factor.
Because a meta description that contains the keywords searched for (in bold) may catch the user’s eye. A well crafted meta description may help ‘sell’ that result to the user. Both can result in additional clicks to your site. As such, it makes sense for the meta description tag to be counted as a success factor.
Be forewarned, having a meta description tag doesn’t guarantee that your description will actually get used. Search engines may create different descriptions based on what they believe is most relevant for a particular query. But having one increases the odds that what you prefer will appear. And it’s easy to do. So do it.
Hh: Header Tags
See the headline up at the top of this page? Behind the scenes HTML code is used to make that a header tag. In this case, an H1 tag.
See the sub-headlines on the page? Those also use header tags. Each of them is the next “level” down, using H2 tags.
Header tags are a formal way to identify key sections of a web page. Search engines have long used them as clues to what a page is about. If the words you want to be found for are in header tags, you have a slightly increased chance of appearing in searches for those words.
Naturally, this knowledge has caused some people to go overboard. They’ll put entire paragraphs in header tags. That doesn’t help. Header tags are as much for making content easy to read for users as it is for search engines.
Header tags are useful when they reflect the logical structure (or outline) of a page. If you have a main headline, use an H1 tag. Relevant subheads should use an H2 tag. Use headers as they make sense and they may reinforce other ranking factors.
Hs: Structured Data
What if you could tell search engines what your content was about in their own “language”? Behind the scenes, sites can use specific mark-up (code) that make it easy for search engines to understand the details of the page content and structure.
The result of structured data often translates into what is called a ‘rich snippet‘, a search listing that has extra bells and whistles that make it more attractive and useful to users. The most common rich snippet you’re likely to encounter is reviews/ratings which usually includes eye-catching stars.
While the use of structured data may not be a direct ranking factor, it is clearly a success factor. All things being equal, a listing with a rich snippet will get more clicks than one without. And search engines are eager for site owners to embrace structured data, providing new and easier ways for less tech-savvy webmasters to participate.
Structured data has been around for quite some time in various forms. But recently search engines have begun to rely on it more with the advent of Google’s Knowledge Graph and Bing’s Snapshot.
This element enters the periodic table for the first time and we suspect it may become more important over time.
Chapter 4: Site Architecture & Search Engine Success Factors
The last major On The Page group in the Periodic Table Of SEO Success Factors is site architecture. The right site structure can help your SEO efforts flourish while the wrong one can cripple them.
Ac: Site Crawlability
Search engines “crawl” web sites, going from one page to another incredibly quickly, acting like hyperactive speed readers. They make copies of your pages that get stored in what’s called an “index,” which is like a massive book of the web.
When someone searches, the search engine flips through this big book, finds all the relevant pages and then picks out what it thinks are the very best ones to show first. To be found, you have to be in the book. To be in the book, you have to be crawled.
Most sites generally don’t have crawling issues, but there are things that can cause problems. For example, JavaScript or Flash can potentially hide links, making the pages those links lead to hidden from search engines. And both can potentially cause the actual words on pages to be hidden.
Each site is given a crawl budget, an approximate amount of time or pages a search engine will crawl each day, based on the relative trust and authority of a site. Larger sites may seek to improve their crawl efficiency to ensure that the ‘right’ pages are being crawled more often. The use of robots.txt, internal link structures and specifically telling search engines to not crawl pages with certain URL parameters can all improve crawl efficiency.
However, for most, crawl problems can be easily avoided. In addition, it’s good practice to use sitemaps, both HTML and XML, to make it easy for search engines to crawl your site.
Ad: Duplication / Canonicalization
Sometimes that big book, the search index, gets messy. Flipping through it, a search engine might find page after page after page of what looks like virtually the same content, making it more difficult for it to figure out which of those many pages it should return for a given search. This is not good.
It gets even worse if people are actively linking to different versions of the same page. Those links, an indicator of trust and authority, are suddenly split between those versions. The result is a distorted (and lower) perception of the true value users have assigned that page. That’s why canonicalization is so important.
You only want one version of a page to be available to search engines.
There are a number of different ways that duplicate versions of a page can creep into existence. A site may have www and non-www versions of the site instead of redirecting one to the other. An ecommerce site may allow search engines to index their paginated pages. But no one is search for “page 9 red dresses”. Or filtering parameters might be appended to a URL, making it look (to a search engine) like a different page.
For as many ways as there are to inadvertently create URL bloat, there are ways to address it. Proper implementation of 301 redirects, the use of rel=canonical tags, managing URL parameters and effective pagination strategies can all help ensure you’re running a tight ship.
As: Site Speed
Google wants to make the web a faster place and has declared that speedy sites get a small ranking advantage over slower sites.
However, making your site blistering fast isn’t a guaranteed express ride to the top of search results. Speed is a minor factor that impacts just 1 in 100 queries according to Google.
But speed can reinforce other factors and may actually improve others. We’re an impatient bunch of folks these days. So engagement (and conversion) on a site may improve based on a speedy load time.
Speed up your site! Search engines and humans will both appreciate it.
Au: Are Your URLs Descriptive?
Yes. Having the words you want to be found for within your domain name or page URLs can help your ranking prospects. It’s not a major factor but if it makes sense to have descriptive words in your URLs, do so.
Aside from helping a bit from a ranking perspective, various research reports over the years have shown that searchers are more likely to select pages with short, descriptive URLs over other pages in search results.
It’s also notable that all major search engines have moved the URL up just under the title in their search listings. And structured data can be used to transform the URL into a breadcrumb, giving users more ways to navigate to a site directly from a search result.
If the URL wasn’t important, why do search engines pay so much attention to it?
Chapter 5: Link Building & Ranking In Search Engines
Links were the first major “Off The Page” ranking factor used by search engines. Google wasn’t the first search engine to count links as “votes,” but it was the first search engine to rely heavily on link analysis (or the Link Graph) as a way to improve relevancy.
Despite the chatter around other signals, links remain the most important external signal for search rankings. But as you’ll find, some links are more valuable than others.
Lq: Link Quality
If you were sick, which would you trust more? The advice from five doctors or from fifty random people who offered their advice as you walked down the street.
Unless you’ve had a really bad experience with doctors, you’ll probably trust the advice from the doctors. Even though you’re getting fewer opinions, you’re getting those opinions from experts. The quality of their opinions carries more weight.
It works the same way with search engines. They’ll count all the links pointing at web sites (except those blocked using nofollow or other methods), but they don’t count them all equally. They give more weight to the links that are considered to be of better quality.
What’s a quality link? It’s one of those “you’ll know it when you see it” types of things in many cases. But a link from any large, respectable site is going to be higher on the quality scale than a link you might get from commenting on a blog. In addition, links from those in your “neighborhood”, sites that are topically relevant to your site, may also count more.
Lt: Link Text / Anchor Text
Amazon has millions of links pointing at it. Yet, it doesn’t rank for “cars.” It does rank for “books.” Why? Many of those links pointing at Amazon say the word “books” within the links while relatively few say “cars,” since Amazon doesn’t sell cars.
The words within a link — the link text or “anchor text” — are seen by search engines as the way one web site is describing another. It’s as if someone’s pointing at you in real life and saying “books” and declaring you to be an expert on that topic.
You often can’t control the words people use to link to you, so capitalize on your opportunities to influence anchor text, within reason.
Link text is a powerful factor but has been decreased from a weight of 3 to 2 in this edition of the table. The downgrade revolves around the efforts Google has made, via the Penguin Updates, to identify over-optimized anchor text and a desire by search engines to see more ‘natural’ link text patterns.
Ln: Number Of Links
Plenty of sites have found that getting a whole lot of links can add up to SEO success. Even more so if your getting a lot of links from many different sites. All things being equal, 1000 links from 1 site will mean far less than 1000 links from 1000 sites.
Long ago, the sheer number of links used to be a far more important, but has decreased steadily as search engines learn how to better evaluate the quality of links.
Tactics such as viral link baiting campaigns, badges and widgets can all be effective at securing large numbers of links and something even search engine representatives have suggested.
But in your quest for links, don’t fire up automated software and begin spamming blogs. That’s a bad thing, in many ways, as we’ll explore later in this guide.
Chapter 6: Social Media & Ranking In Search Results
Using links as an Off The Page ranking factor was a great leap forward for search engines. But over time, links have lost some of their value for a variety of reasons. Some sites are stingy about linking out. Others block links to help fight spam. And links get bought and sold, making them less trustworthy.
Enter social media. If links were a way for people to “vote” in favor of sites, social media sharing represents a way for that voting behavior to continue. Social signals are emerging as ranking factors as search engines determine how to leverage our social interaction and behavior.
Sr: Social Reputation
Just as search engines don’t count all links equally, they don’t view all social accounts as being the same. This makes sense, since anyone can create a new account on a social network. What’s to prevent someone from making 100 different accounts in order to manufacture fake buzz?
Nothing, really, other than fake accounts like these can often be easy to spot. They may only have a handful of “quality” friends in their network and few might pass along material they share.
Ideally, you want to gain references from social accounts with good reputations. Having your own social presence that is well regarded is important. So participate on relevant social platforms in a real, authentic way, just as you would with your web site, or with customers in an offline setting.
Ss: Social Shares
Similar to links, getting quality social shares is ideal, but being shared widely on social networks is still helpful. Good things happen when more people see your site or brand.
Again, participation in social sharing sites is crucial. If you don’t have a Twitter account, a Facebook fan page or Google+ Page you’re missing out. You’re not building up a network that can help spread (aka share) your content, site and brand.
Nowhere is this more apparent than Google+, where
Ripples show how content is shared through the platform and winds up personalizing more and more search results.
Chapter 7: Trust, Authority, Identity & Search Rankings
If search engines can decide to trust links or social accounts, can they learn to trust web sites? Absolutely. Many SEOs believe that site trust plays a big role in whether a site will succeed or fail from a search perspective.
Ta: Authority
Is your site an authority? Is it a widely recognized leader in its field, area, business or in some other way? That’s the goal.
No one knows exactly how search engines calculate authority and, in fact, there are probably multiple “authority” signals. The type of links your site receives (lots of quality or ‘neighborhood’ links?) or social references (from respected accounts?) and engagement metrics (long clicks?) may all play a role in site authority. Of course, negative sentiment and reviews may hurt site authority as covered below:
There’s little doubt that search engines try to assess authority. One only needs to look through the questions Google told publishers to ask themselves in building high-quality sites that should be immune to “Panda” updates.
Th: History
Since search engines are constantly visiting your web site, they can get a sense of what’s “normal” or how you’ve behaved over time.
Are you suddenly linking out to what the search engines euphemistically call “bad neighborhoods?” Are you publishing content about a topic you haven’t typically covered? Such things might raise alarm bells.
Then again, sites do change just like people do, and often for the better. Changes aren’t taken in isolation. Other factors are also assessed to determine if something worrisome has happened.
Similarly, a site with a history of violating guidelines and receiving multiple penalties may find it more difficult to work their way back to search prominence.
In the end, a good overall track record may help you. An older, established site may find it can keep cruising along with search success, while a new site may have to “pay its dues,” so to speak, for weeks, months or even longer to gain respect.
Ti: Identity
Is that the real Benedict Cumberbatch on Twitter or someone impersonating him? Does a site claiming to be the ‘official’ site … really official? Who is the person giving legal advice on their blog? Are they even a lawyer!?
In the offline world it’s easier to figure these things out. We spot Benedict at a local coffee shop (no pictures please), we can tell instantly when we’re in a real Target store and you’ll likely take a look at the degrees on the wall when visiting a lawyer.
Search engines have found it increasingly important to ensure they’re dealing with the ‘right’ data. Amit Singhal, who oversees Google’s search efforts, made the importance of identity clear:
A good product can only be built where we understand who’s who and who is related to whom. Relationships are also important alongside content. To build a good product, we have to do all types of processing. But fundamentally, it’s not just about content. It’s about identity, relationships and content.
Identity takes many forms, from Google’s Authorship program to social profile verification on platforms such as Twitter and Facebook. While there’s clearly a debate around the balance of privacy and anonymity, search engines continue to seek out those willing to stand up and behind the content they produce.
Chapter 8: Personalization & Search Engine Rankings
Years ago, everyone saw exactly the same search results. Today, no one sees exactly the same search results, not on Google, not on Bing. Everyone gets a personalized experience.
Of course, there’s still a lot commonality. It’s not that everyone sees completely different results. Instead, everyone sees many of the same “generic” listings. But there will also be some listings appearing because of where someone is, who they know or how they surf the web.
Pc: Country
One of the easiest personalization ranking factors to understand is that people are shown results relevant to the country they’re in.
Someone in the US searching for “football” will get results about American football; someone in the UK will get results about the type of football that Americans would call soccer.
If your site isn’t deemed relevant to a particular country, then you’ve got no chance of showing up when country personalization happens. If you feel you should be relevant, then you’ll probably have to work on your international SEO.
Pl: Locality
Search engines don’t stop personalizing at the country level. They’ll tailor results to match the city or metropolitan area based on the user’s location.
As with country personalization, if you want to appear when someone gets city-specific results, you need to ensure your site is relevant to that city.
This is increasingly important as search becomes more prevalent on mobile devices and geolocation becomes a primary way of delivering more relevant results. In addition, Google’s Venice Update placed far more emphasis on sites who were physically located in that user’s area.
Today, if you’re looking for a dentist, you’ll find more individual dental practice sites in your area populating your search results rather than national directory sites.
Ph: Personal History
What has someone been searching for and clicking on from their search results? What sites do they regularly visit? Have they “Liked” a site using Facebook, shared it via Twitter or perhaps +1′d it?
This type of personal history is used by both Google and Bing to influence search results. Unlike country or city personalization, there’s no easy way to try and make yourself more relevant.
Instead, it places more importance on first impressions and brand loyalty. When a user clicks on a “regular” search result, you want to ensure you’re presenting a great experience so they’ll come again. Over time, they may seek out your brand in search results, clicking on it despite it being below other listings.
This behavior reinforces your site as one that they should be shown more frequently to that user. Even better if they initiate a social gesture, such as a Like, +1 or Tweet that indicates a greater affinity for your site or brand.
History is even more important in new search interfaces such as
Google Now, which will proactively present “cards” to users based on explicit preferences (i.e. – which sports teams or stocks do you track) and search history.
Ps: Social Connections
What do someone’s friends think about a web site? This is one of the newer ranking factors to impact search results. Someone’s social connections can influence what they see on Google and Bing.
Those connections are what truly matter because search engines view those connections as a user’s personal set of advisors. Offline, you might trust and ask your friends to give you advice on a restaurant or gardening.
Increasingly, when you search today search engines are trying to emulate that offline scenario. So if a user is connected to a friend and that friend has reviewed a restaurant or shared an article on growing tomatoes then that restaurant and article may rank higher for that user.
If someone can follow you, or easily share your content, that helps get your site into their circle of trust and increases the odds that others they know will find you. Nowhere is this more transformative than Google+, where circling a site’s Google+ Page will change the personalized search results for that user.
And if the rising percentage of (not provided) keyword data is any indication, the number of people getting personalized search results on Google is growing fast.
Chapter 9: Violations & Search Engine Spam Penalties
So far, we’ve discussed the positive signals that make up the Periodic Table Of SEO Success Factors. But there are also some negative factors to avoid.
A word of reassurance. Very few people who believe they’ve spammed a search engine have actually done so. It’s hard to accidentally spam and search engines look at a variety of signals before deciding if someone deserves a harsh penalty.
That said, let’s talk about things not to do!
Vt: “Thin” or “Shallow” Content
Responding to a drumbeat of complaints about poor search results, Google rolled out its “Panda” update in February 2011. Panda targets what is described as “thin” or “shallow” content or content that is lacking in substance.
This domain level penalty targets sites with a predominant amount of so-so content and essentially treats it similar to overt spam techniques.
Today it’s no longer a question of whether the content is simply relevant but whether it is valuable to the user.
Va: Ads / Top Heavy Layout
Have you ever been on a site and found it hard to find the actual content amid a slew of ads? Where’s the beef!
That’s what the Page Layout algorithm was meant to address. Matt Cutts, Google’s head of Webspam described it as follows:
… we’ve heard complaints from users that if they click on a result and it’s difficult to find the actual content, they aren’t happy with the experience. Rather than scrolling down the page past a slew of ads, users want to see content right away. So sites that don’t have much content “above-the-fold” can be affected by this change. If you click on a website and the part of the website you see first either doesn’t have a lot of visible content above-the-fold or dedicates a large fraction of the site’s initial screen real estate to ads, that’s not a very good user experience. Such sites may not rank as highly going forward.
Often referred to as Top Heavy, this penalty is reserved for sites that frustrate the user experience by placing an over abundance of ads before content. So don’t make your users search for the content.
Vs: Keyword Stuffing
It’s one of the oldest spam tactics on the books. Search engines say to use words you want to be found for on your pages. OK, I’ll give them those words over and over again! How about 100 times. In a row? That work for you, Google?
Actually, no, it doesn’t. But “keyword stuffing” like this could get you penalized.
How often is too often? There’s no correct answer here, but you’d really have to go to extremes to cause this penalty to kick in. It’s most likely to happen to non-SEOs who just don’t know better and might decide to paste a word many times in a row, typically at the bottom of a web page.
Vh: Hidden Text
Once you decide to keyword stuff, you’re next thought will probably be “Why don’t I hide all this text that no human wants to see.” You might make the text white, so it blends with a page’s background. In doing so, you will have spammed a search engine.
Search engines don’t like anything hidden. They want to see everything that a user sees. Don’t hide text, whether it be using styles, fonts, display:none or any other means that means a typical user can’t see it.
Vc: Cloaking
Let’s talk sophisticated hiding. How about rigging your site so that search engines are shown a completely different version than what humans see?
That’s called cloaking. Search engines really don’t like it. It’s one of the worst things you could do. Heck, Google’s even banned itself for cloaking. Seriously.
While most people are unlikely to accidentally spam a search engine, the opposite is true when it comes to cloaking. That’s why it’s such a heavy penalty, if you’re caught doing it. It’s a bait and switch and seen as a deliberate attempt to manipulate search results.
Vp: Paid Links
Speaking of Google banning itself, it also banned Google Japan, when that division was found to be buying links. For 11 months.
That’s longer than JC Penney was penalized (3 months) in 2011. But JC Penney suffered another penalty after having its paid link purchase splashed across a giant New York Times article. So did several large online florists. And Overstock got hammered via a Wall Street Journal article.
The debate over whether Google should act so aggressively against those who buy and sell links has gone on for years. The bottom line is to rank on Google, you have to follow Google’s rules — and the rules say no buying or selling links in a way that passes on search engine ranking credit.
If you choose to ignore Google’s rules, be prepared for little mercy if caught. And don’t believe programs that tell you they’re paid links are undetectable. They’re not, especially when so many of the cold-call ones are run by idiots.
As for Bing, officially it doesn’t penalize for paid links, but it frowns on the practice.
Vl: Link Spam
Tempted to run around and drop links on forums and blogs, all with highly optimized anchor text (like ‘louis vuitton handbags 2013′), with the help of automated software?
You suck.
You’re also not doing SEO, though sadly, all the people who hate the spam you leave behind get the impression that’s what SEO is about. So SEOs hate you too – with a passion.
If you do go ahead with it, most of the links won’t give you the credit you were thinking they would. On top of that, you can find yourself on the sharp end of a penalty.
This penalty has been given more weight in this version of the table based on the efforts Google has made in neutralizing and penalizing link spam and, in particular, the launch of the “Penguin” update.
If you’ve been caught dabbling on the dark side, or if a fly-by-night “SEO” company got your site in hot water you can disavow those links on both Google and Bing in hopes of redemption and a clean start.
Vd: Piracy / DMCA Takedowns
The “Pirate” update targeted sites infringing on copyright law. Under pressure from the Recording Industry Associate of America (RIAA), Hollywood powerhouses and governments, Google began to penalize sites who received a large number of Digital Millennium Copyright Act (DMCA) “takedown” requests.
It’s unlikely that most sites will have to deal with these issues, but you should handle any DMCA takedown notifications that show up in your Google Webmaster Tools account.