In my experience, severe panda-related hits tend to boil down to a root cause of either duplicate content, thin content, or extremely poor user experience.
As I’ve already covered many of the other areas involved in recovering from panda this month, I wanted to focus on thin content – what it is, how to spot it, and most importantly, how to fix it using Google Analytics.
Here is an illustration of what thin content looks like. When I search for a two bedroom house in East Sussex, and scroll down to the one hundredth result, I begin to see results like this one where I’m directed to a page that offers absolutely no value whatsoever.
So how can you find out whether your site is being affected by thin content?
Well, as of a few days ago you can now find out in the Webmaster Tools ‘manual actions’ tab. However, I have to admit I am skeptical of this – I’ve looked at several sites in WMT that I know have thin content issues, and yet no notifications have come up.
I recently had the challenge of fixing thin content issues on a 1.5m page site with approximately 75,000 pages of what I would describe as low quality thin content.
While I’m sure there are numerous clever ways of coming to the same conclusion, I’m going to share my approach which I hope will at least provide a starting point to help you identify and rectify your thin content.
1. Define thin content quantitatively
The hardest part about identifying thin content is getting past the subjective nature of what is or isn’t considered ‘thin’.
In my analysis I decided to create a weighted formula that shortlisted a page for being considered thin if it had all of the following characteristics:
- A bounce rate between 95 and 99.99% (here’s why I excluded pages with a 100% bounce rate).
- An average time on page between 0.1 and five seconds.
You can use the following formula to work this out
=IF(CELL WITH BOUNCE RATE < 95%, "Not Thin", IF(CELL WITH AVERAGE TIME ON SITE < 5, "Thin", "Not Thin"))
Once you’ve got this shortlist of pages that are performing poorly, you can begin looking for trends. Which types of pages, or sections of your site are causing trouble?
Try to find common patterns in the URL structure, and get an understanding from a user’s perspective why these pages might be causing people to bounce straight away.
2. Rectifying thin content
There is no right or wrong way to rectify thin content, so let me go through various options with an example.
Below are the metrics for a page on MusicJobBoard.com, a site that I use for testing purposes from time to time. As you can see, this page has a combined high bounce rate and low time on page, qualifying this page to be shortlisted as ‘thin content’ by my definition above.
Here is the page, looking rather thin.
Job boards like this one are typically susceptible to thin content, as if no one posts a job in a certain category, the category page can remain indexed despite providing a poor result for someone looking for what the page would usually offer.
In this case, we could apply a rule that would noindex the page if it reached 0 results. I personally don’t like doing this, but on larger sites I’ve seen it work as a pretty effective strategy for keeping low quality results out of the SERPs.
Alternatively, we could design the page in a way that provided value even if no jobs were present – e.g. providing links to see similar jobs in audio production, or even offering some cool information on average salaries for this type of job, as Indeed does.
Another option, in this instance, would be to try and find an ongoing job listing for every category page i.e. a recording studio that is will to receive CVs on an ongoing basis.
One interesting option, which I’ve seen used by several property aggregator services is to redirect people to the next best result i.e. rather than showing me an empty page with 0 property listings in street X, send me to the page on street Y 200 meters away.
A more agreed-upon approach is to merge your pages. Rather than having a page on every single street in the UK, with many being empty, you could merge your street pages to post code pages, or town pages.
This goes against the idea of targeting the long tail with dedicated pages, but without quality in place I think it’s fair to say that that ship has sailed anyway.
One final option is to simply remove your poor performing pages. If there really is no point merging the pages or trying to improve them, it may be worth considering hacking off the low quality content and investing your efforts on improving your best content instead.
Thin content is a tricky issue to define and tackle, which is perhaps why it’s not covered in quite as much detail as more objective site quality issues. I’d love to hear how others are tackling it, and whether there are any other creative solutions that I’ve missed above.
Feel free to leave a comment below, send me a tweet, or drop me an email on marcus (at) ventureharbour.com.
Econsultancy's Crunch - Data, Analytics and the Rise of the Marketing Geek, takes place on October 10 at Truman Brewery, London. Crunch is the event for the analysts, strategists and boffins who turns raw numbers into insight, then revenue. This event is one of five that make up our week-long Festival of Marketing.