At Propecta, we recommend an 80/20 approach to SEO: spend 80% of your time on engagement SEO tasks and 20% of your time on technical SEO tasks. It’s the engagement SEO work that makes the best use of keywords, answers user intent, creates a good user experience, and—overall—drives rankings, clicks, and revenue. No amount of technical SEO work will drive good traffic to bad content.
But certain technical tasks are still vitally important to an effective SEO strategy. Technical SEO issues can undermine all of your engagement SEO efforts if search engines aren’t able to find, crawl, and render all of the high-quality content you’re publishing.
Instead of neglecting engagement SEO to focus on low-value technical SEO tasks, focus your attention on the most important technical problems on your site.
Researching technical SEO tasks, or running an automated technical audit, can result in a daunting list of to-dos that could take much more than 20% of your team’s time. Instead of neglecting engagement SEO to focus on low-value technical SEO tasks, focus your attention on the most important technical problems on your site—the things that will prevent your SEO from really performing.
Before a page can appear in search results, it must be crawled and indexed by search engines. If a site is blocking search engines from finding or crawling important pages, it’s impossible to drive organic traffic to those pages. To determine if the most important pages are being indexed, conduct a search using the following query: site:[your domain name].
The search engine will return results for every page of your site that it has indexed. Browse through the list to make sure all critical pages have results. If they do, you’re golden. If they don’t, you may be unintentionally blocking search crawlers.
Crawl issues are commonly caused by:
If code or server issues aren’t causing the problem, it could be that the navigational structure of your site is the culprit. If important pages are buried deep within your site, it’s possible that search engine crawlers—and users—simply can’t find your important pages. Take time to review the navigational structure of your site, and move links for important pages to more prominent locations.
Slow-loading sites and pages create a cascade of negative impacts:
Two common culprits for delayed page load speeds include excessive tracking code and oversized image files.
To eliminate delays caused by excessive tracking code, put a system in place for tracking code installations. Google Tag Manager is an excellent tool for installing and deploying tracking codes. It allows for viewing and managing all of of the campaigns being tracked—from both Google and third-party systems—from a single source. Review tracking codes regularly and deactivate any that are not being used.
To eliminate delays caused by oversized image files, format image dimensions and sizes before uploading them. Uploading oversized image files and allowing CSS to reduce images to the proper size and resolution can severely impact page load speeds. Ideally, uploaded images should be sized to fit display containers, but when that’s not possible, there are some general guidelines that can be followed to reduce delays in load speeds:
Keep in mind that each site is different. While excessive tracking codes and oversized image files are frequent offenders, there are dozens of other issues that could be impacting page load speeds. To establish a prioritized list of offenders, run your site through GTmetrix. GTmetrix will analyze the speeds at which different page elements load, and will offer a grade and suggestions for improving page speeds.
If suggestions with low or failing grades can be resolved, those should be prioritized highly in your technical SEO to-do list.
Rich snippets, rich cards, and AMP carousels are becoming increasingly prevalent in search results. In 2016, Google added rich cards to search results for image searches and online courses, and added an AMP-only carousel featuring top news stories. Those were in addition to preexisting rich snippets for reviews, videos, and upcoming events—just to name a few.
To display rich snippets in search results or to appear as results in rich cards, sites must use schema markup or implement microdata. However, marking up all pages of a large site is a labor-intensive initiative that may or may not yield justifiable results. Instead, take time to identify important queries your site is currently ranking for. If rich snippets or cards appear for those queries, consider implementing microdata or schema markup on those pages.
Google penalties can be devastating to a site’s SEO. If Google has taken a manual action on your site, your rankings could be penalized, or you could be removed from search results altogether. There are several reasons why Google might penalize your site. Most are related to actions taken to intentionally deceive search engines, but there are a few that could be unintentional as well:
To find out if your site has been penalized, check for manual actions listed in Google Search Console:
If any manual actions are listed for your site, follow Google’s instructions for resolving the issue. You may need to get rid of content created by hackers, disavow unnatural backlinks, delete or revise shallow content pages, or remove spam comments left by site spammers.
Improperly configured redirects can cause severe issues, including loss of pagerank and loss of traffic. Common redirect problems include:
If you’re seeing a loss in traffic and ranking after a site redesign or noticing lots of new crawl errors in Google Search Console, take time to audit your redirects. By transitioning 302 redirects to 301s, resolving redirect loops, and ensuring deleted pages are pointing to related URLs, you may be able to recover lost traffic and rankings.
All entries in your sitemap should point to live pages on your site, not 404 errors, 301 or 302 redirects, or server errors (5XX). While Google says these errors in a sitemap aren’t a problem, Bing has little tolerance for sitemap errors. According to Bing’s Duane Forrester,
“Your Sitemaps need to be clean. We have a 1% allowance for dirt in a Sitemap. If we see more than a 1% level of dirt, we begin losing trust in the Sitemap.”
Perform occasional audits of your sitemap to ensure all pages are listed, and all listed pages are live on your site. Delete entries for any pages that have been removed, and update any entries that have been redirected with the new page URLs.
Review the renderings of your pages to see if all important components are displayed.
If the search engine doesn’t render your display network ads, it’s not an issue. If it doesn’t display important components, tools, or content, it’s not using those elements to determine your ranking. Discuss these issues with your development team to determine if there are more SEO-friendly ways to code important site components.
If you’ve invested in a machine-run technical SEO audit, you may have a long list of suggestions for issues that need to be resolved. Approach these tasks with caution: not all will result in measurable improvements to your traffic or rankings. Some will require a lot of effort but will produce little value. Save those for a rainy day, and spend your time focusing on important issues.
By focusing on the most important and impactful technical SEO tasks, you’ll support the efforts you’re putting into engagement SEO. With clean code that search engines are able to properly index, crawl, and render, technical and engagement SEO tasks work hand-in-hand to increase incoming organic traffic.
Like what you see?
Join marketers around the world to receive leading SEO insights and strategies!
You got it! More is coming your way!