How to balance competing forces in SEO
We all want to demonstrate strong results for our clients or stakeholders. But sometimes, pushing to an extreme can undo our efforts. This is much more obvious between different disciplines and departments.
For example, designers (or UX / CRO specialists) may think they can increase a siteâs conversion rate by 10% by cutting content and giving a more streamlined look. But if that 10% increase in conversion rate comes at the cost of 20% of organic traffic intake, then itâs probably not a good trade.
These conflicts are common, especially between competing disciplines and roles. But even within one discipline, like SEO, similar issues can arise.Â
This article looks at some competing forces in SEO and how to approach them.
- Volume of URLs: Ranking footprint vs. crawl efficiency
- Links and content: Quality vs. quantity
- Keyword optimization: Sparse vs. spam
- User experience: Speed vs. functionality
- Regional deployment: Local focus vs. global reach
- Internal linking: Connected vs. cumbersome
Volume of URLs: Ranking footprint vs. crawl efficiency
Working on a large site with plenty of webpages? Some SEOs might think more pages and content items are synonymous with a broader indexing (and therefore ranking) footprint.Â
But more URLs on your site doesnât always equate to more potential ranking opportunities or organic traffic. This is especially applicable to sites that suffer from poor architecture.Â
For example, ecommerce sites that include a product category within a product-level URL which also allow products to be nested within multiple separate categories. In such a situation, you can end up with:
- Mysite.com/category-1/product-1/
- Mysite.com/category-2/product-1/
- Mysite.com/category-3/product-1/
Since all of the above resolve the same product page (product-1), there are now three URLs for the same page (duplicate content).Â
This means that Google will invariably end up (eventually, over time) crawling the same product three times. Two of those three crawls could have gone to different products or content. That content could then have gone on to rank.
So, in this situation, inefficient use of the crawl budget actually ends up harming the velocity at which new content is ranked.Â
Hopefully, Google will still crawl all the unique, distinct content eventually, but it may take longer. As new content is published, it will take longer to perform.
Several other scenarios can cause this same phenomenon.Â
For example, different filtering combinations on a site with faceted navigation may result in exponentially expanding volumes of parameter URLs spawning on a website.Â
A non-filtered category page may end up with ten or even 100 parameter variations as different filtering is applied.
We can just put canonical tags on the highly-duplicate pages that we donât want Google to index and that will handle the content duplication issues, right?Â
While that is true, Google still has to crawl and visit the non-canonical addresses to see that they are non-canonical (to read their embedded canonical tags).Â
Canonical tags only help to alleviate content duplication, but they donât really help much with crawl efficiency and content discovery.Â
You could argue that this is where you deploy complex wildcard robots.txt rules. Still, you must be careful in that area, as you can easily unintentionally cut off chunks of organic search traffic.
The best practice is to implement correct URL architecture and supporting redirects. If you have those in place, not much can go wrong.Â
Often, canonical tags are deployed as a band-aid solution after issues have already arisen. But theyâre quite a messy patch to a more fundamental problem.
Links and content: Quality vs. quantity
On the surface, this seems like a no-brainer. Google has constantly stated that quality content and links matter more than mass-manufactured spam.Â
SEOs and digital PR specialists can often spend weeks attempting to create great content and ascertain a single high-value placement to knock the competition off their ranking pedestal.
No SEO worth their salt would argue that mass-spun content and spam links are an effective tool. These tactics are ineffective if you expect to maintain a long-term online brand, a foundation of business that you can build upon over time.
So, is there a place for quantity on a quality-first web, where higher quality signals matter more?Â
Yes. If you have worked on large sites for enterprise-tier clients, youâll know that such brands (and those they compete with) already have high-quality links and content.Â
Quality never becomes irrelevant, but quantity once again raises its head.Â
For such clients, the game is about delivering a quantity of quality. In these situations, both dimensions (quantity and quality) matter.
When your site has such powerful ranking equity, every minute youâre not delivering new content, which targets new keywords, is lost time and traffic.Â
Every moment you get three high-value links, as your competitor wins 10, can be a moment of failure.
For such high-caliber sites and clients, the goalposts change entirely. The only way to earn high-value links quickly enough is to do very noticeable things in the real world, like:Â
- Sponsoring charities and speaking at educational institutions.Â
- High-impact PR stunts and the activity types that connect with an audience. (Hint: not web directory submissions!)Â
You stop thinking about building links and individual placements and start thinking, âHow can we go out there and do something newsworthy?â
While quality is sovereign, donât forget that quantity is still required within the most competitive spaces.Â
Quality real-world activity can deliver a quantity of quality links. Thatâs where you want to be.
Dig deeper: How to use digital PR to drive backlinks and business growth
Keyword optimization: Sparse vs. spam
A content gap analysis can present you with two primary findings.Â
- A keyword isnât ranking well enough because the connected content doesnât make enough of that concept (gap âinâ content).
- Thereâs a missing page that you need to create for your website (gap âofâ content).Â
In the former scenario, youâre likely to open an existing page and work out where you could deploy the underperforming keyword.Â
Or you might go further and determine whether an additional content section is required.
Either way, youâre browsing a content page and looking for a keyword deployment opportunity. And whatâs wrong with that? Itâs what weâre paid to tweak content and get each item of content performing to its optimal standard.Â
We want to identify sparse, thin, underperforming content, which doesnât say enough.Â
But itâs a thin line between sparse content with too few referenced topics (too few keywords) and spam content, which is nothing but keyword injections.
Even before Googleâs well-known Panda update, there were attempts to curb the âkeyword enthusiasmâ of SEOs.Â
Content that doesnât contain atopical relevance doesnât have the weight to penetrate Googleâs SERPs. By contrast, content that is too optimization-heavy sinks.
Keep these competing forces in mind when optimizing or reducing the optimization intensity of your content. Your content must be heavy enough to penetrate but not so heavy that it sinks.
Get the daily newsletter search marketers rely on.
See terms.
User experience: Speed vs. functionality
WordPress is known as an SEO-friendly content management system (CMS). But often, some site wonders want more functionality than the default CMS, so they start installing many plugins.Â
Fairly quickly, site performance deteriorates as pages load slower and slower.Â
Shortcode must be queried and transmuted to HTML / CSS, which involves additional calls to various tables from the database.Â
Additional scripts pile up in the browserâs main thread, creating execution bottlenecks.
Getting a good balance of page-loading speeds and functionality was fairly easy in the past.Â
As long as you minified your scripts and sheets, compressed your images and installed a caching plugin, you were good to go.Â
Those days are over.Â
Nowadays, Google wants us to begin interpreting what happens on the clientâs (end-userâs) browserâs main processing thread. Thereâs no point shipping 5-10 scripts to a user really efficiently if all of that JavaScript waits in the browserâs main processing thread to be executed.Â
As such, we now have to consider:Â
- JavaScript code audits.
- Intelligent JavaScript deployment (only call scripts on the pages where they are needed).
- Server-side rendering (SSR).
- JavaScript parallelization.
You can still achieve high functionality combined with high speed. It just takes a lot more work (and intelligence) than previously.Â
Forging an effective critical JavaScript/CSS rendering path is not for the faint of heart.Â
If you can spare the senior development time, you can have a relatively feature-rich and fast site running on mediocre hosting.Â
It will take more time than ever, so be prepared.Â
Regional deployment: Local focus vs. global reach
This is a trap that can spring both ways.Â
You can aim for global reach without sufficient content, architecture and authority. In such a situation, you may wish you had picked a more localized domain (national) instead.
You may wish you had aimed at your local area with NAP signals. Sometimes itâs better to walk before you can run, and over-extending your reach too quickly can lead to failure (success on neither local nor global fronts).
On the other hand, going with a local approach when you have global ambition can really lock you in. For example, itâs unlikely that a .co.uk (UK) domain would rank well in France or Germany.Â
Itâs important to realize that none of these decisions are concrete. If you lock yourself down locally, you can buy new domains and perform site migrations.Â
In such a situation, youâll likely lose at least a little of your ranking power, so you should only jump ship (from one domain to another) once your site has gained critical mass.Â
If youâre only seeing a few hundred organic sessions monthly, itâs probably not time to make that move yet.
Depending on your ambitions, a local or global approach may be better.Â
If youâre a local vacuum repair shop, no one is likely to be critically interested in your business. Aiming for global SEO might be a bit of a reach.Â
If there are only two other vacuum shops in your local area, a locally targeted SEO campaign would almost guarantee you ranking on top of the search results for relevant, local terms. Thatâs much less effort than reaching out to potential consumers across the seas.Â
If youâre a well-known fashion brand and branching out from clothing to other items like scents (cologne), youâd probably expect some business from other countries.
Take the actions that will bring you the fastest revenue with the least effort.Â
If youâre a small business without enough ranking power to rank globally, go local first and circle back later. Otherwise, aim high and chip away.
Internal linking: Connected vs. cumbersome
Adding a few choice links within your content is great, perhaps to support orphaned pages or top-performing products. That said, thereâs such a thing as having too many internal links.Â
Imagine a page where every other word or phrase was linked to a destination URL. How would you determine where to go?Â
It would seem like every snippet of text was competing for your attention equally.
This would be problematic for end users of your website. The same can also be said for search engines.Â
If every text item is a link to somewhere else and every page on your site supplies and receives thousands of links, how can a search engine interpret which pages are more or less valuable?Â
Even contextual analysis and thematic page categorization become much trickier.
Navigating competing forces in SEO
Balancing competing forces in SEO requires a strategic approach.
- Volume of URLs: Prioritize efficient crawling over excessive content. Optimize URL architecture and redirects to prevent duplication and wasted crawl budget.
- Links and content: Emphasize quality while considering the quantity needed in competitive spaces. Focus on delivering high-value content and seeking impactful real-world link building opportunities.
- Keyword optimization: Strive for a balance between sparse content and keyword stuffing. Ensure topical relevance without overwhelming optimization.
- User experience: Aim for a blend of speed and functionality. Optimize JavaScript deployment, consider server-side rendering, and manage execution bottlenecks for a smooth user experience.
- Regional deployment: Tailor your approach based on goals. For local focus, prioritize NAP signals. For global reach, focus on content, architecture, and authority.
- Internal linking: Maintain a connected structure without overwhelming users or search engines. Include relevant links to enhance navigation while avoiding link overload.
Adopting a holistic and adaptable strategy that respects the nuances of each force can help you make your SEO efforts more manageable.
The post How to balance competing forces in SEO appeared first on Search Engine Land.



Recent Comments