DEV Community

Burton Smith
Burton Smith

Posted on

SEO and Web Components - 2023 Edition

tl;dr - Web components are SEO friendly!

I had read a number of articles that discussed SEO and web component compatibility and saw that many of them stated web components were not SEO friendly. These articles were at least a few years old so I thought they were out-of-date, but I didn't see anything recently that contradicted those articles. So, I decided to test some things out and see what the story was for myself.

Creating a Test Site

I created a simple web page with a static "vanilla" web component to test if search providers would render and index content in the shadow root. The site can be viewed here and the source code can be viewed here.

Google Test

Google commands the lion's share of search traffic, so this seemed like a good place to start the test.

After creating my page and adding the property to my Google Search Console account, I went to URL inspection and clicked on VIEW CRAWLED PAGE.

google search console dashboard with arrows indicating where to click

In the preview window, I can see that the <h1> and <h2> tags in the shadow DOM were rendered and can be indexed by Google's web crawler! πŸ‘

HTML markup displaying the rendered web components

Bing Test

Although Bing is not the most popular search engine, its index is used to power many other search engines including Yahoo!, DuckDuckGo, Neeva, and You.com.

I added my page to the Bing Webmaster Tools and selected URL Inspection, clicked the Live URL tab at the top of the page, and clicked the View Tested Page Button.

bing webmaster tools with indicators of where to click

When I look at the tested page's HTML, I notice that none of my custom elements have rendered their shadow DOM contents and I am seeing error messages about missing H1 tags... 😳

HTML markup displaying custom elements without their content being rendered

error message from bing webmaster tools stating there is a missing h1 tag

As a huge proponent of web components, I was extremely concerned about these results, so I reached out to some of my co-workers on the Bing team to find out what was going on and figure out how we could fix this.

The good news is that after looking into it, they told me the issue is a bug in the Bing Webmaster Tools and that the content does get rendered and indexed! πŸ™Œ

The great news was that they have prioritized the bug and are looking to have it resolved by the end of this month (August 2023)!!! πŸŽ‰πŸŽ‰πŸŽ‰

Conclusion

Do to a bug in some of the tools, there have been some misconceptions about how compatible web components are for SEO. Fortunately, that should be resolved soon and teams can continue using web components know their content is searchable!

Top comments (22)

Collapse
Β 
fayzakseo profile image
Itzik (Yitzhak) Fayzak β€’

Great case study! It proves that SEO isn't just about content, it's about Rendering Infrastructure. In 2026, the gap between what the crawler sees and what the tools report is where most SEOs get lost. Web Components are the future of structured UI, and as long as we monitor the Shadow DOM correctly, the 'SEO-friendly' debate is finally over.

Collapse
Β 
quincarter profile image
quincarter β€’

This is great news for web components!

Collapse
Β 
jessaimaya profile image
jessaΓ­ maya β€’

Hey, thank you for sharing; this is very helpful!Β 

Collapse
Β 
diana_malets_c318de3241db profile image
Diana Malets β€’ β€’ Edited

Interesting points on Shadow DOM. Even with perfect web components, Google often struggles to prioritize crawling if the domain authority is low. For a quicker way to improve indexing through existing niche articles without the manual outreach, I highly recommend this: wmlinks.net/en/niche-edits/ β€” it’s a practical tool once the technical side is sorted.

Collapse
Β 
lorenzkahl profile image
Lorenz Kahl β€’

Hi Burton,

thanks for testing this and for sharing. I just cloned your repo, just to see with my own eyes, maybe add some more tests and to check up on Bing.

Unfortunately Bing still gives me the "SEO issue found - H1 tag missing" warning. This still concerns me a bit and I'm still not sure if it's a good idea to build a <my-headline level="2" layoutLevel="3">An important H2 headline<my-headline> component hiding the semantic structure of my page regarding the initial HTML.

On the other hand passing the <h2> with the slotted content exposes my headline to styles coming from the global scope <my-headline layoutLevel="3"><h2>An important H2 headline</h2><my-headline> - which can only be safeguarded against using the ultimate weapon !important

Collapse
Β 
stuffbreaker profile image
Burton Smith β€’

Yeah, I agree. I would love to see that fixed ASAP. It looks like updating the Web Master Tools to use the indexed content rather than the page source is quite the process. I don't have an ETA on when that will be updated, but they have assured me the indexed content does execute and traverse the shadow DOM content.

Collapse
Β 
claudiomedina profile image
ClΓ‘udio Medina β€’

That's indeed good news coming from Bing side.
I've heard that Google Crawler provides limited "website crawling budget" when using the JavaScript enabled web crawler, do you know if the same applies to Bing?
That has been one of my main blockers to the use of Shadow DOM in large websites because it ends up not indexing ALL pages.

Collapse
Β 
stuffbreaker profile image
Burton Smith β€’

I'm not sure to what extent it indexes JS content, but it should execute web components on the client to index the shadow root contents.

Collapse
Β 
webketje profile image
webketje β€’ β€’ Edited

Hey I wondered about this too, thanks for sharing.
Though I have a follow-up concern, maybe source material for another article?
Even if the webcomponent inners are readable, what significance are they given in the document structure? For example, would Google Search generate an enriched search result with main navigation links if the was embedded in a custom webcomponent? If a page has 2 <h1>'s, with one in the shadow DOM of a webcomponent, which one will render as page title?

Collapse
Β 
stuffbreaker profile image
Burton Smith β€’ β€’ Edited

It renders the content and reads it like standard markup on the page, so the behavior should be the same as if you added 2 <h1>'s in the light DOM.

Collapse
Β 
jaredchristensen profile image
Jared Christensen β€’

Do the pages need to be server side rendered or are you testing client side pages?

Collapse
Β 
stuffbreaker profile image
Burton Smith β€’

Great question! These were all rendered client-side.

Collapse
Β 
e0206_healthpa profile image
Bell-Myers, Bruce M β€’

Some of our components are server-side hydrated and some are not. I wonder if Google will get confused and read the page too early, not waiting for client-side hydration to finish. Seems like that might be happening on some pages already...

Collapse
Β 
stuffbreaker profile image
Burton Smith β€’

All of my examples were client-side hydrated and it didn't have any issues with it.

Collapse
Β 
vvenv profile image
ζŽζ–‡ε―Œ β€’

What if loading the scripts with async?

Collapse
Β 
stuffbreaker profile image
Burton Smith β€’

Good question! I guess it would depend on how long it takes to load the async code. I'm not sure how long the bots are willing to wait for that code.

Some comments may only be visible to logged-in visitors. Sign in to view all comments.