DEV Community

Code Pocket
Code Pocket

Posted on • Originally published at westoeast.com

Tracking 47 Reddit comments through Perplexity citation rails

Reddit is one of the most-cited single domains in Perplexity for B2B-adjacent queries. That's not a controversial claim anymore; if you watch Perplexity outputs for any week, you'll see Reddit show up in the citation rail constantly. The harder question is whether you can intentionally contribute Reddit content that gets cited, or whether the citations are essentially a passive function of what already exists on the platform.

Over a five-week window in Q1 2026 we ran a small structured experiment. Forty-seven comments and posts, written by team members on their personal accounts (no astroturfing, no agency branding, no client mentions) on topics adjacent to our 40-prompt baseline. We tracked which of those 47 contributions later appeared in Perplexity citation rails over the following six weeks. The answer was lower than we hoped and more interesting than we expected.

The setup, briefly

I want to be specific about what we did and didn't do, because Reddit content experiments are easy to do unethically and we tried not to.

Each contribution was made by a real team member on an account they actually use. Each was on-topic for the subreddit, added genuine information from our own work or experience, and was upvoted or downvoted on its merits. We did not coordinate upvotes. We did not use ghost accounts. We did not link to client work. We did not link to westOeast properties from these comments. The brief to the team was "if you'd be embarrassed to have this comment quoted back to you in three years, don't post it."

We logged each contribution: subreddit, account, post type (top-level vs. reply), word count, presence of structured information (lists, numbers, named entities), and whether it received upvotes within 48 hours.

What we found

Eleven of the 47 contributions, or about 23%, appeared in at least one Perplexity citation rail during the tracking window. That sounds high. Caveat one: most of those citations were for queries that returned the parent thread (not necessarily our specific comment) as the cited URL. Caveat two: Perplexity citation rails are noisy, and on multiple re-runs of the same query, only six of the eleven cited threads showed up consistently. So the durable hit rate was closer to 13%.

A few patterns we noticed:

Comments with numbers in them cited more often. Comments that named specific tools or products cited more often. Top-level posts cited more often than nested replies, even when the nested reply had more upvotes. Subreddit choice mattered enormously: contributions in subs with high karma thresholds and active moderation cited at maybe 3x the rate of contributions in lower-quality subs, even when the comment quality was held constant in our internal review.

One thing that didn't work

We tried writing a small set of comments (n=8) that were designed specifically to be "Perplexity-friendly" — short, structured, with bulleted lists and named entities front-loaded. Zero of those eight were cited. Our hypothesis is that this style reads as Reddit-unnatural and got either downvoted or ignored by humans before the engine ever surfaced the thread. The lesson, which we should have seen earlier, is that the Reddit community is the upstream filter; if your comment fails on Reddit, it doesn't matter that it's Perplexity-shaped. We stopped doing this within two weeks.

What this means for B2B GEO work

I want to be very careful here. The agency I work with has a clear policy against fabricating Reddit accounts, mass-posting, or any of the patterns that get flagged as astroturfing. Reddit's moderators are good at catching this, and the platform-side cost of getting caught is real and reputation-permanent.

So the question isn't "how do we win Reddit citations." It's something more like: do team members have genuine expertise that, if shared honestly in a relevant community, would be useful enough to upvote and quote? In our testing the answer is sometimes yes, sometimes no. The contributions that got cited were the ones the team members would have wanted to write anyway. The contributions that flopped were the ones we'd manufactured.

That sounds like a soft conclusion, and it is. The hard conclusion is that Reddit citations in Perplexity are largely passive infrastructure: they reflect what's already there. You can contribute to it as a person. You probably can't engineer it as a brand without ethically and operationally crossing lines we won't cross.

A pattern that surprised us: the upvote-citation disconnect

I expected upvote counts to correlate with citation likelihood. They mostly didn't, in our test set.

Among the 47 contributions, the eight that received the highest upvote counts were cited at roughly the same rate as the contributions in the middle of the upvote distribution. The bottom of the distribution (contributions with negative or zero upvotes) cited essentially never, which makes sense; the platform itself was burying them.

What seemed to predict citation more than raw upvotes was a combination of: was the comment surfaced near the top of the thread (which is partly an upvote function but also a recency and reply-density function), was the subreddit one that Perplexity appears to crawl heavily, and did the surrounding thread have multiple substantive comments rather than just one popular reply.

This is a long way of saying: trying to game Reddit citations by chasing upvotes seems to miss the mechanism. The mechanism, as best we can tell, is closer to "is this a thread that an AI engine would consider an authoritative resource on this question," and authority at the thread level is a function of the whole conversation, not any single comment.

Subreddit selection as the dominant variable

If I had to pick one variable from the experiment that explained the most variance in citation outcomes, it would be subreddit choice.

A handful of subreddits in our test (we won't name them specifically because we don't want to create incentives to flood them) accounted for a disproportionate share of the citations we saw. These tended to be subs with active moderation, karma thresholds for posting, low rates of self-promotion, and topics that overlap with the kind of questions B2B users ask AI engines.

Subs without those traits cited at roughly noise-floor rates regardless of comment quality. Some of the team's better contributions, on quality criteria, were in lower-citation subs and went nowhere.

This has uncomfortable implications. It suggests that Reddit citation outcomes are partly a function of which communities your team members already participate in, which is partly a function of who you've hired, which is partly luck. You can't easily build this out by hiring; you'd have to hire people who were already credible in the right communities, and those people generally aren't looking for agency work.

Numbers we're not fully sure of

The 13% durable hit rate is from a small sample. We'd want n=200 contributions before claiming a generalizable rate. The 6-week tracking window may be too short; some Perplexity citations seem to take weeks to surface. And our subreddit selection wasn't randomized; we picked subs the team members already participated in, which biases toward higher-quality contributions.

We also didn't measure the second-order effects: did any of these Reddit contributions drive direct traffic, sign-ups, or other downstream outcomes for the contributing team members or for the clients in their topic areas? We didn't track that, partly because the experiment was scoped to citation tracking and partly because untangling causality on direct traffic from a Reddit comment is hard.

If you've run a similar experiment with cleaner methodology, I'd want to read it. If you're considering running one and you don't have a clear ethical line about astroturfing, please don't run it. The pollution costs everyone.

What was the last comment you saw on Reddit that you'd quote back to a client unprompted? That's the bar.


This field report was published by **westOeast, a B Corp certified marketing agency working on generative engine optimization for B2B SaaS. The methodology, framework, and data described here come from internal audits at westOeast across our client portfolio in 2025-2026. More field notes at westoeast.com.

Top comments (0)