A client walked into a discovery call last year with a spreadsheet containing 340 directory submissions. Her face had that particular expression I’ve come to recognise — part exhaustion, part suspicion that she’d been sold something expensive and useless. She was right on both counts.
What follows is the framework I built reverse-engineering that mess, the 15 signals I now score every directory against before a client spends a penny or a minute on submission, and the constraints under which I’d adjust the weights. Steal it, adapt it, argue with it — but run something like it before your next citation campaign.
The Client That Sparked This Framework
A SaaS founder’s directory spreadsheet nightmare
The company sold project management software for architecture firms. ARR around £1.2M, small team, a previous agency engagement that had ended with the founder quietly uninstalling their Slack app. The spreadsheet had columns for URL, date submitted, login credentials, monthly fee, and — tellingly — a column called “still exists?” with quite a lot of “no” values.
Her brief was simple: “Tell me what to kill and what to keep.” She’d already spent about £8,400 across various pay-to-list schemes over eighteen months, plus untold hours of a junior marketer’s time copying NAP (name, address, phone — the trifecta of local citation data) into forms.
340 listings, 12 citations actually helping
I pulled the domains into Ahrefs, cross-referenced with Google Search Console referral data, and sorted by two things: did the directory actually send humans, and did Google treat the link as anything other than digital dandruff? The result was grim. Of 340 live listings, 12 generated any measurable referral traffic over the prior 90 days. A further 23 appeared to pass some link equity based on movement in target keyword rankings when we later tested with disavow experiments. The remaining 305 were — to use the technical term — noise.
Worse, 61 of them were on domains that Ahrefs flagged as part of interlinked networks, meaning one algorithmic update away from the kind of manual action that ruins your quarter.
Why we started auditing instead of submitting
Most directory engagements start with “where should we submit?” That’s the wrong question when the client’s backlink profile already looks like a landfill. We flipped it: audit everything first, cut ruthlessly, then identify the handful of new targets worth the submission friction.
Did you know? According to business directory, search engines use directory listings as trust signals when determining local search rankings — but only when the information is consistent across multiple respected directories. Inconsistency can actively hurt you.
First Pass: Killing the Obvious Farms
Before the 15-signal scoring even begins, there’s a triage. Roughly 70% of the 340 listings got binned at this stage without ceremony. Four quick heuristics do most of the work.
Footer link ratios above 60%
Open the directory’s homepage. View source. Count the outbound links in the footer versus the body. If the footer is dumping 40+ partner links to unrelated domains (payday loans, casino affiliates, “business services” in six languages) while the body has twelve listings for local plumbers, you’re looking at a link farm wearing a directory costume. I use a crude but effective check:
// In Chrome DevTools console
document.querySelectorAll('footer a').length /
document.querySelectorAll('a').lengthAnything above 0.6 warrants a hard look. Above 0.8, I don’t bother looking — it’s gone.
Directories accepting submissions within minutes
Submit a test listing with deliberately vague category selection and a description containing the phrase “quality services for discerning clients” (meaningless, the sort of thing a human editor should reject). If the confirmation email arrives before you’ve finished making coffee, there is no editorial process. A directory without editorial review is a comment section with better CSS.
The “pay $49 for instant approval” red flag
Paid listings aren’t automatically bad — plenty of legitimate directories charge for inclusion because editorial work costs money. The tell isn’t the fee; it’s the promise. “Instant approval upon payment” means the payment is the approval. That’s a transaction, not an endorsement. Compare this to directories that charge but still reserve editorial rejection rights — the price is for consideration, not publication.
Myth: Free directories are always better than paid ones because Google prefers “natural” links. Reality: Google prefers edited links. A paid directory with genuine editorial review and a rejection rate passes more trust than a free auto-approval farm. The fee often funds the editorial standard.
Reverse-lookup tricks that exposed three networks
Three of the client’s directories had suspiciously similar page templates. A reverse IP lookup via hackertarget.com revealed 47 “different” directories running on the same server, cross-linking aggressively. A WHOIS check on registrant details confirmed a shared owner behind a privacy shield. Same story for two other clusters.
Total kill from first pass: 241 listings queued for disavow, 52 marked for NAP correction and retention consideration, 47 carried forward to detailed scoring.
The 15 Signals, Weighted by Impact
Here’s the scoring rubric. I weight them because not all signals are equal — editorial friction and outbound link hygiene matter far more than, say, whether the directory has a blog.
| Signal | Weight | What Good Looks Like | Red Flag Threshold | How to Check |
|---|---|---|---|---|
| Editorial review depth | 15% | Multi-day review, occasional rejections | Approval under 1 hour | Test submission with marginal listing |
| Outbound link hygiene | 12% | Relevant, curated outbound targets | Footer dumps to unrelated niches | Manual sampling of 20 listings |
| Referral traffic evidence | 12% | Measurable clicks in GSC/GA4 | Zero clicks over 90 days | Search Console referral report |
| Category taxonomy depth | 8% | Genuine sub-categorisation | Flat list of 12 generic categories | Browse category tree manually |
| Geographic specificity | 7% | City/region-level filtering | “Worldwide” with no filters | Test location search functionality |
| Domain age and ownership | 7% | 5+ years, transparent owner | Privacy shield + new domain | WHOIS lookup |
| Indexation rate of listings | 7% | 80%+ of listings indexed | Under 30% indexed | site: operator sampling |
| Organic traffic to directory | 6% | Consistent growth trend | Flat or declining since 2022 | Ahrefs/Semrush traffic graph |
| Review/rating system integrity | 6% | Verified reviewers, moderation | All 5-star, no timestamps | Read 30 recent reviews |
| Listing profile depth | 5% | Photos, hours, descriptions, FAQs | Name + URL + nothing else | Inspect competitor listings |
| Schema markup presence | 4% | LocalBusiness or Organization schema | No structured data at all | Rich Results Test |
| Spam signals in listings | 4% | Real businesses, clean descriptions | Gambling, pharma, adult mixed in | Browse 50 random listings |
| Contact and support responsiveness | 3% | Real human replies within days | No reply or auto-responder only | Email a genuine question |
| Pricing transparency | 2% | Clear tiers, published terms | “Contact us” for basic inclusion | Check pricing page |
| Brand recognition in niche | 2% | Practitioners mention it unprompted | Never heard of it; no mentions | Ask five people in the industry |
Editorial review depth and submission friction
This is the single most durable signal I’ve found. Friction correlates with quality because it’s expensive to maintain — a real editor reading submissions can’t be scaled to a million listings cheaply, which is why farms don’t do it. When Jasmine Directory’s piece on low-quality listings talks about “clear submission guidelines, editorial policies, and evidence that listings are reviewed before publication,” that’s not marketing — that’s the moat.
Traffic patterns that reveal real user intent
Directories exist to be used. If nobody uses them, the link is just a link, and a link without traffic is a trust signal at best, link farm fodder at worst. I pull three data points: total organic traffic trend over 24 months, keyword diversity (are they ranking for hundreds of long-tail queries or ten head terms?), and branded search volume. A real directory has people typing its name into Google.
Category taxonomy and geographic specificity
A good directory thinks hard about how to organise listings. Services > Legal > Family Law > Divorce > Collaborative Divorce” with geographic filters is a library. “Business” as a top-level category with 40,000 listings is a pile.
Outbound linking hygiene across the domain
Sample 20 random listings and check where they link. If the directory’s listings point to coherent, same-niche businesses in plausible geographies, healthy. If page three shows a Birmingham solicitor next to a Philippines casino affiliate next to a CBD vape shop in Nevada, you have your answer.
Myth: Domain Authority (DA) above 50 means a directory is worth pursuing. Reality: DA is a third-party metric Moz invented; Google doesn’t use it. Plenty of DA 60+ directories are expired-domain resurrection projects. Editorial behaviour beats any single third-party score.
Scoring the Remaining 47 Candidates
Building a weighted spreadsheet in two hours
For each of the 47 surviving directories, I scored each signal 0–3 (0 = absent/bad, 3 = excellent), multiplied by the weight, and summed. Maximum score: 300. I set pass thresholds at 180 (keep/pursue), 120–179 (monitor, no further effort), below 120 (disavow or remove listing).
The scoring itself took two hours because most signals are fast to check once you know what you’re looking at. The tempting mistake is to over-engineer this — a half-decent weighted list you actually use beats a perfect one gathering dust.
Quick tip: Don’t score directories in isolation — score three at once in a horizontal comparison. Your brain calibrates faster when you can see “this one’s category tree is actually quite good compared to that one’s” rather than trying to hold absolute standards in your head.
Where Yelp, Clutch, and G2 actually landed
Predictable results for the usual suspects, with some surprises:
- G2: 268/300. Strong editorial, genuine user intent, review system integrity is industry-leading for B2B software. Obvious keep.
- Clutch: 241/300. The verified-review process drags the score up significantly; traffic patterns show real buyer intent in the agency/services space.
- Yelp: 198/300. Marginal for B2B SaaS specifically — Yelp’s audience isn’t buying project management software. Higher for a restaurant client, lower for this one.
- Capterra: 244/300. Same parent as G2’s competitors but a distinct audience; kept.
- Crunchbase: 212/300. Not a directory in the traditional sense but functions as one for SaaS; referral traffic was lower than expected.
The niche directory that outranked three majors
The surprise was a directory I’d never heard of: a specialist platform for AEC (architecture, engineering, construction) software. Small — maybe 400 listings total. But every signal scored high: manual editorial process (I got a polite rejection on my first test submission for insufficient detail), genuine traffic from the target buyer, schema markup throughout, outbound links exclusively to AEC-relevant domains. It scored 261/300.
More importantly, six months after we got listed, that single citation was driving more qualified demo requests than G2, Clutch, and Capterra combined. Not more traffic — more qualified traffic. A CAC calculation showed the one listing paid back 14x in the first quarter.
Did you know? According to Birdeye’s analysis of directory benefits, a well-positioned directory listing boosts online presence, improves search rankings, and drives referral traffic through reviews. The compounding effect of reviews on a small, high-intent directory often outweighs raw traffic volume on a major one.
Cutoff thresholds we defended to the client
The conversation where you tell a client you want to disavow 28 directories she paid for is not a fun one. I brought the scoring sheet, showed the reasoning, and — critically — framed it as risk management rather than pure SEO. “These aren’t helping. Three of them are actively dangerous. The cost of keeping them is low until an algorithmic update makes it catastrophic.” She signed off within 20 minutes.
Results After Six Months of Cleanup
28 disavowed, 19 retained, 4 newly pursued
Final numbers from the 47 that survived first-pass triage: 28 disavowed via Google’s disavow tool, 19 retained (with NAP corrections applied across all of them for consistency), and 4 new submissions to directories that scored above 200 but weren’t currently listed. Total new submission effort: about 14 hours including the two that required editorial back-and-forth.
Referral traffic up 217%, rankings held steady
Six months later, the picture:
- Directory referral traffic: up 217% (from a low base, admittedly — 340 useless listings sent almost nothing)
- Target keyword rankings: held steady for the top-20 priority terms; three moved from page 2 to page 1
- Qualified demo requests attributed to directory sources: up from 3/month average to 11/month
- Monthly directory spend: down from roughly £470 to £180
The ranking stability was the point I was most nervous about. Disavowing 28 links on a domain with a modest profile carries real risk. We staged the disavow file in two batches four weeks apart to isolate any negative signal. None appeared.
What the Ahrefs and GSC data confirmed
Ahrefs showed the obvious: referring domain count dropped, but domain rating actually ticked up by two points over the period because the toxic signals stopped dragging the average down. GSC showed cleaner impression data — fewer irrelevant queries where we’d been ranking accidentally for spam-adjacent terms due to association with the farm directories.
What if… you inherit a backlink profile so bad that disavowing the worst 80% would tank rankings short-term? In practice I’ve seen this twice. The answer is staged disavow over 8-12 weeks, combined with an aggressive campaign to build 3-5 high-quality citations per month during the same period. You’re replacing the foundation while the house is still standing on it. It’s slower and it costs more, but rip-and-replace in one go has burned clients I’ve watched from the sidelines.
Adapting the Checklist to Your Constraints
Running this on a £500 budget
If you can’t afford Ahrefs (£200+/month) or a full audit engagement, the framework still works — you just lean harder on free tools. Google Search Console gives you referral data. The site: operator gives you indexation rates. WHOIS lookups are free. Manual sampling of 20 listings takes 15 minutes per directory.
Skip the signals that require paid tools (organic traffic trend, domain rating) and double the weight on the ones you can check manually (editorial friction, outbound hygiene, listing profile depth). You’ll miss some nuance but catch 80% of the problems. For a small local business with 30-50 existing listings, this approach takes a weekend.
Industries where the weights shift dramatically
The weightings I gave aren’t universal. A few examples of how I’d adjust:
- Legal services: Review system integrity jumps to 15%; Avvo and similar platforms are primary acquisition channels and review authenticity matters enormously. Editorial friction drops slightly because legal directories tend to all have some editorial layer.
- Restaurants and hospitality: Geographic specificity jumps to 15%; review quantity matters more than quality distribution; schema markup becomes near-mandatory because local pack visibility depends on it.
- B2B SaaS: Brand recognition in niche jumps to 8%; category taxonomy matters more because buyers filter heavily; geographic specificity drops to 2% (who cares where the software is “based”?).
- Trades (plumbers, electricians): Listing profile depth and review integrity dominate; outbound linking hygiene matters less because buyers aren’t clicking around the directory.
Local service businesses versus B2B software
The fundamental split is intent. Local service buyers are in emergency or planning mode — they want a nearby provider they can trust, fast. Directory reviews, photos, and hours are the decision inputs. B2B software buyers are in research mode — they want comparison, depth, case studies, and peer validation. Directory category taxonomy and review integrity are the decision inputs.
As noted in Jasmine Directory’s analysis of directory purpose, quality directories allow businesses to showcase experience through detailed profiles, case studies, and portfolio sections — but the value of those features varies wildly by industry. A law firm’s case studies are gold on a legal directory and noise on a general business directory.
When a two-week timeline forces shortcuts
Sometimes the brief is “we’re relaunching in a fortnight, fix the directory situation.” My shortcut protocol:
- Day 1-2: Pull all existing backlinks from GSC. Sort by referral traffic descending.
- Day 3-4: Anything sending traffic gets kept and NAP-corrected. Anything on a flagged domain network gets disavowed without further scoring.
- Day 5-7: For the middle 60%, apply only the top five signals from the full checklist.
- Day 8-14: Identify the 3-5 highest-probability new directories for the niche and submit.
You’ll miss nuance. You’ll probably disavow one or two links you shouldn’t have, and keep one or two you should have killed. But you’ll get 85% of the outcome in 20% of the time, and sometimes that’s the job.
Myth: More directory citations always improve local SEO rankings. Reality: Beyond the 15-20 most relevant citations for your industry and geography, additional listings show diminishing returns and increasing risk. The goal is coverage of the directories your buyers actually use, not maximum count.
Principles That Outlast Algorithm Updates
Editorial friction as the durable signal
Every ranking algorithm I’ve watched evolve — from Panda to Penguin to the link spam updates and the helpful content system — has moved in the same direction: rewarding signals that are expensive to fake. Editorial review is expensive to fake. Genuine user traffic is expensive to fake. Reputation in a niche is expensive to fake. If you only remember one thing from this walkthrough, remember: anything easy to do at scale is something Google eventually discounts.
Why user intent beats domain authority metrics
I’ll say this bluntly: I don’t use DA or DR as primary signals anymore. I look at them, but they’re secondary. A directory with DR 45 that sends me qualified buyers beats a directory with DR 72 that sends me bots and tyre-kickers — every time, in every industry. Third-party authority metrics optimise for correlation with ranking; they don’t measure the thing I actually care about, which is whether a listing moves a business forward.
The best curated directories — places like Jasmine Directory and its peers that maintain editorial review — tend to score well on the signals that matter because they maintain that friction, not because they game a metric. The correlation with DA exists, but the causation runs through editorial behaviour.
The question to ask before any submission
Before submitting to any directory, ask one question: would I send a prospective customer to this page?
If the answer is yes — the page is useful, the directory is coherent, the listings around yours are legitimate competitors or complements — submit. If the answer involves hedging (“well, nobody would actually go there, but Google might count the link…”), you’ve already got your answer. You’re not building authority; you’re renting a cubicle in a digital slum and hoping the neighbourhood doesn’t get raided.
Did you know? The U.S. Small Business Administration recommends proper market research including directory selection as vital for business success. The SBA’s framing is instructive — they treat directory choice as a marketing decision, not an SEO tactic. That reframing alone changes the questions you ask.
Quick tip: Set a calendar reminder to re-run your directory audit every 12 months. Directories decline; ownership changes; editorial standards slip. The list you built this year won’t be the right list in three years, and the cost of finding out the hard way is higher than the hour it takes to re-score.
The client I opened with now runs this checklist herself, quarterly, on any new citation opportunity her team surfaces. That’s the real win — not the 217% traffic lift, not the disavow file, but the fact that she stopped being a mark. If you build the same discipline into your own workflow, the next time a cold emailer offers 500 directory submissions for £99, you’ll already know the answer, and you’ll save yourself the 18 months of cleanup I’ve just described.
Start with your existing citation profile this week. Score ten directories against the rubric. You’ll learn more about your backlink hygiene in two hours than any automated audit tool will tell you in a month.

