The Problem
Cold email at scale requires personalization to maintain deliverability and reply rates, but personalization data from static databases (Apollo, ZoomInfo, Clearbit) ages quickly. Companies pivot, rebrand, launch new products, and hire -- none of which appears in CRM data until months later. Sending emails that reference outdated company info signals laziness and tanks reply rates.
The Scavio Solution
Before each send batch, run Google and Reddit searches for each prospect's company. Extract fresh signals: recent news, product launches, community discussions, hiring posts. Use these as personalization variables. The data is hours old instead of months old, and the cost is $0.005-0.015 per prospect depending on search depth.
Before
Before search signal enrichment, a 2,000-prospect campaign used 3-month-old Apollo data for personalization. 15% of company descriptions were outdated (pivoted product, rebranded, or shut down). Bounce rate: 6.1%. Reply rate: 1.4%. Multiple prospects replied pointing out that the referenced product no longer existed.
After
After adding live search signals, each prospect gets 2-3 searches for fresh data. 2,000 prospects x 3 searches = 6,000 queries at $30 total. Outdated reference rate dropped to under 2%. Reply rate increased to 3.9%. Zero embarrassing replies about outdated product mentions.
Who It Is For
Cold email operators, growth teams, and outbound agencies who need fresh personalization data that static B2B databases cannot provide.
Key Benefits
- Live search signals replace months-old database records
- 3 searches per prospect at $0.015 total cost
- Reply rates improve from 1.4% to 3.9% with fresh personalization
- Reddit data surfaces community pain points for angle development
- Works with any cold email platform via CSV enrichment
Python Example
import requests, os, json
H = {'x-api-key': os.environ['SCAVIO_API_KEY']}
def enrich_prospect(company: str, domain: str) -> dict:
# Google for recent news
g = requests.post('https://api.scavio.dev/api/v1/search', headers=H,
json={'platform': 'google', 'query': f'{company} news 2026'}, timeout=10).json()
# Reddit for community signals
r = requests.post('https://api.scavio.dev/api/v1/search', headers=H,
json={'platform': 'reddit', 'query': company}, timeout=10).json()
google_snippets = [o.get('snippet', '') for o in g.get('organic', [])[:3]]
reddit_posts = [o.get('title', '') for o in r.get('organic', [])[:3]]
return {
'company': company, 'domain': domain,
'latest_news': google_snippets[0] if google_snippets else '',
'reddit_buzz': reddit_posts[0] if reddit_posts else '',
'signal_count': len(google_snippets) + len(reddit_posts),
}
prospects = [('Vercel', 'vercel.com'), ('Supabase', 'supabase.com')]
for company, domain in prospects:
data = enrich_prospect(company, domain)
print(f'{company}: {data["latest_news"][:100]}')
if data['reddit_buzz']:
print(f' Reddit: {data["reddit_buzz"][:80]}')JavaScript Example
const H = { 'x-api-key': process.env.SCAVIO_API_KEY, 'Content-Type': 'application/json' };
async function enrichProspect(company, domain) {
const [g, r] = await Promise.all([
fetch('https://api.scavio.dev/api/v1/search', { method: 'POST', headers: H,
body: JSON.stringify({ platform: 'google', query: `${company} news 2026` }) }).then(r => r.json()),
fetch('https://api.scavio.dev/api/v1/search', { method: 'POST', headers: H,
body: JSON.stringify({ platform: 'reddit', query: company }) }).then(r => r.json()),
]);
return {
company, domain,
latestNews: (g.organic || [])[0]?.snippet?.slice(0, 120) || '',
redditBuzz: (r.organic || [])[0]?.title?.slice(0, 80) || '',
};
}
const data = await enrichProspect('Vercel', 'vercel.com');
console.log(`${data.company}: ${data.latestNews}`);
if (data.redditBuzz) console.log(` Reddit: ${data.redditBuzz}`);Platforms Used
Web search with knowledge graph, PAA, and AI overviews
Community, posts & threaded comments from any subreddit