Web Search
Google, Bing, DuckDuckGo, Brave aggregated and re-ranked into one feed. Cloudflare-clean.
POST /v1/search
One bearer token. Multi-engine web search, RAG answers with citations, Reddit, Google Maps reviews, podcast transcripts — the Cloudflare bypass and proxy rotation are someone else's problem now.
the catalog
Every endpoint returns a stable schema. The proxy tier, JS rendering, cache TTL, and Cloudflare bypass are picked for each source — you just see clean JSON.
Google, Bing, DuckDuckGo, Brave aggregated and re-ranked into one feed. Cloudflare-clean.
POST /v1/search
Search + auto-fetch the top N pages with extracted text. One round-trip for RAG.
POST /v1/search/deep
RAG-style. Search, fetch, summarise with numbered [1][2] citations.
POST /v1/search/answer
Point us at a page. We get past the wall — rotated proxies, headless rendering, Cloudflare challenges — and hand back the rendered HTML and clean text.
POST /v1/scrape/url
28 languages. Flat per-call price — no per-character or per-word billing. A drop-in alternative to DeepL or Google Translate.
POST /v1/translate
Search like a browser. Filter by sub, sort, time. Real posts, ranked.
POST /v1/scrape/reddit/search
One URL, get the post + full ranked comment tree.
POST /v1/scrape/reddit/post
Every recent review for a place_id. Author, rating, owner response, posted_at.
POST /v1/scrape/google-maps/reviews
Timestamped, speaker-tagged. Apple, Overcast. 30-day cache.
POST /v1/scrape/podcast/transcript
Title, price, currency, shipping, seller, rating, variants.
POST /v1/scrape/aliexpress/product
Catalog grows. Vote on the next endpoint or pull anything else through the generic URL fetcher.
anatomy of a request
You write the curl. We do the unglamorous half — rotating proxies, solving Cloudflare challenges, parsing five flavours of broken HTML — and hand back stable, versioned JSON.
Auth verified in ~50 ms. Rate-limited per key, per minute, per day. Failed requests don't bill.
Datacenter for soft targets, residential for hostile ones, a real headless browser for challenge walls. Tier is picked per-domain, not per-customer.
Same request inside the TTL? Cached hit at 100 ms for 1 credit. Otherwise we fetch, extract, normalise, and cache it for next time.
Stable response schema per endpoint. Pre-built scrapers don't make you write selectors. Failures return credits.
curl -X POST https://scrapenest.dev/v1/search \ -H "Authorization: Bearer sn_••••" \ -H "Content-Type: application/json" \ -d '{ "query": "best wireless earbuds 2026", "num_results": 5 }'
{
"query": "best wireless earbuds 2026",
"results": [
{
"url": "https://www.rtings.com/headphones/reviews/best/...",
"title": "The 6 Best Wireless Earbuds - Spring 2026",
"snippet": "Our top pick is the Sony WF-1000XM5...",
"engine": "google"
},
/* 4 more */
],
"took_ms": 640,
"cache_hit": false,
"credits_charged": 3
}
pricing
Every plan has every endpoint, every proxy tier, JS rendering. The higher the bundle, the cheaper each credit. Failed requests refund automatically.
Kick the tires
Side projects
For shipping
For teams
Data teams
No commit. $0.0002 per credit, billed weekly for what you used. Same endpoints, same proxy tiers.
Need 25M+ credits or a dedicated VPC? Talk to sales. Plans cancel anytime, no contracts.
questions
Those are generic scraping APIs — you bring the URL and the selectors and you write the parsers. ScrapeNest centres on pre-built endpoints for specific high-value sources (Google Maps reviews, Reddit posts, podcast transcripts, AliExpress products). Stable schema, no selectors, credit cost matches what the call actually used. The generic scrape/url covers the long tail.
If our proxy chain is exhausted and the source still refuses (CAPTCHA wall, geofence, hard-block), the request returns 502 and the credits return to your balance inside 60 seconds. Failures don't bill.
No. Proxy bandwidth is bundled. We route each call through the tier that fits the source automatically — datacenter for soft targets, residential for hostile ones, headless browser when there's a challenge page. Override per-request with proxy_tier.
Yes — that's a primary use case. Growth and Scale customers run daily ingest jobs into RAG indexes and fine-tuning corpora. Talk to us about volume rates above 500k req/mo.
Blended target is 60–65% across paid traffic. Cached hits bill at 1 credit regardless of which tier originally fetched. Bypass with cache_ttl_seconds: 0 when you need live data.
1,000 free credits. No card. The API key prints once — copy it, paste it, you're done.