Extract clean text, markdown, or structured data from any URL. JavaScript rendering included. No proxies, no browser setup, no headless Chrome. Just send a URL.
// Click "Scrape" to extract contentHeadless browser renders the full page before extraction. SPAs, dynamic content, and lazy-loaded elements are captured.
Get results as clean text, Markdown (preserving headers, lists, links), raw HTML, or structured JSON with metadata.
Automatically extracts page title, meta description, Open Graph tags, canonical URL, and all links and images.
No rotating proxies, no IP bans to manage, no CAPTCHA solving. We handle all the infrastructure.
Markdown format is perfect for feeding into LLMs, RAG pipelines, and AI agents. Clean content without HTML noise.
Most pages scraped in under 3 seconds. Results are cached for repeated lookups of the same URL.
| Method | Endpoint | Description |
|---|---|---|
| GET | /api/scrape?url={url} | Scrape URL (default: markdown) |
| GET | /api/scrape?url={url}&format=text | Plain text extraction |
| GET | /api/scrape?url={url}&format=html | Raw HTML source |
| GET | /api/scrape?url={url}&format=json | Structured JSON with metadata |
# Scrape a URL as markdown
curl "https://agent-gateway-kappa.vercel.app/api/scrape?url=https://example.com" \
-H "Authorization: Bearer YOUR_API_KEY"
# Get structured JSON with metadata
curl "https://agent-gateway-kappa.vercel.app/api/scrape?url=https://example.com&format=json" \
-H "Authorization: Bearer YOUR_API_KEY"
const API_KEY = 'YOUR_API_KEY';
const BASE = 'https://agent-gateway-kappa.vercel.app';
// Scrape URL and get markdown
const target = encodeURIComponent('https://example.com');
const res = await fetch(
`${BASE}/api/scrape?url=${target}`,
{ headers: { 'Authorization': `Bearer ${API_KEY}` } }
);
const data = await res.json();
console.log(data.title); // Page title
console.log(data.content); // Markdown content
import requests
API_KEY = "YOUR_API_KEY"
BASE = "https://agent-gateway-kappa.vercel.app"
# Scrape URL and get markdown
r = requests.get(
f"{BASE}/api/scrape",
params={"url": "https://example.com"},
headers={"Authorization": f"Bearer {API_KEY}"}
)
data = r.json()
print(data["content"]) # Clean markdown
Feed web content into LLMs for question answering, summarization, and retrieval-augmented generation (RAG).
Extract competitor pricing, product details, reviews, and news articles for analysis and reporting.
Build news feeds, content curators, and monitoring dashboards by scraping multiple sources programmatically.
Extract meta tags, headers, links, and structured data from competitor pages for SEO auditing and benchmarking.
200 free credits. No proxies. No browser setup. Just clean content from any URL.
Get Started Free