Skip to main content

Crawlee is a web
scraping and browser
automation library

Crawlee is a web
scraping
and browser
automation
library

It helps you build reliable crawlers. Fast.

pipx run crawlee create my-crawler

Reliable crawling 🏗️

Crawlee won't fix broken selectors for you (yet), but it helps you build and maintain your crawlers faster.

When a website adds JavaScript rendering, you don't have to rewrite everything, only switch to one of the browser crawlers. When you later find a great API to speed up your crawls, flip the switch back.

It keeps your proxies healthy by rotating them smartly with good fingerprints that make your crawlers look human-like. It's not unblockable, but it will save you money in the long run.

Crawlee is built by people who scrape for a living and use it every day to scrape millions of pages. Meet our community on Discord.

JavaScript & TypeScript

We believe websites are best scraped in the language they're written in. Crawlee runs on Node.js and it's built in TypeScript to improve code completion in your IDE, even if you don't use TypeScript yourself. Crawlee supports both TypeScript and JavaScript crawling.

HTTP scraping

Crawlee makes HTTP requests that mimic browser headers and TLS fingerprints. It also rotates them automatically based on data about real-world traffic. Popular HTML parsers Cheerio  and JSDOM are included.

Headless browsers

Switch your crawlers from HTTP to headless browsers in 3 lines of code. Crawlee builds on top of Puppeteer and Playwright and adds its own anti-blocking features and human-like fingerprints. Chrome, Firefox and more.

Automatic scaling and proxy management

Crawlee automatically manages concurrency based on available system resources and smartly rotates proxies. Proxies that often time-out, return network errors or bad HTTP codes like 401 or 403 are discarded.

Queue and Storage

You can save files, screenshots and JSON results to disk with one line of code or plug an adapter for your DB. Your URLs are kept in a queue that ensures their uniqueness and that you don't lose progress when something fails.

Helpful utils and configurability

Crawlee includes tools for extracting social handles or phone numbers, infinite scrolling, blocking unwanted assets and many more. It works great out of the box, but also provides rich configuration options.

Try Crawlee out 👾

before you start
Crawlee requires Python 3.9 or higher.

The fastest way to try Crawlee out is to use the Crawlee CLI and choose the Getting started example. The CLI will install all the necessary dependencies and add boilerplate code for you to play with.

pipx run crawlee create my-crawler

If you prefer adding Crawlee into your own project, try the example below. Because it uses PlaywrightCrawler we also need to install Playwright. It's not bundled with Crawlee to reduce install size.

pip install 'crawlee[playwright]'
import asyncio

from crawlee.playwright_crawler import PlaywrightCrawler, PlaywrightCrawlingContext


async def main() -> None:
# Create a crawler instance and provide a request provider (and other optional arguments)
crawler = PlaywrightCrawler(
# headless=False,
# browser_type='firefox',
)

@crawler.router.default_handler
async def request_handler(context: PlaywrightCrawlingContext) -> None:
data = {
'request_url': context.request.url,
'page_url': context.page.url,
'page_title': await context.page.title(),
'page_content': (await context.page.content())[:10000],
}
await context.push_data(data)

await crawler.run(['https://crawlee.dev'])

# Export the whole dataset to a single file in `./result.csv`.
await crawler.export_data('./result.csv')

# Or work with the data directly.
const data = await crawler.get_data()
print(data.items) # TODO any alternative to `console.table()` in python?


if __name__ == '__main__':
# Add first URL to the queue and start the crawl.
asyncio.run(main())