Skip to main content

Result storage

This guide explains the available result storage options in Crawlee and how to use them to store and retrieve data.

Introduction

Crawlee provides multiple result storage options, each suited to specific use cases. By default, data is saved to a local directory specified by the CRAWLEE_STORAGE_DIR environment variable. If this variable is not set, Crawlee defaults to using ./storage within the current working directory.

Result storages are managed by storage clients, which are subclasses of the BaseStorageClient. For example, the MemoryStorageClient stores data in memory and can also offload them to the local directory. Data are stored in the following directory structure:

{CRAWLEE_STORAGE_DIR}/{result_storage}/{QUEUE_ID}/
note

Local directory is specified by the CRAWLEE_STORAGE_DIR environment variable with default value ./storage. {QUEUE_ID} is the name or ID of the specific storage. The default value is default, unless we override it by setting the CRAWLEE_DEFAULT_DATASET_ID or CRAWLEE_DEFAULT_KEY_VALUE_STORE_ID environment variables.

Dataset

The Dataset is designed for storing structured data, where each entry has a consistent set of attributes, such as products in an online store or real estate listings. Think of a Dataset as a table: each entry corresponds to a row, with attributes represented as columns. Datasets are append-only, allowing you to add new records but not modify or delete existing ones. Every Crawlee project run is associated with a default dataset, typically used to store results specific to that crawler execution. However, using this dataset is optional.

By default, data are stored using the following path structure:

{CRAWLEE_STORAGE_DIR}/datasets/{DATASET_ID}/{INDEX}.json
  • {CRAWLEE_STORAGE_DIR}: The root directory for all storage data specified by the environment variable.
  • {DATASET_ID}: The dataset's ID, "default" by default.
  • {INDEX}: Represents the zero-based index of the record within the dataset.

The following code demonstrates basic operations of the dataset:

import asyncio

from crawlee.storages import Dataset


async def main() -> None:
# Open the dataset, if it does not exist, it will be created.
# Leave name empty to use the default dataset.
dataset = await Dataset.open()

# Push a single row of data.
await dataset.push_data({'foo': 'bar'})

# Push multiple rows of data (anything JSON-serializable can be pushed).
await dataset.push_data([{'foo': 'bar2', 'col2': 'val2'}, {'col3': 123}])

# Fetch all data from the dataset.
data = await dataset.get_data()
# Do something with it...

# Remove the dataset.
await dataset.drop()


if __name__ == '__main__':
asyncio.run(main())

Key-value store

The KeyValueStore is designed to save and retrieve data records or files efficiently. Each record is uniquely identified by a key and is associated with a specific MIME type, making the KeyValueStore ideal for tasks like saving web page screenshots, PDFs, or tracking the state of crawlers.

By default, data are stored using the following path structure:

{CRAWLEE_STORAGE_DIR}/key_value_stores/{STORE_ID}/{KEY}.{EXT}
  • {CRAWLEE_STORAGE_DIR}: The root directory for all storage data specified by the environment variable.
  • {STORE_ID}: The KVS's ID, "default" by default.
  • {KEY}: The unique key for the record.
  • {EXT}: The file extension corresponding to the MIME type of the content.

The following code demonstrates the usage of the KeyValueStore:

import asyncio

from crawlee.storages import KeyValueStore


async def main() -> None:
# Open the key-value store, if it does not exist, it will be created.
# Leave name empty to use the default KVS.
kvs = await KeyValueStore.open()

# Set a value associated with 'some-key'.
await kvs.set_value(key='some-key', value={'foo': 'bar'})

# Get the value associated with 'some-key'.
value = kvs.get_value('some-key')
# Do something with it...

# Delete the value associated with 'some-key' by setting it to None.
await kvs.set_value(key='some-key', value=None)

# Remove the key-value store.
await kvs.drop()


if __name__ == '__main__':
asyncio.run(main())

To see a real-world example of how to get the input from the key-value store, see the Screenshots example.

We offer several helper functions to simplify interactions with result storages:

  • The push_data function allows you to manually add data to the dataset. You can optionally specify the dataset ID or its name.
  • The get_key_value_store function retrieves the key-value store for the current crawler run. If the KVS does not exist, it will be created. You can also specify the KVS's ID or its name.

Cleaning up the storages

Default storages are purged before the crawler starts, unless explicitly configured otherwise. For that case, see Configuration.purge_on_start. This cleanup happens as soon as a storage is accessed, either when you open a storage (e.g. using Dataset.open, KeyValueStore.open) or when interacting with a storage through one of the helper functions (e.g. push_data), which implicitly opens the result storage).

import asyncio

from crawlee.configuration import Configuration
from crawlee.http_crawler import HttpCrawler, HttpCrawlingContext


async def main() -> None:
config = Configuration(purge_on_start=False)
crawler = HttpCrawler(configuration=config)

@crawler.router.default_handler
async def request_handler(context: HttpCrawlingContext) -> None:
context.log.info(f'Processing {context.request.url} ...')

await crawler.run(['https://crawlee.dev/'])


if __name__ == '__main__':
asyncio.run(main())

If you do not explicitly interact with storages in your code, the purging will occur automatically when the BasicCrawler.run method is invoked.

If you need to purge storages earlier, you can call MemoryStorageClient.purge_on_start directly. This method triggers the purging process for the underlying storage implementation you are currently using.

import asyncio

from crawlee.memory_storage_client import MemoryStorageClient


async def main() -> None:
storage_client = MemoryStorageClient()
await storage_client.purge_on_start()


if __name__ == '__main__':
asyncio.run(main())