Skip to main content

MemoryRequestQueueClient

Memory implementation of the request queue client.

No data is persisted between process runs, which means all requests are lost when the program terminates. This implementation is primarily useful for testing, development, and short-lived crawler runs where persistence is not required.

This client provides fast access to request data but is limited by available memory and does not support data sharing across different processes.

Hierarchy

Index

Methods

__init__

  • __init__(*, metadata): None
  • Initialize a new instance.

    Preferably use the MemoryRequestQueueClient.open class method to create a new instance.


    Parameters

    Returns None

add_batch_of_requests

  • Add batch of requests to the queue.

    This method adds a batch of requests to the queue. Each request is processed based on its uniqueness (determined by unique_key). Duplicates will be identified but not re-added to the queue.


    Parameters

    • requests: Sequence[Request]

      The collection of requests to add to the queue.

    • optionalkeyword-onlyforefront: bool = False

      Whether to put the added requests at the beginning (True) or the end (False) of the queue. When True, the requests will be processed sooner than previously added requests.

    Returns AddRequestsResponse

drop

  • async drop(): None
  • Drop the whole request queue and remove all its values.

    The backend method for the RequestQueue.drop call.


    Returns None

fetch_next_request

  • async fetch_next_request(): Request | None
  • Return the next request in the queue to be processed.

    Once you successfully finish processing of the request, you need to call RequestQueue.mark_request_as_handled to mark the request as handled in the queue. If there was some error in processing the request, call RequestQueue.reclaim_request instead, so that the queue will give the request to some other consumer in another call to the fetch_next_request method.

    Note that the None return value does not mean the queue processing finished, it means there are currently no pending requests. To check whether all requests in queue were finished, use RequestQueue.is_finished instead.


    Returns Request | None

get_metadata

get_request

  • async get_request(request_id): Request | None

is_empty

  • async is_empty(): bool

mark_request_as_handled

open

  • Open or create a new memory request queue client.

    This method creates a new in-memory request queue instance. Unlike persistent storage implementations, memory queues don't check for existing queues with the same name or ID since all data exists only in memory and is lost when the process terminates.


    Parameters

    • keyword-onlyid: str | None

      The ID of the request queue. If not provided, a random ID will be generated.

    • keyword-onlyname: str | None

      The name of the request queue. If not provided, the queue will be unnamed.

    Returns MemoryRequestQueueClient

purge

  • async purge(): None
  • Purge all items from the request queue.

    The backend method for the RequestQueue.purge call.


    Returns None

reclaim_request

  • Reclaim a failed request back to the queue.

    The request will be returned for processing later again by another call to RequestQueue.fetch_next_request.


    Parameters

    • request: Request

      The request to return to the queue.

    • optionalkeyword-onlyforefront: bool = False

      Whether to add the request to the head or the end of the queue.

    Returns ProcessedRequest | None