Skip to main content
Version: 3.11

LinkeDOMCrawlingContext <UserData, JSONData>

Hierarchy

  • InternalHttpCrawlingContext<UserData, JSONData, LinkeDOMCrawler>
    • LinkeDOMCrawlingContext

Index

Properties

addRequests

addRequests: (requestsLike: readonly (string | ReadonlyObjectDeep<Partial<RequestOptions<Dictionary>> & { regex?: RegExp; requestsFromUrl?: string }> | ReadonlyObjectDeep<Request<Dictionary>>)[], options?: ReadonlyObjectDeep<RequestQueueOperationOptions>) => Promise<void>

Add requests directly to the request queue.


Type declaration

    • (requestsLike: readonly (string | ReadonlyObjectDeep<Partial<RequestOptions<Dictionary>> & { regex?: RegExp; requestsFromUrl?: string }> | ReadonlyObjectDeep<Request<Dictionary>>)[], options?: ReadonlyObjectDeep<RequestQueueOperationOptions>): Promise<void>
    • Parameters

      • requestsLike: readonly (string | ReadonlyObjectDeep<Partial<RequestOptions<Dictionary>> & { regex?: RegExp; requestsFromUrl?: string }> | ReadonlyObjectDeep<Request<Dictionary>>)[]
      • optionaloptions: ReadonlyObjectDeep<RequestQueueOperationOptions>

        Options for the request queue

      Returns Promise<void>

body

body: string | Buffer

The request body of the web page. The type depends on the Content-Type header of the web page:

  • String for text/html, application/xhtml+xml, application/xml MIME content types
  • Buffer for others MIME content types

contentType

contentType: { encoding: BufferEncoding; type: string }

Parsed Content-Type header: { type, encoding }.


Type declaration

  • encoding: BufferEncoding
  • type: string

crawler

document

document: Document

getKeyValueStore

getKeyValueStore: (idOrName?: string) => Promise<KeyValueStore>

Get a key-value store with given name or id, or the default one for the crawler.


Type declaration

id

id: string

json

json: JSONData

The parsed object from JSON string if the response contains the content type application/json.

log

log: Log

A preconfigured logger for the request handler.

optionalproxyInfo

proxyInfo?: ProxyInfo

An object with information about currently used proxy by the crawler and configured by the ProxyConfiguration class.

request

request: Request<UserData>

The original Request object.

response

response: PlainResponse

optionalsession

session?: Session

useState

useState: <State>(defaultValue?: State) => Promise<State>

Returns the state - a piece of mutable persistent data shared across all the request handler runs.


Type declaration

    • <State>(defaultValue?: State): Promise<State>
    • Type parameters

      • State: Dictionary = Dictionary

      Parameters

      • optionaldefaultValue: State

      Returns Promise<State>

window

window: Window

Methods

enqueueLinks

  • This function automatically finds and enqueues links from the current page, adding them to the RequestQueue currently used by the crawler.

    Optionally, the function allows you to filter the target links' URLs using an array of globs or regular expressions and override settings of the enqueued Request objects.

    Check out the Crawl a website with relative links example for more details regarding its usage.

    Example usage

    async requestHandler({ enqueueLinks }) {
    await enqueueLinks({
    globs: [
    'https://www.example.com/handbags/*',
    ],
    });
    },

    Parameters

    Returns Promise<BatchAddRequestsResult>

    Promise that resolves to BatchAddRequestsResult object.

parseWithCheerio

  • parseWithCheerio(selector?: string, timeoutMs?: number): Promise<CheerioAPI>
  • Returns Cheerio handle, allowing to work with the data same way as with CheerioCrawler. When provided with the selector argument, it will first look for the selector with a 5s timeout.

    Example usage:

    async requestHandler({ parseWithCheerio }) {
    const $ = await parseWithCheerio();
    const title = $('title').text();
    });

    Parameters

    • optionalselector: string
    • optionaltimeoutMs: number

    Returns Promise<CheerioAPI>

pushData

  • pushData(data?: ReadonlyDeep<Dictionary | Dictionary[]>, datasetIdOrName?: string): Promise<void>
  • This function allows you to push data to a Dataset specified by name, or the one currently used by the crawler.

    Shortcut for crawler.pushData().


    Parameters

    • optionaldata: ReadonlyDeep<Dictionary | Dictionary[]>

      Data to be pushed to the default dataset.

    • optionaldatasetIdOrName: string

    Returns Promise<void>

sendRequest

  • sendRequest<Response>(overrideOptions?: Partial<OptionsInit>): Promise<Response<Response>>
  • Fires HTTP request via got-scraping, allowing to override the request options on the fly.

    This is handy when you work with a browser crawler but want to execute some requests outside it (e.g. API requests). Check the Skipping navigations for certain requests example for more detailed explanation of how to do that.

    async requestHandler({ sendRequest }) {
    const { body } = await sendRequest({
    // override headers only
    headers: { ... },
    });
    },

    Type parameters

    • Response = string

    Parameters

    • optionaloverrideOptions: Partial<OptionsInit>

    Returns Promise<Response<Response>>

waitForSelector

  • waitForSelector(selector: string, timeoutMs?: number): Promise<void>
  • Wait for an element matching the selector to appear. Timeout defaults to 5s.

    Example usage:

    async requestHandler({ waitForSelector, parseWithCheerio }) {
    await waitForSelector('article h1');
    const $ = await parseWithCheerio();
    const title = $('title').text();
    });

    Parameters

    • selector: string
    • optionaltimeoutMs: number

    Returns Promise<void>