📄️ Accept user input
This example accepts and logs user input:
📄️ Add data to dataset
This example saves data to the default dataset. If the dataset doesn't exist, it will be created.
📄️ Basic crawler
This is the most bare-bones example of the Apify SDK, which demonstrates some of its building blocks such as the BasicCrawler. You probably don't need to go this deep though, and it would be better to start with one of the full-featured crawlers
📄️ Call actor
This example demonstrates how to start an Apify actor using
📄️ Cheerio crawler
This example demonstrates how to use CheerioCrawler to crawl a list of URLs from an external file, load each URL using a plain HTTP request, parse the HTML using the Cheerio library and extract some data from it: the page title and all h1 tags.
📄️ Crawl all links on a website
This example uses the enqueueLinks() method to add new links to the RequestQueue as the crawler navigates from page to page. If only the
📄️ Crawl multiple URLs
This example crawls the specified list of URLs.
📄️ Crawl a website with relative links
When crawling a website, you may encounter different types of links present that you may want to crawl.
📄️ Crawl a single URL
This example uses the got-scraping npm package
📄️ Crawl a sitemap
This example downloads and crawls the URLs from a sitemap.
📄️ Crawl some links on a website
This CheerioCrawler example uses the pseudoUrls property in the enqueueLinks() method to only add links to the RequestQueue queue if they match the specified regular expression.
This example demonstrates how to use PuppeteerCrawler to
📄️ Dataset Map and Reduce methods
This example shows an easy use-case of the Dataset map
📄️ Playwright crawler
This example demonstrates how to use PlaywrightCrawler
📄️ Capture a screenshot using Puppeteer
To run this example on the Apify Platform, select the apify/actor-node-puppeteer-chrome image for your Dockerfile.
📄️ Puppeteer crawler
This example demonstrates how to use PuppeteerCrawler in combination
📄️ Puppeteer recursive crawl
Run the following example to perform a recursive crawl of a website using PuppeteerCrawler.
📄️ Puppeteer with proxy
This example demonstrates how to load pages in headless Chrome / Puppeteer over Apify Proxy.