Skip to main content
The Olostep Zapier integration brings powerful web scraping capabilities to Zapier’s ecosystem of 8,000+ apps. Build automated workflows that extract, monitor, and process web data without writing any code. Get started with Olostep on Zapier →

Features

The integration provides 4 powerful actions for automated web data extraction:

Scrape Website

Extract content from any single URL in multiple formats (Markdown, HTML, JSON, text)

Batch Scrape URLs

Process up to 100,000 URLs in parallel. Perfect for large-scale data extraction

Create Crawl

Autonomously discover and scrape entire websites by following links

Create Map

Extract all URLs from a website for site structure analysis and content discovery

Installation

1. Find Olostep in Zapier

Search for “Olostep” in the Zapier app directory or when creating a new Zap:
  1. Go to Olostep on Zapier
  2. Click “Create Zap”
  3. Search for “Olostep” in the app selector
  4. Select the Olostep app
Alternatively, visit the Olostep integration page directly to browse available actions and create your first Zap.

2. Connect Your Account

When you first use Olostep in a Zap, you’ll be prompted to connect your account:
  1. Click “Sign in to Olostep”
  2. Enter your Olostep API key
  3. Click “Yes, Continue to Olostep”
Get your API key from the Olostep Dashboard.

Available Actions

Scrape Website

Extract content from a single URL. Supports multiple formats and JavaScript rendering. Use Cases:
  • Monitor specific pages for changes
  • Extract product information from e-commerce sites
  • Gather data from news articles or blog posts
  • Pull content for content aggregation
Configuration:
URL to Scrape
string
required
Website URL to scrape (must include http:// or https://)
Output Format
dropdown
default:"Markdown"
Choose format: Markdown, HTML, JSON, or Plain Text
Country Code
string
Country code for location-specific content (e.g., “US”, “GB”, “CA”)
Wait Before Scraping
integer
Wait time in milliseconds for JavaScript rendering (0-10000)
Parser
string
Optional parser ID for specialized extraction (e.g., “@olostep/amazon-product”)
Output Fields:
  • Scrape ID
  • Scraped URL
  • Markdown Content
  • HTML Content
  • JSON Content
  • Text Content
  • Status
  • Timestamp
  • Screenshot URL (if available)
  • Page Metadata
Example Workflows:
Trigger: Schedule (Every day at 9 AM)Action: Olostep - Scrape Website
  • URL: Competitor product page
  • Format: JSON
  • Parser: @olostep/amazon-product
Action: Google Sheets - Create Row
  • Add price data to tracking spreadsheet
Action: Gmail - Send Email (If price drops)
  • Alert team about price changes
Trigger: RSS by Zapier - New Item in FeedAction: Olostep - Scrape Website
  • URL: {{Post URL}}
  • Format: Markdown
Action: Notion - Create Page
  • Save article content to Notion database
Trigger: Google Sheets - New RowAction: Olostep - Scrape Website
  • URL: Company website from sheet
  • Format: Markdown
Action: OpenAI - Complete Text
  • Extract company information using AI
Action: Google Sheets - Update Row
  • Add enriched data back to sheet

Batch Scrape URLs

Process multiple URLs in parallel (up to 100,000 at once). Perfect for large-scale data extraction. Use Cases:
  • Scrape entire product catalogs
  • Extract data from multiple search results
  • Process lists of URLs from spreadsheets
  • Bulk content extraction
Configuration:
URLs to Scrape
text
required
JSON array of objects with url and custom_id fields.Example: [{"url":"https://example.com","custom_id":"site1"}]
Output Format
dropdown
default:"Markdown"
Choose format for all URLs: Markdown, HTML, JSON, or Plain Text
Country Code
string
Country code for location-specific scraping
Wait Before Scraping
integer
Wait time in milliseconds for JavaScript rendering
Parser
string
Optional parser ID for specialized extraction
Output Fields:
  • Batch ID (use this to retrieve results later)
  • Status
  • Total URLs
  • Created At
  • Requested Format
  • Country Code
  • Parser Used
Example Workflows:
Trigger: Webhook - Receive POST RequestAction: Code by Zapier - Run Python
  • Convert CSV/list to JSON array format
Action: Olostep - Batch Scrape URLs
  • URLs: {{JSON array from previous step}}
  • Format: JSON
  • Parser: @olostep/amazon-product
Action: Webhook - POST
  • Send batch ID to your system for retrieval
Trigger: Schedule - Every day at 6 AMAction: Google Sheets - Get Rows
  • Fetch URLs to monitor
Action: Code by Zapier - Format URLs
  • Convert to batch array format
Action: Olostep - Batch Scrape URLs
  • Process all URLs at once
Action: Slack - Send Message
  • Notify team that scraping is complete

Create Crawl

Autonomously discover and scrape entire websites by following links. Perfect for documentation sites, blogs, and content repositories. Use Cases:
  • Crawl and archive entire documentation sites
  • Extract all blog posts from a website
  • Build knowledge bases from web content
  • Monitor website structure changes
Configuration:
Start URL
string
required
Starting URL for the crawl (must include http:// or https://)
Maximum Pages
integer
default:"10"
Maximum number of pages to crawl
Whether to follow links found on pages
Output Format
dropdown
default:"Markdown"
Format for scraped content
Country Code
string
Optional country code for location-specific crawling
Parser
string
Optional parser ID for specialized content extraction
Output Fields:
  • Crawl ID (use this to retrieve results later)
  • Object Type
  • Status
  • Start URL
  • Maximum Pages
  • Follow Links
  • Created Timestamp
  • Formats
Example Workflows:
Trigger: Schedule - Monthly on 1st at 12 AMAction: Olostep - Create CrawlAction: Webhook - POST
  • Send crawl ID to your archive system
Action: Slack - Send Message
  • Notify team that crawl is in progress
Trigger: Schedule - Weekly on Monday at 9 AMAction: Olostep - Create Crawl
  • Start URL: Competitor blog URL
  • Max Pages: 100
  • Format: Markdown
Action: Delay - For 10 minutes
  • Wait for crawl to complete
Action: Airtable - Create Records
  • Store crawl data for analysis

Create Map

Extract all URLs from a website for content discovery and site structure analysis. Use Cases:
  • Build sitemaps and site structure diagrams
  • Discover all pages before batch scraping
  • Find broken or missing pages
  • SEO audits and analysis
Configuration:
Website URL
string
required
Website URL to extract links from (must include http:// or https://)
Search Query
string
Optional search query to filter URLs (e.g., “blog”)
Top N URLs
integer
Limit the number of URLs returned
Include URL Patterns
string
Glob patterns to include specific paths (e.g., “/blog/**”)
Exclude URL Patterns
string
Glob patterns to exclude specific paths (e.g., “/admin/**”)
Output Fields:
  • Map ID
  • Object Type
  • Website URL
  • Total URLs Found
  • URLs (JSON array)
  • Search Query
  • Top N Limit
Example Workflows:
Trigger: Button Click in ZapierAction: Olostep - Create MapAction: Code by Zapier - Extract URLs
  • Parse URLs from map result
Action: Olostep - Batch Scrape URLs
  • URLs: {{URLs from map}}
  • Format: JSON
Action: Google Sheets - Create Rows
  • Add all product data to spreadsheet
Trigger: Schedule - MonthlyAction: Olostep - Create Map
  • URL: Your website
  • Top N: 1000
Action: Airtable - Create Records
  • Store all URLs for tracking
Action: Slack - Send Message
  • Report total pages found

E-commerce Price Monitoring

Monitor competitor prices and get instant alerts:
Trigger: Schedule (Hourly)

Action: Olostep - Scrape Website
  - URL: Competitor product page
  - Format: JSON
  - Parser: @olostep/amazon-product

Action: Filter (Only continue if price changed)

Action: Slack - Send Message
  - Alert: "Price changed to $\{\{price\}\}"

Content Aggregation

Aggregate content from multiple sources:
Trigger: Google Sheets - New Row

Action: Olostep - Scrape Website
  - URL: \{\{URL from sheet\}\}
  - Format: Markdown

Action: OpenAI - Summarize
  - Summarize the content

Action: Airtable - Create Record
  - Store article with summary

Lead Enrichment Pipeline

Enrich lead data with web information:
Trigger: HubSpot - New Contact

Action: Olostep - Scrape Website
  - URL: \{\{Company website\}\}
  - Format: Markdown

Action: OpenAI - Extract Data
  - Extract: company size, industry, products

Action: HubSpot - Update Contact
  - Add enriched data to contact

Research Automation

Automate research from multiple sources:
Trigger: Airtable - New Record

Action: Olostep - Create Map
  - URL: Research target website
  - Include: /research/**

Action: Code - Parse URLs

Action: Olostep - Batch Scrape URLs
  - URLs: \{\{Discovered URLs\}\}
  - Format: Markdown

Action: Notion - Create Pages
  - Create research database

Social Media Monitoring

Track mentions and content:
Trigger: Schedule (Every 6 hours)

Action: Olostep - Scrape Website
  - URL: News site search page
  - Format: HTML

Action: Code - Extract Mentions
  - Find brand mentions

Action: Google Sheets - Create Row
  - Log mentions with timestamp

Multi-Step Workflows

Complete Product Scraping Pipeline

Build a comprehensive product data pipeline:
1

Discover Product URLs

Use Create Map to find all product pages on the target website
  • Include patterns: /products/**
  • Exclude patterns: /cart/**, /checkout/**
2

Batch Process Products

Use Batch Scrape URLs to extract all product data
  • Format: JSON
  • Parser: Product-specific parser if available
3

Store in Database

Send batch ID to your system or wait and retrieve results
  • Use Airtable, Google Sheets, or your database
4

Monitor for Changes

Schedule daily scrapes to track price/availability changes
  • Compare with existing data
  • Alert on significant changes

SEO Content Strategy

Analyze competitors and plan content:
1

Map Competitor Sites

Use Create Map on competitor websites
  • Extract all blog posts and content pages
2

Scrape Content

Use Batch Scrape URLs to get full content
  • Format: Markdown for easy analysis
3

AI Analysis

Use OpenAI to analyze topics and keywords
  • Identify content gaps
  • Find trending topics
4

Create Content Calendar

Add insights to Notion or Airtable
  • Plan your content strategy

Specialized Parsers

Olostep provides pre-built parsers for popular websites. Use them with the Parser field:

Amazon Product

@olostep/amazon-productExtract: title, price, rating, reviews, images, variants

LinkedIn Profile

@olostep/linkedin-profileExtract: name, title, company, location, experience

LinkedIn Company

@olostep/linkedin-companyExtract: company info, employee count, industry, description

Google Search

@olostep/google-searchExtract: search results, titles, snippets, URLs

Google Maps

@olostep/google-mapsExtract: business info, reviews, ratings, location

Instagram Profile

@olostep/instagram-profileExtract: profile info, followers, posts, bio

Using Parsers

Simply add the parser ID to the Parser field:
Action: Olostep - Scrape Website
  - URL: https://www.amazon.com/dp/PRODUCT_ID
  - Format: JSON
  - Parser: @olostep/amazon-product
The parser automatically extracts structured data specific to that website type.

Google Sheets

Perfect for data collection and tracking:
1. Olostep scrapes website
2. Filter or transform data
3. Google Sheets - Create/Update Row
Use Cases:
  • Price tracking spreadsheets
  • Lead enrichment databases
  • Content inventory
  • Competitor analysis sheets

Airtable

Build powerful databases with scraped data:
1. Olostep scrapes or crawls
2. Code - Format data
3. Airtable - Create Records
Use Cases:
  • Product catalogs
  • Research databases
  • Content calendars
  • Link databases

Slack

Get instant notifications:
1. Olostep monitors page
2. Filter - Check for changes
3. Slack - Send Message
Use Cases:
  • Price drop alerts
  • Content update notifications
  • Error monitoring
  • Daily digests

HubSpot / Salesforce

Enrich CRM data automatically:
1. New contact added
2. Olostep scrapes company website
3. OpenAI extracts key info
4. CRM - Update contact
Use Cases:
  • Lead enrichment
  • Company research
  • Competitive intelligence
  • Account mapping

Notion

Build knowledge bases:
1. Olostep crawls documentation
2. Code - Parse content
3. Notion - Create Pages
Use Cases:
  • Documentation mirrors
  • Research repositories
  • Content libraries
  • Team wikis

Best Practices

When scraping more than 3-5 URLs, use Batch Scrape URLs instead of multiple Scrape Website actions. Batch processing is:
  • Much faster (parallel processing)
  • More cost-effective
  • Easier to manage
  • Better for rate limits
For JavaScript-heavy sites, use the “Wait Before Scraping” parameter:
  • Simple sites: 0-1000ms
  • Dynamic sites: 2000-3000ms
  • Heavy JavaScript: 5000-8000ms
Test with different values to find the optimal wait time.
For popular websites (Amazon, LinkedIn, Google), use pre-built parsers:
  • Get structured data automatically
  • More reliable extraction
  • No need for custom parsing
  • Maintained by Olostep
Use Zapier’s Filter action to avoid unnecessary scrapes:
  • Check if URL has changed
  • Verify data hasn’t been scraped recently
  • Apply business logic before scraping
This saves API credits and execution time.
Batch, Crawl, and Map operations are asynchronous:
  • Store the returned ID (batch_id, crawl_id, map_id)
  • Use a Delay action if retrieving immediately
  • Consider webhook callbacks for completion
  • Set up separate Zaps for retrieval
Choose the right storage based on your needs:
  • Google Sheets: Simple tracking, team collaboration
  • Airtable: Relational data, rich formatting
  • Database: Large-scale, complex queries
  • Notion: Knowledge base, documentation
Set up monitoring for your scraping workflows:
  • Use Error paths in Zaps
  • Send alerts to Slack/Email on failures
  • Track API usage in Olostep dashboard
  • Log important metrics

Common Use Cases by Industry

E-commerce

  • Price Monitoring: Track competitor pricing in real-time
  • Product Research: Discover trending products and market gaps
  • Inventory Tracking: Monitor stock availability
  • Review Analysis: Aggregate and analyze customer reviews

Marketing & SEO

  • Content Discovery: Find content opportunities
  • Competitor Analysis: Track competitor strategies
  • Backlink Research: Discover link opportunities
  • Keyword Research: Extract keyword data from search results

Sales & Lead Generation

  • Lead Enrichment: Enhance CRM data with web information
  • Company Research: Gather company intelligence
  • Contact Discovery: Find decision-makers
  • Competitive Intelligence: Track competitor moves

Research & Analytics

  • Data Collection: Gather data from multiple sources
  • Market Research: Track industry trends
  • Academic Research: Collect research data
  • Price Intelligence: Analyze pricing strategies

Media & Publishing

  • Content Aggregation: Curate content from multiple sites
  • News Monitoring: Track news and mentions
  • Social Media: Monitor social platforms
  • Trend Detection: Identify trending topics

Troubleshooting

Error: “Invalid API key”Solutions:
  • Check API key from dashboard
  • Ensure no extra spaces in API key
  • Reconnect your Olostep account in Zapier
  • Verify API key is active
Error: Content fields are emptySolutions:
  • Increase “Wait Before Scraping” time
  • Check if website requires login
  • Try different format (HTML vs Markdown)
  • Verify URL is accessible
  • Check if site blocks automated access
Error: “Invalid JSON format for batch array”Solutions:
  • Use format: [{"url":"https://example.com","custom_id":"id1"}]
  • Ensure proper JSON syntax
  • Use Code step to format URLs correctly
  • Test JSON with online validator
Error: “Rate limit exceeded”Solutions:
  • Space out Zap runs with delays
  • Use batch processing instead of individual scrapes
  • Upgrade your Olostep plan
  • Check rate limit in dashboard
Error: Specific URLs fail to scrapeSolutions:
  • Verify URL format (include http:// or https://)
  • Check if URL requires authentication
  • Test URL in browser first
  • Try with country parameter
  • Contact support for blocked domains

Zapier Limitations & Workarounds

Task Limits

Zapier has task limits based on your plan. Each Olostep action counts as 1 task. Workaround: Use batch processing to scrape multiple URLs as a single task.

Execution Time

Zaps timeout after 30 seconds. Crawls and large batches may take longer. Workaround: Store the ID and retrieve results in a separate Zap or use webhooks.

Data Size

Zapier has limits on data size per task. Workaround: Use hosted URLs in the output to retrieve large content separately.

Polling vs Instant

Triggers are polling-based (check every 5-15 minutes). Workaround: Use webhooks for instant notifications or schedule at specific times.

Pricing

Olostep charges based on API usage, independent of Zapier:
  • Scrapes: Pay per scrape
  • Batches: Pay per URL in batch
  • Crawls: Pay per page crawled
  • Maps: Pay per map operation
Check current pricing at olostep.com/pricing. Zapier Plan: You also need an active Zapier plan to run Zaps.

Support

Need help with the Zapier integration?

Get Started

Ready to automate your web scraping workflows?

Create Your First Zap

Start building automated workflows with Olostep and Zapier
Connect Olostep with 8,000+ apps and automate your web data extraction today!