Scrape websites without writing a single line of code.
📚 Check out the docs for a comprehensive quickstart guide and detailed information.
- XPath-Based Extraction: Precisely target page elements
- Queue Management: Submit and manage multiple scraping jobs
- Domain Spidering: Option to scrape all pages within the same domain
- Custom Headers: Add JSON headers to your scraping requests
- Media Downloads: Automatically download images, videos, and other media
- Results Visualization: View scraped data in a structured table format
- Data Export: Export your results in markdown and csv formats
- Notifcation Channels: Send completion notifcations, through various channels
make up
Refer to the docs for helm deployment: https://scraperr-docs.pages.dev/guides/helm-deployment
When using Scraperr, please remember to:
- Respect
robots.txt
: Always check a website'srobots.txt
file to verify which pages permit scraping - Terms of Service: Adhere to each website's Terms of Service regarding data extraction
- Rate Limiting: Implement reasonable delays between requests to avoid overloading servers
Disclaimer: Scraperr is intended for use only on websites that explicitly permit scraping. The creator accepts no responsibility for misuse of this tool.
Get support, report bugs, and chat with other users and contributors.
This project is licensed under the MIT License. See the LICENSE file for details.
Development made easier with the webapp template.
To get started, simply run make build up-dev
.