pyppeteer
scraper
Our great sponsors
pyppeteer | scraper | |
---|---|---|
17 | 12 | |
3,393 | 98 | |
2.3% | - | |
5.6 | 0.0 | |
13 days ago | about 1 year ago | |
Python | TypeScript | |
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
pyppeteer
-
Pyppeteer Tutorial: The Ultimate Guide to Using Puppeteer with Python
The latest version of Pyppeteer, i.e., 1.0.2, can also be installed by executing pip3 install -U git+https://github.com/pyppeteer/pyppeteer@dev on the terminal.
-
Thoughts on AsyncIO
Having async baked into Python both as a keyword and standard library has vastly opened up the number of use cases async will be used in where it is appropriate, one of my favorite libraries pyppeteer would have not existed in the easy nice way it can be used without asyncio: https://github.com/pyppeteer/pyppeteer
-
Do you have a tip to bypass cookie expiration when scraping a webpage?
Yes, this is scraping. You need a legit new cookie issued by the site. The solution is to automate the going to get the cookies part. The most straight forward way to do this is probably using a headless browser through something like Selenium or pyppeteer.
-
will requests-html library work as selenium
Last I checked, pyppeteer wasn't a thing anymore, and I haven't tried Playwright but if it has a headless mode, thats what you want so you don't have a browser open.
- What have you automated with python?
-
Getting this error while installing pyppeteer
pip install -U git+https://github.com/pyppeteer/pyppeteer@dev
- Note, the first time you ever run the render() method, it will download Chromium into your home directory (e.g. ~/.pyppeteer/). This only happens once.
-
Trying to find a way to automate button clicking on work program without image use
The normal Puppeteer package is JavaScript, but I do see that there's a Python port called pyppeteer. I can't vouch for it specifically, but I imagine it's similarly easy to use as the JS version.
- Scrape JSON from Network Traffic using Selenium
- How to start Web scraping with python?
scraper
- Most Used Individual JavaScript Libraries - jQuery still leads
-
Most Used JavaScript Libraries (percentage) - June 2022 [OC]
Additional info and source code for generating the dataset, summarizing it and rendering the chart are available at https://github.com/get-set-fetch/scraper/tree/main/datasets/javascript-libs-from-top-1mm-sites
-
How to collaborate on web scraping?
Store the scrape progress (to-be-scraped / in-progress / scraped / in-error URLs) in a database shared by all participants and scrape in parallel with as many machines as the db load permits. Got a connection timeout / IP is blocked on one machine ? Update the scrape status for the corresponding URL and let another machine retry it. https://github.com/get-set-fetch/scraper (written in typescript) follows this idea. Using Terraform from a simple config file you can adjust the number of scraper instances to be deployed in cloud at startup and during the scraping process. In benchmarks a PostgreSQL server running on a DigitalOcean vm with 4vCPU, 8GB memory allows for ~2000 URLs to be scraped per second (synthetic data, no external traffic). From my own experience this is almost never the bottleneck. Obeying robots.txt crawl-delay will surely put you under this limit. Disclaimer: I'm the npm package author.
-
How to serve scrapped data?
Written in typescript https://github.com/get-set-fetch/scraper stores scraped content directly in a database (sqlite, mysql, postgresql). Each URL represents a Resource. You can implement your own IResourceStorage and define the exact db columns you need.
-
How to scrape entire blogs with content?
You can use https://github.com/get-set-fetch/scraper with a custom plugin based on the mozilla/readability as detailed in https://getsetfetch.org/node/custom-plugins.html (extracting news article content). I think it's a close match to your use case.
-
A simple solution to rotate proxies or how to spin up your own rotation proxy server with Puppeteer and only a few lines of JS code
I'm currently implementing concurrency conditions at project/proxy/domain/session level in https://github.com/get-set-fetch/scraper . On each level you can define the maximum number of requests and the delay between two consecutive requests.
-
Web scraping content into postgresql? Scheduling web scrapers into a pipeline with airflow?
If you're familiar with nodejs give https://github.com/get-set-fetch/scraper a try. Scraped content can be stored in sqlite, mysql or postgresql. It also supports puppeteer, playwright, cheerio or jsdom for the actual content extraction. No scheduler though.
-
Web Scraping 101 with Python
I'm using this exact strategy to scrape content directly from DOM using APIs like document.querySelectorAll. You can use the same code in both headless browser clients like Puppeteer or Playwright and DOM clients like cheerio or jsdom (assuming you have a wrapper over document API). Depending on the way a web page was fetched (opened in a browser tab or fetched via nodejs http/https requests), ExtractHtmlContentPlugin, ExtractUrlsPlugin use different DOM wrappers (native, cheerio, jsdom) to scrape the content.
[1] https://github.com/get-set-fetch/scraper/blob/main/src/plugi...
-
What is your “I don't care if this succeeds” project?
https://github.com/get-set-fetch/scraper - I've been working (intermittently :) ) on a nodejs or browser extension scraper for the last 3 years, see the other projects under the get-set-fetch umbrella. Putting a lot more effort lately as I really want to do those Alexa top 1 million analysis like top js libraries, certificate authorities and so on. A few weeks back I've posted on Show:HN as you can do basic/intermediate? scraping with it.
Not capable of handling 1 mil+ pages as it still limited to puppeteer or playwright. Working on adding cheerio/jsdom support right now.
What are some alternatives?
puppeteer - Node.js API for Chrome
puppeteer-cluster - Puppeteer Pool, run a cluster of instances in parallel
Scrapy - Scrapy, a fast high-level web crawling & scraping framework for Python.
playwright-recaptcha-solver - ReCaptcha V2 solver for Playwright
Playwright - Playwright is a framework for Web Testing and Automation. It allows testing Chromium, Firefox and WebKit with a single API.
playwright-python - Python version of the Playwright testing and automation library.
Twitch-Drops-Bot - A Node.js bot that will automatically watch Twitch streams and claim drop rewards.
selenium-python-helium - Lighter web automation for Python [Moved to: https://github.com/mherrmann/helium]
vopono - Run applications through VPN tunnels with temporary network namespaces
requests - A simple, yet elegant, HTTP library.
listudy - Listudy - chess training server