selectolax
colly
Our great sponsors
selectolax | colly | |
---|---|---|
6 | 39 | |
967 | 22,165 | |
- | 1.8% | |
7.7 | 6.0 | |
about 2 months ago | 7 days ago | |
Cython | Go | |
MIT License | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
selectolax
-
GitHub – GSA/code-gov: An informative repo for all Code.gov repos
https://github.com/rushter/selectolax#simple-benchmark )
(Apache Nutch is a Java-based web crawler which supports e.g. CommonCrawl (which backs various foundational LLMs)) https://en.wikipedia.org/wiki/Apache_Nutch#Search_engines_bu... . But extruct extracts more types of metadata and data than Nutch AFAIU: https://github.com/scrapinghub/extruct )
datasette-graphql adds a GraphQL HTTP API to a SQLite database:
-
8 Most Popular Python HTML Web Scraping Packages with Benchmarks
selectolax
- High performance code in Python
-
Web Scraping with Python: Everything you need to know to get started (2022)
try this... https://github.com/rushter/selectolax
-
The State of Web Scraping in 2021
Lazyweb link: https://github.com/rushter/selectolax
although I don't follow the need to have what appears to be two completely separate HTML parsing C libraries as dependencies; seeing this in the readme for Modest gives me the shivers because lxml has _seen some shit_
> Modest is a fast HTML renderer implemented as a pure C99 library with no outside dependencies.
although its other dep seems much more cognizant about the HTML5 standard, for whatever that's worth: https://github.com/lexbor/lexbor#lexbor
---
> It looks like the author of the article just googled some libraries for each language and didn't research the topic
Heh, oh, new to the Internet, are you? :-D
- Show HN: Fast HTML5 parser for Python with multiple backends
colly
-
Scraping the full snippet from Google search result
SerpApi focuses on scraping search results. That's why we need extra help to scrape individual sites. We'll use GoColly package.
-
Show HN: Flyscrape – A standalone and scriptable web scraper in Go
Interesting. Can you compare it to colly? [0]
Last time I looked it was the most popular choice for scraping in Go and I have some projects using it.
Is it similar? Does it have more/less features or is it more suited for a different use case? (Which one?)
[0] https://github.com/gocolly/colly
- Colly: Elegant Scraper and Crawler Framework for Golang
-
New modern web crawling tool
Sounds cool, but how is this different from Colly: https://github.com/gocolly/colly?
-
colly VS scrapemate - a user suggested alternative
2 projects | 15 Apr 2023
-
Web Scraping in Python: Avoid Detection Like a Ninja
We could write some snippets mixing all these, but the best option in real life is to use a tool with it all, like Scrapy, pyspider, node-crawler (Node.js), or Colly (Go).
- Web scraping with Go
-
Web scraper help
Unless you're specifically trying to do it using net/http, I recommend using colly. I've used it in a few scrappers and I love it!
-
Web Scraping in Golang
In this blog, we will be covering the basics of web scraping in Go using the Fiber and Colly frameworks. Colly is an open-source web scraping framework written in Go. It provides a simple and flexible API for performing web scraping tasks, making it a popular choice among Go developers. Colly uses Go's concurrency features to efficiently handle multiple requests and extract data from websites. It offers a wide range of customization options, including the ability to set request headers, handle cookies, follow redirects, and more
-
Learn how to scrape Trustpilot reviews using Go
github.com/gocolly/colly - popular and widely-used library for web scraping in Go. It provides a higher-level API than net/http and makes it easier to extract information from websites. It also provides features such as concurrency, automatic request retries, and support for cookies and sessions.
What are some alternatives?
lxml - The lxml XML toolkit for Python
GoQuery - A little like that j-thing, only in Go.
lexbor - Lexbor is development of an open source HTML Renderer library. https://lexbor.com
Scrapy - Scrapy, a fast high-level web crawling & scraping framework for Python.
html5lib - Standards-compliant library for parsing and serializing HTML documents and fragments in Python
xpath - XPath package for Golang, supports HTML, XML, JSON document query.
pyppeteer - Headless chrome/chromium automation library (unofficial port of puppeteer)
rod - A Devtools driver for web automation and scraping
pyquery - A jquery-like library for python
Geziyor - Geziyor, blazing fast web crawling & scraping framework for Go. Supports JS rendering.
gazpacho - 🥫 The simple, fast, and modern web scraping library
Ferret - Declarative web scraping