fitter
croncert-config
fitter | croncert-config | |
---|---|---|
15 | 3 | |
98 | 10 | |
- | - | |
8.9 | 9.3 | |
16 days ago | 5 days ago | |
Go | ||
MIT License | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
fitter
- Show HN: Fitter – configurable open-source scraper
- PxyUp/fitter: New way for collect information from the API's/Websites
- Show HN: Fitter – next generation web-scraper
- Show HN: Library for scrape internet like GQL
-
No code command line webscraper
I came to similar idea when i was scrape flight tickets: https://github.com/PxyUp/fitter
- Fitter - GQL for web scrawling, multiple sources, defining fields, emulating browser and more other
- CLI/Go Library Tool for scrape website. Emulate browsers, parse xpath/html/json, multithreaded, combine sources and more in one tool
- Show HN: Fitter – cross platform library for parsing websites released
- Fitter CLI – uniq way for scrape information. Different sources one result
croncert-config
-
No code command line webscraper
I actually started this scraping project because of an idea I wanted to try, which is scraping concert data from as many websites as possible with as little effort as possible, see https://github.com/jakopako/croncert-config This seems to work better and better. Still I am wondering if there are any other valid use cases for such a terminal based scraper or if it's rather niche. What do you think?
-
Crowdsourced concert scraping project
I am currently working on a configurable command line webscraper, called goskyr and my first use case is collecting as much concert data as possible for this website idea I had, croncert.ch I am hoping that people other than me are willing to contribute to the scraper configuration file in this repository, https://github.com/jakopako/croncert-config, which also contains a github action to regularly run the scraper. What do you think? Could this work? How should I spread the word?
-
New concert website
croncert.ch is a website that lists concerts worldwide (currently, ‘worldwide’ is more of a euphemism), focussing on smaller venues. An automated process regularly scrapes the underlying concert data. The idea is that anyone can contribute by extending the scraper configuration with new concert venues. Feel free to check out https://github.com/jakopako/croncert-config for more details!
What are some alternatives?
rod - A Devtools driver for web automation and scraping
requests-html - Pythonic HTML Parsing for Humans™
lux - 👾 Fast and simple video download library and CLI tool written in Go
goskyr - A configurable command-line web scraper written in go with auto configuration capability
open-dictionary - 🦄 An initiative to create a dictionary which is free for everyone 🚀
csv2api - 🏄♂️ The parser reads the CSV file with the raw data, filters the records, identifies fields to be changed, and sends a request to update the data to the specified endpoint of your REST API.
Ferret - Declarative web scraping
osdg-data - The OSDG Community Dataset (OSDG-CD) is a public dataset of thousands of text excerpts, validated by OSDG Community Platform (OSDG-CP) citizen scientists with respect to the Sustainable Development Goals (SDGs). The dataset is updated every quarter and published on Zenodo.
Crawly - Crawly, a high-level web crawling & scraping framework for Elixir.