osdg-data
croncert-config
osdg-data | croncert-config | |
---|---|---|
3 | 3 | |
25 | 10 | |
- | - | |
3.2 | 9.3 | |
8 months ago | 5 days ago | |
GNU General Public License v3.0 only | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
osdg-data
-
[self-promotion] OSDG Community Dataset | Labelled text excerpts with relation to the UN SDGs
We are sharing an open OSDG Community Dataset (OSDG-CD) on our GitHub. The dataset contains thousands of text excerpts labelled by citizen scientists from around the world with respect to the UN Sustainable Development Goals (SDGs).
-
OSDG Community Dataset | First results of a citizen science initiative on UN SDGs
Today we are sharing the OSDG Community Dataset (OSDG-CD), a direct result of the work of hundreds of citizen scientists and UN Volunteers who have contributed to our understanding of the SDGs. The dataset contains thousands of text excerpts which were labelled by the community with respect to SDGs. The data can be used to derive insights into the nature of SDGs using either ontology-based or machine learning approaches. The OSDG-CP dataset will be updated on a quarterly basis.
-
First results of a fully-online citizen science initiative on UN SDGs
The dataset is available on our GitHub: https://github.com/osdg-ai/osdg-data
croncert-config
-
No code command line webscraper
I actually started this scraping project because of an idea I wanted to try, which is scraping concert data from as many websites as possible with as little effort as possible, see https://github.com/jakopako/croncert-config This seems to work better and better. Still I am wondering if there are any other valid use cases for such a terminal based scraper or if it's rather niche. What do you think?
-
Crowdsourced concert scraping project
I am currently working on a configurable command line webscraper, called goskyr and my first use case is collecting as much concert data as possible for this website idea I had, croncert.ch I am hoping that people other than me are willing to contribute to the scraper configuration file in this repository, https://github.com/jakopako/croncert-config, which also contains a github action to regularly run the scraper. What do you think? Could this work? How should I spread the word?
-
New concert website
croncert.ch is a website that lists concerts worldwide (currently, ‘worldwide’ is more of a euphemism), focussing on smaller venues. An automated process regularly scrapes the underlying concert data. The idea is that anyone can contribute by extending the scraper configuration with new concert venues. Feel free to check out https://github.com/jakopako/croncert-config for more details!
What are some alternatives?
sdg-004 - This project is inspired to provide a swift response to tackle this under reported cases. It is a reporting platform aimed to help provide help and solutions to this ravaging under reported issues baffling our society. SpeakUp work by following up reported cases, giving guides and tips on possible way out based on the report and help asked for.
requests-html - Pythonic HTML Parsing for Humans™
carbon-registry - National Carbon Credit Registry Digital Public Good (DPG) by Digital For Climate (D4C) collaboration. Code coordinated by ExO/CDO & BPPS/Climate.
goskyr - A configurable command-line web scraper written in go with auto configuration capability
fitter - New way for collect information from the API's/Websites
open-dictionary - 🦄 An initiative to create a dictionary which is free for everyone 🚀
Ferret - Declarative web scraping
Crawly - Crawly, a high-level web crawling & scraping framework for Elixir.