scrapingfish-blog-projects
uule_grabber
scrapingfish-blog-projects | uule_grabber | |
---|---|---|
2 | 2 | |
3 | 18 | |
- | - | |
10.0 | 3.0 | |
over 1 year ago | 11 months ago | |
Jupyter Notebook | Python | |
- | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
scrapingfish-blog-projects
-
Ask HN: What have you created that deserves a second chance on HN?
This is definitely possible. I’m not sure if we’re going to have time for this as we’re occupied by work on Scraping Fish but we shared the code for scraping nutrition facts data from Walmart on github: https://github.com/pawelkobojek/scrapingfish-blog-projects/t.... Feel free to take it and build such app/website on top of it.
-
Scraping Google SERP with Geolocation
A script with command-line interface is shared in this GitHub repository: https://github.com/pawelkobojek/scrapingfish-blog-projects/tree/main/google-serp. To run it, you will need Scraping Fish API key which you can get here: https://scrapingfish.com/buy
uule_grabber
-
Scraping Google SERP with Geolocation
There's a fairly simple algorithm behind generating uule parameter. It involves a fixed prefix, a special key which depends on location string length, and base64 encoding. If you're interested, here is a blog post from MOZ (a SEO tool) which describes it in detail: https://moz.com/blog/geolocation-the-ultimate-tip-to-emulate-local-search. If you need to automate uule parameter generation, there's a python package for this: https://github.com/ogun/uule_grabber. It generates uule value for you given the location string. Example usage:
-
Google web scrape from excel list?
from urllib.parse import quote_plus import requests API_KEY = "YOUR SCRAPING FISH API KEY" # https://scrapingfish.com/buy url_prefix = f"https://scraping.narf.ai/api/v1/?api_key={API_KEY}&render_js=true&url=" # to get uule for location you can use: https://github.com/ogun/uule_grabber # or https://site-analyzer.pro/services-seo/uule/ uule_usa = "w+CAIQICIDVVNB" keyword = "kitchen sink" search_url = f"https://www.google.com/search?q={quote_plus(keyword)}&uule={uule_usa}&gl=us&hl=en" response = requests.get(f"{url_prefix}{quote_plus(search_url)}", timeout=90) # add your response processing/parsing logic with open("google.html", "wb") as f: f.write(response.content)
What are some alternatives?
hckrweb - Hcker News mobile web app
PyPSI - Python Implementation of Google PageSpeed Insights
metron - A C++ to Verilog translation tool with some basic guarantees that your code will work.
UrlChecker - Android app by TrianguloY: URLCheck
donutdns - Block ads, trackers, and malicious sites with donutdns - simple alternative to pihole. Run as a docker container, standalone executable or core DNS plugin. Supply custom domain block/allow lists in addition to builtin lists maintained by the ad-blocking community.
codebase-visualizer-action - Visualize your codebase during CI.
scheme-for-max - Max/MSP external for scripting and live coding Max with s7 Scheme Lisp
formkiq-core - A full-featured Document Layer for your application, providing the functionality of a flexible document management system, including storage, discovery, processing, and retrieval. Deploys directly into your Amazon Web Services Cloud. 🌟 Star to support our work!
pytkml - Write tests for machine learning models