DownloadNet
ripgrep-all
DownloadNet | ripgrep-all | |
---|---|---|
20 | 43 | |
3,653 | 6,200 | |
2.1% | - | |
6.1 | 7.6 | |
18 days ago | 1 day ago | |
JavaScript | Rust | |
GNU General Public License v3.0 or later | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
DownloadNet
-
ArchiveBox: Open-source self-hosted web archiving
For anyone who uses Chrome and wants to view their archived pages in the browser as if they were still online (URL and everything intact), and also full-text search through their browsing history that was archived (like AB plans to add in future, I think, right nikki?) you can check out DownloadNet: https://github.com/dosyago/DownloadNet
You can have multiple archives, and even use a mode where you only archive pages you bookmark rather than everything.
-
Show HN: Rem: Remember Everything (open source)
This does look cool. It reminds me of a recent discovery I made. The other day, while trying to recover some disk space, I found a giant file on my hard disk. It turned out to be a nine-hour screen recording from almost a year ago. I had no idea it existed, so I must’ve accidentally left the screen recording on. Watching it was fascinating; it was like a window into my thought process at that time. You could see how I was researching something online. It was almost like a play-by-play, akin to re-watching a sports performance – very instructive and surprisingly useful.
In a similar vein to what you’ve done, but focusing specifically on web browsing, I’ve created a tool called ‘DownloadNet.’ It archives for offline use and fully indexes every page you visit. Additionally, it can be configured to archive only the pages you bookmark, offering another mode of operation. It’s an open-source tool, so feel free to check it out: https://github.com/dosyago/DownloadNet
-
You're Gonna Need a Bigger Browser
Given that I directly work in this space I found the article's synthesis of a range of ideas about browser innovation to be highly relevant.
More generally, the article is actually extremely interesting and examines a bunch of ideas worthy of consideration if you're interested in the future of web browsing.
Perhaps none of the ideas are new in isolation, but it's encouraging that people are doing this foundational conceptual work and imagining where a synthesis of them would go.
Despite being interesting somehow on the page it was not so easy to read. Here's a summary of key ideas:
Stagnation in Browser Evolution: Berjon notes that despite being central to the web's architecture, browsers haven't changed much in their fundamental design for a long time. They have undergone incremental changes but the core concept remains largely the same as it was decades ago.
Reimagining Browsers: He suggests that to increase user agency—a principle that the web should empower users—we need to consider major overhauls to what a browser is and how it operates.
Integration of Search and Social: Berjon challenges the traditional separation of browsers, search engines, and social platforms. He advocates for an integrated approach where the browser encompasses these functions, aligning more closely with users' experiences and expectations.
Shift From Client to Agent: The author proposes rethinking the browser not just as a client for retrieving documents but as an "agent" that provides a variety of services, potentially including server-like functions, to empower users.
User Agency and Personal Data Servers: By incorporating elements such as Personal Data Servers (PDS), users could manage their own data and services like recommendations, identity, and subscriptions, which currently rely on third-party providers.
Tab Management: Berjon critiques the use of tabs, suggesting that they are an ineffective method for organizing and interacting with web content, and advocates for better UI solutions.
Business Models: He delves into the financial aspects of browsers, highlighting the significant profits derived from setting search engine defaults. Berjon argues for reinvestment of these profits into the web as a public good and for developing business models that truly benefit user agency.
Potential for Change: Despite the challenges, Berjon is optimistic about the possibility of change, noting that there is room for product differentiation and that financial incentives can drive innovation in the browser space.
I found the one about User Agency and Personal Data Servers particularly fascinating. I've been exploring the idea of a federated search engine, where a person curates their own search through their browsing history (and ultimately could share it socially), in DownloadNet: https://github.com/dosyago/DownloadNet
And my company has been developing a platform for building extended and customized browsing experiences and delivering them anywhere. It's my hope that BrowserBox will play a part in the future direction of the browser as user agent. It's open source so if you care about the future of the web, get involved: https://github.com/BrowserBox/BrowserBox :)
-
Google Chrome pushes browser history-based ad targeting
If you're interested in utilizing your history information for something in your intentional interests, consider saving an archive of pages you browse to make a search engine you can query back through later.
You can save the full content for indexing with full text search, and you can even export archives as tarballs by zipping up the directory. Many people find this a useful way to "mine" their own browser history to create a curated search engine aligned with your interests. Or simply to save the pages they browse for review offline--either to save bandwidth, or just because they're actually "offline"--at a remote site, or on an airplane.
Everything is saved in a fully interactive way. Personally tho, I find search the most useful feature. Also, we're open source so if you want to get involved, please do so!
https://github.com/dosyago/DiskerNet
-
Show HN: Linkwarden – An open source collaborative bookmark manager
If you want full-text-search with archiving check out my project, DiskerNet. https://github.com/dosyago/DiskerNet --> also well done on LinkWarden! Looks like a great product! :)
- Show HN: DiskerNet – Browse the Internet from Your Disk, Now Open Source
-
Wayback: Self-hosted archiving service integrated with Internet Archive
For archiving, look into https://github.com/dosyago/DiskerNet
It's real next gen thinking on this topic.
As for the featured tool wayback... If HN readers can't figure out what it does after reading docs, its likely the thinking behind it is equally unclear.
- DiskerNet - Save and index web content locally
- Show HN: DiskerNet – save and index web content locally
ripgrep-all
- Ripgrep-all: rga: ripgrep, but also search PDFs, E-Books, Office documents, zip
-
Ripgrep is faster than {grep, ag, Git grep, ucg, pt, sift}
I searched in portage, and it seems there is another version working also with other documents like PDFs and doc.
https://github.com/phiresky/ripgrep-all
-
Calibre – New in Calibre 7.0
If you want even faster search across different formats, you can try ripgrep-all ( https://github.com/phiresky/ripgrep-all ). It can search across epub, docx, pdf, zip, mp4 etc. If you are handy with the tool, you can write custom adaptor to search across images using OCR with tesseract.
- Rga: Ripgrep, but also search in PDF, ebooks, office documents, zip, tar.gz etc.
-
Show HN: Khoj – Chat Offline with Your Second Brain Using Llama 2
1. If you want better adoption especially among corporations, GPL-3 wont cut it. Maybe think of some business friendly licenses (MIT etc)
2. I understand the excitement about llm's. But how about making something more accessible. I use rip-grep-all (rga) along with fzf [1] that can search all files including pdfs in a specific folders. However, I would like a GUI tool to search across multiple folders, provide priority of results across folders and store and search histories where I can do a meta-search. This is sufficient for 95% of my usecases to search locally and I dont need LLM. If khoj can enable such search as default without LLM that will be a gamechanger for many people without a heavy compute machine or who dont want to use OpenAI.
[1] https://github.com/phiresky/ripgrep-all/wiki/fzf-Integration
-
How to make file paths clickable?
I use `rga` to search through multiple PDF files for work. The tool returns a list of files and I would like to make those file paths clickable.
- Burgr – Books in Your Terminal
-
Is there a way to searching multiple epub and pdf?
rga, aka ripgrep-all
-
Internet Archive Scholar
I wanted to say 'au contrer' to your 'screenshots are not searchable' and link this[0] but I don't actually see images in the readme.. I swear it was there, maybe it's a buried extra flag..
[0] https://github.com/phiresky/ripgrep-all
- Recoll – Full-text search for your desktop
What are some alternatives?
min - A fast, minimal browser that protects your privacy
pdfgrep - PDFGrep is a GNU/Emacs module providing grep comparable facilities but for PDF files
SingleFileZ - Web Extension to save a faithful copy of an entire web page in a self-extracting ZIP file
OCRmyPDF - OCRmyPDF adds an OCR text layer to scanned PDF files, allowing them to be searched
BackstopJS - Catch CSS curve balls.
notational-fzf-vim - Notational velocity for vim.
hamsterbase - self-hosted, local-first web archive application.
InvoiceNet - Deep neural network to extract intelligent information from invoice documents.
ZAP - The ZAP core project
fd - A simple, fast and user-friendly alternative to 'find'
Archiver - a streaming interface for archive generation
ripgrep - ripgrep recursively searches directories for a regex pattern while respecting your gitignore