node-gtk
marker
node-gtk | marker | |
---|---|---|
2 | 33 | |
511 | 26,307 | |
0.4% | 3.2% | |
4.9 | 9.8 | |
10 months ago | 3 days ago | |
C++ | Python | |
MIT License | GNU General Public License v3.0 only |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
node-gtk
-
CTA: We need Web Developers to Contribute to GNOME!
I'll say it because it kinda saddens me, but contributing to gnome is not a fun experience. Even though I would be normally super excited about such a request and would be happy to contibute to a FOSS project that I like (I did https://github.com/romgrk/web-toolkit and https://github.com/romgrk/node-gtk after all), my experience with many of the long term contributors to gnome has simply been too disheartening. You guys should think about why so many people are put off from contributing to gnome.
-
GTK 4.2.0 is out! New GL renderer, input hints, and a whole new API reference
Look, all I'm saying is GTK needs to do better in terms of documentation, it has been needing it for years. I've worked on nodejs bindings for years and I've more than once stopped worked on it altogether simply out of frustration due to the absence of proper documentation for the whole ecosystem. Please, don't take all this as an attack, I'd be happy to help. But I think you're limiting the project by applying restrictions that chase those who could help.
marker
-
Launch HN: Reducto Studio (YC W24) – Build accurate document pipelines, fast
Congrats on the launch! How do you guys compare with Datalab with regards to accuracy?
https://www.datalab.to/
- Marker: Convert PDF to Markdown and JSON
-
Open-source 3B param model better than Mistral OCR
How does it compare to Datalab/Marker https://github.com/datalab-to/marker ? We evaluated many PDF->MD converters and this one performed the best, though it is not perfect.
- Using Docling’s OCR features with RapidOCR
-
German parliament votes as a Git contribution graph
I usually convert to markdown from PDF local laws when I need them as a reference for the functional specification. That way its easier to pinpoint to exact section of the thing.
Its not easy to convert general law to markdown, it involved online converters and manual fixes. Currently experimenting with marker [1] on local LLM hardware and so far it is the best out there.
[1]: https://github.com/VikParuchuri/marker
-
Mistral OCR
> with LLM as a judge
For anyone else interested, prompt is here [0]. The model used was gemini-2.0-flash-001.
Benchmarks are hard, and I understand the appeal of having something that seems vaguely deterministic rather than having a human in the loop, but I have a very hard time accepting any LLM-judged benchmarks at face value. This is doubly true when we're talking about something like OCR which, as you say, is a very hard problem for computers of any sort.
I'm assuming you've given this some thought—how did you arrive at using an LLM to benchmark OCR vs other LLMs? What limitations with your benchmark have you seen/are you aware of?
[0] https://github.com/VikParuchuri/marker/blob/master/benchmark...
-
OlmOCR, Ai2's open-source tool to extract clean plain text from PDFs
I'm a fan of the team of Allen AI and their work. Unfortunately, the benchmarking of olmocr against marker (https://github.com/VikParuchuri/marker) is quite flawed.
Throughput - they benchmarked marker API cost vs local inference cost for olmocr. In our testing, marker locally gets 20 - 120 pages per second on an H100 (without custom kernels, etc). Olmocr in our testing gets between .4 (unoptimized) and 4 (sglang) pages per second on the same machine.
Accuracy - their quality benchmarks are based on win rate with only 75 samples - which are different between each tool pair. The samples were filtered down from a set of ~2000 based on opaque criteria. They then asked researchers at Allen AI to judge which output was better. When we benchmarked with our existing set and LLM as a judge, we got a 56% win rate for marker across 1,107 documents. We had to filter out non-English docs, since olmocr is English-only (marker is not).
Hallucinations/other problems - we noticed a lot of missing text and hallucinations with olmocr in our benchmark set. You can see sample output and llm ratings here - https://huggingface.co/datasets/datalab-to/marker_benchmark_... .
You can see all benchmark code at https://github.com/VikParuchuri/marker/tree/master/benchmark... .
Happy to chat more with anyone at Allen AI who wants to discuss this. I think olmocr is a great contribution - happy to help you benchmark marker more fairly.
- Show HN: Benchmarking VLMs vs. Traditional OCR
- Ask HN: What is the best method for turning a scanned book as a PDF into text?
-
Parsing PDFs (and more) in Elixir using Rust
Very cool, any plans for a dockerized API of market similar to what Unstructured released? I know you have a very attractively priced serverless offering (https://www.datalab.to) but having something to develop against locally would be great (for those of us not in the Python world).
What are some alternatives?
gnome-keyring-yubikey-unlock - This is a read-only mirror for https://git.recolic.net/root/gnome-keyring-yubikey-unlock
PyMuPDF - PyMuPDF is a high performance Python library for data extraction, analysis, conversion & manipulation of PDF (and other) documents.
SwiftGtk - A Swift wrapper around gtk-3.x and gtk-4.x that is largely auto-generated from gobject-introspection
llmsherpa - Developer APIs to Accelerate LLM Projects
Marker - 🖊 A gtk3 markdown editor
surya - OCR, layout analysis, reading order, table recognition in 90+ languages