sqlite-utils
compare-go-json
Our great sponsors
sqlite-utils | compare-go-json | |
---|---|---|
34 | 5 | |
1,454 | 18 | |
- | - | |
8.4 | 0.0 | |
12 days ago | almost 2 years ago | |
Python | Go | |
Apache License 2.0 | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
sqlite-utils
-
Why you should probably be using SQLite
Sounds like your problem is with SQLAlchemy, not with SQLite.
My https://sqlite-utils.datasette.io library might be a better fit for you. It's a much thinner abstraction than SQLAlchemy.
-
Welcome to Datasette Cloud
There are a few things you can do here.
SQLite is great at JSON - so I often dump JSON structures in a TEXT column and query them using https://www.sqlite.org/json1.html
I also have plugins for running jq() functions directly in SQL queries - https://datasette.io/plugins/datasette-jq and https://github.com/simonw/sqlite-utils-jq
I've been trying to drive the cost of turning semi-structured data into structured SQL queries down as much as possible with https://sqlite-utils.datasette.io - see this tutorial for more: https://datasette.io/tutorials/clean-data
This is also an area that I'm starting to explore with LLMs. I love the idea that you could take a bunch of messy data, tell Datasette Cloud "I want this imported into a table with this schema"... and it does that.
I have a prototype of this working now, I hope to turn it into an open source plugin (and Datasette Cloud feature) pretty soon. It's using this trick: https://til.simonwillison.net/gpt3/openai-python-functions-d...
-
SQLite Functions for Working with JSON
I've baked a ton of different SQLite tricks - including things like full-text indexing support and advanced alter table methods - into my sqlite-utils CLI tool and Python library: https://sqlite-utils.datasette.io
My Datasette project provides tools for exploring, analyzing and publishing SQLite databases, plus ways to expose them via a JSON API: https://datasette.io
I've also written a ton of stuff about SQLite on my two blogs:
-
Show HN: Trogon – An automatic TUI for command line apps
This is really fun. I have an experimental branch of my sqlite-utils CLI tool (which has dozens of sub-commands) running with this now and it really did only take 4 lines of code - I'm treating Trogon as an optional dependency because people using my package as a Python library rather than a CLI tool may not want the extra installed components:
https://github.com/simonw/sqlite-utils/commit/ec12b780d5dcd6...
There's an animated GIF demo of the result here: https://github.com/simonw/sqlite-utils/issues/545#issuecomme...
-
I'm sure I'm being stupid.. Copying data from an API and making a database
My project https://datasette.io/ is ideal for this kind of thing. You can use https://sqlite-utils.datasette.io/ to load JSON data into a SQLite database, then publish it with Datasette.
-
Just: A Command Runner
I've been using this for about six months now and I absolutely love it.
Make never stuck for me - I couldn't quite get it to fit inside my head.
Just has the exact set of features I want.
Here's one example of one of my Justfiles: https://github.com/simonw/sqlite-utils/blob/fc221f9b62ed8624... - documented here: https://sqlite-utils.datasette.io/en/stable/contributing.htm...
I also wrote about using Just with Django in this TIL: https://til.simonwillison.net/django/just-with-django
-
Ask HN: What Do You Use for a Personal Database
SQLite with the open source toolchain I've been building over the past five years:
https://datasette.io as the interface for running queries against (and visualizing) my data.
https://sqlite-utils.datasette.io/ as a set of tools for creating and modifying my databases (inserting JSON or CSV data, enabling full text search text)
https://dogsheep.github.io as a suite of tools for importing my personal data - see also this talk I gave about that project: https://simonwillison.net/2020/Nov/14/personal-data-warehous...
-
The Perfect Commit
> After identifying about 7 commits (with pretty basic/useless messages, and no PR link!), I then had to find the corresponding PRs based on timestamps, and search the PR history for PRs merged around those timestamps.
Not sure if this would save any time, but it is possible to search PRs by commit. For example, say git blame led me to this commit: https://github.com/simonw/sqlite-utils/commit/129141572f249e...
I could have found PR #373 via this search: https://github.com/simonw/sqlite-utils/pulls?q=bb16f52681b6d...
> I thus treat PRs as ephemeral
I think I see what you're saying but as others have pointed out, sometimes you want to add screenshots etc to the context, and you can't capture this kind of info in commit messages. So then you have two choices: issues or PRs.
> Then any review comments are preferably not addressed directly in the PR
I would think that sometimes you really do want to have a back and forth conversation in the PR, rather than just a "make this change" -> "ok done" type of feedback loop.
I view the PR as an decent place for all of this because it's basically a commit of commits, capturing the related changes/conversation/context all in a single place at the point of merge.
Is this example too large or about the right size for you? https://github.com/simonw/sqlite-utils/commit/ab8d4aad0c42f9...
-
Big data, CSV file exploration. How do you do it? Here's my way
You can use https://sqlite-utils.datasette.io/ to load data into SQLite. It's also possible to import CSVs directly using the `sqlite3` command line tool - I wrote about that here: https://til.simonwillison.net/sqlite/import-csv
compare-go-json
-
The fastest tool for querying large JSON files is written in Python (benchmark)
For me OjG (https://github.com/ohler55/ojg) has been great. I regularly use it on files that can not be loaded into memory. The best JSON file format for multiple record is one JSON document per record all in the same file. OjG doesn't care if they are on different lines. It is fast (https://github.com/ohler55/compare-go-json) and uses a fairly complete JSONPath implementation for searches. Similar to jq but using JSONPath instead of a proprietary query language.
I am biased though as I wrote OjG to handle what other tools were not able to do.
-
OjG now has a tokenizer that is almost 10 times faster than json.Decode
jsoniter is json-iterator/go. It is the 3rd column at https://github.com/ohler55/compare-go-json
The title says it all. The new tokenizer here https://github.com/ohler55/ojg and the benchmarks and comparison to other JSON packages is here https://github.com/ohler55/compare-go-json.
You'll find some examples in second link that does the benchmark. https://github.com/ohler55/compare-go-json.
What are some alternatives?
jsoniter - A high-performance 100% compatible drop-in replacement of "encoding/json"
sqlmodel - SQL databases in Python, designed for simplicity, compatibility, and robustness.
orjson - Fast, correct Python JSON library supporting dataclasses, datetimes, and numpy
sqliteviz - Instant offline SQL-powered data visualisation in your browser
ImportExcel - PowerShell module to import/export Excel spreadsheets, without Excel
octosql - OctoSQL is a query tool that allows you to join, analyse and transform data from multiple databases and file formats using SQL.
q - q - Run SQL directly on delimited files and multi-file sqlite databases
Scoop - A command-line installer for Windows.
datasette - An open source multi-tool for exploring and publishing data
pocketbase - Open Source realtime backend in 1 file
yt-eve - A CLI utility tool
sqlitebrowser - Official home of the DB Browser for SQLite (DB4S) project. Previously known as "SQLite Database Browser" and "Database Browser for SQLite". Website at: