sqlmodel
sqlite-utils
Our great sponsors
sqlmodel | sqlite-utils | |
---|---|---|
23 | 35 | |
12,894 | 1,498 | |
- | - | |
8.6 | 8.4 | |
8 days ago | 14 days ago | |
Python | Python | |
MIT License | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
sqlmodel
-
SQLModel with the same relationship column twice
Seems like this is a known bug in SQLModel: https://github.com/tiangolo/sqlmodel/issues/10
-
Best ORM to use with FastAPI?
I have not used it myself but the creator of fastapi has made https://github.com/tiangolo/sqlmodel
-
SQLAlchemy: Parent instance is not bound to a Session; lazy load operation of attribute cannot proceed
I have already posted this question in Stack Overflow and GitHub and have been ignored in both 😢. You guys are my last hope.
-
I wrote okjson - A fast, simple, and pythonic JSON Schema Validator
I had a requirement to process and validate large payloads of JSON concurrently for a web service, initially I implemented it using jsonschema and fastjsonschema but I found the whole JSON Schema Specification to be confusing at times and on top of that wanted better performance. Albeit there are ways to compile/cache the schema, I wanted to move away from the schema specification so I wrote a validation library inspired by the design of tiangolo/sqlmodel (type hints) to solve this problem easier.
- Django Ninja – Fast Django REST Framework for Building APIs
-
Trending Python Projects of the Week
Github Repository
-
Tuesday Daily Thread: Advanced questions
I would say as long as your current solution works and is easy to maintain keep it. If you want to switch I would recommend FastAPI, it is new(ish), but definitely old enough to have been tested and used in a large variety of production usecases. In your case it might be interesting to have a look at SQLModel (works with FastAPI, same author), especially if the API endpoints match closely to the database objects. https://github.com/tiangolo/sqlmodel
-
The hand-picked selection of the best Python libraries released in 2021
SQLModel.
-
Pydbantic - A single model ( DB & Pydantic) with automatic migrations
Sounds similar to https://github.com/tiangolo/sqlmodel
-
tiangolo/SQLModel DoA?
There was a lot of hype and excitement around the release of SQLModel, a Pydantic + SQLAlchemy hybrid Model library with native integration for FastAPI. I pulled it out just now and there hasn't been any update beyond the initial anouncment on August 25th.
sqlite-utils
-
Ask HN: High quality Python scripts or small libraries to learn from
https://github.com/simonw/sqlite-utils
So, his code might not be a good place to find best patterns (for ex, I don't think they are fully typed), but his repos are very pragmatic, and his development process is super insightful (well documented PRs for personal repos!). Best part, he blogs about every non-trivial update, so you get all the context!
-
Why you should probably be using SQLite
Sounds like your problem is with SQLAlchemy, not with SQLite.
My https://sqlite-utils.datasette.io library might be a better fit for you. It's a much thinner abstraction than SQLAlchemy.
-
Welcome to Datasette Cloud
There are a few things you can do here.
SQLite is great at JSON - so I often dump JSON structures in a TEXT column and query them using https://www.sqlite.org/json1.html
I also have plugins for running jq() functions directly in SQL queries - https://datasette.io/plugins/datasette-jq and https://github.com/simonw/sqlite-utils-jq
I've been trying to drive the cost of turning semi-structured data into structured SQL queries down as much as possible with https://sqlite-utils.datasette.io - see this tutorial for more: https://datasette.io/tutorials/clean-data
This is also an area that I'm starting to explore with LLMs. I love the idea that you could take a bunch of messy data, tell Datasette Cloud "I want this imported into a table with this schema"... and it does that.
I have a prototype of this working now, I hope to turn it into an open source plugin (and Datasette Cloud feature) pretty soon. It's using this trick: https://til.simonwillison.net/gpt3/openai-python-functions-d...
-
SQLite Functions for Working with JSON
I've baked a ton of different SQLite tricks - including things like full-text indexing support and advanced alter table methods - into my sqlite-utils CLI tool and Python library: https://sqlite-utils.datasette.io
My Datasette project provides tools for exploring, analyzing and publishing SQLite databases, plus ways to expose them via a JSON API: https://datasette.io
I've also written a ton of stuff about SQLite on my two blogs:
- https://simonwillison.net/tags/sqlite/
- https://til.simonwillison.net/sqlite
-
Show HN: Trogon – An automatic TUI for command line apps
This is really fun. I have an experimental branch of my sqlite-utils CLI tool (which has dozens of sub-commands) running with this now and it really did only take 4 lines of code - I'm treating Trogon as an optional dependency because people using my package as a Python library rather than a CLI tool may not want the extra installed components:
https://github.com/simonw/sqlite-utils/commit/ec12b780d5dcd6...
There's an animated GIF demo of the result here: https://github.com/simonw/sqlite-utils/issues/545#issuecomme...
-
I'm sure I'm being stupid.. Copying data from an API and making a database
My project https://datasette.io/ is ideal for this kind of thing. You can use https://sqlite-utils.datasette.io/ to load JSON data into a SQLite database, then publish it with Datasette.
-
Just: A Command Runner
I've been using this for about six months now and I absolutely love it.
Make never stuck for me - I couldn't quite get it to fit inside my head.
Just has the exact set of features I want.
Here's one example of one of my Justfiles: https://github.com/simonw/sqlite-utils/blob/fc221f9b62ed8624... - documented here: https://sqlite-utils.datasette.io/en/stable/contributing.htm...
I also wrote about using Just with Django in this TIL: https://til.simonwillison.net/django/just-with-django
-
Ask HN: What Do You Use for a Personal Database
SQLite with the open source toolchain I've been building over the past five years:
https://datasette.io as the interface for running queries against (and visualizing) my data.
https://sqlite-utils.datasette.io/ as a set of tools for creating and modifying my databases (inserting JSON or CSV data, enabling full text search text)
https://dogsheep.github.io as a suite of tools for importing my personal data - see also this talk I gave about that project: https://simonwillison.net/2020/Nov/14/personal-data-warehous...
-
The Perfect Commit
Here's an example: https://github.com/simonw/sqlite-utils/pull/468
> After identifying about 7 commits (with pretty basic/useless messages, and no PR link!), I then had to find the corresponding PRs based on timestamps, and search the PR history for PRs merged around those timestamps.
Not sure if this would save any time, but it is possible to search PRs by commit. For example, say git blame led me to this commit: https://github.com/simonw/sqlite-utils/commit/129141572f249e...
I could have found PR #373 via this search: https://github.com/simonw/sqlite-utils/pulls?q=bb16f52681b6d...
> I thus treat PRs as ephemeral
I think I see what you're saying but as others have pointed out, sometimes you want to add screenshots etc to the context, and you can't capture this kind of info in commit messages. So then you have two choices: issues or PRs.
> Then any review comments are preferably not addressed directly in the PR
I would think that sometimes you really do want to have a back and forth conversation in the PR, rather than just a "make this change" -> "ok done" type of feedback loop.
I view the PR as an decent place for all of this because it's basically a commit of commits, capturing the related changes/conversation/context all in a single place at the point of merge.
What are some alternatives?
pydantic-sqlalchemy - Tools to convert SQLAlchemy models to Pydantic models
sqliteviz - Instant offline SQL-powered data visualisation in your browser
pydantic - Data validation using Python type hints
ImportExcel - PowerShell module to import/export Excel spreadsheets, without Excel
SQLAlchemy - The Database Toolkit for Python
octosql - OctoSQL is a query tool that allows you to join, analyse and transform data from multiple databases and file formats using SQL.
ormar - python async orm with fastapi in mind and pydantic validation
q - q - Run SQL directly on delimited files and multi-file sqlite databases
geojson-pydantic - Pydantic data models for the GeoJSON spec
Scoop - A command-line installer for Windows.
sqlalchemy-hana - SQLAlchemy Dialect for SAP HANA
datasette - An open source multi-tool for exploring and publishing data