Fast JSON schema for Python
ultrajson
Our great sponsors
Fast JSON schema for Python | ultrajson | |
---|---|---|
1 | 3 | |
398 | 4,131 | |
- | 0.5% | |
0.0 | 5.4 | |
19 days ago | 12 days ago | |
Python | C | |
BSD 3-clause "New" or "Revised" License | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Fast JSON schema for Python
-
I wrote okjson - A fast, simple, and pythonic JSON Schema Validator
I had a requirement to process and validate large payloads of JSON concurrently for a web service, initially I implemented it using jsonschema and fastjsonschema but I found the whole JSON Schema Specification to be confusing at times and on top of that wanted better performance. Albeit there are ways to compile/cache the schema, I wanted to move away from the schema specification so I wrote a validation library inspired by the design of tiangolo/sqlmodel (type hints) to solve this problem easier.
ultrajson
-
Processing JSON 2.5x faster than simdjson with msgspec
ujson
-
Benchmarking Python JSON serializers - json vs ujson vs orjson
For most cases, you would want to go with python’s standard json library which removes dependencies on other libraries. On other hand you could try out ujsonwhich is simple replacement for python’s json library. If you want more speed and also want dataclass, datetime, numpy, and UUID instances and you are ready to deal with more complex code, then you can try your hands on orjson
-
The fastest tool for querying large JSON files is written in Python (benchmark)
I asked about this on the Github issue regarding these benchmarks as well.
I'm curious as to why libraries like ultrajson[0] and orjson[1] weren't explored. They aren't command line tools, but neither is pandas right? Is it perhaps because the code required to implement the challenges is large enough that they are considered too inconvenient to use through the same way pandas was used (ie, `python -c "..."`)?
What are some alternatives?
marshmallow - A lightweight library for converting complex objects to and from simple Python datatypes.
jsonschema - JSON Schema validation library
cattrs - Complex custom class converters for attrs.
greenpass-covid19-qrcode-decoder - An easy tool for decoding Green Pass Covid-19 QrCode
python-rapidjson - Python wrapper around rapidjson
serpy - ridiculously fast object serialization
pysimdjson - Python bindings for the simdjson project.
PyLD - JSON-LD processor written in Python
Trafaret - Ultimate transformation library that supports validation, contexts and aiohttp.