schematic VS pronto

Compare schematic vs pronto and see what are their differences.

InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
schematic pronto
2 4
84 6
- -
0.0 0.0
over 3 years ago over 3 years ago
Haskell Java
GNU General Public License v3.0 or later MIT License
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

schematic

Posts with mentions or reviews of schematic. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2022-04-10.
  • A review of JSON Schema libraries for Haskell
    3 projects | dev.to | 10 Apr 2022
    schematic: Last updated in 2021. "It can be thought of as a subset of JSON Schema", "Schematic schemas can be exported to json-schema".
  • Buf raises $93M to deprecate REST/JSON
    6 projects | news.ycombinator.com | 8 Dec 2021
    Thanks for Cap'n Proto. I think the article is clearly indicating the issues that a wider community of conventional type systems in their mainstream languages is not fully aware of. And I disagree with your comments. Firstly, I don't like that you are labelling the author of the article as a "PL design theorist who doesn't have a clue":

    > his article appears to be written by a programming language design theorist who, unfortunately, does not understand (or, perhaps, does not value) practical software engineering.

    I'm not the author, but they mention their prior industrial experience at Google with protobufs.

    I'm not a PL theorist either, and I see that you don't fully understand the problems of composability, compatibility, and versioning and are too eager to dismiss them based on your prior experience with inferior type suystems. And here's why I think it is:

    > > This is especially true when it comes to protocols, because in a distributed system, you cannot update both sides of a protocol simultaneously. I have found that type theorists tend to promote "version negotiation" schemes where the two sides agree on one rigid protocol to follow, but this is extremely painful in practice: you end up needing to maintain parallel code paths, leading to ugly and hard-to-test code. Inevitably, developers are pushed towards hacks in order to avoid protocol changes, which makes things worse.

    You are conflating your experience with particular conventional tooling with the general availability of superior type systems and toolings out there. There's a high demand in utilising their properties in protocol design today.

    Version negotiation is not the only option available to a protocol designer. It is possible to implicitly migrate and up/down-cast schemas in semi-automated way. Example [1]

    > This seems to miss the point of optional fields. Optional fields are not primarily about nullability but about compatibility. Protobuf's single most important feature is the ability to add new fields over time while maintaining compatibility.

    There are at least two ways to achieve compatibility, and the optional fields that expand a domain type to the least common denominator of all encompassing possibilities is the wrong solution to this. Schema evolition via unions and versioning is the proper approach that allows for automatic resolution of compatibility issues.

    > Real-world practice has also shown that quite often, fields that originally seemed to be "required" turn out to be optional over time, hence the "required considered harmful" manifesto. In practice, you want to declare all fields optional to give yourself maximum flexibility for change.

    This is false. In practice I want a schema versioning and deprecation policies, and not ever-growing domain expansion to the blob of all-optional data.

    > It's that way because the "oneof" pattern long-predates the "oneof" language construct. A "oneof" is actually syntax sugar for a bunch of "optional" fields where exactly one is expected to be filled in.

    this is not true either, and it doesn't matter what pattern predates which other pattern. Tagged unions are neither a language construct nor a syntax sugar, it's a property of Type Algebra where you have union- and product-compositions. Languages that implement Type Algebra don't do it to just add another fancy construct, they do it to benefit from mathematical foundations of these concepts.

    > How do you make this change without breaking compatibility?

    you version it, and migrate over time at your own pace without bothering your clients too often [1]

    [1] https://github.com/typeable/schematic#migrations

pronto

Posts with mentions or reviews of pronto. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2021-12-08.
  • Buf raises $93M to deprecate REST/JSON
    6 projects | news.ycombinator.com | 8 Dec 2021
    5. Message streaming (gRPC streams are amazing)

    I can think of a whole host of features that can be built off of protos (I've even built ORMs off of protobuffs for simple things [0]). The value prop is there IMO. HTTP + json APIs are a local minima. The biggest concerns "I want to be able to view the data that is being sent back and forth" is a tooling consideration (curl ... isn't showing you the voltages from the physical layer, it is decoded). Buff is building that tooling.

    [0] - https://github.com/CaperAi/pronto

  • Parsing Gigabytes of JSON per Second
    7 projects | news.ycombinator.com | 23 Oct 2021
    I've written translation layers for such systems and it's not too bad. See this project from $job - 1: https://github.com/CaperAi/pronto

    It allowed us to have a single model for storage in the DB, for sending between services, and syncing to edge devices.

  • gRPC for Microservices Communication
    5 projects | news.ycombinator.com | 23 Sep 2021
    There's no reason you couldn't use gRPC with json as a serialized message format. For example grpc-gateway [0] provides a very effective way of mapping a gRPC concept to HTTP/JSON. The thing is, after moving to gRPC, I've never really felt a desire to move back to JSON. While it may be correct to say "parsing json is fast enough" it's important to note that there's a "for most use cases" after that. Parsing protos is fast enough for even more use cases. You also get streams which are amazing for APIs where you have to sync some large amounts of data (listing large collections from a DB for example) across two services.

    With gRPC you also have a standardized middleware API that is implemented for "all" languages. The concepts cleanly map across multiple languages and types are mostly solved for you.

    Adding to that you can easily define some conventions for a proto and make amazing libraries for your team. At a previous job I made this: https://github.com/CaperAi/pronto/

    Made it super easy to prototype multiple services as if you mock a service backed by memory we could plop it into a DB with zero effort.

    I think this "gRPC vs X" method of thinking isn't appropriate here because protos are more like a Object.prototype in JavaScript. They're a template for what you're sending. If you have the Message you want to send you can serialize that to JSON or read from JSON or XML or another propriety format and automatically get a host of cool features (pretty printing, serialization to text/binary, sending over the network, etc).

    [0] - https://github.com/grpc-ecosystem/grpc-gateway

  • We Went All in on Sqlc/Pgx for Postgres and Go
    31 projects | news.ycombinator.com | 8 Sep 2021
    I attempted to make something similar to this except the opposite direction at a previous job. It was called Pronto: https://github.com/CaperAi/pronto/

    It allowed us to store and query Protos into MongoDB. It wasn't perfect (lots of issues) but the idea was rather than specifying custom models for all of our DB logic in our Java code we could write a proto and automatically and code could import that proto and read/write it into the database. This made building tooling to debug issues very easy and make it very simple to hide a DB behind a gRPC API.

    The tool automated the boring stuff. I wish I could have extended this to have you define a service in a .proto and "compile" that into an ORM DAO-like thing automatically so you never need to worry about manually wiring that stuff ever again.

What are some alternatives?

When comparing schematic and pronto you can also consider the following projects:

rules_proto - Bazel build rules for protobuf / gRPC (now with gazelle)

simd-json - Rust port of simdjson

gRPC - The Java gRPC implementation. HTTP/2 based RPC

pike - Generate CRUD gRPC backends from single YAML description.

gRPC - The C based gRPC (C++, Python, Ruby, Objective-C, PHP, C#)

sqlparser-rs - Extensible SQL Lexer and Parser for Rust

awesome-jsonschema - A curated list of awesome JSON Schema resources, tutorials, tools, and more.

grpc-gateway - gRPC to JSON proxy generator following the gRPC HTTP spec

pggen - A database first code generator focused on postgres

sqlite

ccgo

proteus - A simple tool for generating an application's data access layer.