bytewax
Flask
Our great sponsors
bytewax | Flask | |
---|---|---|
18 | 135 | |
1,144 | 66,350 | |
8.2% | 0.8% | |
9.8 | 8.7 | |
6 days ago | 6 days ago | |
Python | Python | |
Apache License 2.0 | BSD 3-clause "New" or "Revised" License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
bytewax
- Building a streaming SQL engine with Arrow and DataFusion
-
Near Real Time Ingestion to DB using Python
You can probably use Python to solve your problem, there are many ways you can speed up your deserialization/flattening. I work on Bytewax (https://github.com/bytewax/bytewax) and I wouldn't mention it if it wasn't a good fit, but I think it's worth looking at here. It is a stream processor that makes it easy to scale, maintain order, track progress, and you just write native Python.
-
Stream processing framework for a new project in Python
Disclaimer: I work on Bytewax, but it feels like this could be a good fit and would save you some time looking around. If you need to do stateful operations (reduce, window, etc.) then you can use bytewax - https://github.com/bytewax/bytewax with pub/sub, but you would need to build a custom connector. There are some guides on how to do that - https://www.bytewax.io/blog/custom-input-connector.
- What are your favorite tools or components in the Kafka ecosystem?
-
A Python package for streaming synthetic data
This is great, definitely see the utility here. I have had to hack this together so many times while building streaming workflows with github.com/bytewax/bytewax and other tools.
-
Snowflake - what are the streaming capabilities it provides?
When low latency matters you should always consider an ETL approach rather than ELT, e.g. collect data in Kafka and process using Kafka Streams/Flink in Java or Quix Streams/Bytewax in Python, then sink it to Snowflake where you can handle non-critical workloads (as is the case for 99% of BI/analytics). This way you can choose the right path for your data depending on how quickly it needs to be served.
-
Sunday Daily Thread: What's everyone working on this week?
Working on how to use https://github.com/bytewax/bytewax to create embeddings in real-time for ML use cases. I want to make a small library for embedding pipelines, but still learning about vector dbs and the tradeoffs between the different solutions.
-
Arroyo: A distributed stream processing engine written in Rust
Project looks cool! Glad you open sourced it. It could use some comments in the code base to help contributors ;). I also like the datafusion usage, that is awesome. BTW I work on github.com/bytewax/bytewax, which is based on https://github.com/TimelyDataflow/timely-dataflow another Rust dataflow computation engine.
-
Launch HN: BuildFlow (YC W23) – The FastAPI of data pipelines
Cool, nice idea. Can you sub in different backend like bytewax (https://github.com/bytewax/bytewax) for stateful processing?
-
Kafka Stream Processing in Java or Scala
If you want to keep in your Python/SQL area of expertise and by all means I don't mean to promote not learning a new language, but just as an FYI. There are some non-Java/Scala tools between streaming databases like risingwave and materialize, streaming platforms like fluvio and redpanda, and stream processors like bytewax and faust.
Flask
-
Ask HN: High quality Python scripts or small libraries to learn from
I'd suggest Flask or some of the smaller projects in the Pallets ecosystem:
https://github.com/pallets/flask
-
Rapid Prototyping with Flask, Bootstrap and Secutio
#!/usr/bin/python # # https://flask.palletsprojects.com/en/3.0.x/installation/ # from flask import Flask, jsonify, request contacts = [ { "id": "1", "firstname": "Lorem", "lastname": "Ipsum", "email": "[email protected]", }, { "id": "2", "firstname": "Mauris", "lastname": "Quis", "email": "[email protected]", }, { "id": "3", "firstname": "Donec Purus", "lastname": "Purus", "email": "[email protected]", } ] app = Flask(__name__, static_url_path='', static_folder='public',) @app.route("/contact//save", methods=["PUT"]) def save_contact(id): data = request.json contacts[id - 1] = data return jsonify(contacts[id - 1]) @app.route("/contact/", methods=["GET"]) @app.route("/contact//edit", methods=["GET"]) def get_contact(id): return jsonify(contacts[id - 1]) @app.route('/') def root(): return app.send_static_file('index.html') if __name__ == '__main__': app.run(debug=True)
- Microdot "The impossibly small web framework for Python and MicroPython"
-
Why do all the popular projects use relative imports in __init__ files if PEP 8 recommends absolute?
I was looking at all the big projects like numpy, pytorch, flask, etc.
-
10 Github repositories to achieve Python mastery
Explore here.
-
Ask HN: What would you use to build a mostly CRUD back end today?
I may use Flask-Admin initially to offload the "CRUD" operations to have an initial prototype fast but then drop it ASAP because I don't want to write a "flask-admin application" to fight against later on. If the application is mainly "CRUD", then Flask-Admin is suitable.
Now...
Would you do a breakdown/list of all the jobs you've done by sector/vertical and by function/role and by application functionality?
- [0]: https://flask.palletsprojects.com
- [1]: https://flask-admin.readthedocs.io/en/latest
- [2]: https://flask.palletsprojects.com/en/2.3.x/patterns/celery
- [3]: https://sentry.io
- [4]: https://posthog.com
- [5]: https://www.docker.com
-
Implementing continuous delivery pipelines with GitHub Actions
In the lab to follow, we will be setting up an end-to-end DevOps workflow for a Flask microservice with GitHub Actions, using a self-managed custom runner for maximal control over the pipeline execution environment and automating deployments to a local Kubernetes cluster. Furthermore, we will construct separate pipelines for our "development" and "production" environments to further elaborate on the concepts of continuous deployment and delivery.
- How do you iterate on a library built locally?
-
Flask Application Load Balancing using Docker Compose and Nginx
Flask Micro web Framework: You will use Flask to build a Flask web application.
-
Open Source Flask-based web applications
In an earlier post I mentioned a bunch of Open Source web applications. Let's now focus on the ones written in Python using Flask the light-weight web framework.
What are some alternatives?
timely-dataflow - A modular implementation of timely dataflow in Rust
fastapi - FastAPI framework, high performance, easy to learn, fast to code, ready for production
arroyo - Distributed stream processing engine in Rust
Django - The Web framework for perfectionists with deadlines.
2022-bytewax-redpanda-air-quality-monitoring
AIOHTTP - Asynchronous HTTP client/server framework for asyncio and Python
django-unicorn - The magical reactive component framework for Django ✨
starlette - The little ASGI framework that shines. 🌟
quart - An async Python micro framework for building web applications.
Pyramid - Pyramid - A Python web framework
Tornado - Tornado is a Python web framework and asynchronous networking library, originally developed at FriendFeed.