dbx
spark-fast-tests
dbx | spark-fast-tests | |
---|---|---|
5 | 6 | |
434 | 418 | |
2.3% | - | |
4.6 | 0.0 | |
2 months ago | 7 days ago | |
Python | Scala | |
GNU General Public License v3.0 or later | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
dbx
-
Snowpark equivalent on Databricks?
Pyspark is the python API for spark. You can write code in a notebook on databricks and run it on a cluster or you can write code in an IDE and run it using dbx through the dbx execute command. If you’re more familiar with Pandas API, you can use Koalas which is a pandas API on Spark
- how/where do you define your databricks jobs, tasks and workflows?
-
Unit & integration testing in Databricks
Hey, Databricks person here. Check out DBX for a template on how to do unit and integration tests: https://github.com/databrickslabs/dbx
-
My top 5 learnings from driving an OSS project
Approximately 1 year ago I've released the first version of dbx - a CLI tool for simple and efficient development and deployment of Databricks jobs.
- Anyone use Pyspark notebook in production ?
spark-fast-tests
-
Lakehouse architecture in Azure Synapse without Databricks?
I was a Databricks user for 5 years and spent 95% of my time developing Spark code in IDEs. See the spark-daria and spark-fast-tests projects as Scala examples. I developed internal libraries with all the business logic. The Databricks notebooks would consist of a few lines of code that would invoke a function in the proprietary Spark codebase. The proprietary Spark codebase would depend on the OSS libraries I developed in parallel.
-
Well designed scala/spark project
https://github.com/MrPowers/spark-fast-tests https://github.com/97arushisharma/Scala_Practice/tree/master/BigData_Analysis_with_Scala_and_Spark/wikipedia
-
Unit & integration testing in Databricks
If the majority of your stuff is not UDF-based there is an OS solution to run assertion tests against full data frames called spark-fast-tests. The idea here is similar in that you have a it notebook that calls your actual notebook against a staged input reads the output and compares it to a prefabed expected output. This does take a bit of setup and trial and error but it’s the closest I’ve been able to get to proper automated regression testing in databricks
-
Show dataengineering: beavis, a library for unit testing Pandas/Dask code
I am the author of spark-fast-tests and chispa, libraries for unit testing Scala Spark / PySpark code.
-
Ask HN: What are some tools / libraries you built yourself?
I built daria (https://github.com/MrPowers/spark-daria) to make it easier to write Spark and spark-fast-tests (https://github.com/MrPowers/spark-fast-tests) to provide a good testing workflow.
quinn (https://github.com/MrPowers/quinn) and chispa (https://github.com/MrPowers/chispa) are the PySpark equivalents.
Built bebe (https://github.com/MrPowers/bebe) to expose the Spark Catalyst expressions that aren't exposed to the Scala / Python APIs.
Also build spark-sbt.g8 to create a Spark project with a single command: https://github.com/MrPowers/spark-sbt.g8
-
Open source contributions for a Data Engineer?
I've built popular PySpark (quinn, chispa) and Scala Spark (spark-daria, spark-fast-tests) libraries.
What are some alternatives?
databricks-cli - The missing command line client for Databricks SQL
Prefect - The easiest way to build, run, and monitor data pipelines at scale.
cicd-templates - Manage your Databricks deployments and CI with code.
chispa - PySpark test helper methods with beautiful error messages
jupyterlab-integration - DEPRECATED: Integrating Jupyter with Databricks via SSH
soda-sql - Data profiling, testing, and monitoring for SQL accessible data.
nutter - Testing framework for Databricks notebooks
airbyte - The leading data integration platform for ETL / ELT data pipelines from APIs, databases & files to data warehouses, data lakes & data lakehouses. Both self-hosted and Cloud-hosted.
fastdbfs - fastdbfs - An interactive command line client for Databricks DBFS.
sqlfluff - A modular SQL linter and auto-formatter with support for multiple dialects and templated code.
databricks-nutter-projects-demo - Demo of using the Nutter for testing of Databricks notebooks in the CI/CD pipeline [Moved to: https://github.com/alexott/databricks-nutter-repos-demo]
spark-daria - Essential Spark extensions and helper methods ✨😲