dagger
dockertest
dagger | dockertest | |
---|---|---|
93 | 48 | |
10,228 | 3,967 | |
2.4% | 1.0% | |
9.9 | 3.0 | |
7 days ago | 29 days ago | |
Go | Go | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
dagger
- Dagger: Programmable open source CI/CD engine that runs pipelines in containers
-
Nix is a better Docker image builder than Docker's image builder
The fact that I couldn't point to one page on the docs that shows the tl;dr or the what problem is this solving
https://docs.dagger.io/quickstart/562821/hello just emits "Hello, world!" which is fantastic if you're writing a programming language but less helpful if you're trying to replace a CI/CD pipeline. Then, https://docs.dagger.io/quickstart/292472/arguments doubles down on that fallacy by going whole hog into "if you need printf in your pipline, dagger's got your back". The subsequent pages have a lot of english with little concrete examples of what's being shown.
I summarized my complaint in the linked thread as "less cowsay in the examples" but to be honest there are upteen bazillion GitHub Actions out in the world, not the very least of which your GHA pipelines use some https://github.com/dagger/dagger/blob/v0.10.2/.github/workfl... https://github.com/dagger/dagger/blob/v0.10.2/.github/workfl... so demonstrate to a potential user how they'd run any such pipeline in dagger, locally, or in Jenkins, or whatever by leveraging reusable CI functions that setup go or run trivy
Related to that, I was going to say "try incorporating some of the dagger that builds dagger" but while digging up an example, it seems that dagger doesn't make use of the functions yet <https://github.com/dagger/dagger/tree/v0.10.2/ci#readme> which is made worse by the perpetual reference to them as their internal codename of Zenith. So, even if it's not invoked by CI yet, pointing to a WIP PR or branch or something to give folks who have CI/CD problems in their head something concrete to map into how GHA or GitLabCI or Jenkins or something would go a long way
-
Testcontainers
> GHA has "service containers", but unfortunately the feature is too basic to address real-world use cases: it assumes a container image can just β¦ boot! β¦ and only talk to the code via the network. Real world use cases often require serialized steps between the test & the dependencies, e.g., to create or init database dirs, set up certs, etc.)
My biased recommendation is to write a custom Dagger function, and run it in your GHA workflow. https://dagger.io
If you find me on the Dagger discord, I will gladly write a code snippet summarizing what I have in mind, based on what you explained of your CI stack. We use GHA ourselves and use this pattern to great effect.
Disclaimer: I work there :)
-
BuildKit in depth: Docker's build engine explained
Dagger (https://dagger.io) is a great way to use BuildKit through language SDKs. It's such a better paradigm, I cannot imagine going back.
Dagger is by the same folks that brought us Docker. This is their fresh take on solving the problem of container building and much more. BuildKit can more than build images and Dagger unlocks it for you.
-
Cloud, why so difficult? π€·ββοΈ
And suddenly, it's almost painfully obvious where all the pain came from. Cloud applications today are simply a patchwork of disconnected pieces. I have a compiler for my infrastructure, another for my functions, another for my containers, another for my CI/CD pipelines. Each one takes its job super seriously, and keeps me safe and happy inside each of these machines, but my application is not running on a single machine anymore, my application is running on the cloud.
-
Share your DevOps setups
That said I've been moving my CI/CD to https://dagger.io/ which has been FANTASTIC. It's code based so you can define all your pipelines in Go, Python, or Javascript and they all run on containers so I can run actions locally without any special setup. Highly recommended.
-
Whatβs with DevOps engineers using `make` of all things?
You are right make is arcane. But it gets the job done. There are new exciting things happening in this area. Check out https://dagger.io.
-
Shellcheck finds bugs in your shell scripts
> but I'm not convinced it's ready to replace Gitlab CI.
The purpose of Dagger it's not to replace your entire CI (Gitlab in your case). As you can see from our website (https://dagger.io/engine), it works and integrates with all the current CI providers. Where Dagger really shines is to help you and your teams move all the artisanal scripts encoded in YAML into actual code and run them in containers through a fluent SDK which can be written in your language of choice. This unlocks a lot of benefits which are detailed in our docs (https://docs.dagger.io/).
> Dagger has one very big downside IMO: It does not have native integration with Gitlab, so you end up having to use Docker-in-Docker and just running dagger as a job in your pipeline.
This is not correct. Dagger doesn't depend on Docker. We're just conveniently using Docker (and other container runtimes) as it's generally available pretty much everywhere by default as a way to bootstrap the Dagger Engine. You can read more about the Dagger architecture here: https://github.com/dagger/dagger/blob/main/core/docs/d7yxc-o...
As you can see from our docs (https://docs.dagger.io/759201/gitlab-google-cloud/#step-5-cr...), we're leveraging the *default* Gitlab CI `docker` service to bootstrap the engine. There's no `docker-in-docker` happening there.
> It clumps all your previously separated steps into a single step in the Gitlab pipeline.
This is also not the case, we should definitely improve our docs to reflect that. You can organize your dagger pipelines in multiple functions and call them in separate Gitlab jobs as you're currently doing. For example, you can do the following:
```.gitlab-ci.yml
-
Cicada β A FOSS, Cross-Platform Version of GitHub Actions and Gitlab CI
Check out https://dagger.io/. Write declarative pipelines in code, reproducibly run anywhere.
-
Show HN: Togomak β declarative pipeline orchestrator based on HCL and Terraform
Is this similar to Dagger[1] ?
[1] https://dagger.io
dockertest
-
Testcontainers
I am using https://github.com/ory/dockertest for tests, specifically for databases. Is there any advantage to use Testcontainers?
-
Level UP your RDBMS Productivity in GO
Now, let's run the tests. For this purpose, we are going to use dockertest, but test containers is also a good solution.
-
Golang testing using docker services via dockertest
During my path learning go so far I have come across some amazing libraries and utilites, one of my favourite for integration testing is dockertest.
-
How to start a Go project in 2023
Things I can't live without in a new Go project in no particular order:
- https://github.com/golangci/golangci-lint - meta-linter
- https://goreleaser.com - automate release workflows
- https://magefile.org - build tool that can version your tools
- https://github.com/ory/dockertest/v3 - run containers for e2e testing
- https://github.com/ecordell/optgen - generate functional options
- https://golang.org/x/tools/cmd/stringer - generate String()
- https://mvdan.cc/gofumpt - stricter gofmt
- https://github.com/stretchr/testify - test assertion library
- https://github.com/rs/zerolog - logging
- https://github.com/spf13/cobra - CLI framework
FWIW, I just lifted all the tools we use for https://github.com/authzed/spicedb
We've also written some custom linters that might be useful for other folks: https://github.com/authzed/spicedb/tree/main/tools/analyzers
-
Beginner-friendly API made with Go following hexagonal architecture.
I've used dockertest a bunch and it is really amazing.
-
How to unit test your database interactions with Docker
Reminds me of https://github.com/ory/dockertest
-
When to mock and what to mock in a Web API?
If your project is relatively simple and you can get away with recreating your scenarios against a real mock database and run other related services locally. It would be good to setup docker containers for your test scripts and write e2e tests. I believe e2e tests are harder but more useful in understanding/reasoning how users are impacted.
-
Don't Mock the Database
Just a heads up, the repository in your comment is invalid, the correct link is https://github.com/ory/dockertest
-
Mocking database calls without a library?
Don't mock. Use https://github.com/ory/dockertest to actually run tests against a dockerized DB.
-
Different SQL drivers for test and production
Use a library like ory/dockertest to spin up a test database for integration tests. It's easy to use, and tests are still fast. It'll take a minute to download the mysql docker image the first time. But, once it's been downloaded, starting the db, running migrations, and running the tests is still pretty quick.
What are some alternatives?
earthly - Super simple build framework with fast, repeatable builds and an instantly familiar syntax β like Dockerfile and Makefile had a baby.
testcontainers-go - Testcontainers for Go is a Go package that makes it simple to create and clean up container-based dependencies for automated integration/smoke tests. The clean, easy-to-use API enables developers to programmatically define containers that should be run as part of a test and clean up those resources when the test is done.
pipeline - A cloud-native Pipeline resource.
fake-gcs-server - Google Cloud Storage emulator & testing library.
gitlab-ci-local - Tired of pushing to test your .gitlab-ci.yml?
mockaroo - Mock-A-π¦ (mock-aa-roo) a comprehensive HTTP/HTTPS interface mocking tool for all your development and testing needs!
act - Run your GitHub Actions locally π
venom - π Manage and run your integration tests with efficiency - Venom run executors (script, HTTP Request, web, imap, etc... ) and assertions
aws-cdk - The AWS Cloud Development Kit is a framework for defining cloud infrastructure in code
steampipe - Zero-ETL, infinite possibilities. Live query APIs, code & more with SQL. No DB required.
dagster - An orchestration platform for the development, production, and observation of data assets.
go-sqlmock - Sql mock driver for golang to test database interactions