cargo-update
vector
Our great sponsors
cargo-update | vector | |
---|---|---|
11 | 96 | |
1,125 | 16,427 | |
- | 5.2% | |
6.6 | 9.9 | |
about 1 month ago | 5 days ago | |
Rust | Rust | |
MIT License | Mozilla Public License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
cargo-update
-
Zellij 0.35.1 brings stacked panes to your terminal
Personally, I like cargo-update
- Segfault on network request in Alpine
-
Rust 1.66
Speaking of cargo remove, see also cargo-edit [0] from which adding and removing originally came, as well as cargo-binstall [1] which installs binaries rather than compiling from source every time. The binaries are updatable with cargo-update [2].
The latter two can replace a package manager for Rust related utilities, as I often find that those in OS package repositories are often not as up to date as directly from cargo.
[0] https://github.com/killercup/cargo-edit
[1] https://github.com/cargo-bins/cargo-binstall
[2] https://github.com/nabijaczleweli/cargo-update
-
`cargo audit` can now scan compiled binaries
Would be nice if this worked with cargo-update somehow.
-
Hey Rustaceans! Got a question? Ask here! (26/2022)!
There is cargo install-update plugin: https://github.com/nabijaczleweli/cargo-update
-
go-global-update - the missing command for updating globally installed go executables
I didn't find any command or package to update those packages, and given that npm has npm -g update and cargo has cargo install-update, I decided to create go-global-update for go.
- cargo-update - A cargo subcommand for checking and applying updates to installed executables
-
I just realised Monday is now my favourite day of the week, because in my timezone it’s the day new rust-analyzer releases come out!
rust-analyzer isn't a rust component (like rust-src, etc. which will update with rustup update), nor a cargo binary (where you could use cargo install-update - https://github.com/nabijaczleweli/cargo-update ).
-
Git-cliff: generate changelog files from the Git history
I initially was interested in Rust because of performance + speed + safety, but now I have to say that cargo is a big selling point for me.
I always used to be scared of compiling software myself because I never seemed to be able to get it to work without endless headaches. Now, I generally find it easy to compile Rust programs if they aren't in my package manager, and with cargo install-update https://github.com/nabijaczleweli/cargo-update I find it easy to keep the software up to date. I have higher confidence that I can get hobbyist Rust software working, and the more Rust software I use, the more familiar I am with the ecosystem and the more comfortable I am.
If this was written in some obscure language I wasn't familiar with, I'd be less confident I would be able to run it at all, let alone keep it updated, and I may not bother even trying to install it.
-
DoorDash: Migrating From Python to Kotlin for Our Backend Services
So while it may take a while for some, it's already absolutely fine for me to compile my projects in a few seconds or a minute. I install all my related tooling via cargo install and update it via cargo install-update -a ( https://github.com/nabijaczleweli/cargo-update ) so I frequently/daily build different Rust projects and I'm quite ok with the compilation times.
vector
-
Docker Log Observability: Analyzing Container Logs in HashiCorp Nomad with Vector, Loki, and Grafana
job "vector" { datacenters = ["dc1"] # system job, runs on all nodes type = "system" group "vector" { count = 1 network { port "api" { to = 8686 } } ephemeral_disk { size = 500 sticky = true } task "vector" { driver = "docker" config { image = "timberio/vector:0.30.0-debian" ports = ["api"] volumes = ["/var/run/docker.sock:/var/run/docker.sock"] } env { VECTOR_CONFIG = "local/vector.toml" VECTOR_REQUIRE_HEALTHY = "false" } resources { cpu = 100 # 100 MHz memory = 100 # 100MB } # template with Vector's configuration template { destination = "local/vector.toml" change_mode = "signal" change_signal = "SIGHUP" # overriding the delimiters to [[ ]] to avoid conflicts with Vector's native templating, which also uses {{ }} left_delimiter = "[[" right_delimiter = "]]" data=<
- FLaNK AI Weekly 18 March 2024
- Vector: A high-performance observability data pipeline
-
Hacks to reduce cloud spend
we are doing something similar with OTEL but we are looking at using https://vector.dev/
-
About reading logs
We don't pull logs, we forward logs to a centralized logging service.
-
Self hosted log paraer
opensearch - amazon fork of Elasticsearch https://opensearch.org/docs/latestif you do this an have distributed log sources you'd use logstash for, bin off logstash and use vector (https://vector.dev/) its better out of the box for SaaS stuff.
-
creating a centralize syslog server with elastic search
I have done something similar in the past: you can send the logs through a centralized syslog servers (I suggest syslog-ng) and from there ingest into ELK. For parsing I am advice to use something like Vector, is a lot more faster than logstash. When you have your logs ingested correctly, you can create your own dashboard in Kibana. If this fit your requirements, no need to install nginx (unless you want to use as reverse proxy for Kibana), php and mysql.
-
Show HN: Homelab Monitoring Setup with Grafana
I think there's nothing currently that combines both logging and metrics into one easy package and visualizes it, but it's also something I would love to have.
Vector[1] would work as the agent, being able to collect both logs and metrics. But the issue would then be storing it. I'm assuming the Elastic Stack might now be able to do both, but it's just to heavy to deal with in a small setup.
A couple of months ago I took a brief look at that when setting up logging for my own homelab (https://pv.wtf/posts/logging-and-the-homelab). Mostly looking at the memory usage to fit it on my synology. Quickwit[2] and Log-Store[3] both come with built in web interfaces that reduce the need for grafana, but neither of them do metrics.
- [1] https://vector.dev
-
Retaining Logs generated by service running in pod.
Log to stdout/stderr and collect your logs with a tool like vector (vector.dev) and send it to something like Grafana Loki.
-
Lightweight logging on RPi?
I would recommend that you run vector as a systems service so you don't have to worry about managing it. Here is a basic config to do that - https://github.com/vectordotdev/vector/blob/master/distribution/systemd/vector.service .
What are some alternatives?
Clippy - A bunch of lints to catch common mistakes and improve your Rust code. Book: https://doc.rust-lang.org/clippy/
graylog - Free and open log management
Rustup - The Rust toolchain installer
Fluentd - Fluentd: Unified Logging Layer (project under CNCF)
cargo-deb - A cargo subcommand that generates Debian packages from information in Cargo.toml
agent - Vendor-neutral programmable observability pipelines.
cargo-ebuild - cargo extension that can generate ebuilds using the in-tree eclasses
syslog-ng - syslog-ng is an enhanced log daemon, supporting a wide range of input and output methods: syslog, unstructured text, queueing, SQL & NoSQL.
crate-deps
OpenSearch - 🔎 Open source distributed and RESTful search engine.
git-cliff - A highly customizable Changelog Generator that follows Conventional Commit specifications ⛰️
tracing - Application level tracing for Rust.