ml-compiler-opt
bloaty
ml-compiler-opt | bloaty | |
---|---|---|
7 | 15 | |
668 | 4,956 | |
2.5% | 1.6% | |
9.1 | 6.8 | |
1 day ago | 6 months ago | |
Python | C++ | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
ml-compiler-opt
-
Large Language Models for Compiler Optimization
I did a bit of work on this last summer on (much) smaller models [1] and it was briefly discussed towards the end of last year's MLGO panel [2]. For heuristic replacements specifically, you might be able to glean some things (or just use interpretable models like decision trees), but something like a neural network works fundamentally differently than the existing heuristics, so you probably wouldn't see most of the performance gains. For just tuning heuristics, the usual practice is to make most of the parameters configurable and then use something like bayesian optimization to try and find an optimal set, and this is sometimes done as a baseline in pieces of ML-in-compiler research.
1. https://github.com/google/ml-compiler-opt/pull/109
-
How to make smaller C and C++ binaries
If you're using Clang/LLVM you can also enable ML inlining[1] (assuming you build from source) which can save up to around 7% if all goes well.
There are also talks of work on just brute forcing the inlining for size problem for embedded releases for smallish applications. It's definitely feasible if the problem is important enough to you to throw some compute at it [2].
1. https://github.com/google/ml-compiler-opt
2. https://doi.org/10.1145/3503222.3507744
-
A code optimization Ai?
LLVM's inlining-for-size and register-allocation-for-performance optimizations are both implemented using machine learning models trained by Google.
-
Google AI Proposes ‘MLGO’: A Machine Learning Guided Compiler Optimization Python Framework
Continue reading | Checkout the paper, github, demo and ref article.
-
Google ML Compiler Inlining Achieves 3-7% Reduction in Size
Looks like they do have a pretrained model:
https://github.com/google/ml-compiler-opt/releases/download/...
The code will by default auto-download it during the build process. It's about 800 kbytes, which seems very reasonable for something that will reduce the generated code size by gigabytes for a large codebase.
bloaty
-
ESP32-C3 Wireless Adventure: A Comprehensive Guide to IoT [pdf]
ESP32s aren't really ‘lower level’ in the sense that anyone is likely to write assembly code for them (compared to, say, 8051 or PIC), other than maybe some driver author at Espressif. The big win from using RISC-V, other than name recognition, is mainstream compiler support (which is nothing to sneeze at, especially when it's largely funded by someone else).
When I worked on Matter¹, the Xtensa and RISC-V versions were basically fungible from the software point of view. (And really, so were other vendors' various ARMs.) We did find that Bloaty McBloatface² didn't support Xtensa, so I had to write an alternative.
¹ https://github.com/project-chip/connectedhomeip/
² https://github.com/google/bloaty
-
How to make smaller C and C++ binaries
I’ve gotten good insight into what takes up space in binaries by profiling with Bloaty (https://github.com/google/bloaty). My last profiling session showed that clang’s ThinLTO was inlining too aggressively in some cases, causing functions that should be tiny to be 75 kB+.
-
Reducing Tailscale’s binary size on macOS
I'm surprised they didn't go for the binary size analysis tools like
https://github.com/google/bloaty
Or goweight.
- C extension making everything bigger
- Template code bloat - how to measure, and what does that even mean?
-
Do you usually use periphery (or other code optimization tools) so that your final built release app is fast/ small?
I was able to shave a few % off our app binary with Bloaty. It’s pretty hard to use but once you figure out how to make regular expressions to properly classify things from your codebase, you can really visually analyze what your binary is composed of.
- how to compare two .so(shared lib) files for size
-
Debugging/optimizing/diagnostic tools for C++
Bloaty
- Bloaty McBloatface: a size profiler for binaries
- Bloaty McBloatface
What are some alternatives?
LOGICGUIDE - Plug in and Play implementation of "Certified Reasoning with Language Models" that elevates model reasoning by 40%
prometheus-cpp - Prometheus Client Library for Modern C++
capstone - Capstone disassembly/disassembler framework for ARM, ARM64 (ARMv8), Alpha, BPF, Ethereum VM, HPPA, LoongArch, M68K, M680X, Mips, MOS65XX, PPC, RISC-V(rv32G/rv64G), SH, Sparc, SystemZ, TMS320C64X, TriCore, Webassembly, XCore and X86.
periphery - A tool to identify unused code in Swift projects.
connectedhomeip - Matter (formerly Project CHIP) creates more connections between more objects, simplifying development for manufacturers and increasing compatibility for consumers, guided by the Connectivity Standards Alliance.