Our great sponsors
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
Hi, I don't have much experience with Nvidia Jetsons. You can find some examples on GitHub (here https://github.com/NVIDIA/cuda-samples). You can find CUDA implementations of most functions on the internet though, you just have to look for the specific thing you are looking for. Cuda kernels are not platform specific, they should work on GPUs and embedded developer boards without problems as long as you respect the limits imposed by the "compute capability" of your device, you just have to compile your code using the right architecture flag. The biggest limit you have to deal with when developing for Jetson nano is the low amount of memory.
Since the ability to use C++ parallel algorithms on the GPU is a relatively new thing, some applications have used other C++ abstraction libraries instead, such as Kokkos (https://kokkos.org/) and RAJA (https://github.com/LLNL/RAJA). These both have multiple backends that support GPUs and CPUs without needing to change your application code.