-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
Quantum computing and supercomputers are two different things. Quantum computers are currently an area of research, there isn't a version ready for use apart from some prototypes, and it will probably stay that way for while. Also, quantum computing will most likely not be a completly new architecture, that all of the chips we use will adopt, but an addition to current chipsets for some important but special tasks. Supercomputers, or HPC (High performance clusters) are classic computers, just that they are huge. They use derivatives of "off-the-shelf", but high end, hardware. There is a lot of interesting work in designing such systems, a lot of challenging problems in distributed systems theory, but they aren't a complete detached industry. Using them for work, not designing them, doesn't require a EECS type degree, they guy who sit's next to me in the office, uses a supercomputer to predict protein folding, he is by training a doctor and now does computational microbiology. The applications for massive compute power (often times "just brute force the solution instead of spending years in the lab") are almost endless, but to use them it's not that important to understand the full details of how they are constructed, domain knowledge in the application domain is much more important. If you know how your cluster is structured, and knowledge of slurm etc. will enable you to use the supercomputer just fine, again, they aren't that different from regular computers, just that you workstation might have 1 CPU and your supercomputer has 500. Hiding this complexity is done by slurm or any other resource manager. It's open source as well :) https://github.com/SchedMD/slurm