cubical VS nqthm

Compare cubical vs nqthm and see what are their differences.

nqthm

nqthm - the original Boyer-Moore theorem prover, from 1992 (by John-Nagle)
Our great sponsors
  • WorkOS - The modern identity platform for B2B SaaS
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • SaaSHub - Software Alternatives and Reviews
cubical nqthm
3 4
420 43
2.1% -
8.6 10.0
9 days ago over 7 years ago
Agda Common Lisp
GNU General Public License v3.0 or later GNU General Public License v3.0 only
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

cubical

Posts with mentions or reviews of cubical. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-06-30.
  • Will Computers Redefine the Roots of Math?
    6 projects | news.ycombinator.com | 30 Jun 2023
    For those interested in formalisation of homotopy type theory, there are several (more or less) active and developed libraries. To mention a few:

    UniMath (https://github.com/UniMath/UniMath, mentioned in the article)

    Coq-HoTT (https://github.com/HoTT/Coq-HoTT)

    agda-unimath (https://unimath.github.io/agda-unimath/)

    cubical agda (https://github.com/agda/cubical)

    All of these are open to contributions, and there are lots of useful basic things that haven't been done and which I think would make excellent semester projects for a cs/math undergrad (for example).

  • Homotopy Type Theory
    3 projects | news.ycombinator.com | 23 Jun 2021
  • Cubical Type Theory?
    2 projects | /r/dependent_types | 11 Feb 2021
    In the case of transpension, it seems like one of the uses is proving something about a path in inductive types by cases on an abstract point along that path. For instance, right now, the way that you prove that a path in A + B is either a path in A or a path in B is to define a family by cases and then transport like here. But I think transpension might let you just do cases on a formal intermediate point directly, which would be much simpler.

nqthm

Posts with mentions or reviews of nqthm. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-02-14.
  • Why Is Common Lisp Not the Most Popular Programming Language?
    8 projects | news.ycombinator.com | 14 Feb 2024
    This is a generic problem with macro systems, of course, which is why C deliberately had a weak macro system.

    LISP is a blast from the path. It's fun for retro reasons, but things have moved on.

    [1] https://github.com/John-Nagle/nqthm

    [2] https://github.com/John-Nagle/pasv/tree/master/src/CPC4

  • Will Computers Redefine the Roots of Math?
    6 projects | news.ycombinator.com | 30 Jun 2023
    > In the 70's, this wasn't considered a 'real' proof.

    I ran into that decades ago. We used the original Boyer-Moore theorem prover [1] as part of a program verification system. The system had two provers, the Nelson-Oppen simplifier (the first SAT solver) to automatically handle the easy proofs, and the Boyer-Moore system for the hard ones. To make sure that both had consistent theories, I used the Boyer-Moore prover to prove the "axioms" of the Nelson-Oppen system, especially what are usually called McCarthy's axioms (the ones that use Select and Store) for arrays.

    The Boyer-Moore system uses a strictly constructive approach to mathematics. It starts from something like Peano arithmetic (there is a number zero, and an operation add 1) and builds up number theory. So I added a concept of arrays, represented as (index, value) tuples in sorted order, and was able to prove the usual "axioms" for arrays as theorems.

    The machine proofs were long and involved much case analysis.[2] I submitted a paper to JACM in the early 1980s and got back reviews saying that it was just too long and inelegant to be at the fundamentals of computer science. That might not be the case today.

    A few years back, I put the Boyer-Moore prover on Github, after getting it to work with Gnu Common LISP. So you can still run all this 1980s stuff. It's much faster today. It took about 45 minutes to grind through these proofs on a VAX 11/780 in the early 1980s. Now it takes about a second.

    The proof log [2] is amusing. It's building up number theory from a very low level, starting by proving that X + 0 = X. Each theorem proved can be used as a lemma by later theorems, so you guide the process by giving it problems to solve in the right order. By line 1900, it's proving that multiplication distributes over addition. Array theory, the new stuff, starts around line 2994.

    The reason this is so complicated and ugly is that there's no use of axiomatic set theory. Arrays are easy if you have sets. But there are no sets here. Sets don't fit well into this strict constructive theory, because EQUAL means identical. You can't create weaker definitions of equality which say that two sets are equal if they contain the same elements regardless of order, because that introduces a risk of unsoundness. Effort must be put into keeping the tuples of the array representation in ascending order by subscript, which implies much case analysis. Mathematicians hate case analysis. Computers are good at it.

    [1] https://github.com/John-Nagle/nqthm

    [2] https://github.com/John-Nagle/pasv/blob/master/src/work/temp...

  • Grothendieck's Approach to Equality [pdf]
    2 projects | news.ycombinator.com | 30 May 2022
    which proves that the storing operation always produces a validly ordered array. That's essentially a code proof of correctness for a recursive function The Boyer-Moore prover was able to grind out a proof of that without help. That was a long proof, too.

    I submitted this to JACM. It was rejected, mostly for uglyness. The concept that you needed all this heavy machine-driven case analysis to prove a nice simple "axiom" upset mathematicians. Today it would be less of an issue. People are now more used to proofs that take a lot of grinding through cases.

    You could build up set theory this way, via ordered lists, if you wanted.

    So that's a classic of what happens if you take "equal" seriously.

    [1] http://www-formal.stanford.edu/jmc/towards.pdf

    [2] https://theory.stanford.edu/~arbrad/papers/arrays.pdf

    [3] https://github.com/John-Nagle/pasv/blob/master/src/work/temp...

    [4] https://github.com/John-Nagle/nqthm

  • Exploring Standard ML's robustness to time and interoperability | Request for "analysis" of (or rebuttle for) Common Lisp
    2 projects | /r/lisp | 6 May 2022
    I have played with nqthm, which was last modified in 1992. Only a few modifications were needed to get it to load from Quicklisp, one amusingly was because LaTeX changed over time, so the LaTeX generator had to be changed slightly.

What are some alternatives?

When comparing cubical and nqthm you can also consider the following projects:

Coq-HoTT - A Coq library for Homotopy Type Theory

UniMath - This coq library aims to formalize a substantial body of mathematics using the univalent point of view.

redtt - "Between the darkness and the dawn, a red cube rises!": a proof assistant for cartesian cubical type theory

hott3 - HoTT in Lean 3

mathlib - Lean 3's obsolete mathematical components library: please use mathlib4

pasv - The Pascal-F Verifier