HumesGuillotine
awesome-causality-algorithms
HumesGuillotine | awesome-causality-algorithms | |
---|---|---|
3 | 1 | |
1 | 2,796 | |
- | - | |
4.8 | 3.5 | |
about 1 month ago | 9 months ago | |
- | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
HumesGuillotine
-
Learning Universal Predictors
As the guy who suggested to Marcus a lossless compression prize to replace the Turing Test, I've got to confess that all this pedantic sophistry "critiquing" algorithmic information is there for a good reason. In the immortal words of Mel Brooks: "We've got to protect our phoney baloney jobs gentlemen!"
https://youtu.be/bpJNmkB36nE
There is actually more at stake here than machine learning. This gets to the root of "bias" in the scientific method. Imagine what horrors, what risks, what chaos would be ours if a truly objective information criterion for causal model selection were to exist! Why, virtually every "sociologist" would be hauled to Hume's Guillotine in a Reign of Terror!
https://github.com/jabowery/HumesGuillotine
But to be clear, Marcus and I have a disagreement about pragmatics of such an approach to dispute processing in the natural sciences. He believes, for example, that the dispute over climate change should be handled by the standard processes in place with academia. My approach differs, based on my hard won experience with reform reforming institutional incentives:
https://jimbowery.blogspot.com/2018/04/necessity-and-incenti...
When it comes to multi-trillion dollar scientific questions, the conflicts of interest become so intense that you really need to apply a gold standard for objectivity and that is the single number: How big is your executable archive of the data in evidence.
While I understand the machine learning world looms as a rival for "unbiased" academic research, it nevertheless remains true that even in this emerging "marketplace of ideas", there is no formal definition of "bias" that disciplines discourse and thereby guides development at the institutional, let alone technical level. Everyone is weighing in with their fuzzy notions of "bias" that betray intense motivations when there has been, for over 50 years, a very clear and present mathematical definition.
-
Elon Musk proposes that a new version of quantum mechanics/cosmology, will be derived, possibly by using his version of artificial intelligence "xAI".
See Hume's Guillotine at github for what Musk should be pursuing.
-
Market price of power, as produced by the Suncell will not be very low, for a long time
This is one of the reasons I've been advocating a philanthropic prize for macrosocial modeling: Ockham's Guillotine: Beheading the social pseudosciences.
awesome-causality-algorithms
-
Why the world needs computational social science
https://github.com/rguo12/awesome-causality-algorithms
"The limits of graphical causal discovery" (2021)
What are some alternatives?
causalnex - A Python library that helps data scientists to infer causation rather than observing correlation.
looper - A resource list for causality in statistics, data science and physics
Rath - Next generation of automated data exploratory analysis and visualization platform.
genome_integration - MR-link and genome integration. genome_integration is a repository for the analysis of genomic data. Specifically, the repository implements the causal inference method MR-link, as well as other Mendelian randomization methods.
pgmpy - Python Library for learning (Structure and Parameter), inference (Probabilistic and Causal), and simulations in Bayesian Networks.
LightFM - A Python implementation of LightFM, a hybrid recommendation algorithm.
causal-learn - Causal Discovery in Python. It also includes (conditional) independence tests and score functions.
spotlight - Deep recommender models using PyTorch.
dowhy - DoWhy is a Python library for causal inference that supports explicit modeling and testing of causal assumptions. DoWhy is based on a unified language for causal inference, combining causal graphical models and potential outcomes frameworks.
Eliot - Eliot: the logging system that tells you *why* it happened