alibi-detect

Algorithms for outlier, adversarial and drift detection (by SeldonIO)

Alibi-detect Alternatives

Similar projects and alternatives to alibi-detect

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a better alibi-detect alternative or higher similarity.

alibi-detect reviews and mentions

Posts with mentions or reviews of alibi-detect. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-12-13.
  • Exploring Open-Source Alternatives to Landing AI for Robust MLOps
    18 projects | dev.to | 13 Dec 2023
    Numerous tools exist for detecting anomalies in time series data, but Alibi Detect stood out to me, particularly for its capabilities and its compatibility with both TensorFlow and PyTorch backends.
  • Looking for recommendations to monitor / detect data drifts over time
    3 projects | /r/datascience | 15 Apr 2023
  • [D] Distributions to represent an Image Dataset
    1 project | /r/MachineLearning | 24 Feb 2023
    That is, to see whether a test image belongs in the distribution of the training images and to provide a routine for special cases. After a bit of reading Ive found that this is related to the field of drift detection in which I tried out alibi-detect . Whereby the training images are trained by an autoencoder and any subsequent drift will be flagged by the AE.
  • [D] Which statistical test would you use to detect drift in a dataset of images?
    1 project | /r/MachineLearning | 24 Aug 2022
    Wasserstein distance is not very suitable for drift detection on most problems given that the sample complexity (and estimation error) scales with O(n^(-1/d)) with n the number of instances (100k-10m in your case) and d the feature dimension (192 in your case). More interesting will be to use for instance a detector based on the maximum mean discrepancy (MMD) with estimation error of O(n^(-1/2)). Notice the absence of the feature dimension here. You can find scalable implementations in Alibi Detect (disclosure: I am a contributor): MMD docs, image example. We just added the KeOps backend for the MMD detector to scale and speed up the drift detector further, so if you install from master, you can leverage this backend and easily scale the detector to 1mn instances on e.g. 1 RTX2080Ti GPU. Check this example for more info.
  • Ask HN: Who is hiring? (January 2022)
    28 projects | news.ycombinator.com | 3 Jan 2022
    Seldon | Multiple positions | London/Cambridge UK | Onsite/Remote | Full time | seldon.io

    At Seldon we are building industry leading solutions for deploying, monitoring, and explaining machine learning models. We are an open-core company with several successful open source projects like:

    * https://github.com/SeldonIO/seldon-core

    * https://github.com/SeldonIO/mlserver

    * https://github.com/SeldonIO/alibi

    * https://github.com/SeldonIO/alibi-detect

    * https://github.com/SeldonIO/tempo

    We are hiring for a range of positions, including software engineers(go, k8s), ml engineers (python, go), frontend engineers (js), UX designer, and product managers. All open positions can be found at https://www.seldon.io/careers/

  • What Machine Learning model monitoring tools can you recommend?
    1 project | /r/mlops | 2 Dec 2021
  • Ask HN: Who is hiring? (December 2021)
    37 projects | news.ycombinator.com | 1 Dec 2021
  • [D] How do you deal with covariate shift and concept drift in production?
    2 projects | /r/MachineLearning | 28 Oct 2021
    I work in this area and also contribute to outlier/drift detection library https://github.com/SeldonIO/alibi-detect. To tackle this type of problem, I would strongly encourage following a more principled, fundamentally (statistically) sound approach. So for instance measuring metrics such as the KL-divergence (or many other f-divergences) will not be that informative since it has a lot of undesirable properties for the problem at hand (in order to be informative requires already overlapping distributions P and Q, it is asymmetric, not a real distance metric, will not scale well with data dimensionality etc). So you should probably look at Integral Probability Metrics (IPMs) such as the Maximum Mean Discrepancy (MMD) instead which have much nicer behaviour to monitor drift. I highly recommend the Interpretable Comparison of Distributions and Models NeurIPS workshop talks for more in-depth background.
  • [D] Is this a reasonable assumption in machine learning?
    1 project | /r/MachineLearning | 5 Jul 2021
    All of the above functionality and more can be easily used under a simple API in https://github.com/SeldonIO/alibi-detect.
  • A note from our sponsor - InfluxDB
    www.influxdata.com | 25 Apr 2024
    Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality. Learn more →

Stats

Basic alibi-detect repo stats
9
2,082
7.6
7 days ago

Sponsored
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com