lit VS markup

Compare lit vs markup and see what are their differences.

lit

The Learning Interpretability Tool: Interactively analyze ML models to understand their behavior in an extensible and framework agnostic interface. (by PAIR-code)
SurveyJS - Open-Source JSON Form Builder to Create Dynamic Forms Right in Your App
With SurveyJS form UI libraries, you can build and style forms in a fully-integrated drag & drop form builder, render them in your JS app, and store form submission data in any backend, inc. PHP, ASP.NET Core, and Node.js.
surveyjs.io
featured
InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
lit markup
3 3
3,409 234
1.3% -
9.3 6.9
14 days ago 5 months ago
TypeScript TypeScript
Apache License 2.0 MIT License
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

lit

Posts with mentions or reviews of lit. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2021-11-24.
  • How to create a broad/representative sample from millions of records?
    1 project | /r/LanguageTechnology | 6 Feb 2022
    I'd also suggest looking at your data sample, and how your model handles it, with some kind of exploratory analysis tool. Google's Language Interpretability Tool might work for your scenario. This can give you a lot of ideas about how to prepare the data better.
  • AWS - NLP newsletter November 2021
    2 projects | dev.to | 24 Nov 2021
    Visualize and understand NLP models with the Language Interpretability Tool The Language Interpretability Tool (LIT) is for researchers and practitioners looking to understand NLP model behavior through a visual, interactive, and extensible tool. Use LIT to ask and answer questions like: What kind of examples does my model perform poorly on? Why did my model make this prediction? Can it attribute it to adversarial behavior, or undesirable priors from the training set? Does my model behave consistently if I change things like textual style, verb tense, or pronoun gender? LIT contains many built-in capabilities but is also customizable, with the ability to add custom interpretability techniques, metrics calculations, counterfactual generators, visualizations, and more.
  • Are there any tools for seeing / understanding what a fine-tuned BERT model is looking at for a downstream task?
    2 projects | /r/MLQuestions | 19 Aug 2021
    Use LIT https://github.com/PAIR-code/lit

markup

Posts with mentions or reviews of markup. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-05-15.
  • Show HN: An annotation tool for ML and NLP
    2 projects | news.ycombinator.com | 15 May 2023
    Hey HN! I'm super excited to share Markup with you, which is a totally free & open-source annotation tool that helps you transform unstructured text (e.g. news articles) into structured data that you can use for building, training, or fine-tuning ML models!

    Check it out: https://github.com/samueldobbie/markup

    2 projects | news.ycombinator.com | 19 Jun 2021
    Just to preface this summary, it's all a bit hacked together at the moment, and I'm in the process of rewriting the tool from scratch so this description is privy to change.

    To generate the suggestions there's an active learner with an underlying random forest classifier, that has been fed ~60 seed sentences [1], to classify positive sentences (e.g. contains a prescription) and negative sentences (e.g. doesn't contain a prescription).

    All positive sentences are fed into a sequence-to-sequence RNN model, that has been trained on ~50k synthetic rows of data [2] which maps unstructured sentences (e.g. patient is on pheneturide 250mg twice a day) to a structured output with the desired features (e.g. name: pheneturide; dose: 285; unit: g; frequency: 2). These synthetic sentences were generated with the in-built data generator [3].

    The outputs of the RNN are validated to ensure they meet the expected structure and are valid for the sentence (e.g. the predicted drug name must exist somewhere within the sentence).

    All non-junk predictions are shown to the user who can accept, edit, or reject each. Based on the users' response, the active learner is refined (currently nothing is fed back into the RNN).

    [1] https://github.com/samueldobbie/markup/blob/master/data/text...

    [2] https://raw.githubusercontent.com/samueldobbie/markup/master...

    [3] https://www.getmarkup.com/tools/data-generator/

What are some alternatives?

When comparing lit and markup you can also consider the following projects:

scattertext - Beautiful visualizations of how language differs among document types.

pawls - Software that makes labeling PDFs easy.

bertviz - BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)

xtreme1 - Xtreme1 is an all-in-one data labeling and annotation platform for multimodal data training and supports 3D LiDAR point cloud, image, and LLM.

amazon-sagemaker-examples - Example 📓 Jupyter notebooks that demonstrate how to build, train, and deploy machine learning models using 🧠 Amazon SageMaker.

force-multiplier - Use AI to edit your documents in real-time. Provide feedback and let the AI do all the work.

datalabel - datalabel is a UI-based data editing tool that makes it easy to create labeled text data in a dataframe. With datalabel, you can quickly and effortlessly edit your data without having to write any code. Its intuitive interface makes it ideal for both experienced data professionals and those new to data editing.

langhuan - Light weight labeling engine

refinery - The data scientist's open-source choice to scale, assess and maintain natural language data. Treat training data like a software artifact.