Fine-tuned model consistently producing Precision and Recall scores of 0 from start of training, any suggestions on how to improve?

This page summarizes the projects mentioned and recommended in the original post on /r/MLQuestions

Our great sponsors
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • WorkOS - The modern identity platform for B2B SaaS
  • SaaSHub - Software Alternatives and Reviews
  • electra

    ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators

  • If this is your own implementation of ELECTRA, hopefully you have previous versions you've demonstrated working, you could revert back to a working version, then apply the changes you made one-by-one. If it's open-source code you are using, such as this one, try and find a working example, run it yourself, carefully modify it, preserve it in a working (high performance) state, change it piece-by-piece until it works on your problem.

  • iSarcasmEval

    Datasets used for iSarcasmEval shared-task (Task 6 at SemEval 2022)

  • The labels are extracted and put into their own df which is then fed alongside the text data as tensors to the model. The observations for each class are fairly low due to it being a small but thorough dataset defined and labelled specifically for these tasks, so I can't really change it. However I have been wondering whether I should just generally train the model on sarcasm detection first using a Kaggle dataset or something, then fine tuning again for this subtask (B in the link).

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts