Dice Score not decreasing when doing semantic segmentation

This page summarizes the projects mentioned and recommended in the original post on /r/learnmachinelearning

Our great sponsors
  • WorkOS - The modern identity platform for B2B SaaS
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • SaaSHub - Software Alternatives and Reviews
  • segmentation_models.pytorch

    Segmentation models with pretrained backbones. PyTorch.

  • When i pass the CT-Scans and the masks to the Loss Function, which is the Jaccard-Loss from the segmentation_models.pytorch library, the value does not decrease but stay in the range of 1.0-0.9 over 50 epochs training on only one batch of 32 images. As far as I have understood, my network should overfit and the loss should decrease since I am only training on one batch of a small amount of images. However this does not happen. I also tried more batches with all the data over 100 epochs, but the loss does not decrease either obviously. Does anyone have an idea what I might have done wrong? Do I have to change anything when passing the masks to my loss function?

  • WorkOS

    The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.

    WorkOS logo
NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts