As you continue iterating on your model and your data, you will likely end up with many Model Runs. Every Model Run represents an experiment (i.e., an instance of training a specific model on some specific data).
At this point, machine learning teams typically want to compare the predictions and the performance of their different models. The goal of comparing models is to measure and understand the marginal value of every machine learning iteration (additional labels, re-worked labels, modeling improvements, hyperparameter fine-tuning...).
Models is designed to help you compare Model Runs: you can visualize predictions and compare metrics between Model Runs.
You will need a model with 2 or more Model Runs to use this feature. Visit these pages to get started:
Once you select the first Model Run, use the drop-down along the menu bar labeled Compare against. Then, select the second Model Run to run the comparison. Labelbox automatically assigns each Model Run a different color to distinguish them in metrics and visualization.
From the gallery, you can click on an individual data row to expand it. Using the sidebar you can toggle which ground truth and Model Runs to view. To learn more about how view predictions look here.
We support comparison of scalar and confusion matrix metrics.
Updated 18 days ago