Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

LabelPrecisionRecallF1 ScoreCount
Performance100%100%100%87
Test99.59%100%99.8%245
Question100%97.02%98.49%302
Doc100%90.32%94.92%155
Installation100%84.07%91.35%113
Example100%80.81%89.39%99
Bug100%78.66%88.06%389
Build100%69.87%82.26%156
onnx80%84.21%82.05%23
scala86.67%75%80.41%58
gluon62.28%60.68%61.47%160
flaky96.51%43.46%59.93%194
Feature32.43%98.18%48.76%335
C++55%38.6%45.36%75
ci48.39%40.54%44.12%53
Cuda22.09%100%36.19%86

Data Insights:

Motivations/Conclusion:

We are able to see which labels the model can predict accurately for. Given a certain accuracy threshold, the bot has the potential to label an issue given that it surpasses this value. As a result, we would be able to accurately provide labels to new issues. 

...