Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

LabelPrecisionRecallF1 ScoreCount
Performance100%100%100%87
Test99.59%100%99.8%245
Question100%97.02%98.49%302
Doc100%90.32%94.92%155
Installation100%84.07%91.35%113
Example100%80.81%89.39%99
Bug100%78.66%88.06%389
Build100%69.87%82.26%156
onnx80%84.21%82.05%23
scala86.67%75%80.41%58
gluon62.28%60.68%61.47%160
flaky96.51%43.46%59.93%194
Feature32.43%98.18%48.76%335
C++55%38.6%45.36%75
ci48.39%40.54%44.12%53
Cuda22.09%100%36.19%86


Precision here representing how accurate our classifier was in correctly labelling an issue given all the times it had predicted that label.

Recall here representing how accurate our classifier was in correctly labelling an issue given all the times the issue actually had that label.

Motivations/Conclusion:

We are able to see which labels the model can predict accurately for. Given a certain accuracy threshold, the bot has the potential to label an issue given that it surpasses this value. As a result, we would be able to accurately provide labels to new issues. 

...