Skip to content

Confidence does a complete 180 when provided with longer training data input #46

@lukebarton

Description

@lukebarton

https://colab.research.google.com/drive/1llNY6O2kn1GmMroZ352MHO_9XtwM9MyC

swim, expect zing
[('zing', 0.9998135976528079), ('zoob', 0.0001864023471921245)]
[('zoob', 0.8241350650787352), ('zing', 0.17586493492126465)] longer examples
meow, expect zoob
[('zoob', 0.9998439073858366), ('zing', 0.00015609261416340414)]
[('zing', 0.9763082785095902), ('zoob', 0.02369172149040979)] longer examples

classifier goes from being confidently correct, to being confidently incorrect with longer training inputs.

All three of the issues I've raised today make it very difficult to reason about whether the package is working as expected or not -- the behaviour I'm observing makes extremely little sense to my casual understanding.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions