Why torch.softmax was not needed on logits for ComputerVision Model_0 ON FashionMNIST data? #798
Answered
by
mrdbourke
ankitgooner
asked this question in
Q&A
-
While preparing the accuracy fn parameters on test data argmax was directly used without applying softmax prior to it unlike multi class model training eariler |
Beta Was this translation helpful? Give feedback.
Answered by
mrdbourke
Jan 11, 2024
Replies: 1 comment 1 reply
-
Hey @ankitgooner , In short, softmax is a transformation from logits -> prediction probabilities that range from 0 to 1. Softmax makes the logits easier to understand from a human perspective. However, it's not 100% necessary to use softmax when performing argmax on the logits to get the predictions. See a fuller explanation here: #314 |
Beta Was this translation helpful? Give feedback.
1 reply
Answer selected by
ankitgooner
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Hey @ankitgooner ,
In short, softmax is a transformation from logits -> prediction probabilities that range from 0 to 1.
Softmax makes the logits easier to understand from a human perspective.
However, it's not 100% necessary to use softmax when performing argmax on the logits to get the predictions.
See a fuller explanation here: #314