Learning Path
Question & Answer1
Understand Question2
Review Options3
Learn Explanation4
Explore TopicChoose the Best Answer
A
True
B
False
Understanding the Answer
Let's break down why this is correct
Answer
The claim is incorrect: logistic loss, also known as binary cross‑entropy, is designed for two classes, but it can be extended to many classes by using the softmax function and the categorical cross‑entropy loss. In a multi‑class setting each example gets a probability vector over all classes, and the loss sums the negative log of the correct class probability, which is exactly the same idea as binary logistic loss applied to each class separately. For instance, if an image belongs to class “cat,” the model outputs probabilities for cat, dog, and rabbit; the loss penalizes the probability assigned to cat and rewards higher values. This extension, called cross‑entropy or softmax loss, is widely used for tasks like digit recognition or language modeling. Thus, logistic loss is not limited to binary problems; it is the foundation for multi‑class loss functions.
Detailed Explanation
Logistic loss was first made for two classes, but it can be turned into a multi‑class tool by using a softmax trick. Other options are incorrect because The idea that logistic loss only works for two classes comes from its original design.
Key Concepts
Multi-class classification
Loss functions
Optimization techniques
Topic
Multi-class Loss Functions
Difficulty
easy level question
Cognitive Level
understand
Ready to Master More Topics?
Join thousands of students using Seekh's interactive learning platform to excel in their studies with personalized practice and detailed explanations.