Temperature adjusts the softmax distribution in decoding. Increasing it (e.g., to 2.0) flattens the curve, giving lower-probability words a better chance, thus increasing diversity—Option C is correct. Option A exaggerates—top words still have impact, just less dominance. Option B is backwards—decreasing temperature sharpens, not broadens. Option D is false—temperature directly alters distribution, not speed. This controls output creativity.
OCI 2025 Generative AI documentation likely reiterates temperature effects under decoding parameters.
Contribute your Thoughts:
Chosen Answer:
This is a voting comment (?). You can switch to a simple comment. It is better to Upvote an existing comment if you don't have anything to add.
Submit