NVIDIA Generative AI LLMs NCA-GENL Question # 6 Topic 1 Discussion

NVIDIA Generative AI LLMs NCA-GENL Question # 6 Topic 1 Discussion

NCA-GENL Exam Topic 1 Question 6 Discussion:
Question #: 6
Topic #: 1

In transformer-based LLMs, how does the use of multi-head attention improve model performance compared to single-head attention, particularly for complex NLP tasks?


A.

Multi-head attention reduces the model’s memory footprint by sharing weights across heads.


B.

Multi-head attention allows the model to focus on multiple aspects of the input sequence simultaneously.


C.

Multi-head attention eliminates the need for positional encodings in the input sequence.


D.

Multi-head attention simplifies the training process by reducing the number of parameters.


Get Premium NCA-GENL Questions

Contribute your Thoughts:


Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.