NVIDIA Generative AI LLMs NCA-GENL Question # 14 Topic 2 Discussion

NVIDIA Generative AI LLMs NCA-GENL Question # 14 Topic 2 Discussion

NCA-GENL Exam Topic 2 Question 14 Discussion:
Question #: 14
Topic #: 2

Imagine you are training an LLM consisting of billions of parameters and your training dataset is significantly larger than the available RAM in your system. Which of the following would be an alternative?


A.

Using the GPU memory to extend the RAM capacity for storing the dataset and move the dataset in and out of the GPU, using the PCI bandwidth possibly.


B.

Using a memory-mapped file that allows the library to access and operate on elements of the dataset without needing to fully load it into memory.


C.

Discarding the excess of data and pruning the dataset to the capacity of the RAM, resulting in reduced latency during inference.


D.

Eliminating sentences that are syntactically different by semantically equivalent, possibly reducing the risk of the model hallucinating as it is trained to get to the point.


Get Premium NCA-GENL Questions

Contribute your Thoughts:


Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.