Pass the NVIDIA NVIDIA-Certified Professional NCP-AIO Questions and answers with CertsForce

Viewing page 1 out of 2 pages
Viewing questions 1-10 out of questions
Questions # 1:

An organization has multiple containers and wants to view STDIN, STDOUT, and STDERR I/O streams of a specific container.

What command should be used?

Options:

A.

docker top CONTAINER-NAME


B.

docker stats CONTAINER-NAME


C.

docker logs CONTAINER-NAME


D.

docker inspect CONTAINER-NAME


Expert Solution
Questions # 2:

A Slurm user needs to display real-time information about the running processes and resource usage of a Slurm job.

Which command should be used?

Options:

A.

smap -j


B.

scontrol show job


C.

sstat -j


D.

sinfo -j


Expert Solution
Questions # 3:

An administrator is troubleshooting a bottleneck in a deep learning run time and needs consistent data feed rates to GPUs.

Which storage metric should be used?

Options:

A.

Disk I/O operations per second (IOPS)


B.

Disk free space


C.

Sequential read speed


D.

Disk utilization in performance manager


Expert Solution
Questions # 4:

When troubleshooting Slurm job scheduling issues, a common source of problems is jobs getting stuck in a pending state indefinitely.

Which Slurm command can be used to view detailed information about all pending jobs and identify the cause of the delay?

Options:

A.

scontrol


B.

sacct


C.

sinfo


Expert Solution
Questions # 5:

A GPU administrator needs to virtualize AI/ML training in an HGX environment.

How can the NVIDIA Fabric Manager be used to meet this demand?

Options:

A.

Video encoding acceleration


B.

Enhance graphical rendering


C.

Manage NVLink and NVSwitch resources


D.

GPU memory upgrade


Expert Solution
Questions # 6:

If a Magnum IO-enabled application experiences delays during the ETL phase, what troubleshooting step should be taken?

Options:

A.

Disable NVLink to prevent conflicts between GPUs during data transfer.


B.

Reduce the size of datasets being processed by splitting them into smaller chunks.


C.

Increase the swap space on the host system to handle larger datasets.


D.

Ensure that GPUDirect Storage is configured to allow direct data transfer from storage to GPU memory.


Expert Solution
Questions # 7:

You are managing an on-premises cluster using NVIDIA Base Command Manager (BCM) and need to extend your computational resources into AWS when your local infrastructure reaches peak capacity.

What is the most effective way to configure cloudbursting in this scenario?

Options:

A.

Use BCM's built-in load balancer to distribute workloads evenly between on-premises and cloud resources without any pre-configuration.


B.

Manually provision additional cloud nodes in AWS when the on-premises cluster reaches its limit.


C.

Set up a standby deployment in AWS and manually switch workloads to the cloud during peak times.


D.

Use BCM's Cluster Extension feature to automatically provision AWS resources when local resources are exhausted.


Expert Solution
Questions # 8:

What should an administrator check if GPU-to-GPU communication is slow in a distributed system using Magnum IO?

Options:

A.

Limit the number of GPUs used in the system to reduce congestion.


B.

Increase the system's RAM capacity to improve communication speed.


C.

Disable InfiniBand to reduce network complexity.


D.

Verify the configuration of NCCL or NVSHMEM.


Expert Solution
Questions # 9:

An organization only needs basic network monitoring and validation tools.

Which UFM platform should they use?

Options:

A.

UFM Enterprise


B.

UFM Telemetry


C.

UFM Cyber-AI


D.

UFM Pro


Expert Solution
Questions # 10:

You are tasked with deploying a deep learning framework container from NVIDIA NGC on a stand-alone GPU-enabled server.

What must you complete before pulling the container? (Choose two.)

Options:

A.

Install Docker and the NVIDIA Container Toolkit on the server.


B.

Set up a Kubernetes cluster to manage the container.


C.

Install TensorFlow or PyTorch manually on the server before pulling the container.


D.

Generate an NGC API key and log in to the NGC container registry using docker login.


Expert Solution
Viewing page 1 out of 2 pages
Viewing questions 1-10 out of questions