A cloud engineer is looking to deploy a digital fingerprinting pipeline using NVIDIA
Morpheus and the NVIDIA AI Enterprise Virtual Machine Image (VMI).
Where would the cloud engineer find the VMI?
A. Github and Dockerhub
B. Azure, Google, Amazon Marketplaces
C. NVIDIA NGC
D. Developer Forums
A data scientist is training a deep learning model and notices slower than expected training
times. The data scientist alerts a system administrator to inspect the issue. The system
administrator suspects the disk IO is the issue.
What command should be used?
A. tcpdump
B. iostat
C. nvidia-smi
D. htop
What is the primary purpose of assigning a provisioning role to a node in NVIDIA Base Command Manager (BCM)?
A. To configure the node as a container orchestration manager
B. To enable the node to monitor GPU utilization across the cluster
C. To allow the node to manage software images and provision other nodes
D. To assign the node as a storage manager for certified storage
A Slurm user needs to display real-time information about the running processes and
resource usage of a Slurm job.
Which command should be used?
A. smap -j jobid
B. scontrol show job jobid
C. sstat -j job(.step)
D. sinfo -j jobid
An organization only needs basic network monitoring and validation tools.
Which UFM platform should they use?
A. UFM Enterprise
B. UFM Telemetry
C. UFM Cyber-AI
D. UFM Pro
You are managing an on-premises cluster using NVIDIA Base Command Manager (BCM)
and need to extend your computational resources into AWS when your local infrastructure
reaches peak capacity.
What is the most effective way to configure cloudbursting in this scenario?
A. Use BCM's built-in load balancer to distribute workloads evenly between on-premises and cloud resources without any pre-configuration.
B. Manually provision additional cloud nodes in AWS when the on-premises cluster reaches its limit.
C. Set up a standby deployment in AWS and manually switch workloads to the cloud during peak times.
D. Use BCM's Cluster Extension feature to automatically provision AWS resources when local resources are exhausted.
You have successfully pulled a TensorFlow container from NGC and now need to run it on
your stand-alone GPU-enabled server.
Which command should you use to ensure that the container has access to all available
GPUs?
A. kubectl create pod --gpu=all nvcr.io/nvidia/tensorflow:< tag >
B. docker run nvcr.io/nvidia/tensorflow:< tag >
C. docker start nvcr.io/nvidia/tensorflow:< tag >
D. docker run --gpus all nvcr.io/nvidia/tensorflow:< tag >
A system administrator notices that jobs are failing intermittently on Base Command
Manager due to incorrect GPU configurations in Slurm. The administrator needs to ensure
that jobs utilize GPUs correctly.
How should they troubleshoot this issue?
A. Increase the number of GPUs requested in the job script to avoid using unconfigured GPUs.
B. Check if MIG (Multi-Instance GPU) mode has been enabled incorrectly and reconfigure Slurm accordingly.
C. Verify that non-MIG GPUs are automatically configured in Slurm when detected, and adjust configurations if needed.
D. Ensure that GPU resource limits have been correctly defined in Slurm’s configuration file for each job type.
You are using BCM for configuring an active-passive high availability (HA) cluster for a firewall system. To ensure seamless failover, what is one best practice related to session synchronization between the active and passive nodes?
A. Configure both nodes with different zone names to avoid conflicts during failover.
B. Use heartbeat network for session synchronization between active and passive nodes.
C. Ensure that both nodes use different firewall models for redundancy.
D. Set up manual synchronization procedures to transfer session data when needed.
A system administrator needs to collect the information below:
GPU behavior monitoring
GPU configuration management
GPU policy oversight
GPU health and diagnostics
GPU accounting and process statistics
NVSwitch configuration and monitoring
What single tool should be used?
A. nvidia-smi
B. CUDA Toolkit
C. DCGM
D. Nsight Systems
What must be done before installing new versions of DOCA drivers on a BlueField DPU?
A. Uninstall any previous versions of DOCA drivers.
B. Re-flash the firmware every time.
C. Disable network interfaces during installation.
D. Reboot the host system.
You are managing a deep learning workload on a Slurm cluster with multiple GPU nodes,
but you notice that jobs requesting multiple GPUs are waiting for long periods even though
there are available resources on some nodes.
How would you optimize job scheduling for multi-GPU workloads?
A. Reduce memory allocation per job so more jobs can run concurrently, freeing up resources faster for multi-GPU workloads.
B. Ensure that job scripts use --gres=gpu:< number > and configure Slurm’s backfill scheduler to prioritize multi-GPU jobs efficiently.
C. Set up separate partitions for single-GPU and multi-GPU jobs to avoid resource conflicts between them.
D. Increase time limits for smaller jobs so they don’t interfere with multi-GPU job scheduling.
| Page 1 out of 6 Pages |
| 12 |
Real-World Scenario Mastery: Our NCP-AIO practice exam don't just test definitions. They present you with the same complex, scenario-based problems you'll encounter on the actual exam.
Strategic Weakness Identification: Each practice session reveals exactly where you stand. Discover which domains need more attention, before NVIDIA AI Operations exam day arrives.
Confidence Through Familiarity: There's no substitute for knowing what to expect. When you've worked through our comprehensive NCP-AIO practice exam questions pool covering all topics, the real exam feels like just another practice session.