Month End Sale Special - Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: sntaclus

Nutanix AI officially supports which two LLM packaging formats? (Choose two.)

A.

NVIDIA NIM

B.

ONNX

C.

OLLAMA

D.

Hugging Face

An administrator needs to spot the busiest credentials at a glance.

Which Dashboard widget provides insight into the most frequently used credentials?

A.

Endpoints Summary

B.

API Keys (Top 5)

C.

Infrastructure Summary

D.

API Requests Trends

An administrator is managing a Nutanix AI cluster used for NLP (Natural Language Processing) training. A data scientist reports that training jobs intermittently stall and fail to complete within the expected time window. The administrator reviews the performance data for the VM hosting the job and finds:

    The VM is configured with passthrough access to a dedicated GPU

    Memory ballooning is active, and swap usage is increasing

    CPU utilization is moderate (~60%)

    GPU utilization is stable and high (~85%)

    The VM has 8 vCPUs and 24 GB of RAM assigned

    NCC shows no hardware or driver issues

What is the most appropriate optimization to improve workload stability and performance?

A.

Reduce the number of vCPUs allocated to lower the CPU scheduling overhead.

B.

Add additional vGPUs to the VM to reduce processing time.

C.

Increase the VM's RAM to eliminate memory ballooning and swap usage.

D.

Disable GPU passthrough and use a shared vGPU profile instead.

Which task is an AI/ML User unable to perform in Nutanix Enterprise AI?

A.

Create an API key to access an AI inference endpoint.

B.

Delete an endpoint that they previously created.

C.

Test an AI inference endpoint that they previously created.

D.

View a LLM that was imported by a different user.

What does the Sample Request code include?

A.

Endpoint URL and API key

B.

User credentials and API key

C.

Endpoint URL and GPU type

D.

GPU type and max_tokens

Which deployment type of Nutanix Enterprise AI is supported in Amazon EKS?

A.

Standalone

B.

Cloud Cluster

C.

GPT-in-a-Box 2.0

D.

Bare Metal

An administrator is monitoring the performance of a deployed Large Language Model within the Nutanix Enterprise AI platform. After initial deployment, users report slow inference response times and occasional timeouts when accessing the model through its API endpoint.

The administrator reviews the performance metrics available in the NAI Dashboard and notes the following:

    CPU usage is consistently high across all inference-serving containers.

    Memory utilization is nearing the allocated limits for the model service.

    The request latency graph shows increasing average inference times during peak usage.

Which action should the administrator take to improve performance and reduce latency?

A.

Restart the model container to clear memory cache and allow the system to rebalance performance.

B.

Scale out the number of instances and allocate additional CPU and memory resources.

C.

Disable logging temporarily to reduce resource consumption during peak load periods.

D.

Increase the number of API keys assigned to the endpoint to allow more concurrent access.

In a new instance of Nutanix Enterprise AI, what task must be completed to create an API key?

A.

Download LLM and create endpoint.

B.

Provision an API key in NKP.

C.

Provision a GPU worker node.

D.

Create an endpoint first.

What prerequisites must be skipped when adding GPU nodes to a managed Kubernetes service on an Azure AKS cluster using Azure CLI?

A.

Install the GPU driver

B.

Install the Worker driver

C.

Install the Worker Connector

D.

Install the GPU Connector

Which Nutanix AI deployment platform is supported?

A.

Native Kubernetes

B.

AWS EC2

C.

Nutanix NKE

D.

On-premises NCI