Skip to main content
Version: 1.14.0

Features

The following table describes the key features of the EGS platform.

Main FeatureSub FeatureDescription
Multi-Cluster and Multi-Cloud Fleet ManagementCluster registrationSupports EKS, GKE, OKE, AKS, Akamai, on-prem cluster registration and management
Create or delete workspaceWorkspace have one or more namespaces associated with it.
Users can deploy AI workloads in the Workspace namespaces.
Workspace has a GPU VPC. By default, no GPUs are available for workspace.
A GPR (GPU Provision Request) is required to provision GPUs for a workspace.
Multi-Tenancy: WorkspacesOne or more namespaces per workspaceOne or more namespaces can be associated with a workspace.
One or more worker clusters per workspaceWorkspace can span across clusters. One or more clusters can be associated with a slice workspace.
Admin Access to PortalAdmin Profile and RBACRBAC for Admin access
Access to Admin PortalAn individual portal for an admin.
User Access to PortalUser per workspaceUser per workspace
Access to User PortalAn individual portal for a user to manage GPRs and AI workloads within a specific slice workspace.
User profile and RBACUser Workspace RBAC and manage that workspace.
Inventory ManagementGPU node pools, inventory list, and allocation statusInventory management. Nvidia GPUs. GPU node pools. By default, all GPU pools/nodes are managed by EGS.
Failover and ResilienceAutomatically bursts workloads to secondary clusters during hardware failures or resource exhaustion in the primary cluster.
GPU Provisioning Requests (GPRs)Create/delete/update GPRGPU provisioning request for a workspace. Request to include a number of parameters such as GPU type, number, memory, nodes, cluster, and so on.
Auto Eviction of a GPRAn admin can enable auto eviction of a GPR while registering a worker cluster.
GPR Auto RemediationTo prevent GPU downtime, auto remediation can be configured during GPR creation.
GPR TemplatesGPR TemplatesAn admin can create GPR templates for a workspace so that users and other admins can use them.
Auto GPRAuto GPR can be enabled with a default template to create GPRs automatically in a slice workspace.
GPR Queue ManagementGPR queue/listList/view GPR queues, and GPR status.
Change GPR priorityChange the priority of a queued GPR.
Early release a GPREarly release of a provisioned GPR. Release GPUs provisioned to a workspace.
Multi-Cluster support. GPR scoped to single clusterUsers can select the cluster for GPR.
GPU Allocation ManagementFull node allocationsFull node allocated to only one workspace
Idle timeout of allocationsAfter Idle timeout, return the GPUs to free pool.
Multi-instance GPU node supportUsers can configure the memory for MIG nodes during GPR creation.
GPU Nodes Cost ExplorationThe Admin portal provides cost analysis dashboard and an individual page to explore GPU nodes cost.
EGS Core API SupportAPI TokensUsers can create/manage API tokens. Needed to access Core APIs.
Core APIsEGS core APIs
EGS Core SDK SupportPython SDK packageSupports all the EGS Core APIs
Deep AI Workload ObservabilityAI Workloads visualizationAI workload details in a workspace
Pods/Model detailsPods/jobs running in the workspace. List of pods/jobs and model details.
GPU usage/allocationsGPU/CPU allocations and usage for the workspace.
DashboardsGPU Allocations/Usage and GPRsGPU node/GPUs allocated/available, cost associated with allocations, total cost, leaderboards, different cluster/GPU type/and so on.
Inference EndpointsDeploy and manage Inference EndpointsDeploy one or more Inference Endpoints in a workspace.
Supports HF/Mistral/other Models; Model SelectionSupport for various commercial/open source LLM models.
Supports custom model deploymentUpload custom model for deployment
GPU provisioning for Inference endpointsGPU presets or advanced custom configuration
OpenAI Compatible APIsOpenAI compatible APIs for inference - chat/completions/and so on
EventsGPR EventsVisualization of events
Admin WorkflowsRegister clusters, and workspaces and usersUsing the Admin access token
Manage Workspaces, GPRs, and Inference EndpointsManage Workspaces, GPRs, and Inference Endpoints. Explore GPU costs.
User WorkflowsManage GPRs and Inference EndpointsUsing the User access token