Skip to main content
Version: 1.10.0

Features

The following table describes the key features of the EGS platform.

Main FeatureSub FeatureDescription
Multi-Cluster and Multi-Cloud Fleet ManagementCluster registrationSupports EKS, GKE, OKE, AKS, Akamai, on-prem cluster registration and management
Create or delete slice workspaceSlice workspace has one or more namespaces associated with it.
Users can deploy AI workloads in the Slice Workspace namespaces.
Slice workspace has a GPU VPC. By default, no GPUs are available for slice workspace.
A GPR (GPU Provision Request) is required to provision GPUs for a slice workspace.
Multi-Tenancy: Slice WorkspacesOne or more namespaces per slice workspaceOne or more namespaces can be associated with a workspace.
One or more worker clusters per slice workspaceSlice workspace can span across clusters. One or more clusters can be associated with a slice workspace.
Admin Access to PortalAdmin Profile and RBACRBAC for admin access
Access to Admin PortalAn individual portal for an admin.
User per slice workspaceUser per Slice workspace
User Access to PortalAccess to User PortalAn individual portal for a user to manage GPRs and AI workloads within a specific slice workspace.
User profile and RBACUser Slice RBAC and manage that slice workspace.
Inventory ManagementGPU node pools, inventory list, and allocation statusInventory management. Nvidia GPUs. GPU node pools. By default, all GPU pools/nodes are managed by EGS.
GPU Provisioning Requests (GPRs)Create/delete/update GPRGPU provisioning request for a slice workspace. Request to include a number of parameters such as GPU type, number, memory, nodes, cluster, and so on.
GPR Queue ManagementGPR queue/listList/view GPR queues, and GPR status.
Change GPR priorityChange the priority of a queued GPR.
Early release a GPREarly release of a provisioned GPR. Release GPUs provisioned to a slice workspace.
Multi-Cluster support. GPR scoped to single clusterUsers can select the cluster for GPR.
GPU Allocation ManagementFull node allocationsFull node allocated to only one slice workspace
Idle timeout of allocationsAfter Idle timeout, return the GPUs to free pool.
EGS Core API SupportAPI TokensUsers can create/manage API tokens. Needed to access Core APIs.
Core APIsEGS core APIs
EGS Core SDK SupportPython SDK packageSupports all the EGS Core APIs
Deep AI Workload ObservabilityAI Workloads visualizationAI workload details in a slice workspace
Pods/Model detailsPods/jobs running in the workspace. List of pods/jobs and model details.
GPU usage/allocationsGPU/CPU allocations and usage for the workspace.
DashboardsGPU Allocations/Usage and GPRsGPU node/GPUs allocated/available, cost associated with allocations, total cost, leaderboards, different cluster/GPU type/and so on.
Inference EndpointsDeploy and manage Inference EndpointsDeploy one or more Inference Endpoints in a slice workspace.
Supports HF/Mistral/other Models; Model SelectionSupport for various commercial/open source LLM models.
Supports custom model deploymentUpload custom model for deployment
GPU provisioning for Inference endpointsGPU presets or advanced custom configuration
OpenAI Compatible APIsOpenAI compatible APIs for inference - chat/completions/and so on
EventsGPR EventsVisualization of events
Admin WorkflowsRegister clusters, and manage slice workspaces and usersUsing the Admin access token
Manage Slice Workspaces, GPRs, and Inference EndpointsManage Slice Workspaces, GPRs, and Inference Endpoints.
User WorkflowsManage GPRs and Inference EndpointsUsing the User access token