Skip to main content
Version: 1.10.0

Manage an Inference Endpoint

This topic describes the steps to view, deploy, and delete Inference Endpoints for your slice workspaces.

View the Inference Endpoint

To view the Inference Endpoint:

  1. Go to Inference Endpoints on the left sidebar.

  2. On the Workspaces page, click a slice workspace whose Inference Endpoints you want to view.

    inference-endpoints

  3. On the Inference Endpoints page, you see a list of Inference Endpoints for that slice workspace.

    inference-endpoints

    The following figure illustrates the deployment for the workspace.

    inference-endpoints

Deploy an Inference Endpoint

note

Only the admin has the privilege to add namespaces to a slice. The user (non-admin) with access to a slice must create the inference endpoint with the namespace name added to a slice by the admin.

For example, if the admin adds the inference-1-gpu and inference-1-cpu namespaces to a slice. The non-admin user must create an inference endpoints with name inference-1-gpu or inference-1-cpu only.

To deploy an Inference Endpoint on your slice workspace:

  1. Go to Inference Endpoint on the left sidebar.

  2. On the Workspaces page, go to the slice workspace on which you want to deploy an Inference Endpoint.

  3. On the Inferences Endpoint page, click Deploy Inference Endpoint.

    inference-endpoints

  4. On the Create Inference Endpoint pane, under Basic Specifications:

    1. Enter a name for the Inference Endpoint in the Endpoint Name text box. Ensure that the namespace with the same name as the Inference Endpoint already exists on the worker cluster. If it does not exist, then ask your cluster admin to create it for you.

      warning

      If you try creating an Inference Endpoint with a name without ensuring that a namespace with the same name exists, then you get an error that says Failed to create namespace.

    2. Select the cluster from the Cluster Name drop-down list, on which you want to deploy the Inference Endpoint.

  5. Under Advanced Options, enter the specifications for model deployment. Under Model Specifications, enter the following:

    note

    The following are standard parameters for most of model deployments. However, if these parameters do not meet your model requirements, then select the Specify your own model configuration checkbox and enter your own model configuration.

    1. Enter a name in the Model Format Name text box.
    2. Add the storage URI in the Storage URI text box.
    3. Add the CPU value in the CPU text box.
    4. Add the Memory value in the Memory text box.
    5. Add the arguments in the Args text box.
    6. To add secret key-value pair, click the plus sign against Secret and add them.

    Own Model Configuration

    To add your own model configuration:

    1. Select the Specify your own model configuration checkbox.

    2. On the terminal screen, enter your model configuration yaml file from KServe. For more information, see KServe.

  6. Under GPU Specifications, enter the following:

    info

    Select the Create CPU-only Inference checkbox for CPU only specification for the model.

    inference-endpoints

    1. Select node type from the Node Type drop-down list. After you select node type:

      • The GPU Shape and Memory per GPU values are auto populated. These values are immutable.
      • The GPU Nodes, GPUs Per Node have the default values. Edit these values as per your requirements.
      • The Reserve For have 365 days by default. The duration is in days\hours\mins format. Edit the days as per your requirements. The number of days must be less that 365.
      • The Priority and Priority Number have default values. Edit the values as per your requirements.
    2. Click the Create Inference Endpoint button. The status goes to Pending before it changes to Ready.

      inference-endpoints

Delete an Inference Endpoint

To delete an Inference Endpoint:

  1. On the Workspaces page, click a slice workspace.

  2. On that Inference Endpoint page, click the Delete button.

  3. Enter the name of the Inference Endpoint in the text box and click Delete.

    inference-endpoints