Skip to content

Click Thru Demos

Service ProviderΒΆ

Videos showcasing workflows for service providers/operators

  • Model Deployment


    Workflows to deploy and operate multi tenant deployments of GenAI models.

  • Service Provider Dashboards

    Visualization and analytics of time series data related to end user usage of deployed models.


  • Storage Namespace

    Workflows to create and manage locally available storage resource (backed by local, high speed Object storage) where the operator can upload model weights etc.


  • Inference Endpoints

    Workflow to configure, deploy and operate OpenAI compatible API endpoints where end users can access one/many LLMs


  • Compute Clusters

    Workflow to register Kubernetes clusters based data plane for model deployments



End UserΒΆ

Click through demos showcasing workflows for End Users of Customer Orgs

  • Usage

    Workflows for end user usage of LLMs on service provider hosted inference endpoints.