logo
Alauda AI
  • English
    • English
    • 简体中文
  • Navigation
    Overview
    Architecture
    Introduction
    Quick Start
    Release Notes
    Install
    Pre-installation Configuration
    Install Workbench
    Install Alauda AI Essentials
    Install Alauda AI
    Install Alauda Build of KServe
    Tools Menu Configuration
    Kubeflow Chart Plugins
    Upgrade
    Upgrade Alauda AI
    Uninstall
    Uninstall
    Alauda support for Kubeflow
    Introduction
    how_to
    Use Kubeflow Notebooks
    Use Kubeflow Tensorboards
    Use Kubeflow Volumes
    Use Kubeflow Model Registry
    Use Kubeflow Pipelines
    Infrastructure Management
    Device Management
    About Alauda Build of Hami
    About Alauda Build of NVIDIA GPU Device Plugin
    Multi-Tenant
    Guides
    Namespace Management
    Workbench
    Overview
    Introduction
    Upgrade
    How To
    Create Workbench
    Fine-tuning LLMs using Workbench
    Kubeflow Trainer Quick Start
    Model Deployment & Inference
    Overview
    Introduction
    Features
    Inference Service
    Introduction
    Guides
    Inference Service
    How To
    Create Inference Service using CLI
    Extend Inference Runtimes
    Using KServe Modelcar for Model Storage
    Configure External Access for Inference Services
    Configure Scaling for Inference Services
    Scheduling Inference Services based on the CUDA version
    Set Up Autoscaling for Inference Services with KEDA
    Troubleshooting
    Experiencing Inference Service Timeouts with MLServer Runtime
    Inference Service Fails to Enter Running State
    Model Management
    Introduction
    Guides
    Model Repository
    Model Storage
    How To
    Upload Models Using Notebook
    Share Models
    Monitoring & Ops
    Overview
    Introduction
    Features Overview
    Logging & Tracing
    Introduction
    Guides
    Logging
    Resource Monitoring
    Introduction
    Guides
    Resource Monitoring
    How To
    Add a Monitoring Dashboard
    Troubleshooting
    Monitor Dashboard Stuck at Loading
    Label Studio
    Overview
    Introduction
    Main Features
    Install Label Studio
    Quickstart
    Dify
    Overview
    Introduction
    Main Features
    Install Dify
    Alauda Build of Kueue
    Introduction
    Install
    How To
    Setup RBAC
    Configuring quotas
    Monitoring pending workloads
    Using cohorts
    Configuring fair sharing
    Gang scheduling
    Integrate with Alauda DevOps Pipelines
    Integrate with InferenceService
    LLM Compressor
    Introduction
    How To
    LLM Compressor with Alauda AI
    Alauda Build of Llama Stack
    Overview
    Introduction
    Main Features
    Install Llama Stack
    Quickstart
    Alauda Build of LeaderWorkerSet
    Install
    Alauda Build of TrustyAI
    Introduction
    Install TrustyAI
    Evaluate LLM
    AI Guardrails for LLM safety
    API Reference
    Introduction
    Kubernetes APIs
    Inference Service APIs
    ClusterServingRuntime [serving.kserve.io/v1alpha1]
    InferenceService [serving.kserve.io/v1beta1]
    Workbench APIs
    Workspace Kind [kubeflow.org/v1beta1]
    Workspace [kubeflow.org/v1beta1]
    Manage APIs
    AmlNamespace [manage.aml.dev/v1alpha1]
    Operator APIs
    AmlCluster [amlclusters.aml.dev/v1alpha1]
    Glossary

    #Multi-Tenant

    Guides#

    Edit this page
    Previous pageAbout Alauda Build of NVIDIA GPU Device PluginNext pageGuides