logo
Alauda AI
  • English
    • English
    • 简体中文
  • Navigation
    Overview
    Architecture
    Introduction
    Quick Start
    Release Notes
    Install
    Pre-installation Configuration
    Enable Fine-Tuning and Training Features
    Install Workbench
    Install Alauda AI Essentials
    Install Alauda AI
    Tools Menu Configuration
    Upgrade
    Upgrade Alauda AI
    Uninstall
    Uninstall
    Infrastructure Management
    Device Management
    About Alauda Build of Hami
    About Alauda Build of NVIDIA GPU Device Plugin
    Multi-Tenant
    Guides
    Namespace Management
    Workbench
    Overview
    Introduction
    Upgrade
    How To
    Create Workbench
    Model Deployment & Inference
    Overview
    Introduction
    Features
    Inference Service
    Introduction
    Guides
    Inference Service
    How To
    Extend Inference Runtimes
    Configure External Access for Inference Services
    Configure Scaling for Inference Services
    Configure Accurately Scheduling Inference Services based on the CUDA version
    Troubleshooting
    Experiencing Inference Service Timeouts with MLServer Runtime
    Inference Service Fails to Enter Running State
    Model Management
    Introduction
    Guides
    Model Repository
    How To
    Upload Models Using Notebook
    Share Models
    Fine-tuning
    Introduction
    How To
    Create Fine-tuning Tasks
    Developing Custom Fine-Tuning Templates
    Training
    Introduction
    How To
    Create Training Tasks
    Develop Custom Training Templates
    Monitoring & Ops
    Overview
    Introduction
    Features Overview
    Logging & Tracing
    Introduction
    Guides
    Logging
    Resource Monitoring
    Introduction
    Guides
    Resource Monitoring
    How To
    Add a Monitoring Dashboard
    Troubleshooting
    Monitor Dashboard Stuck at Loading
    Label Studio
    Overview
    Introduction
    Main Features
    Install Label Studio
    Quickstart
    Alauda Build of Kueue
    Introduction
    Install
    How To
    Setup RBAC
    Configuring quotas
    Monitoring pending workloads
    Using cohorts
    Configuring fair sharing
    Gang scheduling
    Integrate with Alauda DevOps Pipelines
    Integrate with InferenceService
    API Reference
    Introduction
    Kubernetes APIs
    Inference Service APIs
    ClusterServingRuntime [serving.kserve.io/v1alpha1]
    InferenceService [serving.kserve.io/v1beta1]
    Workbench APIs
    Workspace Kind [kubeflow.org/v1beta1]
    Workspace [kubeflow.org/v1beta1]
    Manage APIs
    AmlNamespace [manage.aml.dev/v1alpha1]
    Operator APIs
    AmlCluster [amlclusters.aml.dev/v1alpha1]
    Glossary

    #How To

    📝 Edit this page
    Previous pageModel RepositoryNext pageUpload Models Using Notebook