Install Alauda Hyperflux
TOC
Download package and upload to cluster
You can download the app named 'Alauda Hyperflux' from the Marketplace on the Customer Portal website.
The downloaded package is a tarball file named alauda-hyperflux-<version>.tar.gz.
Download the violet command line tool if it is not present on the machine:
- Log into the ACP Web Console and switch to the Administrator view.
- In Marketplace / Upload Packages, click Download Packaging and Listing Tool.
- Select the right OS/CPU arch, and click Download.
- Run
chmod +x ${PATH_TO_THE_VIOLET_TOOL}to make the tool executable.
Save the following script in upload.sh,
then edit the file to fill in the correct configuration values acording to the comments.
Prepare your LLM and rerank service
Before installing Alauda Hyperflux, you need to prepare an LLM service for Alauda Hyperflux to use. You can use Azure OpenAI service, or deploy an On-Premise LLM service like vllm using Alauda AI.
You will use the LLM service endpoint, model name and API key in the Alauda Hyperflux installation step.
Optionally, if you want to enable the rerank feature in Alauda Hyperflux, you also need to prepare a rerank service that supports Cohere Reranker API v2.
Prepare the database dump file (ONLY needed for v1.2.0)
Download the database dump file like docvec_acp_4_1.dump for your current ACP version.
You MUST use the file name like docvec_acp_4_1 as the database name during installation.
Install Alauda Hyperflux cluster plugin
Go to Administrator / Marketplace / Cluster Plugins page,
select "global" cluster from the cluster dropdown list,
then find the Alauda Hyperflux plugin and click Install.
NOTE: Alauda Hyperflux MUST be installed in the
Globalcluster.
Fill in the configurations below:
- Built-in PG database:
- Enabled: will install a single PostgreSQL instance in the cluster for Alauda Hyperflux to use. You need to set:
- storage size: the storage size for PostgreSQL data.
- storage class name: Kubernetes storage class name, e.g.
sc-topolvm
- Disabled: create a secret below to provide external PostgreSQL connection info.
- Enabled: will install a single PostgreSQL instance in the cluster for Alauda Hyperflux to use. You need to set:
- PG database name: the database name for Alauda Hyperflux to use. MUST be the same as the database dump file name without the
.dumpsuffix. - Node Selector(Optional): set the node selector for Alauda Hyperflux pods if needed.
- LLM Model Type: Azure or OpenAI.
- LLM Base URL: the base URL for LLM API calls. When using On-Premise deployment of LLM service like vllm, the url should like
http://<your-vllm-host>:<port>/v1. - Model Name: the model name for LLM API calls.
- API Key: the API key for LLM API calls.
- Azure API Version(Optional): when using Azure OpenAI service, set the API version here.
- Azure Deployment Name(Optional): when using Azure OpenAI service, set the deployment name here.
- Enable Rerank: whether to enable the rerank feature in Alauda Hyperflux using Cohere API. Set below values if enabled:
- Cohere Reranker BaseURL: the base URL for Cohere Reranker API calls.
- Cohere Reranker Model: the model name for Cohere Reranker API calls.
- Cohere API Key: the API key for Cohere Reranker API calls.
- Enable Agent Mode: whether to enable Agent mode to leverage MCP tools to retrieve real-time cluster information.
- NOTE: Agent mode is an experimental feature, please use with caution.
- MCP K8s API Server Address: the K8s API server address of the MCP cluster.
- IMPORTANT: You should set this url to erebus address like
https://erebus.cpaas-system:443/kubernetes/<cluster-name>. - IMPORTANT: the
cluster-nameshould be set to the cluster name you want MCP tools to access.
- IMPORTANT: You should set this url to erebus address like
- Admin User Names: the comma-separated admin user list. Admin users can manage Audit logs in Alauda Hyperflux.
Click Install to start installation.
Import database dump to initialize knowledge base (ONLY needed for v1.2.0)
After the Alauda Hyperflux installation is complete, you need to import the database dump file to initialize the knowledge base. Use the following command to import the database dump file:
NOTE: when using built-in PostgreSQL database, the default password is
alauda-test.
Troubleshooting
If the chat interface fails to respond, you can check the Alauda Hyperflux pod logs for troubleshooting:
In most cases, the issue is caused by incorrect LLM service configurations, or Cohere API configurations when rerank is enabled. Check the error messages in the logs to fix the configuration issues.