×
Container Service

Use Argo Workflows SDK for Python to Create Large-Scale Workflows

This article describes how to use Hera, Argo Workflows SDK for Python, to create large-scale workflows.

Building a Large Language Model Inference Service Optimized by TensorRT-LLM Based on KServe on ASM

This article introduces how to deploy optimized LLM model inference services in a cloud-native environment using the TensorRT-LLM-optimized Llama-2-hf model as an example.

Fluid 1.0: Bridging the Last Mile for Efficient Cloud-Native Data Usage

The article introduces the Fluid 1.0's new features and future plans.

Alibaba Cloud Service Mesh (ASM): Efficient Traffic Management with Gateway API

This article introduces how to efficiently manage cluster traffic using the Gateway API in Alibaba Cloud Service Mesh.

Container Memory Observability: Exploring WorkingSet and PageCache Monitoring

This article introduces WorkingSet and PageCache Monitoring from a new perspective of container memory observability.

Best Practices for AI Model Inference Configuration in Knative

The article introduces best practices for deploying and configuring AI model inference in Knative, focusing on the optimization of GPU resource utilization and rapid scaling.

Analyzing the Distributed Inference Process Using vLLM and Ray from the Perspective of Source Code

This article explores how to implement distributed inference with vLLM and Ray from a source code perspective.

The Age of Cloud-native: Building Efficient CI Pipeline from Jenkins to Argo Workflows

This article explains how to build a large-scale, efficient, and cost-effective CI pipeline using Argo Workflows.

ACK Cloud Native AI Suite | Training and Inference of Open-Source Large Models on Kubernetes

The sixth episode of ACK Cloud Native AI Suite series introduces how to train and infer open-source foundation models based on the ACK Cloud-Native AI suite.

ACK Cloud Native AI Suite | Scaling Distributed Elastic Training for Large Models

The fifth episode of ACK Cloud Native AI Suite series introduces how to perform large-scale distributed elastic training based on the ACK Cloud-Native AI suite.

ACK Cloud Native AI Suite | Elastic Acceleration of Generative AI Model Inference with Fluid

The fourth episode of ACK Cloud Native AI Suite series introduces Fluid, the data orchestration acceleration engine in the ACK Cloud-Native AI suite.

ACK Cloud Native AI Suite | Efficiently Scheduling Large Scale AI Big Data Tasks on Kubernetes

The third episode of ACK Cloud Native AI Suite series introduces how the ACK Cloud-Native AI suite efficiently schedules AI and big data tasks.

ACK Cloud Native AI Suite | Simplifying GPU Cluster Operations and Improving GPU Utilization

The second episode of ACK Cloud Native AI Suite series introduces how to simplify the complexity of GPU cluster operations and improve GPU resource utilization through the ACK Cloud-Native AI suite.

ACK Cloud Native AI Suite | Implementing Cloud Native AI based on Kubernetes

The first episode of ACK Cloud Native AI Suite series introduces Alibaba Cloud's Cloud-Native AI Suite.

ACK One Argo Workflow Clusters: Mastering Container Object Storage Service

This article introduces Argo Workflows, Artifacts, OSS, and the advantages of ACK One Serverless Argo Workflow.

Serverless Cost Optimization: Knative Supports Preemptible Instances

This article describes how to use preemptible instances in Knative.

Alibaba Cloud ACK One GitOps: Streamline Multi-Cluster Application Delivery

This article provides an overview of how Alibaba Cloud's ACK One GitOps facilitates continuous deployment and management of applications across multip...

Deploy Magento on Alibaba Cloud Container Service for Kubernetes (ACK)

This article focuses on deploying Magento on Alibaba Cloud Container Service for Kubernetes (ACK).

AI Container Image Deployment: Stable Diffusion

This article introduces how to quickly build a personal text-based image generation service based on Alibaba Cloud AMD servers and OpenAnolis AI container service.

AI Container Image Deployment: Qwen-Audio-Chat

This article introduces how to quickly build an AI voice assistant service based on Alibaba Cloud AMD servers and OpenAnolis AI container service.