Follow
This article describes how to use TorchScript custom C++ operators to build the post-processing network of an object detection model and use PAI-Blade to optimize the model.
This article describes how to use PAI-Blade provided by Machine Learning Platform for AI (PAI) to optimize a RetinaNet model that is in the Detectron2 framework.
This article describes how to use PAI-Blade to optimize a detection model whose post-processing network is built by using TensorRT plug-ins.
This article describes how to associate a RAG-based LLM chatbot with an ApsaraDB RDS for PostgreSQL instance when you deploy the RAG-based LLM chatbot.
This article describes how to deploy and call MLLM inference services by using PAI-EAS.
This article provides a detailed exploration of applying agents in operations and maintenance (O&M) diagnosis.
This article explores a Large Language Model (LLM)-based data warehouse solution that addresses the challenges of traditional data warehouses, including high costs, complexity, and accuracy concerns.
This article explores the potential integration of AI foundation models with DevOps, focusing on the concept of "Agent + Tool" in AI and its application in the LangChain framework.
This article describes an overview of the implementation principles and best practices of Hologres Binlog.
This article describes the technical principles of Hologres' JSONB semi-structured data and highlights the exceptional analysis performance of JSON semi-structured data.
This article describes how to deploy a RAG-based LLM chatbot and how to perform model inference.
This article describes how to use the data processing, model training, and model inference components of Large Language Model (LLM) provided by PAI to complete end-to-end development and use of LLM.
This article describes how to fine-tune the parameters of a Llama 3 model in DSW to enable the model to better align with and adapt to specific scenarios.
This article uses llama-2-7b-chat as an example to describe how to use QuickStart to deploy a model as a service in Elastic Algorithm Service (EAS) and call the service.
This article describes how to quickly deploy a Llama 3 model and use the deployed web application in Elastic Algorithm Service (EAS) of Platform for AI (PAI).
This article describes how to deploy an LLM in EAS and call the model.
This article describes how to deploy a web application based on the open source model Tongyi Qianwen and perform model inference on the web page or using API operations in EAS of PAI.
This article describes how to deploy a Hugging Face model in PAI EAS.
This article describes how to deploy an AI video generation application, related inference services and answers to FAQ during the deployment.
This article describes how to deploy a Stable Diffusion model and use the deployed application to perform model inference and generate images.