search
HomeTechnology peripheralsAIVolcano Engine elastic container instance: a powerful tool for transforming from a node-centric architecture to a serverless architecture

As Kubernetes and container technology become increasingly mature, more and more enterprises are beginning to migrate their business applications to the cloud. By adopting cloud-native architecture, these enterprises can better support the rapid development and stable operation of their businesses.

As cloud computing develops in-depth into the development process, the maturity of Serverless architecture has brought cloud native technology to a new stage - Serverless takes the public cloud's elastic scalability, operation-free maintenance, fast access and other features to the extreme. It greatly reduces the user's cost of use, allowing users and enterprises to only focus on their business logic and achieve agile development in the true sense.

In order to better support enterprises in realizing the new serverless cloud native concept in business applications, the Volcano Engine cloud native team is comprehensively upgrading and innovating products, covering many aspects such as concept, system design and architecture design. . They are committed to giving full play to the advantages and value of Serverless technology to help enterprises run business applications more efficiently.

From Node Center to Serverless Architecture

The traditional Kubernetes architecture usually takes the node as the core, which means that the team It is necessary to build clusters on resource nodes such as cloud servers, and perform operation and maintenance management around these nodes. The traditional Kubernetes cluster architecture requires managing the deployment, monitoring, and maintenance of container applications on different nodes. This approach requires more manpower and time investment to maintain the stability and reliability of the entire cluster. However, with the development of technology, a new type of Kubernetes architecture has emerged, namely an architecture based on service mesh. In an architecture based on a service grid, network functions are abstracted and unified through specialized service grid components. As the business scale expands and the number of nodes increases, this architecture becomes increasingly important in container applications. There are usually some challenges in deployment and operation, mainly including the following aspects:

    Complex node operation and maintenance
  • : Users need to manage and maintain nodes by themselves, including node configuration. Complex and tedious node operation and maintenance such as initialization, operating system update and upgrade, security management, operation monitoring and log collection and analysis, fault recovery, network management, storage management, expansion and contraction, resource planning and allocation, and resource cost management Work;
  • Low resource utilization
  • : Nodes reserve fixed resources. Even if the actual workload requires less resources, the node resource usage cannot be reduced; the node resource fragmentation rate is high, Some unused resources distributed on various nodes are difficult to effectively aggregate and cannot be fully utilized; resources are over-allocated. In order to avoid performance bottlenecks, resources are often planned according to the maximum load, which makes it difficult to predict actual demand and may cause a large number of idle resources;
  • Slow resource expansion
  • : Container applications may experience bursts of traffic and need to scale quickly in response to increased load. In traditional node-centric architecture, the expansion of computing resources often requires the creation and configuration of new cloud server nodes. This process may take several minutes and cannot achieve instant expansion.
  • Volcage Engine
Elastic Container Instance

(Volcengine Container Instance, referred to as VCI) is a new cloud native solution launched for the above situation. It is abstracted from ByteDance’s internal cloud native technology for many years. Accumulation is a serverless and containerized computing service - Through Virtual Kubelet technology and the Volcano Engine container service VKE, the elastic container instance VCI can provide users with flexible computing and Kubernetes orchestration functions, supporting Fast startup, high-concurrency creation, and secure sandbox container isolation.

By integrating the advantages of Serverless and containers, Elastic Container Instance (VCI) can provide enterprises with an efficient, reliable and secure operating environment, allowing users to focus on developing and running cloud-native architecture applications without having to worry about The management and maintenance of the underlying infrastructure, and only paying for the resources required for actual business operation, help enterprises more reasonably control cloud costs.

No operation and maintenance of infrastructure, no planning of computing resources

In the traditional self-built Kubernetes cluster architecture, whether it is the control plane or the data plane, as a container group Pod The running carrier and nodes are located at the core of the Kubernetes architecture. When node computing resources are insufficient, computing resources need to be supplemented by adding nodes to support the deployment or expansion of business workloads.

On the cloud, various managed Kubernetes services represented by Volcano Engine Container Service VKE will replace users in managing and operating the Kubernetes control plane, providing managed services with SLA guarantees. For the data plane that runs users' actual business workloads, users usually need to operate and maintain working nodes by themselves, and the computing resources provided by the nodes need to be planned in advance according to business characteristics. For example, when business peaks require advance expansion, Business downturns require downsizing.

Therefore, in the traditional node-centered Kubernetes architecture, node operation and maintenance often requires a lot of energy from the technical team; at the same time, how to balance the issue of advanced planning of resources and resource costs has also become a "refined application for enterprises. Questions that we have to think about after the era of "cloud ".

Elastic Container Instance VCI provides Serverless and containerized computing services. Each VCI will only provide the operating environment and computing resources for a single Pod. Users do not need to care about the operation and maintenance and capacity planning of the underlying nodes. They only need to Just deploy the container application. When using VCI to deploy container applications, you can use native Kubernetes semantics, which is no different from the node center architecture.

At the resource level, the elastic container instance VCI uses the Volcano Engine computing infrastructure as a resource pool to provide container computing resources of multiple resource types to facilitate users to flexibly choose according to business needs. At the same time, users do not need to pay attention to resource capacity planning issues. VCI provides massive computing resources and will not cause phenomena such as Pod drift and rescheduling that are common in traditional node-centric architectures, which will further cause service interruptions, performance fluctuations, Problems such as data inconsistency and scheduling delays.

Finally, the seamless integration of container service VKE and elastic container instance VCI not only realizes full hosting of the Kubernetes cluster control plane, but also provides users with full Serverless data plane (i.e. elastic container instance VCI). Hosting, operation-free, and secure isolation greatly reduce users’ operation and maintenance costs on Kubernetes infrastructure and eliminate the complexity of capacity planning, allowing users to focus more on their business applications.

Ultimate elasticity, you get what you need

Flexibility and savings

According to Flexera in the "2022 State of the Cloud Report" According to the disclosed survey data, the self-assessed cloud cost waste of the interviewed companies reached 32%, and "cost" has become one of the cloud challenges that companies are most concerned about for three consecutive years - "refined cloud usage" has begun to gain traction. More and more enterprises are paying attention to how to save cloud costs while meeting business development needs, which has become an important challenge that enterprises need to solve when using the cloud.

Elastic Container Instance VCI focuses on the enterprise's "cloud cost control" needs from the beginning of its product design, emphasizing refined billing based on the actual amount of resources used, supporting multiple billing models, and helping enterprises truly make good use of the cloud. .

Refined billing: Elastic container instance VCI is billed according to the resources actually used by the user container group (Pod). Billing resources include vCPU, memory, GPU, etc. At the same time, the billing time of the VCI instance is its running time, which is the time used by the user Pod from downloading the container image to stopping running. It is accurate to the second level, truly realizing billing based on actual usage.

Improve the packing rate: Under the cloud native architecture, compared with traditional computing resources, another advantage of elastic container instances is to reduce idle resources and improve the packing rate, thereby reducing user costs. computing resource usage costs. Specifically, when a Kubernetes cluster uses cloud server ECS as a working node, it needs to run the system components required by Kubernetes on the node. Therefore, some computing resources need to be reserved, and there is a certain additional cost; and if an elastic container instance VCI is used as the computing resource , you only need to pay for the resources actually used by the business Pod:

  • Example: Running 3 business Pods, their resource requirements for vCPU are divided into 16C, 8C, 8C
  • Use ECS: At least 2 32C vCPU ECS nodes are required (resources need to be reserved for the system on each node)
  • Use VCI: Create 3 VCI instances equal to the resources required by the Pod, that is, 32 C computing resources (no resources are idle)

火山引擎弹性容器实例:从节点中心转型 Serverless 化架构的利器

Rich billing methods: The current elastic container instance VCI has officially supported pay-as-you-go billing, and will launch an invitation test event for the preemptive instance (Spot) payment model, as well as reserved instances and flexible reservations. Prepaid models such as instances are also supported by plans. By providing multiple payment models, we hope to help users further optimize cost management and budgeting of computing resources, and choose different payment models for different business scenarios, thereby better coordinating business resource needs and cloud cost planning.

Bounce out

Elastic Container Instance VCI fully integrates the computing resource infrastructure of Volcano Engine to provide computing power, and provides a variety of instance specification families for users to choose based on different underlying hardware capabilities, providing differentiated computing, storage, and network performance. , supporting coverage of a variety of business applications and service scenarios.

CPU General-purpose specification family: Such as general-purpose computing u1, general-purpose n3i, etc., which provide balanced vCPU, memory, and network capabilities to meet the needs of large-scale users. Service requirements in most scenarios.

GPU Computing type specification family: Such as GPU computing type gni2 (equipped with NVIDIA A10 GPU graphics card), GPU computing type g1v (equipped with NVIDIA V100 GPU graphics card), etc. It can provide a cost-effective use experience in various AI computing scenarios such as large model training, text and image generation, and task reasoning.

The latest general computing type u1 Instance specification family relies on the Volcano engine resource pooling technology and intelligent scheduling algorithm for dynamic resource management , can provide enterprises with a stable supply of computing power, and support a variety of processors and flexible processor memory ratios, from fine-grained small specifications such as 0.25C-0.5Gi, 0.5C-1Gi, 1C-2Gi to 24C-48Gi , 32C-256Gi and other large specifications are available for selection, with extremely high cost performance!

Backed by a huge supply of computing resources, the elastic container instance VCI also provides users with an industry-leading elastic resource priority scheduling strategy : Combined with the container service VKE, we support users to use their own Define the resource policy (ResourcePolicy) to elastically schedule to different types of computing resource pools such as cloud servers ECS and elastic container instances VCI, and perform scheduling and allocation according to percentage thresholds, priority control and other strategies according to actual business scenarios.

This means that users can combine the cloud server ECS computing resource pool and the elastic container instance VCI computing resource pool to enjoy massive Volcano Engine computing resources; at the same time, combine the Cluster Autoscaler capabilities and VCI in the container cluster , even if computing resources are not requested in advance in the container cluster, users can obtain elastic resources in a timely manner to efficiently support business applications.

bounce fast

Faced with unpredictable business traffic peaks in online business, it is crucial to ensure stable business operation and user experience; at the critical moment of business development, Quickly popping up computing resources and quickly enabling business applications are also critical. Elastic container instances are also naturally suitable for these scenarios.

火山引擎弹性容器实例:从节点中心转型 Serverless 化架构的利器

On the one hand, the elastic capability of the Volcano Engine Elastic Container Instance (VCI) can meet the needs of tens of thousands of cores per minute vCPU computing resources, ensuring that When business needs it, sufficient computing resources are quickly available. When the traffic peak is over and the business workload decreases, elastic computing resources can be quickly released, making cloud costs more economical.

On the other hand, container image pulling often takes a long time, which is even more serious when large-scale Pods are started concurrently. Elastic container instance VCI not only has a variety of container image acceleration capabilities provided by VKE, such as container image lazy loading (image lazy loading), P2P container image distribution, etc., but also has unique acceleration capabilities such as container image caching capabilities. According to actual test data, Container image caching can control the overall startup time of a Pod with a container image (data volume within 100 G) to the level of ten seconds, helping users greatly improve efficiency and reduce cloud costs.

Conclusion

Since its official launch in 2022, Elastic Container Instance VCI has served many large and medium-sized enterprise customers and has been fully utilized in Bytedance’s internal and external business scenarios. practical test. The Volcano Engine cloud native team also continuously polishes product capabilities based on customer service experience and continues to improve the quality and stability of product services:

  • In terms of ecological integration, Elastic Container Instance (VCI) not only continues to enhance its integration with the Kubernetes ecosystem, but can also be seamlessly integrated with the Volcano Engine Container Service (VKE), making it easier for customers to deploy and manage containers in a cloud-native environment. Application;
  • In terms of distributed container image startup acceleration, VCI collaborates with Volcano Engine Image Warehouse CR and other products to improve the Dragonfly architecture, further enhancing the efficiency and reliability of image distribution to VCI;
  • VCI's unique container image cache acceleration capability better solves the problems of slow Pod startup and low image pull success rate in large-scale concurrency and large container image scenarios, and reduces the pressure on the image warehouse;
  • In terms of diversity of computing resource specifications, VCI will launch more optional resources in 2023, such as n3i, a new general-purpose computing instance based on Volcano Engine’s self-developed DPU fully offloaded architecture, which shields back-end server differences and has unlimited theoretical capabilities The general computing power instance u1 of the resource pool, etc.

The Volcano Engine cloud native team is mainly responsible for the construction of the PaaS product system in the Volcano Engine public cloud and privatized scenarios. Combined with ByteDance’s many years of cloud native technology stack experience and best practices, it helps Enterprises accelerate digital transformation and innovation. Products include container services, image warehouses, distributed cloud native platforms, function services, service grids, continuous delivery, observable services, etc.

Facing 2024, Elastic Container Instance VCI will continue to improve and explore under the Serverless Kubernetes product architecture and form, providing more internal and external customers with no operation and maintenance of infrastructure, no planning of computing resources, and ultimate flexibility. Serverless containerized products and services can better support customers' various GPU and CPU computing power needs in AIGC, bioinformatics scientific computing, social e-commerce and other business scenarios.

The above is the detailed content of Volcano Engine elastic container instance: a powerful tool for transforming from a node-centric architecture to a serverless architecture. For more information, please follow other related articles on the PHP Chinese website!

Statement
This article is reproduced at:机器之心. If there is any infringement, please contact admin@php.cn delete
Can't use ChatGPT! Explaining the causes and solutions that can be tested immediately [Latest 2025]Can't use ChatGPT! Explaining the causes and solutions that can be tested immediately [Latest 2025]May 14, 2025 am 05:04 AM

ChatGPT is not accessible? This article provides a variety of practical solutions! Many users may encounter problems such as inaccessibility or slow response when using ChatGPT on a daily basis. This article will guide you to solve these problems step by step based on different situations. Causes of ChatGPT's inaccessibility and preliminary troubleshooting First, we need to determine whether the problem lies in the OpenAI server side, or the user's own network or device problems. Please follow the steps below to troubleshoot: Step 1: Check the official status of OpenAI Visit the OpenAI Status page (status.openai.com) to see if the ChatGPT service is running normally. If a red or yellow alarm is displayed, it means Open

Calculating The Risk Of ASI Starts With Human MindsCalculating The Risk Of ASI Starts With Human MindsMay 14, 2025 am 05:02 AM

On 10 May 2025, MIT physicist Max Tegmark told The Guardian that AI labs should emulate Oppenheimer’s Trinity-test calculus before releasing Artificial Super-Intelligence. “My assessment is that the 'Compton constant', the probability that a race to

An easy-to-understand explanation of how to write and compose lyrics and recommended tools in ChatGPTAn easy-to-understand explanation of how to write and compose lyrics and recommended tools in ChatGPTMay 14, 2025 am 05:01 AM

AI music creation technology is changing with each passing day. This article will use AI models such as ChatGPT as an example to explain in detail how to use AI to assist music creation, and explain it with actual cases. We will introduce how to create music through SunoAI, AI jukebox on Hugging Face, and Python's Music21 library. Through these technologies, everyone can easily create original music. However, it should be noted that the copyright issue of AI-generated content cannot be ignored, and you must be cautious when using it. Let’s explore the infinite possibilities of AI in the music field together! OpenAI's latest AI agent "OpenAI Deep Research" introduces: [ChatGPT]Ope

What is ChatGPT-4? A thorough explanation of what you can do, the pricing, and the differences from GPT-3.5!What is ChatGPT-4? A thorough explanation of what you can do, the pricing, and the differences from GPT-3.5!May 14, 2025 am 05:00 AM

The emergence of ChatGPT-4 has greatly expanded the possibility of AI applications. Compared with GPT-3.5, ChatGPT-4 has significantly improved. It has powerful context comprehension capabilities and can also recognize and generate images. It is a universal AI assistant. It has shown great potential in many fields such as improving business efficiency and assisting creation. However, at the same time, we must also pay attention to the precautions in its use. This article will explain the characteristics of ChatGPT-4 in detail and introduce effective usage methods for different scenarios. The article contains skills to make full use of the latest AI technologies, please refer to it. OpenAI's latest AI agent, please click the link below for details of "OpenAI Deep Research"

Explaining how to use the ChatGPT app! Japanese support and voice conversation functionExplaining how to use the ChatGPT app! Japanese support and voice conversation functionMay 14, 2025 am 04:59 AM

ChatGPT App: Unleash your creativity with the AI ​​assistant! Beginner's Guide The ChatGPT app is an innovative AI assistant that handles a wide range of tasks, including writing, translation, and question answering. It is a tool with endless possibilities that is useful for creative activities and information gathering. In this article, we will explain in an easy-to-understand way for beginners, from how to install the ChatGPT smartphone app, to the features unique to apps such as voice input functions and plugins, as well as the points to keep in mind when using the app. We'll also be taking a closer look at plugin restrictions and device-to-device configuration synchronization

How do I use the Chinese version of ChatGPT? Explanation of registration procedures and feesHow do I use the Chinese version of ChatGPT? Explanation of registration procedures and feesMay 14, 2025 am 04:56 AM

ChatGPT Chinese version: Unlock new experience of Chinese AI dialogue ChatGPT is popular all over the world, did you know it also offers a Chinese version? This powerful AI tool not only supports daily conversations, but also handles professional content and is compatible with Simplified and Traditional Chinese. Whether it is a user in China or a friend who is learning Chinese, you can benefit from it. This article will introduce in detail how to use ChatGPT Chinese version, including account settings, Chinese prompt word input, filter use, and selection of different packages, and analyze potential risks and response strategies. In addition, we will also compare ChatGPT Chinese version with other Chinese AI tools to help you better understand its advantages and application scenarios. OpenAI's latest AI intelligence

5 AI Agent Myths You Need To Stop Believing Now5 AI Agent Myths You Need To Stop Believing NowMay 14, 2025 am 04:54 AM

These can be thought of as the next leap forward in the field of generative AI, which gave us ChatGPT and other large-language-model chatbots. Rather than simply answering questions or generating information, they can take action on our behalf, inter

An easy-to-understand explanation of the illegality of creating and managing multiple accounts using ChatGPTAn easy-to-understand explanation of the illegality of creating and managing multiple accounts using ChatGPTMay 14, 2025 am 04:50 AM

Efficient multiple account management techniques using ChatGPT | A thorough explanation of how to use business and private life! ChatGPT is used in a variety of situations, but some people may be worried about managing multiple accounts. This article will explain in detail how to create multiple accounts for ChatGPT, what to do when using it, and how to operate it safely and efficiently. We also cover important points such as the difference in business and private use, and complying with OpenAI's terms of use, and provide a guide to help you safely utilize multiple accounts. OpenAI

See all articles

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

Video Face Swap

Video Face Swap

Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Article

Hot Tools

Notepad++7.3.1

Notepad++7.3.1

Easy-to-use and free code editor

SecLists

SecLists

SecLists is the ultimate security tester's companion. It is a collection of various types of lists that are frequently used during security assessments, all in one place. SecLists helps make security testing more efficient and productive by conveniently providing all the lists a security tester might need. List types include usernames, passwords, URLs, fuzzing payloads, sensitive data patterns, web shells, and more. The tester can simply pull this repository onto a new test machine and he will have access to every type of list he needs.

MantisBT

MantisBT

Mantis is an easy-to-deploy web-based defect tracking tool designed to aid in product defect tracking. It requires PHP, MySQL and a web server. Check out our demo and hosting services.

ZendStudio 13.5.1 Mac

ZendStudio 13.5.1 Mac

Powerful PHP integrated development environment

SublimeText3 Chinese version

SublimeText3 Chinese version

Chinese version, very easy to use