


This article will take you to understand the memory management and garbage collection algorithm of the V8 engine. I hope it will be helpful to you!
As we all know, JS automatically manages garbage collection, and developers do not need to care about memory allocation and recycling. Moreover, the garbage collection mechanism is also a commonly tested part in front-end interviews. This article mainly explains the generational garbage collection algorithm of V8. I hope that after reading this article, friends can have a painful
about the V8
garbage collection mechanism (haha, it is painful
!!!), the article mainly covers the following content:
-
V8
Memory limitations and solutions - New generation memory objects
Scavenge
Algorithm - Based on
reachability analysis algorithm
Logic and optimization methods for marking surviving objects - Promotion conditions for new generation memory objects,
-
Scavenge
The depth/breadth first difference of the algorithm - Cross-generation memory write barrier
- Mark clearing/organizing algorithm for old generation memory objects
-
GC
STW
Causes and optimization strategies
V8 memory limitations and solutions
V8 initially Designed for browsers, there are fewer scenarios where large memory is used. There are restrictions on memory usage by default in the design, and only part of the memory is allowed to be used. The 64-bit system allows about 1.4g of memory, and the 32-bit system allows about 0.7g of memory. As shown in the following code, check the memory limit method of the dependent V8 engine in Node:
process.memoryUsage(); // 返回内存的使用量,单位字节 { rss: 22953984, // 申请的总的堆内存 heapTotal: 9682944, // 已使用的堆内存 heapUsed: 5290344, external: 9388 }
V8
There is another way to limit the memory usage size The important reason is that when the heap memory is too large, V8
takes a long time to perform garbage collection (1.5g
takes 50ms
), and it takes longer to do non-incremental garbage collection. Long time (1.5g
to 1s
). After explaining the garbage collection mechanism of V8
later, I believe everyone will be able to relate to it more.
Although the V8
engine has limits on memory usage, the method to modify the memory limit is also exposed, which is to add relevant parameters when starting the V8
engine. The following code is demonstrated in Modify the dependent V8
engine memory limit in Node
:
# 更改老生代的内存限制,单位mb node --max-old-space-size=2048 index.js # 更改新生代的内存限制,单位mb node --max-semi-space-size=1024=64 index.js
It should be noted here that the syntax of the changed new generation memory has been changed to the above writing method, and The unit has also changed from kb
to mb
. The old way of writing is node --max-new-space-size
. You can query the current ## through the following command. #NodeEnvironment to modify the syntax of the new generation memory:
node --v8-options | grep max
V8 garbage collection strategy
Automatic garbage collection in the engine In the historical evolution of the recycling mechanism, people found that there is no universal algorithm that can solve garbage collection in any scenario. Therefore, modern garbage collection algorithmsdivide memory garbage into generations based on the survival time of objects. Generational garbage collection algorithms implement different methods for different types of memory garbage. recycling algorithm.
V8 Divide the memory into two types: New generation
and Old generation
:
- The generation objects in the old generation memory have a longer survival time or they are resident in memory
- The new generation memory is stored in the new generation memory space (
), the old generation memory is stored in the old generation memory space (oldspace
), as shown in the following figure:
- Scavenge
- algorithm
Old generation memory uses
Mark-Sweep - and
Mark-Compact
algorithms Let’s take a look at the algorithm logic of
!
For memory recycling of new generation memory, the
Scavenge algorithm is used. The specific implementation of Scavenge
is Cheney
Algorithm. Cheney
The algorithm is to divide the new generation memory space into two, one space is in use (FromSpace
), and the other space is in idle state (called ToSpace
) .
在内存开始分配时,首先在FromSpace
中进行分配,垃圾回收机制执行时会检查FromSpace
中的存活对象,存活对象会被会被复制到ToSpace
,非存活对象所占用的空间将被释放,复制完成后FromSpace
和ToSpace
的角色将翻转。当一个对象多次复制后依然处于存活状态,则认为其是长期存活对象,此时将发生晋升,然后该对象被移动到老生代空间oldSpace
中,采用新的算法进行管理。
Scavenge
算法其实就是在两个空间内来回复制存活对象,是典型的空间换时间做法,所以非常适合新生代内存,因为仅复制存活的对象且新生代内存中存活对象是占少数的。但是有如下几个重要问题需要考虑:
- 引用避免重复拷贝
假设存在三个对象temp1、temp2、temp3
,其中temp2、temp3
都引用了temp1
,js代码示例如下:
var temp2 = { ref: temp1, } var temp3 = { ref: temp1, } var temp1 = {}
从FromSpace
中拷贝temp2
到ToSpace
中时,发现引用了temp1
,便把temp1
也拷贝到ToSpace
,是一个递归的过程。但是在拷贝temp3
时发现也引用了temp1
,此时再把temp1
拷贝过去则重复了。
要避免重复拷贝,做法是拷贝时给对象添加一个标记visited
表示该节点已被访问过,后续通过visited
属性判断是否拷贝对象。
- 拷贝后保持正确的引用关系
还是上述引用关系,由于temp1
不需要重复拷贝,temp3
被拷贝到ToSpace
之后不知道temp1
对象在ToSpace
中的内存地址。
做法是temp1
被拷贝过去后该对象节点上会生成新的field
属性指向新的内存空间地址,同时更新到旧内存对象的forwarding
属性上,因此temp3
就可以通过旧temp1
的forwarding
属性找到在ToSpace
中的引用地址了。
内存对象同时存在于新生代和老生代之后,也带来了问题:
- 内存对象跨代(跨空间)后如何标记
const temp1 = {} const temp2 = { ref: temp1, }
比如上述代码中的两个对象temp1
和temp2
都存在于新生代,其中temp2
引用了temp1
。假设在经过GC
之后temp2
晋升到了老生代,那么在下次GC
的标记阶段,如何判断temp1
是否是存活对象呢?
在基于可达性分析算法中要知道temp1
是否存活,就必须要知道是否有根对象引用
引用了temp1
对象。如此的话,年轻代的GC
就要遍历所有的老生代对象判断是否有根引用对象引用了temp1
对象,如此的话分代算法就没有意义了。
解决版本就是维护一个记录所有的跨代引用的记录集,它是写缓冲区
的一个列表。只要有老生代中的内存对象指向了新生代内存对象时,就将老生代中该对象的内存引用记录到记录集中。由于这种情况一般发生在对象写的操作,顾称此为写屏障,还一种可能的情况就是发生在晋升时。记录集的维护只要关心对象的写操作和晋升操作即可。此是又带来了另一个问题:
- 每次写操作时维护记录集的额外开销
优化的手段是在一些Crankshaft
操作中是不需要写屏障的,还有就是栈上内存对象的写操作是不需要写屏障的。还有一些,更多的手段就不在这里过多讨论。
- 缓解
Scavenge
算法内存利用率不高问题
新生代内存中存活对象占比是相对较小的,因此可以在分配空间时,ToSpace
可以分配的小一些。做法是将ToSpace
空间分成S0
和S1
两部分,S0
用作于ToSpace
,S1
与原FromSpace
合并当成FromSpace
。
Scavenge算法中深度/广度优先的区别
垃圾回收算法中,识别内存对象是否是垃圾的机制一般有两种:引用计数和基于可达性分析。
Based on reachability analysis, it is to find all root references (such as global variables, etc.), traverse all root references, all references on the recursive root reference, everything that is traversed is Live objects and mark them. At this time, other memory objects in the space are dead objects, thus constructing a directed graph.
Taking into account the limitations of recursion, recursive logic generally adopts non-recursive implementation. Common ones include breadth-first and depth-first algorithms. The difference between the two is:
- changes the order of memory objects when copying depth-first to
ToSpace
, making objects with reference relationships closer. The reason is that after copying itself, the object referenced by itself is copied directly, so the related objects are closer inToSpace
- Depth priority is just the opposite
Because of the CPU's caching strategy, when reading a memory object, there is a high probability that the objects behind it will be read together, in order to hit the cache faster. Because a very common scenario during code development is obj1.obj2.obj3
. At this time, when the CPU reads obj1
, if the following obj2
, If obj3
is read together, it will be very helpful to hit the cache.
So the depth-first algorithm is more conducive to business logic hitting the cache, but its implementation requires additional stack-assisted implementation of the algorithm, which consumes memory space. Breadth first, on the contrary, cannot improve cache hits, but its implementation can use pointers to cleverly avoid space consumption, and the algorithm execution efficiency is high.
Promotion conditions for new generation memory objects
If memory objects in the new generation want to be promoted to the old generation, they need to meet the following conditions:
- Whether the object has been experienced
Scavenge
Recycling -
ToSpace
The memory usage ratio cannot exceed the limit
Judge whether it has been experienced# The logic of ##Scavenge's GC is to give the
age attribute of the surviving object
1 every time
GC, and when
GC## is done again Just judge the age
attribute when #. The basic promotion diagram is as follows:
There are many long-term surviving objects in the old generation memory, and the reason why the
algorithm cannot be recycled is:
- Wasted half of the memory space
Garbage collection of the old generation memory space uses
Mark-Sweep(Mark-Sweep
) and Mark-Sweep
(Mark -Compact
) combination method. Marking and clearing is divided into two parts:
- Clearing phase (if it is marking and sorting, it is the sorting phase)
- Traverse the old students in the marking phase All memory objects in the heap memory are generated and live objects are marked. The cleanup phase only cleans up unmarked objects. The reason is: non-survival objects account for a minority in the old generation memory.
As shown in the figure above, one problem with mark clearing is that after clearing, there is a discontinuous space that cannot be used anymore, so memory cleaning of the old generation memory space is required. A solution combined with tagging. This solution is to move the living objects to one side during the marking process, and then clean up and remove all non-living objects outside the boundary after the movement is completed.
It is necessary to pause the application execution logic during garbage collection, and then resume the application after the garbage collection mechanism is completed Execution logic, this behavior is called "
Full Pause", which is often called Stop The World, referred to as STW
. Garbage collection of young generation memory has little impact on application execution, but because there are many surviving objects in old generation memory, the impact of full pauses caused by garbage collection of old generation memory is very large.
In order to optimize the full pause time of GC, V8 also introduces
, concurrency markers
,parallel Mark
, Incremental cleaning
, Parallel cleaning
, Delayed cleaning
and other methods.
An important metric to measure the time spent in garbage collection is the amount of time the main thread is paused while executing
GC. The impact of STW is unacceptable, so V8 has also adopted many optimization methods. <ul><li>Parallel GC</li></ul>
<p>The GC process needs to do a lot of things, which will lead to the STW phenomenon on the main thread. The method of parallel GC is to open multiple auxiliary threads to share the GC work. This approach still cannot avoid the STW phenomenon, but it can reduce the total time of STW, depending on the number of auxiliary threads enabled. </p>
<p><img src="/static/imghwm/default1.png" data-src="https://img.php.cn/upload/image/985/850/332/1651063178901894.png?x-oss-process=image/resize,p_40" class="lazy" title="165106317370481Lets talk about V8s memory management and garbage collection algorithm" alt="1Lets talk about V8s memory management and garbage collection algorithm"></p>
<ul><li>Incremental<code>GC
Incremental GC splits the GC work and executes it on the main thread Intermittent step-by-step execution. This approach will not reduce the GC time. On the contrary, it will cost a little, but it will also reduce the total STW time of the GC.
- Concurrency
GC
Concurrent GC means that GC runs in the background and no longer runs on the main thread . This approach will avoid the STW phenomenon.
- Idle time
GC
The rendering of animations in Chrome
is approx. ##60 frames (each frame is about
16ms), if the current rendering time reaches
16.6ms, then there will be free time to do other things, such as part
GCTask.
Reduce the impact of garbage collection
If you want to improve execution efficiency, you must minimize the execution and consumption of garbage collection:
- Be careful when using memory as a cache and be careful about using objects as a cache. To reasonably limit the expiration time and infinite growth issues, you can use the lru strategy
- Avoid using memory to store user sessions in
Node
. Otherwise, storing a large number of user session objects in the memory will cause the memory of the old generation to surge, affecting cleanup performance and thus affecting application execution performance and memory overflow. Improved ways to use redis, etc. Benefits of moving the cache externally:
- Reduce the number of resident memory objects and make garbage collection more efficient
- The cache can be shared between processes
nodejs tutorial!
The above is the detailed content of Let's talk about V8's memory management and garbage collection algorithm. For more information, please follow other related articles on the PHP Chinese website!

node、nvm与npm的区别:1、nodejs是项目开发时所需要的代码库,nvm是nodejs版本管理工具,npm是nodejs包管理工具;2、nodejs能够使得javascript能够脱离浏览器运行,nvm能够管理nodejs和npm的版本,npm能够管理nodejs的第三方插件。

Vercel是什么?本篇文章带大家了解一下Vercel,并介绍一下在Vercel中部署 Node 服务的方法,希望对大家有所帮助!

node怎么爬取数据?下面本篇文章给大家分享一个node爬虫实例,聊聊利用node抓取小说章节的方法,希望对大家有所帮助!

node导出模块的两种方式:1、利用exports,该方法可以通过添加属性的方式导出,并且可以导出多个成员;2、利用“module.exports”,该方法可以直接通过为“module.exports”赋值的方式导出模块,只能导出单个成员。

安装node时会自动安装npm;npm是nodejs平台默认的包管理工具,新版本的nodejs已经集成了npm,所以npm会随同nodejs一起安装,安装完成后可以利用“npm -v”命令查看是否安装成功。

node中没有包含dom和bom;bom是指浏览器对象模型,bom是指文档对象模型,而node中采用ecmascript进行编码,并且没有浏览器也没有文档,是JavaScript运行在后端的环境平台,因此node中没有包含dom和bom。

Node.js 如何实现异步资源上下文共享?下面本篇文章给大家介绍一下Node实现异步资源上下文共享的方法,聊聊异步资源上下文共享对我们来说有什么用,希望对大家有所帮助!


Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

AI Hentai Generator
Generate AI Hentai for free.

Hot Article

Hot Tools

Zend Studio 13.0.1
Powerful PHP integrated development environment

Notepad++7.3.1
Easy-to-use and free code editor

Atom editor mac version download
The most popular open source editor

SAP NetWeaver Server Adapter for Eclipse
Integrate Eclipse with SAP NetWeaver application server.

MinGW - Minimalist GNU for Windows
This project is in the process of being migrated to osdn.net/projects/mingw, you can continue to follow us there. MinGW: A native Windows port of the GNU Compiler Collection (GCC), freely distributable import libraries and header files for building native Windows applications; includes extensions to the MSVC runtime to support C99 functionality. All MinGW software can run on 64-bit Windows platforms.
