简体   繁体   English

Hadoop集群中最大的数据节点数是多少?

[英]What is the maximum number of datanodes in Hadoop cluster?

I'm using 8 node hadoop cluster and namenode memory usage shows as 7%. 我正在使用8节点hadoop集群,namenode的内存使用量显示为7%。 I feel this cluster going to reach maximum limits of namenode memory. 我觉得这个群集将达到namenode内存的最大限制。

Current cluster storage size is 5.6TB and namenode heap size is 4GB. 当前群集存储大小为5.6TB,namenode堆大小为4GB。 Further block size is 128MB. 进一步的块大小为128MB。

What is the way of calculating maximum number of datanodes that can have for Hadoop cluster (with single namenode)? 计算Hadoop集群(带有单个namenode)的最大数据节点数的方法是什么?

Namenode memory usage is proportional to the number of blocks also the guideline is 1 million blocks take 1 GB memory...and with you 5.6 TB and 128 MB block size i calculated you can have around half million blocks without even replication(you can recalculate). Namenode的内存使用量与块的数量成正比,而且指导原则是1百万个块占用1 GB的内存...据我计算,如果您拥有5.6 TB和128 MB的块大小,则即使没有复制也可以拥有大约50万个块(您可以重新计算)。 So, my conclusion is you have enough memory unless there are lots of small files... 因此,我的结论是,除非有很多小文件,否则您将有足够的内存...

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM