简体   繁体   English

大数据集群(Hadoop,Spark等)中的副本是否由同一系统管理员管理?

[英]Are replicas in a big data cluster (Hadoop, Spark etc) managed by the same sysadmin?

I am a security researcher and would like to know the standard practice about managing replicas in a big data cluster. 我是一名安全研究员,并且想了解有关在大数据集群中管理副本的标准实践。 Specifically, I would like to know if platforms like Hadoop, Spark etc [or] providers like Google, Amazon etc usually employ the same sysadmin to manage replica nodes? 具体来说,我想知道Hadoop,Spark等[]或Google,Amazon等的提供程序是否通常使用相同的sysadmin来管理副本节点? If so, will it affect the performance so bad if the replicas are managed by different sysadmins? 如果是这样,如果副本由不同的系统管理员管理,是否会对性能造成如此严重的影响?

I know that the scope of this question is too broad, so any insight into big data security standards about replica management is appreciated. 我知道这个问题的范围太广了,因此对有关副本管理的大数据安全标准的任何见解都值得赞赏。 Thanks. 谢谢。

The sysadmin of a slave node does not matters as most of the communication within the cluster happens over the network. 从属节点的sysadmin无关紧要,因为群集中的大多数通信都是通过网络进行的。 And in my view sysadmin cannot impact the performance of a Hadoop Cluster. 我认为sysadmin不会影响Hadoop集群的性能。

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM