简体   繁体   English

融合HDFS Sink连接器:如何配置自定义hadoop用户和组?

[英]Confluent HDFS Sink Connector: How to configure custom hadoop user and group?

We are currently using Confluent HDFS Sink Connector platform within docker container to write data from Kafka(separate Kafka cluster) to HDFS(separate Hadoop cluster). 我们当前在Docker容器中使用Confluent HDFS Sink Connector平台将数据从Kafka(独立的Kafka集群)写入HDFS(独立的Hadoop集群)。 By default the connector platform writes data to HDFS with root user and wheel group. 默认情况下,连接器平台使用root用户和wheel组将数据写入HDFS。 How can i configure connector to use a specific hadoop user/group ? 如何配置连接器以使用特定的hadoop用户/组? Is there an environment variable I need to set in docker ? 我需要在docker中设置环境变量吗? Thanks. 谢谢。

The Java process in the Docker container runs as root. Docker容器中的Java进程以root身份运行。

You need to either make your own container with your own user account or run the Connect Workers as a different Unix account in some other way. 您需要使用自己的用户帐户创建自己的容器,或者以其他方式将Connect Workers作为不同的Unix帐户运行。

You could try setting HADOOP_IDENT_USER or HADOOP_USER_NAME environment variables, but I think these are only pulled by the Hadoop scripts, not the Java API 您可以尝试设置HADOOP_IDENT_USERHADOOP_USER_NAME环境变量,但是我认为这些变量仅由Hadoop脚本而不是Java API提取

Keep in mind that user accounts in Hadoop don't really matter if you're not using a Kerberized cluster 请记住,如果您未使用Kerberized集群,则Hadoop中的用户帐户并不重要

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM