简体   繁体   English

Hadoop-将数据从HTTP上传(PUT)直接流式传输到HDFS

[英]Hadoop - streaming data from HTTP upload (PUT) into HDFS directly

I have the following application deployment 我有以下应用程序部署

  • web front-end taking data from client through HTTP/FTP Web前端通过HTTP / FTP从客户端获取数据
  • hadoop cluster Hadoop集群

I need to store client's data on HDFS. 我需要在HDFS上存储客户端的数据。 What is the best way of doing that? 最好的方法是什么? Is it possible to stream data to HDFS directly, without consuming all data from the client on local drive, and then put it into the HDFS? 是否可以直接将数据流传输到HDFS,而无需使用本地驱动器上来自客户端的所有数据,然后将其放入HDFS?

The feasible options which I can think of right now are : 我现在可以想到的可行方案是:

Choosing the "best" one is totally upto you, based on your convenience and ease. 根据您的方便和轻松,选择“最佳”解决方案完全取决于您。

Personally, if you want low latency access to HDFS, your best bet is HBase. 就个人而言,如果要低延迟访问HDFS,最好的选择是HBase。 You can put and get values very easily since it is just a key value store. 您可以轻松地放置和获取值,因为它只是键值存储。 We are using the same thing in our application(s) and it works fabulously. 我们在我们的应用程序中使用了相同的东西,并且效果很好。

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM