简体   繁体   中英

How to convert hadoop avro, parquet, as well as text file to csv without spark

I have hdfs versions of avro, parquet, and text file. Unfortunately, I can't use spark to convert them to csv. I saw from an earlier so question that this doesn't seem to be possible. How to convert HDFS file to csv or tsv. Is this possible, and if so, how do I do this?

This will help you to read Avro files (just avoid schema evolution/modifications...). Example .

As to Parquet, you can use parquet-mr , take a look at ParquetReader. Example : ignore the Spark usage, they just use it in order to create a Parquet file to be used later on with ParquetReader.

Hope it helps

The technical post webpages of this site follow the CC BY-SA 4.0 protocol. If you need to reprint, please indicate the site URL or the original address.Any question please contact:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM