I have a CSV on the hadoop file system hdfs that I want to convert into multiple serialized java objects using this framework:
https://github.com/clarkduvall/serpy
I heard of avro and parquet. I don't want to use those and want to output serialized binary files. My csv file contains records like:
Name, Age, Date
Jordan, 1, 1/1/2017
John, 5, 2/2/2017
Is this possible using Hadoop or Spark? The output objects should be readable by a normal non-hadoop/spark related Java program. Any example would be appreciated!