I have a class ImageInputFormat
in Hadoop which reads images from HDFS. How to use my InputFormat in Spark?
Here is my ImageInputFormat
:
images all be stored in hadoopRDD ?
yes, everything that will be saved in spark is as rdds
can set the RDD capacity and when the RDD is full, the rest data will be stored in disk?
Default storage level in spark is (StorageLevel.MEMORY_ONLY) ,use MEMORY_ONLY_SER, which is more space efficient. Please refer spark documentation > scala programming > RDD persistance
Will the performance be influenced if the data is too big?
As data size increases , it will effect performance too.