i want to read a very large dataset of xml files ( each xml file size = 1TB) on spark and start a parsing process on each file so that in the end I get csv files as tables.