In R which packages for loading larger data quickly

后端 未结 2 1303
陌清茗
陌清茗 2021-01-06 18:55

In R usually data is loaded in RAM. Are there any packages which load the data in disk rather than RAM

相关标签:
2条回答
  • 2021-01-06 19:40

    Yes, the ff package can do that.

    You may want to look at the Task View on High-Performance Computing for more details.

    0 讨论(0)
  • 2021-01-06 19:43

    Check out the bigmemory package, along with related packages like bigtabulate, bigalgebra, biganalytics, and more. There's also ff, though I don't find it as user-friendly as the bigmemory suite. The bigmemory suite was reportedly partially motivated by the difficulty of using ff. I like it because it required very few changes to my code to be able to access a bigmatrix object: it can be manipulated in almost exactly the same ways as a standard matrix, so my code is very reusable.

    There's also support for HDF5 via NetCDF4, in packages like RNetCDF and ncdf. This is a popular, multi-platform, multi-language method for efficient storage and access of large data sets.

    If you want basic memory mapping functionality, look at the mmap package.

    0 讨论(0)
提交回复
热议问题