TransWikia.com

R: reading big data files in R

Data Science Asked by Minaj on April 12, 2021

In R, there are libraries which help speed up the process of reading huge data files. Examples of these libraries include sqldf and ff.

What would be the disadvantage of using these packages to read a small file that one would typically read with read.csv?

If their is no disadvantage of using them to read small files, does this mean functions such as read.csv might not have much use in future given the emerging big-data-targeted readers?

One Answer

In R many great improvements are typically made by creating new libraries rather than changing the actual functions in base R itself. datatable, read_csv in readr, tibbles, etc are very good examples.

read.csv vs fread as answered by Matt Dowle himself

read.csv coerces strings as factors by default which is not done by read_csv and fread. So if that is what you need, read.csv will work well. Now answering your question,there is no disadvantage using the fast readers to read small files apart from a small speedup that you might get.

Answered by Vipin on April 12, 2021

Add your own answers!

Ask a Question

Get help from others!

© 2024 TransWikia.com. All rights reserved. Sites we Love: PCI Database, UKBizDB, Menu Kuliner, Sharing RPP