Data Science Asked by Minaj on April 12, 2021
In R, there are libraries which help speed up the process of reading huge data files. Examples of these libraries include sqldf and ff.
What would be the disadvantage of using these packages to read a small file that one would typically read with read.csv?
If their is no disadvantage of using them to read small files, does this mean functions such as read.csv might not have much use in future given the emerging big-data-targeted readers?
In R many great improvements are typically made by creating new libraries rather than changing the actual functions in base R itself. datatable, read_csv in readr, tibbles, etc are very good examples.
read.csv vs fread as answered by Matt Dowle himself
read.csv coerces strings as factors by default which is not done by read_csv and fread. So if that is what you need, read.csv will work well. Now answering your question,there is no disadvantage using the fast readers to read small files apart from a small speedup that you might get.
Answered by Vipin on April 12, 2021
Get help from others!
Recent Answers
Recent Questions
© 2024 TransWikia.com. All rights reserved. Sites we Love: PCI Database, UKBizDB, Menu Kuliner, Sharing RPP