I have a fairly large dataframe in R that I would like to export to SPSS.
This file has caused me hours of headaches trying to import it to R in the first place, however I got successful using read.fwf()
using the options comment.char="%"
(a character not appearing in the file) and fill= TRUE
(it was a fixed-width ASCII file with some rows lacking all variables, causing error messages).
Anyway, my data frame currently consists of 3,9 mill observations and 48 variables (all character). I can write it to file fairly quickly by splitting it into 4 x 1 mill obs sets with df2 <- df[1:1000000,]
followed by write.table(df2)
etc., but can't write the entire file in one sweep without the computer locking up and needing a hard reset to come back up.
After hearing anecdotal stories about how R is unsuited for large datasets for years, this is the first time I have actually encountered a problem of this kind. I wonder whether there are other approaches(low-level "dumping" the file directly to disk?) or whether there is some package unknown to me that can handle export of large files of this type efficiently?
sqldf()
as suggested by JD Long since I'm using it a lot in my analyses. Thanks for the pointer! – Talos