i originally asked this question about performing this task with the httr
package, but i don't think it's possible using httr
. so i've re-written my code to use RCurl
instead -- but i'm still tripping up on something probably related to the writefunction
.. but i really don't understand why.
you should be able to reproduce my work by using the 32-bit version of R, so you hit memory limits if you read anything into RAM. i need a solution that downloads directly to the hard disk.
to start, this code to works -- the zipped file is appropriately saved to the disk.
library(RCurl)
filename <- tempfile()
f <- CFILE(filename, "wb")
url <- "http://www2.census.gov/acs2011_5yr/pums/csv_pus.zip"
curlPerform(url = url, writedata = f@ref)
close(f)
# 2.1 GB file successfully written to disk
now here's some RCurl
code that does not work. as stated in the previous question, reproducing this exactly will require creating an extract on ipums.
your.email <- "[email protected]"
your.password <- "password"
extract.path <- "https://usa.ipums.org/usa-action/downloads/extract_files/some_file.csv.gz"
library(RCurl)
values <-
list(
"login[email]" = your.email ,
"login[password]" = your.password ,
"login[is_for_login]" = 1
)
curl = getCurlHandle()
curlSetOpt(
cookiejar = 'cookies.txt',
followlocation = TRUE,
autoreferer = TRUE,
ssl.verifypeer = FALSE,
curl = curl
)
params <-
list(
"login[email]" = your.email ,
"login[password]" = your.password ,
"login[is_for_login]" = 1
)
html <- postForm("https://usa.ipums.org/usa-action/users/validate_login", .params = params, curl = curl)
dl <- getURL( "https://usa.ipums.org/usa-action/extract_requests/download" , curl = curl)
and now that i'm logged in, try the same commands as above, but with the curl
object to keep the cookies.
filename <- tempfile()
f <- CFILE(filename, mode = "wb")
this line breaks--
curlPerform(url = extract.path, writedata = f@ref, curl = curl)
close(f)
# the error is:
Error in curlPerform(url = extract.path, writedata = f@ref, curl = curl) :
embedded nul in string: [[binary jibberish here]]
the answer to my previous post referred me to this c-level writefunction answer, but i'm clueless about how to re-create that curl_writer C program (on windows?)..
dyn.load("curl_writer.so")
writer <- getNativeSymbolInfo("writer", PACKAGE="curl_writer")$address
curlPerform(URL=url, writefunction=writer)
..or why it's even necessary, given that the five lines of code at the top of this question work without anything crazy like getNativeSymbolInfo
. i just don't understand why passing in that extra curl
object that stores the authentication/cookies and tells it not to verify SSL would cause code that otherwise works.. to break?
curl = getCurlHandle()
andcurlPerform(url = url, writedata = f@ref, curl = curl)
? and, are you able to download some other content once session has started? for example, usingcurlPerform
andwritedata
to savehttps://usa.ipums.org/usa-action/extract_requests/download
– Brunelldyn.load("curl_writer.dll")
– Brunellz <- getBinaryURL( extract.path , curl = curl )
works, but it reads everything into RAM and so doesn't solve my problem. 3) is it possible to do this within R on windows? thanks!! :) – Unciform