Read huge csv file in R from Hadoop -
I am trying to read large (250 MB, 0.6 million lines, 60 columns) CSV files in R in Hadoop . Field separator I am using "@" and I am using R code
PC2009 & lt; - as.matrix (from.dfs ("hdfs: //..../ apps / g00008 / moloy / fuzzy / TUSOpsSODetails_2009.csv", make.input.format (format = "csv", sep = "@", Fill = TRUE)) $ val) But the data in the 5th column is getting worse. Should I "Data" in HDFS by "Cats" Has anyone educated me how to read large files in RP? Please note that I am new to MapReduce.
Comments
Post a Comment