[R] different outcomes using read.table vs read.csv

Barry Rowlingson b.rowlingson at lancaster.ac.uk
Fri Mar 13 21:47:04 CET 2009


2009/3/13 jatwood <jatwood at montana.edu>:
> Good Afternoon
> I have noticed results similar to the following several times as I have used
> R over the past several years.
> My .csv file has a header row and 3073 rows of data.
>
>> rskreg<-read.table('D:/data/riskregions.csv',header=T,sep=",")
>> dim(rskreg)
> [1] 2722   13
>> rskreg<-read.csv('D:/data/riskregions.csv',header=T)
>> dim(rskreg)
> [1] 3073   13
>>
>
> Does someone know what could be causing the read.table and read.csv
> functions to give different results on some occasions?  The riskregions.csv
> file was generated with and saved from MS.Excel.

 read.table has 'comment.char="#"', so if a line starts with # it gets
ignored. read.csv doesn't have this set, so it might explain why
read.csv gets more than read.table...

 Do you have lines starting with #? Try read.table with
comment.char="" and see if you get the right number. See the help for
read.table for more info.

 I'd not seen this before, hope it hasn't bitten me...

Barry




More information about the R-help mailing list