github.com/twelsh-aw/go/src@v0.0.0-20230516233729-a56fe86a7c81/encoding/csv/reader.go (about) 1 // Copyright 2011 The Go Authors. All rights reserved. 2 // Use of this source code is governed by a BSD-style 3 // license that can be found in the LICENSE file. 4 5 // Package csv reads and writes comma-separated values (CSV) files. 6 // There are many kinds of CSV files; this package supports the format 7 // described in RFC 4180. 8 // 9 // A csv file contains zero or more records of one or more fields per record. 10 // Each record is separated by the newline character. The final record may 11 // optionally be followed by a newline character. 12 // 13 // field1,field2,field3 14 // 15 // White space is considered part of a field. 16 // 17 // Carriage returns before newline characters are silently removed. 18 // 19 // Blank lines are ignored. A line with only whitespace characters (excluding 20 // the ending newline character) is not considered a blank line. 21 // 22 // Fields which start and stop with the quote character " are called 23 // quoted-fields. The beginning and ending quote are not part of the 24 // field. 25 // 26 // The source: 27 // 28 // normal string,"quoted-field" 29 // 30 // results in the fields 31 // 32 // {`normal string`, `quoted-field`} 33 // 34 // Within a quoted-field a quote character followed by a second quote 35 // character is considered a single quote. 36 // 37 // "the ""word"" is true","a ""quoted-field""" 38 // 39 // results in 40 // 41 // {`the "word" is true`, `a "quoted-field"`} 42 // 43 // Newlines and commas may be included in a quoted-field 44 // 45 // "Multi-line 46 // field","comma is ," 47 // 48 // results in 49 // 50 // {`Multi-line 51 // field`, `comma is ,`} 52 package csv 53 54 import ( 55 "bufio" 56 "bytes" 57 "errors" 58 "fmt" 59 "io" 60 "unicode" 61 "unicode/utf8" 62 ) 63 64 // A ParseError is returned for parsing errors. 65 // Line numbers are 1-indexed and columns are 0-indexed. 66 type ParseError struct { 67 StartLine int // Line where the record starts 68 Line int // Line where the error occurred 69 Column int // Column (1-based byte index) where the error occurred 70 Err error // The actual error 71 } 72 73 func (e *ParseError) Error() string { 74 if e.Err == ErrFieldCount { 75 return fmt.Sprintf("record on line %d: %v", e.Line, e.Err) 76 } 77 if e.StartLine != e.Line { 78 return fmt.Sprintf("record on line %d; parse error on line %d, column %d: %v", e.StartLine, e.Line, e.Column, e.Err) 79 } 80 return fmt.Sprintf("parse error on line %d, column %d: %v", e.Line, e.Column, e.Err) 81 } 82 83 func (e *ParseError) Unwrap() error { return e.Err } 84 85 // These are the errors that can be returned in ParseError.Err. 86 var ( 87 ErrBareQuote = errors.New("bare \" in non-quoted-field") 88 ErrQuote = errors.New("extraneous or missing \" in quoted-field") 89 ErrFieldCount = errors.New("wrong number of fields") 90 91 // Deprecated: ErrTrailingComma is no longer used. 92 ErrTrailingComma = errors.New("extra delimiter at end of line") 93 ) 94 95 var errInvalidDelim = errors.New("csv: invalid field or comment delimiter") 96 97 func validDelim(r rune) bool { 98 return r != 0 && r != '"' && r != '\r' && r != '\n' && utf8.ValidRune(r) && r != utf8.RuneError 99 } 100 101 // A Reader reads records from a CSV-encoded file. 102 // 103 // As returned by NewReader, a Reader expects input conforming to RFC 4180. 104 // The exported fields can be changed to customize the details before the 105 // first call to Read or ReadAll. 106 // 107 // The Reader converts all \r\n sequences in its input to plain \n, 108 // including in multiline field values, so that the returned data does 109 // not depend on which line-ending convention an input file uses. 110 type Reader struct { 111 // Comma is the field delimiter. 112 // It is set to comma (',') by NewReader. 113 // Comma must be a valid rune and must not be \r, \n, 114 // or the Unicode replacement character (0xFFFD). 115 Comma rune 116 117 // Comment, if not 0, is the comment character. Lines beginning with the 118 // Comment character without preceding whitespace are ignored. 119 // With leading whitespace the Comment character becomes part of the 120 // field, even if TrimLeadingSpace is true. 121 // Comment must be a valid rune and must not be \r, \n, 122 // or the Unicode replacement character (0xFFFD). 123 // It must also not be equal to Comma. 124 Comment rune 125 126 // FieldsPerRecord is the number of expected fields per record. 127 // If FieldsPerRecord is positive, Read requires each record to 128 // have the given number of fields. If FieldsPerRecord is 0, Read sets it to 129 // the number of fields in the first record, so that future records must 130 // have the same field count. If FieldsPerRecord is negative, no check is 131 // made and records may have a variable number of fields. 132 FieldsPerRecord int 133 134 // If LazyQuotes is true, a quote may appear in an unquoted field and a 135 // non-doubled quote may appear in a quoted field. 136 LazyQuotes bool 137 138 // If TrimLeadingSpace is true, leading white space in a field is ignored. 139 // This is done even if the field delimiter, Comma, is white space. 140 TrimLeadingSpace bool 141 142 // ReuseRecord controls whether calls to Read may return a slice sharing 143 // the backing array of the previous call's returned slice for performance. 144 // By default, each call to Read returns newly allocated memory owned by the caller. 145 ReuseRecord bool 146 147 // Deprecated: TrailingComma is no longer used. 148 TrailingComma bool 149 150 r *bufio.Reader 151 152 // numLine is the current line being read in the CSV file. 153 numLine int 154 155 // offset is the input stream byte offset of the current reader position. 156 offset int64 157 158 // rawBuffer is a line buffer only used by the readLine method. 159 rawBuffer []byte 160 161 // recordBuffer holds the unescaped fields, one after another. 162 // The fields can be accessed by using the indexes in fieldIndexes. 163 // E.g., For the row `a,"b","c""d",e`, recordBuffer will contain `abc"de` 164 // and fieldIndexes will contain the indexes [1, 2, 5, 6]. 165 recordBuffer []byte 166 167 // fieldIndexes is an index of fields inside recordBuffer. 168 // The i'th field ends at offset fieldIndexes[i] in recordBuffer. 169 fieldIndexes []int 170 171 // fieldPositions is an index of field positions for the 172 // last record returned by Read. 173 fieldPositions []position 174 175 // lastRecord is a record cache and only used when ReuseRecord == true. 176 lastRecord []string 177 } 178 179 // NewReader returns a new Reader that reads from r. 180 func NewReader(r io.Reader) *Reader { 181 return &Reader{ 182 Comma: ',', 183 r: bufio.NewReader(r), 184 } 185 } 186 187 // Read reads one record (a slice of fields) from r. 188 // If the record has an unexpected number of fields, 189 // Read returns the record along with the error ErrFieldCount. 190 // If the record contains a field that cannot be parsed, 191 // Read returns a partial record along with the parse error. 192 // The partial record contains all fields read before the error. 193 // If there is no data left to be read, Read returns nil, io.EOF. 194 // If ReuseRecord is true, the returned slice may be shared 195 // between multiple calls to Read. 196 func (r *Reader) Read() (record []string, err error) { 197 if r.ReuseRecord { 198 record, err = r.readRecord(r.lastRecord) 199 r.lastRecord = record 200 } else { 201 record, err = r.readRecord(nil) 202 } 203 return record, err 204 } 205 206 // FieldPos returns the line and column corresponding to 207 // the start of the field with the given index in the slice most recently 208 // returned by Read. Numbering of lines and columns starts at 1; 209 // columns are counted in bytes, not runes. 210 // 211 // If this is called with an out-of-bounds index, it panics. 212 func (r *Reader) FieldPos(field int) (line, column int) { 213 if field < 0 || field >= len(r.fieldPositions) { 214 panic("out of range index passed to FieldPos") 215 } 216 p := &r.fieldPositions[field] 217 return p.line, p.col 218 } 219 220 // InputOffset returns the input stream byte offset of the current reader 221 // position. The offset gives the location of the end of the most recently 222 // read row and the beginning of the next row. 223 func (r *Reader) InputOffset() int64 { 224 return r.offset 225 } 226 227 // pos holds the position of a field in the current line. 228 type position struct { 229 line, col int 230 } 231 232 // ReadAll reads all the remaining records from r. 233 // Each record is a slice of fields. 234 // A successful call returns err == nil, not err == io.EOF. Because ReadAll is 235 // defined to read until EOF, it does not treat end of file as an error to be 236 // reported. 237 func (r *Reader) ReadAll() (records [][]string, err error) { 238 for { 239 record, err := r.readRecord(nil) 240 if err == io.EOF { 241 return records, nil 242 } 243 if err != nil { 244 return nil, err 245 } 246 records = append(records, record) 247 } 248 } 249 250 // readLine reads the next line (with the trailing endline). 251 // If EOF is hit without a trailing endline, it will be omitted. 252 // If some bytes were read, then the error is never io.EOF. 253 // The result is only valid until the next call to readLine. 254 func (r *Reader) readLine() ([]byte, error) { 255 line, err := r.r.ReadSlice('\n') 256 if err == bufio.ErrBufferFull { 257 r.rawBuffer = append(r.rawBuffer[:0], line...) 258 for err == bufio.ErrBufferFull { 259 line, err = r.r.ReadSlice('\n') 260 r.rawBuffer = append(r.rawBuffer, line...) 261 } 262 line = r.rawBuffer 263 } 264 readSize := len(line) 265 if readSize > 0 && err == io.EOF { 266 err = nil 267 // For backwards compatibility, drop trailing \r before EOF. 268 if line[readSize-1] == '\r' { 269 line = line[:readSize-1] 270 } 271 } 272 r.numLine++ 273 r.offset += int64(readSize) 274 // Normalize \r\n to \n on all input lines. 275 if n := len(line); n >= 2 && line[n-2] == '\r' && line[n-1] == '\n' { 276 line[n-2] = '\n' 277 line = line[:n-1] 278 } 279 return line, err 280 } 281 282 // lengthNL reports the number of bytes for the trailing \n. 283 func lengthNL(b []byte) int { 284 if len(b) > 0 && b[len(b)-1] == '\n' { 285 return 1 286 } 287 return 0 288 } 289 290 // nextRune returns the next rune in b or utf8.RuneError. 291 func nextRune(b []byte) rune { 292 r, _ := utf8.DecodeRune(b) 293 return r 294 } 295 296 func (r *Reader) readRecord(dst []string) ([]string, error) { 297 if r.Comma == r.Comment || !validDelim(r.Comma) || (r.Comment != 0 && !validDelim(r.Comment)) { 298 return nil, errInvalidDelim 299 } 300 301 // Read line (automatically skipping past empty lines and any comments). 302 var line []byte 303 var errRead error 304 for errRead == nil { 305 line, errRead = r.readLine() 306 if r.Comment != 0 && nextRune(line) == r.Comment { 307 line = nil 308 continue // Skip comment lines 309 } 310 if errRead == nil && len(line) == lengthNL(line) { 311 line = nil 312 continue // Skip empty lines 313 } 314 break 315 } 316 if errRead == io.EOF { 317 return nil, errRead 318 } 319 320 // Parse each field in the record. 321 var err error 322 const quoteLen = len(`"`) 323 commaLen := utf8.RuneLen(r.Comma) 324 recLine := r.numLine // Starting line for record 325 r.recordBuffer = r.recordBuffer[:0] 326 r.fieldIndexes = r.fieldIndexes[:0] 327 r.fieldPositions = r.fieldPositions[:0] 328 pos := position{line: r.numLine, col: 1} 329 parseField: 330 for { 331 if r.TrimLeadingSpace { 332 i := bytes.IndexFunc(line, func(r rune) bool { 333 return !unicode.IsSpace(r) 334 }) 335 if i < 0 { 336 i = len(line) 337 pos.col -= lengthNL(line) 338 } 339 line = line[i:] 340 pos.col += i 341 } 342 if len(line) == 0 || line[0] != '"' { 343 // Non-quoted string field 344 i := bytes.IndexRune(line, r.Comma) 345 field := line 346 if i >= 0 { 347 field = field[:i] 348 } else { 349 field = field[:len(field)-lengthNL(field)] 350 } 351 // Check to make sure a quote does not appear in field. 352 if !r.LazyQuotes { 353 if j := bytes.IndexByte(field, '"'); j >= 0 { 354 col := pos.col + j 355 err = &ParseError{StartLine: recLine, Line: r.numLine, Column: col, Err: ErrBareQuote} 356 break parseField 357 } 358 } 359 r.recordBuffer = append(r.recordBuffer, field...) 360 r.fieldIndexes = append(r.fieldIndexes, len(r.recordBuffer)) 361 r.fieldPositions = append(r.fieldPositions, pos) 362 if i >= 0 { 363 line = line[i+commaLen:] 364 pos.col += i + commaLen 365 continue parseField 366 } 367 break parseField 368 } else { 369 // Quoted string field 370 fieldPos := pos 371 line = line[quoteLen:] 372 pos.col += quoteLen 373 for { 374 i := bytes.IndexByte(line, '"') 375 if i >= 0 { 376 // Hit next quote. 377 r.recordBuffer = append(r.recordBuffer, line[:i]...) 378 line = line[i+quoteLen:] 379 pos.col += i + quoteLen 380 switch rn := nextRune(line); { 381 case rn == '"': 382 // `""` sequence (append quote). 383 r.recordBuffer = append(r.recordBuffer, '"') 384 line = line[quoteLen:] 385 pos.col += quoteLen 386 case rn == r.Comma: 387 // `",` sequence (end of field). 388 line = line[commaLen:] 389 pos.col += commaLen 390 r.fieldIndexes = append(r.fieldIndexes, len(r.recordBuffer)) 391 r.fieldPositions = append(r.fieldPositions, fieldPos) 392 continue parseField 393 case lengthNL(line) == len(line): 394 // `"\n` sequence (end of line). 395 r.fieldIndexes = append(r.fieldIndexes, len(r.recordBuffer)) 396 r.fieldPositions = append(r.fieldPositions, fieldPos) 397 break parseField 398 case r.LazyQuotes: 399 // `"` sequence (bare quote). 400 r.recordBuffer = append(r.recordBuffer, '"') 401 default: 402 // `"*` sequence (invalid non-escaped quote). 403 err = &ParseError{StartLine: recLine, Line: r.numLine, Column: pos.col - quoteLen, Err: ErrQuote} 404 break parseField 405 } 406 } else if len(line) > 0 { 407 // Hit end of line (copy all data so far). 408 r.recordBuffer = append(r.recordBuffer, line...) 409 if errRead != nil { 410 break parseField 411 } 412 pos.col += len(line) 413 line, errRead = r.readLine() 414 if len(line) > 0 { 415 pos.line++ 416 pos.col = 1 417 } 418 if errRead == io.EOF { 419 errRead = nil 420 } 421 } else { 422 // Abrupt end of file (EOF or error). 423 if !r.LazyQuotes && errRead == nil { 424 err = &ParseError{StartLine: recLine, Line: pos.line, Column: pos.col, Err: ErrQuote} 425 break parseField 426 } 427 r.fieldIndexes = append(r.fieldIndexes, len(r.recordBuffer)) 428 r.fieldPositions = append(r.fieldPositions, fieldPos) 429 break parseField 430 } 431 } 432 } 433 } 434 if err == nil { 435 err = errRead 436 } 437 438 // Create a single string and create slices out of it. 439 // This pins the memory of the fields together, but allocates once. 440 str := string(r.recordBuffer) // Convert to string once to batch allocations 441 dst = dst[:0] 442 if cap(dst) < len(r.fieldIndexes) { 443 dst = make([]string, len(r.fieldIndexes)) 444 } 445 dst = dst[:len(r.fieldIndexes)] 446 var preIdx int 447 for i, idx := range r.fieldIndexes { 448 dst[i] = str[preIdx:idx] 449 preIdx = idx 450 } 451 452 // Check or update the expected fields per record. 453 if r.FieldsPerRecord > 0 { 454 if len(dst) != r.FieldsPerRecord && err == nil { 455 err = &ParseError{ 456 StartLine: recLine, 457 Line: recLine, 458 Column: 1, 459 Err: ErrFieldCount, 460 } 461 } 462 } else if r.FieldsPerRecord == 0 { 463 r.FieldsPerRecord = len(dst) 464 } 465 return dst, err 466 }