_Visit the [main page](../README.md)_ # Comparing TSV and CSV formats The differences between TSV and CSV formats can be confusing. The obvious distinction is the default field delimiter: TSV uses TAB, CSV uses comma. Both use newline as the record delimiter. By itself, using different field delimiters is not especially significant. Far more important is the approach to delimiters occurring in the data. CSV uses an escape syntax to represent commas and newlines in the data. TSV takes a different approach, disallowing TABs and newlines in the data. The escape syntax enables CSV to fully represent common written text. This is a good fit for human edited documents, notably spreadsheets. This generality has a cost: reading it requires programs to parse the escape syntax. While not overly difficult, it is still easy to do incorrectly, especially when writing one-off programs. It is good practice to use a CSV parser when processing CSV files. Traditional Unix tools like `cut`, `sort`, `awk`, and `diff` do not process CSV escapes, alternate tools are needed. By contrast, parsing TSV data is simple. Records can be read using the typical `readline` routines found in most programming languages. The fields in each record can be found using `split` routines. Unix utilities can be called by providing the correct field delimiter, e.g. `awk -F "\t"`, `sort -t $'\t'`. No special parser is needed. This is much more reliable. It is also faster, no CPU time is used parsing the escape syntax. The speed advantages are especially pronounced for record oriented operations. Record counts (`wc -l`), deduplication (`uniq`, [tsv-uniq](tool_reference/tsv-uniq.md)), file splitting (`head`, `tail`, `split`), shuffling (GNU `shuf`, [tsv-sample](tool_reference/tsv-sample.md)), etc. TSV is faster because record boundaries can be found using highly optimized newline search routines (e.g. `memchr`). Identifying CSV record boundaries requires fully parsing each record. These characteristics make TSV format well suited for the large tabular data sets common in data mining and machine learning environments. These data sets rarely need TAB and newline characters in the fields. The most common CSV escape format uses quotes to delimit fields containing delimiters. Quotes must also be escaped, this is done by using a pair of quotes to represent a single quote. Consider the data in this table: | Field-1 | Field-2 | Field-3 | | ------- | -------------------- | ------- | | abc | hello, world! | def | | ghi | Say "hello, world!" | jkl | In Field-2, the first value contains a comma, the second value contains both quotes and a comma. Here is the CSV representation, using escapes to represent commas and quotes in the data. ``` Field-1,Field-2,Field-3 abc,"hello, world!",def ghi,"Say ""hello, world!""",jkl ``` In the above example, only fields with delimiters are quoted. It is also common to quote all fields whether or not they contain delimiters. The following CSV file is equivalent: ``` "Field-1","Field-2","Field-3" "abc","hello, world!","def" "ghi","Say ""hello, world!""","jkl" ``` Here's the same data in TSV. It is much simpler as no escapes are involved: ``` Field-1 Field-2 Field-3 abc hello, world! def ghi Say "hello, world!" jkl ``` The similarity between TSV and CSV can lead to confusion about which tools are appropriate. Furthering this confusion, it is somewhat common to have data files using comma as the field delimiter, but without comma, quote, or newlines in the data. No CSV escapes are needed in these files, with the implication that traditional Unix tools like `cut` and `awk` can be used to process these files. Such files are sometimes referred to as "simple CSV". They are equivalent to TSV files using comma as the field delimiter. Traditional Unix tools and [tsv-utils](../README.md) tools can process these files correctly by specifying the field delimiter. However, "simple csv" is a very ad hoc and ill defined notion. A simple precaution when working with these files is to run a CSV-to-TSV converter like [csv2tsv](tool_reference/csv2tsv.md) prior to other processing steps. Note that many CSV-to-TSV conversion tools don't actually remove the CSV escapes. Instead, many tools replace comma with TAB as the record delimiter, but still use CSV escapes to represent TAB, newline, and quote characters in the data. Such data cannot be reliably processed by Unix tools like `sort`, `awk`, and `cut`. The [csv2tsv](tool_reference/csv2tsv.md) tool in [tsv-utils](../README.md) avoids escapes by replacing TAB and newline with a space (customizable). This works well in the vast majority of data mining scenarios. To see what a specific CSV-to-TSV conversion tool does, convert CSV data containing quotes, commas, TABs, newlines, and double-quoted fields. For example: ``` $ echo $'Line,Field1,Field2\n1,"Comma: |,|","Quote: |""|"\n"2","TAB: |\t|","Newline: |\n|"' | ``` Approaches that generate CSV escapes will enclose a number of the output fields in double quotes. References: - [Wikipedia: Tab-separated values](https://en.wikipedia.org/wiki/Tab-separated_values) - Useful description of TSV format. - [IANA TSV specification](https://www.iana.org/assignments/media-types/text/tab-separated-values) - Formal definition of the tab-separated-values mime type. - [Wikipedia: Comma-separated-values](https://en.wikipedia.org/wiki/Comma-separated_values) - Describes CSV and related formats. - [RFC 4180](https://tools.ietf.org/html/rfc4180) - IETF CSV format description, the closest thing to an actual standard for CSV. - [brendano/tsvutils: The philosophy of tsvutils](https://github.com/brendano/tsvutils#the-philosophy-of-tsvutils) - Brendan O'Connor's discussion of the rationale for using TSV format in his open source toolkit. - [So You Want To Write Your Own CSV code?](http://thomasburette.com/blog/2014/05/25/so-you-want-to-write-your-own-CSV-code/) - Thomas Burette's humorous, and accurate, blog post describing the troubles with ad-hoc CSV parsing. Of course, you could use TSV and avoid these problems!