Skip to main content

Command line tools for deduplicating and merging csv files

Project description

# csvdedupe

Command line tools for using the [dedupe python library](https://github.com/open-city/dedupe/) for deduplicating CSV files.

`csvdedupe` take a messy input file or STDIN pipe and identify duplicates

`csvlink` take two CSV files and find matches between them

[Read more about csvdedupe on OpenNews Source](http://source.mozillaopennews.org/en-US/articles/introducing-cvsdedupe/)


[![Build Status](https://travis-ci.org/datamade/csvdedupe.png?branch=master)](https://travis-ci.org/datamade/csvdedupe)

## Installation and dependencies

csvdedupe requires [numpy](http://numpy.scipy.org/), which can be complicated to install.
If you are installing numpy for the first time,
[follow these instructions](http://docs.scipy.org/doc/numpy/user/install.html). You'll need to version 1.6 of numpy or higher.

After numpy is set up, then install the following:
* [fastcluster](http://math.stanford.edu/~muellner/fastcluster.html)
* [hcluster](http://code.google.com/p/scipy-cluster/)
* [networkx](http://networkx.github.com/)

```bash
git clone git@github.com:datamade/csvdedupe.git
cd csvdedupe
pip install "numpy>=1.6"
pip install -r requirements.txt
python setup.py install
```

## csvdedupe usage

Take a messy input file or STDIN pipe and identify duplicates

Provide an input file and field names
```bash
csvdedupe examples/csv_example_messy_input.csv \
--field_names "Site name" Address Zip Phone \
--output_file output.csv
```

__or__

Pipe it, UNIX style
```bash
cat examples/csv_example_messy_input.csv | csvdedupe --skip_training \
--field_names "Site name" Address Zip Phone > output.csv
```

__or__

Define everything in a config file
```bash
csvdedupe examples/csv_example_messy_input.csv \
--config_file=config.json
```

### Example config file

```json
{
"field_names": ["Site name", "Address", "Zip", "Phone"],
"field_definition" : {"Site name" : {"type" : "String"},
"Address" : {"type" : "String"},
"Zip" : {"type" : "String",
"Has Missing" : true},
"Phone" : {"type" : "String",
"Has Missing" : true}},
"output_file": "examples/output.csv",
"skip_training": false,
"training_file": "training.json",
"sample_size": 150000,
"recall_weight": 2
}
```

### Arguments:

#### Required

* `input` a CSV file name or piped CSV file to deduplicate

Either
* `--config_file` Path to configuration file.

Or
* `--field_names` List of column names for dedupe to pay attention to

#### Optional
* `--output_file OUTPUT_FILE`
CSV file to store deduplication results (default:
None)
* `--destructive` Output file will contain unique records only
* `--skip_training` Skip labeling examples by user and read training from
training_file only (default: False)
* `--training_file TRAINING_FILE`
Path to a new or existing file consisting of labeled
training examples (default: training.json)
* `--sample_size SAMPLE_SIZE`
Number of random sample pairs to train off of
(default: 150000)
* `--recall_weight RECALL_WEIGHT`
Threshold that will maximize a weighted average of our
precision and recall (default: 2)
* `-h`, `--help` show help message and exit


## csvlink usage

Take two CSV files and find matches between them

Provide an input file and field names
```bash
csvlink examples/restaurant-1.csv examples/restaurant-2.csv \
--field_names name address city cuisine \
--output_file output.csv
```

Line up different field names from each file
```bash
csvlink examples/restaurant-1.csv examples/restaurant-2.csv \
--field_names_1 name address city cuisine \
--field_names_2 restaurant street city type \
--output_file output.csv
```

Pipe the output to STDOUT
```bash
csvlink examples/restaurant-1.csv examples/restaurant-2.csv \
--field_names name address city cuisine \
> output.csv
```

__or__

Define everything in a config file
```bash
csvdedupe examples/restaurant-1.csv examples/restaurant-2.csv \
--config_file=config.json
```

### Example config file

```json
{
"field_names_1": ["name", "address", "city", "cuisine"],
"field_names_2": ["restaurant", "street", "city", "type"],
"field_definition" : {"name": {"type" : "String"},
"address": {"type" : "String"},
"city": {"type" : "String",
"Has Missing" : true},
"cuisine": {"type" : "String",
"Has Missing" : true}},
"output_file": "examples/output.csv",
"skip_training": false,
"training_file": "training.json",
"sample_size": 150000,
"recall_weight": 2
}
```

### Arguments:

#### Required

* `input` two CSV file names to join together

Either
* `--config_file` Path to configuration file.

Or
* `--field_names_1` List of column names in first file for dedupe to pay attention to
* `--field_names_2` List of column names in second file for dedupe to pay attention to

#### Optional
* `--output_file OUTPUT_FILE`
CSV file to store deduplication results (default:
None)
* `--inner_join` Only return matches between datasets
* `--skip_training` Skip labeling examples by user and read training from
training_file only (default: False)
* `--training_file TRAINING_FILE`
Path to a new or existing file consisting of labeled
training examples (default: training.json)
* `--sample_size SAMPLE_SIZE`
Number of random sample pairs to train off of
(default: 150000)
* `--recall_weight RECALL_WEIGHT`
Threshold that will maximize a weighted average of our
precision and recall (default: 2)
* `-h`, `--help` show help message and exit

## Training

The _secret sauce_ of csvdedupe is human input. In order to figure out the best rules to deduplicate a set of data, you must give it a set of labeled examples to learn from.

The more labeled examples you give it, the better the deduplication results will be. At minimum, you should try to provide __10 positive matches__ and __10 negative matches__.

The results of your training will be saved in a JSON file ( __training.json__, unless specified otherwise with the `--training-file` option) for future runs of csvdedupe.

Here's an example labeling operation:

```bash
Phone : 2850617
Address : 3801 s. wabash
Zip :
Site name : ada s. mckinley st. thomas cdc

Phone : 2850617
Address : 3801 s wabash ave
Zip :
Site name : ada s. mckinley community services - mckinley - st. thomas

Do these records refer to the same thing?
(y)es / (n)o / (u)nsure / (f)inished
```

## Output
`csvdedupe` attempts to identify all the rows in the csv that refer to the same thing. Each group of
such records are called a cluster. `csvdedupe` returns your input file with an additional column called `Cluster ID`,
that either is the numeric id (zero-indexed) of a cluster of grouped records or an `x` if csvdedupe believes
the record doesn't belong to any cluster.

`csvlink` operates in much the same way as `csvdedupe`, but will flatten both CSVs in to one
output file similar to a SQL [OUTER JOIN](http://stackoverflow.com/questions/38549/difference-between-inner-and-outer-join) statement. You can use the `--inner_join` flag to exclude rows that don't match across the two input files, much like an INNER JOIN.


## Preprocessing
csvdedupe attempts to convert all strings to ASCII, ignores case, new lines, and padding whitespace. This is all
probably uncontroversial except the conversion to ASCII. Basically, we had to choose between two ways of handling
extended characters.

```
distance("Tomas", "Tomás') = distance("Tomas", "Tomas")
```

__or__

```
distance("Tomas, "Tomás") = distance("Tomas", "Tomzs")
```

We chose the first option. While it is possible to do something more sophisticated, this option seems to work pretty well
for Latin alphabet languages.

## Testing

Unit tests of core csvdedupe functions
```bash
pip install -r requirements-test.txt
nosetests
```

## Community
* [Dedupe Google group](https://groups.google.com/forum/?fromgroups=#!forum/open-source-deduplication)
* IRC channel, #dedupe on irc.freenode.net

## Recipes

### Combining and deduplicating files from different sources.

Lets say we have a few sources of early childhood programs in Chicago and we'd like to get a canonical list.
Let's do it with `csvdedupe`, `csvkit`, and some other common command line tools.

#### Alignment and stacking
Our first task will be to align the files and have the same data in the same columns for stacking.

First let's look at the headers of the files

File 1
```console
> head -1 CPS_Early_Childhood_Portal_Scrape.csv
"Site name","Address","Phone","Program Name","Length of Day"
```

File 2
```console
> head -1 IDHS_child_care_provider_list.csv
"Site name","Address","Zip Code","Phone","Fax","IDHS Provider ID"
```

So, we'll have to add "Zip Code", "Fax", and "IDHS Provider ID"
to ```CPS_Early_Childhood_Portal_Scrape.csv```, and we'll have to add "Program Name",
"Length of Day" to ```IDHS_child_care_provider_list.csv```.

```console
> cd examples
> sed '1 s/$/,"Zip Code","Fax","IDHS Provider ID"/' CPS_Early_Childhood_Portal_Scrape.csv > input_1a.csv
> sed '2,$s/$/,,,/' input_1a.csv > input_1b.csv
```

```console
> sed '1 s/$/,"Program Name","Length of Day"/' IDHS_child_care_provider_list.csv > input_2a.csv
> sed '2,$s/$/,,/' input_2a.csv > input_2b.csv
```

Now, we reorder the columns in the second file to align to the first.

```console
> csvcut -c "Site name","Address","Phone","Program Name","Length of Day","Zip Code","Fax","IDHS Provider ID" \
input_2b.csv > input_2c.csv
```

And we are finally ready to stack.

```console
> csvstack -g CPS_Early_Childhood_Portal_Scrape.csv,IDHS_child_care_provider_list.csv \
-n source \
input_1b.csv input_2c.csv > input.csv
```

#### Dedupe it!
And now we can dedupe

```console
> cat input.csv | csvdedupe --field_names "Site name" Address "Zip Code" Phone > output.csv
```

Let's sort the output by duplicate IDs, and we are ready to open it in your favorite spreadsheet program.

```console
> csvsort -c "Cluster ID" output.csv > sorted.csv
```

[![githalytics.com alpha](https://cruel-carlota.pagodabox.com/88cda639ab635a100d23de5948ffbef5 "githalytics.com")](http://githalytics.com/datamade/csvdedupe)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

csvdedupe-0.1.1.tar.gz (14.1 kB view hashes)

Uploaded Source

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page