You can do this in multiple ways.

in Shell script

sort <filename> | uniq -c

If you have to use Python, then the faster way to do it would be to load
the csv into a Pandas dataframe, which should allow you to use

If you don't want to use pandas then you can loop through the csv and
create a set or hashmap with the row as the key and count as the value

Hope this helps.

On Sun, Feb 11, 2018 at 9:07 AM, Saravanan Muthu <>

> Hello All,
>      I have a csv with multiple column , and I need to figure out the
> duplicates entry ,I have imported csv and assigned the row to dictionary
> ,please share a logic to find the duplicates , sample data is below ,
> Name age employer
> Kumar 28 133678
> Kumar 28 133678
> Anil. 42.   133567
> Kumar entry need to be finded out
> _______________________________________________
> Chennaipy mailing list
Chennaipy mailing list

Reply via email to