It’s fairly common to get a list of URLs or a csv dump from an SEO tool or even parsed server logs. When trying to analyze the data, the first reaction is typically to load the data into Excel or Google Docs where the columns can be sorted, deduped, or filtered to narrow the data set to something more manageable. However, a program like Excel can sometimes take several steps to import and selectively filter the data how you might want it. On top of it all, when dealing with large data sets, Excel can sometimes be very slow to simply open the file, let alone filter or sort columns & rows.
Using the command line, we can quickly whittle down larger data sets into something much more manageable before loading it into a spreadsheet program.