Bash tool to get nth line from a file

前端 未结 19 2102
刺人心
刺人心 2020-11-22 08:07

Is there a \"canonical\" way of doing that? I\'ve been using head -n | tail -1 which does the trick, but I\'ve been wondering if there\'s a Bash tool that speci

19条回答
  •  花落未央
    2020-11-22 09:06

    I have a unique situation where I can benchmark the solutions proposed on this page, and so I'm writing this answer as a consolidation of the proposed solutions with included run times for each.

    Set Up

    I have a 3.261 gigabyte ASCII text data file with one key-value pair per row. The file contains 3,339,550,320 rows in total and defies opening in any editor I have tried, including my go-to Vim. I need to subset this file in order to investigate some of the values that I've discovered only start around row ~500,000,000.

    Because the file has so many rows:

    • I need to extract only a subset of the rows to do anything useful with the data.
    • Reading through every row leading up to the values I care about is going to take a long time.
    • If the solution reads past the rows I care about and continues reading the rest of the file it will waste time reading almost 3 billion irrelevant rows and take 6x longer than necessary.

    My best-case-scenario is a solution that extracts only a single line from the file without reading any of the other rows in the file, but I can't think of how I would accomplish this in Bash.

    For the purposes of my sanity I'm not going to be trying to read the full 500,000,000 lines I'd need for my own problem. Instead I'll be trying to extract row 50,000,000 out of 3,339,550,320 (which means reading the full file will take 60x longer than necessary).

    I will be using the time built-in to benchmark each command.

    Baseline

    First let's see how the head tail solution:

    $ time head -50000000 myfile.ascii | tail -1
    pgm_icnt = 0
    
    real    1m15.321s
    

    The baseline for row 50 million is 00:01:15.321, if I'd gone straight for row 500 million it'd probably be ~12.5 minutes.

    cut

    I'm dubious of this one, but it's worth a shot:

    $ time cut -f50000000 -d$'\n' myfile.ascii
    pgm_icnt = 0
    
    real    5m12.156s
    

    This one took 00:05:12.156 to run, which is much slower than the baseline! I'm not sure whether it read through the entire file or just up to line 50 million before stopping, but regardless this doesn't seem like a viable solution to the problem.

    AWK

    I only ran the solution with the exit because I wasn't going to wait for the full file to run:

    $ time awk 'NR == 50000000 {print; exit}' myfile.ascii
    pgm_icnt = 0
    
    real    1m16.583s
    

    This code ran in 00:01:16.583, which is only ~1 second slower, but still not an improvement on the baseline. At this rate if the exit command had been excluded it would have probably taken around ~76 minutes to read the entire file!

    Perl

    I ran the existing Perl solution as well:

    $ time perl -wnl -e '$.== 50000000 && print && exit;' myfile.ascii
    pgm_icnt = 0
    
    real    1m13.146s
    

    This code ran in 00:01:13.146, which is ~2 seconds faster than the baseline. If I'd run it on the full 500,000,000 it would probably take ~12 minutes.

    sed

    The top answer on the board, here's my result:

    $ time sed "50000000q;d" myfile.ascii
    pgm_icnt = 0
    
    real    1m12.705s
    

    This code ran in 00:01:12.705, which is 3 seconds faster than the baseline, and ~0.4 seconds faster than Perl. If I'd run it on the full 500,000,000 rows it would have probably taken ~12 minutes.

    mapfile

    I have bash 3.1 and therefore cannot test the mapfile solution.

    Conclusion

    It looks like, for the most part, it's difficult to improve upon the head tail solution. At best the sed solution provides a ~3% increase in efficiency.

    (percentages calculated with the formula % = (runtime/baseline - 1) * 100)

    Row 50,000,000

    1. 00:01:12.705 (-00:00:02.616 = -3.47%) sed
    2. 00:01:13.146 (-00:00:02.175 = -2.89%) perl
    3. 00:01:15.321 (+00:00:00.000 = +0.00%) head|tail
    4. 00:01:16.583 (+00:00:01.262 = +1.68%) awk
    5. 00:05:12.156 (+00:03:56.835 = +314.43%) cut

    Row 500,000,000

    1. 00:12:07.050 (-00:00:26.160) sed
    2. 00:12:11.460 (-00:00:21.750) perl
    3. 00:12:33.210 (+00:00:00.000) head|tail
    4. 00:12:45.830 (+00:00:12.620) awk
    5. 00:52:01.560 (+00:40:31.650) cut

    Row 3,338,559,320

    1. 01:20:54.599 (-00:03:05.327) sed
    2. 01:21:24.045 (-00:02:25.227) perl
    3. 01:23:49.273 (+00:00:00.000) head|tail
    4. 01:25:13.548 (+00:02:35.735) awk
    5. 05:47:23.026 (+04:24:26.246) cut

提交回复
热议问题