Synchronizing four shell scripts to run one after another in unix

前端 未结 3 653
慢半拍i
慢半拍i 2021-01-15 22:39

I have 4 shell script to generate a file(let\'s say param.txt) which is used by another tool(informatica) and as the tool is done with processing, it deletes param.txt.

相关标签:
3条回答
  • 2021-01-15 22:52

    You can use GNU Parallel as a counting semaphore or a mutex, by invoking it as sem instead of as parallel. Scroll down to Mutex on this page.

    So, you could use:

    sem --id myGlobalId 'create input file; run informatica'
    sem --id myGlobalId 'create input file; run informatica'
    sem --id myGlobalId 'create input file; run informatica'
    sem --id myGlobalId 'create input file; run informatica'
    

    Note I have specified a global id in case you run the jobs from different terminals or cron. This is not necessary if you are starting all jobs from one terminal.

    0 讨论(0)
  • Thanks for your valuable suggestions. It did help me to think from other dimension. However I missed to mention that I am using Solaris UNIX where I couldn't find equivalent of flock or similar function. I could have asked team to install one utility but in mean time I found a workaround for this issue.

    I read about the mkdir function being atomic in nature where as 'touch' command to create a file is not(still don't have complete explanation on how it works). That means at a time only 1 script can create/delete directory 'lockdir' out of 4 and other 3 has to wait.

    while true;
    do 
    if mkdir "$lockdir"; then
       < create param file >
    
    break;
    fi
        Sleep 30
    done
    
    0 讨论(0)
  • 2021-01-15 23:12

    You are experiencing a classical race condition. To solve this issue, you need a shared "lock" (or similar) between your 4 scripts.

    There are several ways to implement this. One way to do this in bash is by using the flock command, and an agreed-upon filename to use as a lock. The flock man page has some usage examples which resemble this:

    (
        flock -x 200  # try to acquire an exclusive lock on the file
        # do whatever check you want. You are guaranteed to be the only one
        # holding the lock
        if [ -f "$paramfile" ]; then
            # do something
        fi
    ) 200>/tmp/lock-life-for-all-scripts
    # The lock is automatically released when the above block is exited
    

    You can also ask flock to fail right away if the lock can't be acquired, or to fail after a timeout (e.g. to print "still trying to acquire the lock" and restart).

    Depending on your use case, you could also put the lock on the 'informatica' binary (be sure to use 200< in that case, to open the file for reading instead of (over)writing)

    0 讨论(0)
提交回复
热议问题