Hi Eric and Jeff,
what we are trying to do is to use the Linux Host Adapter to launch an app (Desktop in this case) on the LHA. As mentioned earlier, we have an OOD server which has a local home for the user, and Linux remote host that also has a local home. We have set up a shared NFS mount that is not home, but is R/W accessible on both the OOD server and the remote host, and use the OOD_DATAROOT to point to this shared file system to store the Desktop app job’s data files:
Once the job is launched, we see all the data files in the correct staging directory, but, the output.log just has:
/bin/bash: /nfs/ood_data/pierce/batch_connect/sys/bc_desktop/kc_host/output/2c74aad3-2375-4c04-915b-2567ec5755bf/tmp.nSrb8fAufD_sing: No such file or directory
Looking at what happens on our production system where the home dirs are NFS mounted and shared by the OOD server and the remote host, and no OOD_DATAROOT, the tmp.sing and tmp.tmux files are put to $HOME. I have a feeling that the tmux and singularity commands that start the session with those tmp files are putting the tmp files somewhere else, and they can’t be found by subsequent commands that need them. I.e., that there’s some inconsistency in using the OOD_DATAROOT in the LHA.
Can you please comment on this, and/or, point us to the source code of the LHA to understand better what’s the syntax and sequence of the tmux and singularity execution when starting the LHA job?