<> = Commands in SGE and equivalents in SLURM = ||<#ffffcc>'''User Command''' ||<#ffffcc>'''SGE''' ||<#ffffcc>'''SLURM''' || || Interactive login || qlogin || srun -p lipq -q inter --pty bash -i || || Job submission || qsub [script_file] || sbatch [script_file] || || Job status || qstat || squeue || || Job status by job id || qstat -j [job_id] || squeue -j [job_id] || || Job status by user || qstat -u [username]|| squeue -u [username] || || Job deletion || qdel [job_id] || scancel [job_id] || || Job hold || qhold [job_id] || scontrol hold [job_id] || || Job release || qrls [job_id] || scontrol release [job_id] || || Queue list || qconf -sql || sinfo || || Cluster status || qhost -q || sinfo || || Nodes list || qhost || sinfo -Nl || || || || scontrol show nodes || || GUI || qmon || sview || = Common environment = ||<#ffffcc>'''User Command''' ||<#ffffcc>'''SGE''' ||<#ffffcc>'''SLURM''' || || Job ID || $JOB_ID || $SLURM_JOBID || || Submit directory || $SGE_O_WORKDIR || $SLURM_SUBMIT_DIR || || Submit host || $SGE_O_HOST || $SLURM_SUBMIT_HOST || || Node list || $PE_HOSTFILE || $SLURM_JOB_NODELIST || || Job array index || $SGE_TASK_ID || $SLURM_ARRAY_TASK_ID || = Job directives = ||<#ffffcc>'''User Command''' ||<#ffffcc>'''SGE''' ||<#ffffcc>'''SLURM''' || || queue/partition || #$ -q [queue] || #SBATCH -p [partition] || || count of nodes || N/A || #SBATCH -N [min[-max]] || || CPU count || #$ -pe [PE] [count] || #SBATCH -n [count] || || Wall clock limit || #$ -l h_rt=[seconds]|| #SBATCH -t [min] || || || || #SBATCH -t [days-hh:mm:ss] || || Standard output file || #$ -o [file_name] || #SBATCH -o [file_name] || || Standard error file || #$ -e [file_name] || '''#SBATCH -e [file_name]*''' || || Combine stdout and stderr || #$ -j yes || (use -o withou -e) || || Copy environment || #$ -V || #SBATCH --export=[ALL/NONE/varnames]|| || Job stage in transfers || #$ -V SGEIN=name[:name]|| #IN=name[:name] || || || || # IN name[:name] || || Job stage out transfers || #$ -V SGEOUT=name[:name]|| #OUT=name[:name] || || || || # OUT name[:name] || || Job name || #$ -N [name] || #SBATCH --job-name=[name] || || Restart job || #$ -r [yes/no] || #SBATCH --requeue || || || || #SBATCH --no-requeue (default) || || Set working directory || #$ -wd [dir_name] || '''#SBATCH --workdir=[dir_name]**'''|| || Resource sharing || #$ -l exclusive || #SBATCH --exclusive || || || || #SBATCH --shared || || Memory size || #$ -l mem_free=[mem(KMG]|| #SBATCH #SBATCH --mem=[mem(KMG)]|| || || || #SBATCH --mem-per-cpu=[mem(KMG)] || || Tasks per node || (fixed in PE) || #SBATCH --tasks-per-node=[count] || || || || #SBATCH --cpus-per-task=[count] || ||'''*''' || On '''lipq''' partition, for the time being, the standard error is always merged with standard output || ||'''**'''|| On '''lipq''' partition the starting working subdirectory on workernodes is always a volatil uniq directory on local home ||