如何从SLURM获取详细的作业运行信息(例如,为#34生成的标准输出"由LSF生成)?

时间:2015-04-28 20:08:27

标签: slurm lsf

bsub与LSF一起使用时,-o选项提供了许多详细信息,例如作业开始和结束的时间以及作业占用的内存和CPU时间。使用SLURM,我获得的是与运行没有LSF的脚本相同的标准输出。

例如,给出这个Perl 6脚本:

warn  "standard error stream";
say  "standard output stream";

这样提交:

sbatch -o test.o%j -e test.e%j -J test_warn --wrap 'perl6 test.p6'

导致文件test.o34380

Testing standard output

和文件test.e34380

Testing standard Error  in block <unit> at test.p6:2


使用LSF,我可以在标准输出文件中获得各种细节,例如:

Sender: LSF System <lsfadmin@my_node>
Subject: Job 347511: <test> Done

Job <test> was submitted from host <my_cluster> by user <username> in cluster <my_cluster_act>.
Job was executed on host(s) <my_node>, in queue <normal>, as user <username> in cluster <my_cluster_act>.
</home/username> was used as the home directory.
</path/to/working/directory> was used as the working directory.
Started at Mon Mar 16 13:10:23 2015
Results reported at Mon Mar 16 13:10:29 2015

Your job looked like:

------------------------------------------------------------
# LSBATCH: User input
perl6 test.p6

------------------------------------------------------------

Successfully completed.

Resource usage summary:

    CPU time   :    0.19 sec.
    Max Memory :    0.10 MB
    Max Swap   :    0.10 MB

    Max Processes  :         2
    Max Threads    :         3

The output (if any) follows:

standard output stream

PS:

Read file <test.e_347511> for stderr output of this job.

更新

-v的一个或多个sbatch标记提供了更多初步信息,但不会更改标准输出。

1 个答案:

答案 0 :(得分:3)

在我用来插入

的每项工作结束时

sstat -j $SLURM_JOB_ID.batch --format=JobID,MaxVMSize

将RAM使用量添加到标准输出。