Skip to content

Status of queue jobs in HPC #1854

@rajorshichat

Description

@rajorshichat

Hello,

I am trying to install pyiron in a HPC system to submit jobs from my local computer. I am following the instructions in the documentation and am trying to test the installation in the remote HPC. While jobs with run_mode=modal runs flawlessly, I see a problem in the status of jobs with run_mode=queue.

Once the jobs are submitted, the status in the job table shows submitted. This status remains the same even after the job is finished and all the required output files are generated in the file system. However, the status under job['status'] shows initialized. Also the output group is not available in the HDF5 job file although the output files are generated. I run the following lines of code -

pr = Project("test_lammps")
lmp = pr.create.job.Lammps(job_name='my_job_6')
structure = pr.create.structure.bulk('Al', cubic=True).repeat(3)
lmp.structure = structure
pot = lmp.list_potentials()[0]
lmp.potential = pot
lmp.calc_md()
lmp.run(run_mode="queue", delete_existing_job=True)
lmp
{'groups': ['executable', 'input'], 'nodes': ['DICT_VERSION', 'HDF_VERSION', 'NAME', 'OBJECT', 'TYPE', 'VERSION', 'job_id', 'server', 'status']}
lmp=pr.load('my_job_6')
lmp['status']
'initialized'

Inspection of the file system reveals that unlike modal jobs, tar files are not being generated for queue jobs. Am I getting something wrong here ? I thank you for your suggestions.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions