Debugging Open OnDemand
When you run into issues with jobs in Open OnDemand, review the output files from your session for important clues on what went wrong.
Review Session Output Files
Go to My Interactive Sessions.
Find the interactive session of interest from the list and click the Session ID link.
Browse the list of output files for clues about what went wrong. The
user_defined_context.json
andoutput.log
files are highlighted below.
user_defined_context.json
The user_defined_context.json
file shows the options you input to launch the session.
Did you request the correct amount of RAM in the correct format? Use Slurm format (for example, 4096M or 10G). If this field is left blank, 1000 MB will be allocated per CPU core requested.
The following user_defined_context.json
example is for a 30 minute, 1 CPU, JupyterLab job request with default memory.
bc_account "bbka-delta-cpu"
bc_partition "cpu-interactive"
bc_duration "00-00:30:00"
bc_reservation ""
bc_num_slots "1"
bc_num_memory ""
bc_num_gpus "0"
bc_email_on_started "0"
working_dir ""
output.log
The output.log
file shows what happened when the script.sh
file was run. The last line of output.log
should tell you why the job ended.
The following example is the last line of the output.log
file of a job that ended because it reached its requested duration.
slurmstepd: error: *** JOB 4214572 ON cn001 CANCELLED AT 2024-07-24T14:06:53 DUE TO TIME LIMIT ***