[slurm-users] slurm-17.11.3-2 - Redhat Linux 7.2 - Not able to start slurmd.service

Yogesh Aggarwal Yogesh.Aggarwal at alkermes.com
Wed Feb 28 13:33:18 MST 2018


Thanks Geert for your response. I was able to start slurmd by changing ProctrackType setting to proctrack/linuxproc.


Thanks,
Yogesh Aggarwal

-----Original Message-----
From: slurm-users [mailto:slurm-users-bounces at lists.schedmd.com] On Behalf Of Geert Geurts
Sent: Wednesday, February 28, 2018 3:29 PM
To: Slurm User Community List
Subject: EXT: Re: [slurm-users] slurm-17.11.3-2 - Redhat Linux 7.2 - Not able to start slurmd.service

[EXTERNAL]

I think you'll be fine with  uncomenting cgroupaautomount=yes in cgoup.conf file.
Could you try it like that?

Regards,
Geert


________________________________
From: Yogesh Aggarwal <Yogesh.Aggarwal at alkermes.com>
Sent: Wednesday, February 28, 2018 20:09
To: slurm-users at lists.schedmd.com
Subject: [slurm-users] slurm-17.11.3-2 - Redhat Linux 7.2 - Not able to start slurmd.service

Hi All,

I am trying to install slurm-17.11.3-2 on Redhat Linux 7.2 system. I have completed the installation and configuration but not able to start slurmd.service. Below are the logs from /var/log/slurmd.log. I have also pasted slurm.conf, cgroup.conf and cgroup_allowed_devices_file.conf for quick reference. I am able to successfully start slurmctld service. Can anyone please guide me on finding the particular settings which is causing this issue?

[2018-02-28T12:55:34.973] Message aggregation disabled [2018-02-28T12:55:34.974] error: cgroup namespace 'freezer' not mounted. aborting [2018-02-28T12:55:34.974] error: unable to create freezer cgroup namespace [2018-02-28T12:55:34.974] error: Couldn't load specified plugin name for proctrack/cgroup: Plugin init() callback failed [2018-02-28T12:55:34.974] error: cannot create proctrack context for proctrack/cgroup [2018-02-28T12:55:34.974] error: slurmd initialization failed


[root at montecarlo01 etc]# cat /usr/local/etc/slurm.conf | grep -v ^#
ControlMachine=montecarlo01
ControlAddr=<IP Address removed from email.> GresTypes=gpu MpiDefault=none ProctrackType=proctrack/cgroup
ReturnToService=1
SlurmctldPidFile=/var/run/slurmctld.pid
SlurmdPidFile=/var/run/slurmd.pid
SlurmdSpoolDir=/var/spool/slurmd
SlurmUser=slurm
StateSaveLocation=/var/spool/slurmctld
SwitchType=switch/none
TaskPlugin=task/cgroup
FastSchedule=1
SchedulerType=sched/backfill
SelectType=select/linear
AccountingStorageType=accounting_storage/none
ClusterName=montecarlo01
JobAcctGatherType=jobacct_gather/linux
SlurmctldLogFile=/var/log/slurmctld.log
SlurmdLogFile=/var/log/slurmd.log
NodeName=montecarlo01 NodeAddr=<IP Address removed from email.>  Sockets=1 CoresPerSocket=8 ThreadsPerCore=2 Gres=gpu:8 State=UNKNOWN PartitionName=AMBER_GPU Nodes=montecarlo01 Default=YES MaxTime=INFINITE State=UP
[root at montecarlo01 etc]#


[root at montecarlo01 slurm]# cat /usr/local/etc/slurm/cgroup.conf #CgroupMountpoint="/sys/fs/cgroup"
#CgroupAutomount=yes
#iCgroupReleaseAgentDir="/usr/local/etc/slurm/cgroup"
#AllowedDevicesFile="/usr/local/etc/slurm/cgroup_allowed_devices_file.conf"
ConstrainCores=no
TaskAffinity=no
ConstrainRAMSpace=no
ConstrainSwapSpace=no
ConstrainDevices=no
AllowedRamSpace=no
AllowedSwapSpace=no
MaxRAMPercent=100
MaxSwapPercent=100
MinRAMSpace=30
[root at montecarlo01 slurm]#

[root at montecarlo01 slurm]# cat /usr/local/etc/slurm/cgroup_allowed_devices_file.conf
/dev/null
/dev/urandom
/dev/zero
/dev/sda*
/dev/cpu/*/*
/dev/pts/*
[root at montecarlo01 slurm]#



Thanks,
Yogesh Aggarwal

This message may contain confidential or privileged information. If you are not the intended recipient, please advise the sender immediately by reply e-mail and delete this message and any attachments without retaining a copy. Thank you for your cooperation.

This message may contain confidential or privileged information. If you are not the intended recipient, please advise the sender immediately by reply e-mail and delete this message and any attachments without retaining a copy. Thank you for your cooperation.



More information about the slurm-users mailing list