<div dir="ltr">Dear Ole,<div><br></div><div>Thank you for your response.</div><div>I am doing it again using your suggested link.</div><div><br clear="all"><div><div dir="ltr" class="gmail_signature" data-smartmail="gmail_signature"><div dir="ltr"><div dir="ltr"><div dir="ltr"><div dir="ltr">Best Regards,</div><div dir="ltr"><span style="font-family:arial;font-size:small">Nousheen Parvaiz</span><br style="font-family:arial;font-size:small"><br></div></div></div></div></div></div><br></div></div><div hspace="streak-pt-mark" style="max-height:1px"><img alt="" style="width:0px;max-height:0px;overflow:hidden" src="https://mailfoogae.appspot.com/t?sender=abm91c2hlZW5wYXJ2YWl6QGdtYWlsLmNvbQ%3D%3D&type=zerocontent&guid=9a84a710-e6c1-4912-a461-6103eb630f96"><font color="#ffffff" size="1">ᐧ</font></div><br><div class="gmail_quote"><div dir="ltr" class="gmail_attr">On Mon, Jan 31, 2022 at 2:07 PM Ole Holm Nielsen <<a href="mailto:Ole.H.Nielsen@fysik.dtu.dk">Ole.H.Nielsen@fysik.dtu.dk</a>> wrote:<br></div><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex">Hi Nousheen,<br>
<br>
I recommend you again to follow the steps for installing Slurm on a CentOS <br>
7 cluster:<br>
<a href="https://wiki.fysik.dtu.dk/niflheim/Slurm_installation" rel="noreferrer" target="_blank">https://wiki.fysik.dtu.dk/niflheim/Slurm_installation</a><br>
<br>
Maybe you will need to start installation from scratch, but the steps are <br>
guaranteed to work if followed correctly.<br>
<br>
IHTH,<br>
Ole<br>
<br>
On 1/31/22 06:23, Nousheen wrote:<br>
> The same error shows up on compute node which is as follows:<br>
> <br>
> [root@c103008 ~]# systemctl enable slurmd.service<br>
> [root@c103008 ~]# systemctl start slurmd.service<br>
> [root@c103008 ~]# systemctl status slurmd.service<br>
> ● slurmd.service - Slurm node daemon<br>
>     Loaded: loaded (/etc/systemd/system/slurmd.service; enabled; vendor <br>
> preset: disabled)<br>
>     Active: failed (Result: exit-code) since Mon 2022-01-31 00:22:42 EST; <br>
> 2s ago<br>
>    Process: 11505 ExecStart=/usr/local/sbin/slurmd -D -s $SLURMD_OPTIONS <br>
> (code=exited, status=203/EXEC)<br>
>   Main PID: 11505 (code=exited, status=203/EXEC)<br>
> <br>
> Jan 31 00:22:42 c103008 systemd[1]: Started Slurm node daemon.<br>
> Jan 31 00:22:42 c103008 systemd[1]: slurmd.service: main process exited, <br>
> code=exited, status=203/EXEC<br>
> Jan 31 00:22:42 c103008 systemd[1]: Unit slurmd.service entered failed state.<br>
> Jan 31 00:22:42 c103008 systemd[1]: slurmd.service failed.<br>
> <br>
> <br>
> Best Regards,<br>
> Nousheen Parvaiz<br>
> <br>
> <br>
> ᐧ<br>
> <br>
> On Mon, Jan 31, 2022 at 10:08 AM Nousheen <<a href="mailto:nousheenparvaiz@gmail.com" target="_blank">nousheenparvaiz@gmail.com</a> <br>
> <mailto:<a href="mailto:nousheenparvaiz@gmail.com" target="_blank">nousheenparvaiz@gmail.com</a>>> wrote:<br>
> <br>
>     Dear Jeffrey,<br>
> <br>
>     Thank you for your response. I have followed the steps as instructed.<br>
>     After the copying the files to their respective locations "systemctl<br>
>     status slurmctld.service" command gives me an error as follows:<br>
> <br>
>     (base) [nousheen@exxact system]$ systemctl daemon-reload<br>
>     (base) [nousheen@exxact system]$ systemctl enable slurmctld.service<br>
>     (base) [nousheen@exxact system]$ systemctl start slurmctld.service<br>
>     (base) [nousheen@exxact system]$ systemctl status slurmctld.service<br>
>     ● slurmctld.service - Slurm controller daemon<br>
>         Loaded: loaded (/etc/systemd/system/slurmctld.service; enabled;<br>
>     vendor preset: disabled)<br>
>         Active: failed (Result: exit-code) since Mon 2022-01-31 10:04:31<br>
>     PKT; 3s ago<br>
>        Process: 18114 ExecStart=/usr/local/sbin/slurmctld -D -s<br>
>     $SLURMCTLD_OPTIONS (code=exited, status=1/FAILURE)<br>
>       Main PID: 18114 (code=exited, status=1/FAILURE)<br>
> <br>
>     Jan 31 10:04:31 exxact systemd[1]: Started Slurm controller daemon.<br>
>     Jan 31 10:04:31 exxact systemd[1]: slurmctld.service: main process<br>
>     exited, code=exited, status=1/FAILURE<br>
>     Jan 31 10:04:31 exxact systemd[1]: Unit slurmctld.service entered<br>
>     failed state.<br>
>     Jan 31 10:04:31 exxact systemd[1]: slurmctld.service failed.<br>
> <br>
>     Kindly guide me. Thank you so much for your time.<br>
> <br>
>     Best Regards,<br>
>     Nousheen Parvaiz<br>
> <br>
>     ᐧ<br>
> <br>
>     On Thu, Jan 27, 2022 at 8:25 PM Jeffrey R. Lang <<a href="mailto:JRLang@uwyo.edu" target="_blank">JRLang@uwyo.edu</a><br>
>     <mailto:<a href="mailto:JRLang@uwyo.edu" target="_blank">JRLang@uwyo.edu</a>>> wrote:<br>
> <br>
>         The missing file error has nothing to do with slurm.  The<br>
>         systemctl command is part of the systems service management.____<br>
> <br>
>         __ __<br>
> <br>
>         The error message indicates that you haven’t copied the<br>
>         slurmd.service file on your compute node to /etc/systemd/system or<br>
>         /usr/lib/systemd/system.  /etc/systemd/system is usually used when<br>
>         a user adds a new service to a machine.____<br>
> <br>
>         __ __<br>
> <br>
>         Depending on your version of Linux you may also need to do a<br>
>         systemctl daemon-reload to activate the slurmd.service within<br>
>         system.____<br>
> <br>
>         __ __<br>
> <br>
>         Once slurmd.service is copied over, the systemctld command should<br>
>         work just fine.____<br>
> <br>
>         __ __<br>
> <br>
>         Remember:____<br>
> <br>
>                          slurmd.service     -  Only on compute nodes____<br>
> <br>
>                          slurmctld.service – Only on your cluster<br>
>         management node____<br>
> <br>
>                        slurmdbd.service – Only on your cluster management<br>
>         node____<br>
> <br>
>         __ __<br>
> <br>
>         *From:* slurm-users <<a href="mailto:slurm-users-bounces@lists.schedmd.com" target="_blank">slurm-users-bounces@lists.schedmd.com</a><br>
>         <mailto:<a href="mailto:slurm-users-bounces@lists.schedmd.com" target="_blank">slurm-users-bounces@lists.schedmd.com</a>>> *On Behalf Of<br>
>         *Nousheen<br>
>         *Sent:* Thursday, January 27, 2022 3:54 AM<br>
>         *To:* Slurm User Community List <<a href="mailto:slurm-users@lists.schedmd.com" target="_blank">slurm-users@lists.schedmd.com</a><br>
>         <mailto:<a href="mailto:slurm-users@lists.schedmd.com" target="_blank">slurm-users@lists.schedmd.com</a>>><br>
>         *Subject:* [slurm-users] systemctl enable slurmd.service Failed to<br>
>         execute operation: No such file or directory____<br>
> <br>
>         __ __<br>
> <br>
>         ◆ This message was sent from a non-UWYO address. Please exercise<br>
>         caution when clicking links or opening attachments from external<br>
>         sources.____<br>
> <br>
>         __ __<br>
> <br>
>         __ __<br>
> <br>
>         Hello everyone,____<br>
> <br>
>         __ __<br>
> <br>
>         I am installing slurm on Centos 7 following tutorial:<br>
>         <a href="https://www.slothparadise.com/how-to-install-slurm-on-centos-7-cluster/" rel="noreferrer" target="_blank">https://www.slothparadise.com/how-to-install-slurm-on-centos-7-cluster/</a><br>
>         <<a href="https://www.slothparadise.com/how-to-install-slurm-on-centos-7-cluster/" rel="noreferrer" target="_blank">https://www.slothparadise.com/how-to-install-slurm-on-centos-7-cluster/</a>>____<br>
> <br>
>         __ __<br>
> <br>
>         I am at the step where we start slurm but it gives me the<br>
>         following error:____<br>
> <br>
>         __ __<br>
> <br>
>         [root@exxact slurm-21.08.5]# systemctl enable slurmd.service____<br>
> <br>
>         Failed to execute operation: No such file or directory____<br>
> <br>
>         __ __<br>
> <br>
>         I have run the command to check if slurm is configured properly____<br>
> <br>
>         __ __<br>
> <br>
>         [root@exxact slurm-21.08.5]# slurmd -C<br>
>         NodeName=exxact CPUs=12 Boards=1 SocketsPerBoard=1<br>
>         CoresPerSocket=6 ThreadsPerCore=2 RealMemory=31889<br>
>         UpTime=19-16:06:00____<br>
> <br>
>         __ __<br>
> <br>
>         I am new to this and unable to understand the problem. Kindly help<br>
>         me resolve this.____<br>
> <br>
>         __ __<br>
> <br>
>         My slurm.conf file is as follows:____<br>
> <br>
>         __ __<br>
> <br>
>         # slurm.conf file generated by configurator.html.<br>
>         # Put this file on all nodes of your cluster.<br>
>         # See the slurm.conf man page for more information.<br>
>         #<br>
>         ClusterName=cluster194<br>
>         SlurmctldHost=192.168.60.194<br>
>         #SlurmctldHost=<br>
>         #<br>
>         #DisableRootJobs=NO<br>
>         #EnforcePartLimits=NO<br>
>         #Epilog=<br>
>         #EpilogSlurmctld=<br>
>         #FirstJobId=1<br>
>         #MaxJobId=67043328<br>
>         #GresTypes=<br>
>         #GroupUpdateForce=0<br>
>         #GroupUpdateTime=600<br>
>         #JobFileAppend=0<br>
>         #JobRequeue=1<br>
>         #JobSubmitPlugins=lua<br>
>         #KillOnBadExit=0<br>
>         #LaunchType=launch/slurm<br>
>         #Licenses=foo*4,bar<br>
>         #MailProg=/bin/mail<br>
>         #MaxJobCount=10000<br>
>         #MaxStepCount=40000<br>
>         #MaxTasksPerNode=512<br>
>         MpiDefault=none<br>
>         #MpiParams=ports=#-#<br>
>         #PluginDir=<br>
>         #PlugStackConfig=<br>
>         #PrivateData=jobs<br>
>         ProctrackType=proctrack/cgroup<br>
>         #Prolog=<br>
>         #PrologFlags=<br>
>         #PrologSlurmctld=<br>
>         #PropagatePrioProcess=0<br>
>         #PropagateResourceLimits=<br>
>         #PropagateResourceLimitsExcept=<br>
>         #RebootProgram=<br>
>         ReturnToService=1<br>
>         SlurmctldPidFile=/var/run/slurmctld.pid<br>
>         SlurmctldPort=6817<br>
>         SlurmdPidFile=/var/run/slurmd.pid<br>
>         SlurmdPort=6818<br>
>         SlurmdSpoolDir=/var/spool/slurmd<br>
>         SlurmUser=nousheen<br>
>         #SlurmdUser=root<br>
>         #SrunEpilog=<br>
>         #SrunProlog=<br>
>         StateSaveLocation=/home/nousheen/Documents/SILICS/slurm-21.08.5/slurmctld<br>
>         SwitchType=switch/none<br>
>         #TaskEpilog=<br>
>         TaskPlugin=task/affinity<br>
>         #TaskProlog=<br>
>         #TopologyPlugin=topology/tree<br>
>         #TmpFS=/tmp<br>
>         #TrackWCKey=no<br>
>         #TreeWidth=<br>
>         #UnkillableStepProgram=<br>
>         #UsePAM=0<br>
>         #<br>
>         #<br>
>         # TIMERS<br>
>         #BatchStartTimeout=10<br>
>         #CompleteWait=0<br>
>         #EpilogMsgTime=2000<br>
>         #GetEnvTimeout=2<br>
>         #HealthCheckInterval=0<br>
>         #HealthCheckProgram=<br>
>         InactiveLimit=0<br>
>         KillWait=30<br>
>         #MessageTimeout=10<br>
>         #ResvOverRun=0<br>
>         MinJobAge=300<br>
>         #OverTimeLimit=0<br>
>         SlurmctldTimeout=120<br>
>         SlurmdTimeout=300<br>
>         #UnkillableStepTimeout=60<br>
>         #VSizeFactor=0<br>
>         Waittime=0<br>
>         #<br>
>         #<br>
>         # SCHEDULING<br>
>         #DefMemPerCPU=0<br>
>         #MaxMemPerCPU=0<br>
>         #SchedulerTimeSlice=30<br>
>         SchedulerType=sched/backfill<br>
>         SelectType=select/cons_tres<br>
>         SelectTypeParameters=CR_Core<br>
>         #<br>
>         #<br>
>         # JOB PRIORITY<br>
>         #PriorityFlags=<br>
>         #PriorityType=priority/basic<br>
>         #PriorityDecayHalfLife=<br>
>         #PriorityCalcPeriod=<br>
>         #PriorityFavorSmall=<br>
>         #PriorityMaxAge=<br>
>         #PriorityUsageResetPeriod=<br>
>         #PriorityWeightAge=<br>
>         #PriorityWeightFairshare=<br>
>         #PriorityWeightJobSize=<br>
>         #PriorityWeightPartition=<br>
>         #PriorityWeightQOS=<br>
>         #<br>
>         #<br>
>         # LOGGING AND ACCOUNTING<br>
>         #AccountingStorageEnforce=0<br>
>         #AccountingStorageHost=<br>
>         #AccountingStoragePass=<br>
>         #AccountingStoragePort=<br>
>         AccountingStorageType=accounting_storage/none<br>
>         #AccountingStorageUser=<br>
>         #AccountingStoreFlags=<br>
>         #JobCompHost=<br>
>         #JobCompLoc=<br>
>         #JobCompPass=<br>
>         #JobCompPort=<br>
>         JobCompType=jobcomp/none<br>
>         #JobCompUser=<br>
>         #JobContainerType=job_container/none<br>
>         JobAcctGatherFrequency=30<br>
>         JobAcctGatherType=jobacct_gather/none<br>
>         SlurmctldDebug=info<br>
>         SlurmctldLogFile=/var/log/slurmctld.log<br>
>         SlurmdDebug=info<br>
>         SlurmdLogFile=/var/log/slurmd.log<br>
>         #SlurmSchedLogFile=<br>
>         #SlurmSchedLogLevel=<br>
>         #DebugFlags=<br>
>         #<br>
>         #<br>
>         # POWER SAVE SUPPORT FOR IDLE NODES (optional)<br>
>         #SuspendProgram=<br>
>         #ResumeProgram=<br>
>         #SuspendTimeout=<br>
>         #ResumeTimeout=<br>
>         #ResumeRate=<br>
>         #SuspendExcNodes=<br>
>         #SuspendExcParts=<br>
>         #SuspendRate=<br>
>         #SuspendTime=<br>
>         #<br>
>         #<br>
>         # COMPUTE NODES<br>
>         NodeName=linux[1-32] CPUs=11 State=UNKNOWN____<br>
> <br>
>         PartitionName=debug Nodes=ALL Default=YES MaxTime=INFINITE<br>
>         State=UP ____<br>
> <br>
>         __ __<br>
> <br>
> <br>
>         ____<br>
> <br>
>         Best Regards,____<br>
> <br>
>         Nousheen Parvaiz____<br>
> <br>
>         ᐧ____<br>
<br>
</blockquote></div>