[slurm-users] CPUSpecList confusion

Paul Raines raines at nmr.mgh.harvard.edu
Thu Dec 15 18:24:37 UTC 2022


Turns out on that new node I was running hwloc in a cgroup restricted
to cores 0-13 so that was causing the issue.

In an unrestricted cgroup shell, "hwloc-ls --only pu" works properly
and gives me the correct SLURM mapping.

-- Paul Raines (http://help.nmr.mgh.harvard.edu)



On Thu, 15 Dec 2022 10:00am, Paul Raines wrote:

>
> Nice find!
>
> Unfortunately this does not work on the original box this whole
> issue started on where I found the "alternating scheme"
>
> # slurmd -C
> NodeName=foobar CPUs=64 Boards=1 SocketsPerBoard=2 CoresPerSocket=16 
> ThreadsPerCore=2 RealMemory=256312
> UpTime=5-14:55:31
>
> # hwloc-ls --only pu
> PU L#0 (P#0)
> PU L#1 (P#1)
> PU L#2 (P#2)
> PU L#3 (P#3)
> PU L#4 (P#4)
> PU L#5 (P#5)
> PU L#6 (P#6)
> PU L#7 (P#7)
> PU L#8 (P#8)
> PU L#9 (P#9)
> PU L#10 (P#10)
> PU L#11 (P#11)
> PU L#12 (P#12)
> PU L#13 (P#13)
> # grep ^proc /proc/cpuinfo  | wc
>      64     192     950
>
> Which is really strange since it is listing only 14 lines on a machine
> with 16 cores in each of 2 sockets and hyperthreading on.
>
> It is one of my newest boxes (Xeon Gold 6226R) so maybe hwloc just doesn't 
> support it
>
> I have hwloc-2.2.0 on Rocky 8 but just built the latest from git
> and it fails to.  Going to submit a bug in the hwloc git.
>
> This command does work on all my other boxes so I do think using hwloc-ls is 
> the "best" answer for getting the mapping on most hardware out there.
>
> -- Paul Raines (http://help.nmr.mgh.harvard.edu)
>
>
>
> On Thu, 15 Dec 2022 1:24am, Marcus Wagner wrote:
>
>>  Hi Paul,
>>
>>  as Slurm uses hwloc, I was looking into these tools a little bit deeper.
>>  Using your script, I saw e.g. the following output on one node:
>>
>>  === 31495434
>>  CPU_IDs=21-23,25
>>  21-23,25
>>  === 31495433
>>  CPU_IDs=16-18,20
>>  10-11,15,17
>>  === 31487399
>>  CPU_IDs=15
>>  9
>>
>>  That does not match your schemes and on first sight seems to be more
>>  random.
>>
>>  It seems, Slurm uses hwlocs logical indices, whereas cgroups uses the
>>  OS/physical indices.
>>  According to the example above (excerpt of the full output of hwloc-ls)
>>
>>        NUMANode L#1 (P#1 47GB)
>>        L2 L#12 (1024KB) + L1d L#12 (32KB) + L1i L#12 (32KB) + Core L#12 +
>>        PU
>>        L#12 (P#3)
>>        L2 L#13 (1024KB) + L1d L#13 (32KB) + L1i L#13 (32KB) + Core L#13 +
>>        PU
>>        L#13 (P#4)
>>        L2 L#14 (1024KB) + L1d L#14 (32KB) + L1i L#14 (32KB) + Core L#14 +
>>        PU
>>        L#14 (P#5)
>>        L2 L#15 (1024KB) + L1d L#15 (32KB) + L1i L#15 (32KB) + Core L#15 +
>>        PU
>>        L#15 (P#9)
>>        L2 L#16 (1024KB) + L1d L#16 (32KB) + L1i L#16 (32KB) + Core L#16 +
>>        PU
>>        L#16 (P#10)
>>        L2 L#17 (1024KB) + L1d L#17 (32KB) + L1i L#17 (32KB) + Core L#17 +
>>        PU
>>        L#17 (P#11)
>>        L2 L#18 (1024KB) + L1d L#18 (32KB) + L1i L#18 (32KB) + Core L#18 +
>>        PU
>>        L#18 (P#15)
>>        L2 L#19 (1024KB) + L1d L#19 (32KB) + L1i L#19 (32KB) + Core L#19 +
>>        PU
>>        L#19 (P#16)
>>        L2 L#20 (1024KB) + L1d L#20 (32KB) + L1i L#20 (32KB) + Core L#20 +
>>        PU
>>        L#20 (P#17)
>>        L2 L#21 (1024KB) + L1d L#21 (32KB) + L1i L#21 (32KB) + Core L#21 +
>>        PU
>>        L#21 (P#21)
>>        L2 L#22 (1024KB) + L1d L#22 (32KB) + L1i L#22 (32KB) + Core L#22 +
>>        PU
>>        L#22 (P#22)
>>        L2 L#23 (1024KB) + L1d L#23 (32KB) + L1i L#23 (32KB) + Core L#23 +
>>        PU
>>        L#23 (P#23)
>> 
>>
>>  That does seem to match.
>>
>>  and in short, to get the mapping, one can use
>>  $> hwloc-ls --only pu
>>  ...
>>  PU L#10 (P#19)
>>  PU L#11 (P#20)
>>  PU L#12 (P#3)
>>  PU L#13 (P#4)
>>  PU L#14 (P#5)
>>  PU L#15 (P#9)
>>  PU L#16 (P#10)
>>  PU L#17 (P#11)
>>  PU L#18 (P#15)
>>  PU L#19 (P#16)
>>  PU L#20 (P#17)
>>  PU L#21 (P#21)
>>  PU L#22 (P#22)
>>  PU L#23 (P#23)
>>  ...
>> 
>>
>>  Best
>>  Marcus
>>
>>  Am 14.12.2022 um 18:11 schrieb Paul Raines:
>>>   Ugh.  Guess I cannot count.  The mapping on that last node DOES work
>>>   with
>>>   the "alternating" scheme where we have
>>>
>>>     0  0
>>>     1  2
>>>     2  4
>>>     3  6
>>>     4  8
>>>     5 10
>>>     6 12
>>>     7 14
>>>     8 16
>>>     9 18
>>>   10 20
>>>   11 22
>>>   12  1
>>>   13  3
>>>   14  5
>>>   15  7
>>>   16  9
>>>   17 11
>>>   18 13
>>>   19 15
>>>   20 17
>>>   21 19
>>>   22 21
>>>   23 23
>>>
>>>   so CPU_IDs=8-11,20-23 does correspond to cgroup 16-23
>>>
>>>   Using the script
>>>
>>>   cd /sys/fs/cgroup/cpuset/slurm
>>>   for d in $(find -name 'job*') ; do
>>>      j=$(echo $d | cut -d_ -f3)
>>>      echo === $j
>>>      scontrol -d show job $j | grep CPU_ID | cut -d' ' -f7
>>>      cat $d/cpuset.effective_cpus
>>>   done
>>>
>>>   === 1967214
>>>   CPU_IDs=8-11,20-23
>>>   16-23
>>>   === 1960208
>>>   CPU_IDs=12-19
>>>   1,3,5,7,9,11,13,15
>>>   === 1966815
>>>   CPU_IDs=0
>>>   0
>>>   === 1966821
>>>   CPU_IDs=6
>>>   12
>>>   === 1966818
>>>   CPU_IDs=3
>>>   6
>>>   === 1966816
>>>   CPU_IDs=1
>>>   2
>>>   === 1966822
>>>   CPU_IDs=7
>>>   14
>>>   === 1966820
>>>   CPU_IDs=5
>>>   10
>>>   === 1966819
>>>   CPU_IDs=4
>>>   8
>>>   === 1966817
>>>   CPU_IDs=2
>>>   4
>>>
>>>   On all my nodes I see just two schemes.  The alternating odd/even one
>>>   above and one that is does not alternate like on this box with
>>>
>>>   CPUs=32 Boards=1 SocketsPerBoard=2 CoresPerSocket=16 ThreadsPerCore=1
>>>
>>>   === 1966495
>>>   CPU_IDs=0-2
>>>   0-2
>>>   === 1966498
>>>   CPU_IDs=10-12
>>>   10-12
>>>   === 1966502
>>>   CPU_IDs=26-28
>>>   26-28
>>>   === 1960064
>>>   CPU_IDs=7-9,13-25
>>>   7-9,13-25
>>>   === 1954480
>>>   CPU_IDs=3-6
>>>   3-6
>>> 
>>>
>>>   On Wed, 14 Dec 2022 9:42am, Paul Raines wrote:
>>> 
>>>>
>>>>   Yes, I see that on some of my other machines too.  So apicid is
>>>>   definitely not what SLURM is using but somehow just lines up that way
>>>>   on
>>>>   this one machine I have.
>>>>
>>>>   I think the issue is cgroups counts starting at 0 all the cores on the
>>>>   first socket, then all the cores on the second socket.  But SLURM (on a
>>>>   two socket box) counts 0 as the first core on the first socket, 1 as
>>>>   the
>>>>   first core on the second socket, 2 as the second core on the first
>>>>   socket,
>>>>   3 as the second core on the second socket, and so on. (Looks like I am
>>>>   wrong: see below)
>>>>
>>>>   Why slurm does this instead of just using the assignments cgroups uses
>>>>   I have no idea.  Hopefully one of the SLURM developers reads this
>>>>   and can explain
>>>>
>>>>   Looking at another SLURM node I have (where cgroups v1 is still in use
>>>>   and HT turned off) with definition
>>>>
>>>>   CPUs=24 Boards=1 SocketsPerBoard=2 CoresPerSocket=12 ThreadsPerCore=1
>>>>
>>>>   I find
>>>>
>>>>   [root at r440-17 ~]# egrep '^(apicid|proc)' /proc/cpuinfo  | tail -4
>>>>   processor       : 22
>>>>   apicid          : 22
>>>>   processor       : 23
>>>>   apicid          : 54
>>>>
>>>>   So apicid's are NOT going to work
>>>>
>>>>   # scontrol -d show job 1966817 | grep CPU_ID
>>>>       Nodes=r17 CPU_IDs=2 Mem=16384 GRES=
>>>>   # cat /sys/fs/cgroup/cpuset/slurm/uid_3776056/job_1966817/cpuset.cpus
>>>>   4
>>>>
>>>>   If Slurm has '2' this should be the second core on the first socket so
>>>>   should be '1' in cgroups, but it is 4 as we see above which is the
>>>>   fifth
>>>>   core on the first socket.  So I guess I was wrong above.
>>>>
>>>>   But in /proc/cpuinfo the apicid for processor 4 is 2!!!  So is apicid
>>>>   right after all?  Nope, on the same machine I have
>>>>
>>>>   # scontrol -d show job 1960208 | grep CPU_ID
>>>>       Nodes=r17 CPU_IDs=12-19 Mem=51200 GRES=
>>>>   # cat /sys/fs/cgroup/cpuset/slurm/uid_5164679/job_1960208/cpuset.cpus
>>>>   1,3,5,7,9,11,13,15
>>>>
>>>>   and in /proc/cpuinfo the apcid for processor 12 is 16
>>>>
>>>>   # scontrol -d show job 1967214 | grep CPU_ID
>>>>       Nodes=r17 CPU_IDs=8-11,20-23 Mem=51200 GRES=
>>>>   # cat /sys/fs/cgroup/cpuset/slurm/uid_5164679/job_1967214/cpuset.cpus
>>>>   16-23
>>>>
>>>>   I am totally lost now. Seems totally random. SLURM devs?  Any insight?
>>>> 
>>>>
>>>>   -- Paul Raines
>>>>   (http://secure-web.cisco.com/1fV8ajgaXNCaGTVPBmAhVRhk_lAbggJJxlkfKaTPxKwraiXNDFL8Fa_YXc4PH0ZxSP_aU5b9WWH4ds_d25tVNxYy_fxPvlt0lNnunFuneVQhgZjxQYnwtjHzFPP0hz2gaRgHh6zYz37fQzxkhLnkdrY_zEjnNiSCoIXIx4dOOceDVvgZ4-b-3zQMW9wOgsCLz7V4xs9fqysZ1dfuuN9mSWw7cAsm-WNWk0RWG9bDwrNm7YjdwJ5JZURQXckZ0qb4kZAnlJx5-Ihy_EqgkkoTMZeIP7rM_NAG0ejF3SI1yquf9Wi-cFgI6FHEz5ICB53zD/http%3A%2F%2Fhelp.nmr.mgh.harvard.edu)
>>>>
>>>> 
>>>>
>>>>   On Wed, 14 Dec 2022 1:33am, Marcus Wagner wrote:
>>>>
>>>>>    Hi Paul,
>>>>>
>>>>>    sorry to say, but that has to be some coincidence on your system.
>>>>>   I've
>>>>>    never seen Slurm reporting using corenumbers, which are higher than
>>>>>   the
>>>>>    total number of cores.
>>>>>
>>>>>    I have e.g. a intel Platinum 8160 here. 24 Cores per Socket, no
>>>>>    HyperThreading activated.
>>>>>    Yet here the last lines of /proc/cpuinfo:
>>>>>
>>>>>    processor       : 43
>>>>>    apicid          : 114
>>>>>    processor       : 44
>>>>>    apicid          : 116
>>>>>    processor       : 45
>>>>>    apicid          : 118
>>>>>    processor       : 46
>>>>>    apicid          : 120
>>>>>    processor       : 47
>>>>>    apicid          : 122
>>>>>
>>>>>    Never seen Slurm reporting corenumbers for a job > 96
>>>>>    Nonetheless, I agree, the cores reported by Slurm mostly have nothing
>>>>>   to
>>>>>    do with the cores reported e.g. by cgroups.
>>>>>    Since Slurm creates the cgroups, I wonder, why they report some kind
>>>>>   of
>>>>>    abstract coreid, because they should know, which cores are used, as
>>>>>   they
>>>>>    create the cgroups for the jobs.
>>>>>
>>>>>    Best
>>>>>    Marcus
>>>>>
>>>>>    Am 13.12.2022 um 16:39 schrieb Paul Raines:
>>>>>>
>>>>>>     Yes, looks like SLURM is using the apicid that is in /proc/cpuinfo
>>>>>>     The first 14 cpus in /proc/cpu (procs 0-13) have apicid
>>>>>>     0,2,4,6,8,10,12,14,16,20,22,24,26,28 in /proc/cpuinfo
>>>>>>
>>>>>>     So after setting CpuSpecList=0,2,4,6,8,10,12,14,16,18,20,22,24,26
>>>>>>     in slurm.conf it appears to be doing what I want
>>>>>>
>>>>>>     $ echo $SLURM_JOB_ID
>>>>>>     9
>>>>>>     $ grep -i ^cpu /proc/self/status
>>>>>>     Cpus_allowed:   000f0000,000f0000
>>>>>>     Cpus_allowed_list:      16-19,48-51
>>>>>>     $ scontrol -d show job 9 | grep CPU_ID
>>>>>>           Nodes=larkin CPU_IDs=32-39 Mem=25600 GRES=
>>>>>>
>>>>>>     apcid=32 is processor=16 and apcid=33 is processor=48 in
>>>>>>   /proc/cpuinfo
>>>>>>
>>>>>>     Thanks
>>>>>>
>>>>>>     -- Paul Raines
>>>>>>   (http://secure-web.cisco.com/1fV8ajgaXNCaGTVPBmAhVRhk_lAbggJJxlkfKaTPxKwraiXNDFL8Fa_YXc4PH0ZxSP_aU5b9WWH4ds_d25tVNxYy_fxPvlt0lNnunFuneVQhgZjxQYnwtjHzFPP0hz2gaRgHh6zYz37fQzxkhLnkdrY_zEjnNiSCoIXIx4dOOceDVvgZ4-b-3zQMW9wOgsCLz7V4xs9fqysZ1dfuuN9mSWw7cAsm-WNWk0RWG9bDwrNm7YjdwJ5JZURQXckZ0qb4kZAnlJx5-Ihy_EqgkkoTMZeIP7rM_NAG0ejF3SI1yquf9Wi-cFgI6FHEz5ICB53zD/http%3A%2F%2Fhelp.nmr.mgh.harvard.edu)
>>>>>>
>>>>>> 
>>>>>>
>>>>>>     On Tue, 13 Dec 2022 9:52am, Sean Maxwell wrote:
>>>>>>
>>>>>>>            External Email - Use Caution
>>>>>>>     In the slurm.conf manual they state the CpuSpecList ids are
>>>>>>>   "abstract",
>>>>>>>     and
>>>>>>>     in the CPU management docs they enforce the notion that the
>>>>>>>   abstract
>>>>>>>     Slurm
>>>>>>>     IDs are not related to the Linux hardware IDs, so that is probably
>>>>>>>   the
>>>>>>>     source of the behavior. I unfortunately don't have more
>>>>>>>   information.
>>>>>>>
>>>>>>>     On Tue, Dec 13, 2022 at 9:45 AM Paul Raines
>>>>>>>     <raines at nmr.mgh.harvard.edu>
>>>>>>>     wrote:
>>>>>>> 
>>>>>>>>
>>>>>>>>     Hmm.  Actually looks like confusion between CPU IDs on the system
>>>>>>>>     and what SLURM thinks the IDs are
>>>>>>>>
>>>>>>>>     # scontrol -d show job 8
>>>>>>>>     ...
>>>>>>>>           Nodes=foobar CPU_IDs=14-21 Mem=25600 GRES=
>>>>>>>>     ...
>>>>>>>>
>>>>>>>>     # cat
>>>>>>>>    
>>>>>>>>   /sys/fs/cgroup/system.slice/slurmstepd.scope/job_8/cpuset.cpus.effective
>>>>>>>>     7-10,39-42
>>>>>>>> 
>>>>>>>>
>>>>>>>>     -- Paul Raines
>>>>>>>>    
>>>>>>>>   (http://secure-web.cisco.com/1w33sdTB1gUzmmNOl1cd8t7VHLUOemWW6ExRIq2AHSLm0BwRxhnfCCHDdln0LWn7IZ3IUYdxeX2HzyDj7CeKHq7B1H5ek2tow-D_4Q81mK8_x_AKf6cHYOSqHSBelLikTijDZJGsJYKSleSUlZMG1mqkU4e4TirhUu0qTLKUcvqLxsvi1WCbBbyUaDUxd-c7kE2_v4XzvhBtdEqrkKAWOQF2WoJwhmTJlMhanBk-PdjHDsuDcdOgfHrmIAiRC-T8hB094Y1WvEuOjL4o2Kbx28qp4eUSPu8jSOxPEKoWsHpSDE7fWyjrlcVAsEyOpPgp4/http%3A%2F%2Fhelp.nmr.mgh.harvard.edu)
>>>>>>>>
>>>>>>>> 
>>>>>>>>
>>>>>>>>     On Tue, 13 Dec 2022 9:40am, Paul Raines wrote:
>>>>>>>> 
>>>>>>>> > >     Oh but that does explain the CfgTRES=cpu=14.  With the 
>>>>>>>> > >  CpuSpecList
>>>>>>>> >     below and SlurmdOffSpec I do get CfgTRES=cpu=50 so that makes 
>>>>>>>> >  sense.
>>>>>>>> > >     The issue remains that thought the number of cpus in 
>>>>>>>> > >  CpuSpecList
>>>>>>>> >     is taken into account, the exact IDs seem to be ignored.
>>>>>>>> > > >     -- Paul Raines >   
>>>>>>>> > > >  (http://secure-web.cisco.com/1w33sdTB1gUzmmNOl1cd8t7VHLUOemWW6ExRIq2AHSLm0BwRxhnfCCHDdln0LWn7IZ3IUYdxeX2HzyDj7CeKHq7B1H5ek2tow-D_4Q81mK8_x_AKf6cHYOSqHSBelLikTijDZJGsJYKSleSUlZMG1mqkU4e4TirhUu0qTLKUcvqLxsvi1WCbBbyUaDUxd-c7kE2_v4XzvhBtdEqrkKAWOQF2WoJwhmTJlMhanBk-PdjHDsuDcdOgfHrmIAiRC-T8hB094Y1WvEuOjL4o2Kbx28qp4eUSPu8jSOxPEKoWsHpSDE7fWyjrlcVAsEyOpPgp4/http%3A%2F%2Fhelp.nmr.mgh.harvard.edu)
>>>>>>>> > > > >     On Tue, 13 Dec 2022 9:34am, Paul Raines wrote:
>>>>>>>> > >> >>      I have tried it both ways with the same result.  The 
>>>>>>>> > >> >>  assigned CPUs
>>>>>>>> >>      will be both in and out of the range given to CpuSpecList
>>>>>>>> >> >>      I tried setting using commas instead of ranges so used
>>>>>>>> >> >>      CpuSpecList=0,1,2,3,4,5,6,7,8,9,10,11,12,13
>>>>>>>> >> >>      But still does not work
>>>>>>>> >> >>      $ srun -p basic -N 1 --ntasks-per-node=1 --mem=25G \
>>>>>>>> >>      --time=10:00:00 --cpus-per-task=8 --pty /bin/bash
>>>>>>>> >>      $ grep -i ^cpu /proc/self/status
>>>>>>>> >>      Cpus_allowed:   00000780,00000780
>>>>>>>> >>      Cpus_allowed_list:      7-10,39-42
>>>>>>>> >> >> >>      -- Paul Raines >>  
>>>>>>>> >> >> >>  (http://secure-web.cisco.com/1w33sdTB1gUzmmNOl1cd8t7VHLUOemWW6ExRIq2AHSLm0BwRxhnfCCHDdln0LWn7IZ3IUYdxeX2HzyDj7CeKHq7B1H5ek2tow-D_4Q81mK8_x_AKf6cHYOSqHSBelLikTijDZJGsJYKSleSUlZMG1mqkU4e4TirhUu0qTLKUcvqLxsvi1WCbBbyUaDUxd-c7kE2_v4XzvhBtdEqrkKAWOQF2WoJwhmTJlMhanBk-PdjHDsuDcdOgfHrmIAiRC-T8hB094Y1WvEuOjL4o2Kbx28qp4eUSPu8jSOxPEKoWsHpSDE7fWyjrlcVAsEyOpPgp4/http%3A%2F%2Fhelp.nmr.mgh.harvard.edu)
>>>>>>>> >> >> >> >>      On Mon, 12 Dec 2022 10:21am, Sean Maxwell wrote:
>>>>>>>> >> >>>       Hi Paul,
>>>>>>>> >>> >>>       Nodename=foobar \
>>>>>>>> >>>>          CPUs=64 Boards=1 SocketsPerBoard=2 CoresPerSocket=16
>>>>>>>> >>>>          ThreadsPerCore=2
>>>>>>>> >>>>          \
>>>>>>>> >>>>          RealMemory=256312 MemSpecLimit=32768 CpuSpecList=14-63 
>>>>>>>> >>>>  \
>>>>>>>> >>>>          TmpDisk=6000000 Gres=gpu:nvidia_rtx_a6000:1
>>>>>>>> >>>> >>>>       The slurm.conf also has:
>>>>>>>> >>>> >>>>       ProctrackType=proctrack/cgroup
>>>>>>>> >>>>       TaskPlugin=task/affinity,task/cgroup
>>>>>>>> >>>>       TaskPluginParam=Cores,*SlurmdOf**fSpec*,Verbose
>>>>>>>> >>>> >>> >>>       Doesn't setting SlurmdOffSpec tell Slurmd that is 
>>>>>>>> >>>> >>> >>>  should NOT use >>>  the
>>>>>>>> >>>       CPUs
>>>>>>>> >>>       in the spec list? (
>>>>>>>> >>> >>>    
>>>>>>>> >>> >>>  https://secure-web.cisco.com/1V9Fskh4lCAx_XrdlCr8o1EtnePELf-1YK4TerT47ktLxy3fO9FaIpaGXVA8ODhMAdhmXSqToQstwAilA71r7z1Q4jDqPSKEsJQNUhJYYRtxFnZIO49QxsYrVo9c3ExH89cIk_t7H5dtGEjpme2LFKm23Z52yK-xZ3fEl_LyK61uCzkas6GKykzPCPyoNXaFgs32Ct2tDIVL8vI6JW1_-1uQ9gUyWmm24xJoBxLEui7tSTVwMtiVRu8C7pU1nJ8qr6ghBlxrqx-wQiRP4XBCjhPARHa2KBqkUogjEVRAe3WdAbbYBxtXeVsWjqNGmjSVA/https%3A%2F%2Fslurm.schedmd.com%2Fslurm.conf.html%23OPT_SlurmdOffSpec)
>>>>>>>> >>>       In this case, I believe it uses what is left, which is the 
>>>>>>>> >>>  0-13. >>>  We
>>>>>>>>     are
>>>>>>>> >>>       just starting to work on this ourselves, and were looking 
>>>>>>>> >>>  at 
>>>>>>>> >>> >>>   this
>>>>>>>> >>>       setting.
>>>>>>>> >>> >>>       Best,
>>>>>>>> >>> >>>       -Sean
>>>>>>>> >>> >> >
>>>>>>>>     The information in this e-mail is intended only for the person to
>>>>>>>>   whom
>>>>>>>>     it
>>>>>>>>     is addressed.  If you believe this e-mail was sent to you in
>>>>>>>>   error
>>>>>>>>   and
>>>>>>>>     the
>>>>>>>>     e-mail contains patient information, please contact the Mass
>>>>>>>>   General
>>>>>>>>     Brigham Compliance HelpLine at
>>>>>>>>    
>>>>>>>>   https://secure-web.cisco.com/11OmVChs0jRoe-4AH2iRxvEdMN0dxZcFsunG07PJ0sXxdW7tj7-BUiDwEEi3gkqOms_qFRdQbCLHJQW0jD6cG8-griFmte8mXIoPZSDzIE8dHcew9yMCpQxJnYVVs8mK5aB-9o4ospPlPqxo3FA0LN8gpJSrsBKOxr5m7T3Jd7FY04zJnehrYc0FQwfWAPJx523fZTqVTTmwZgdEFZAQtURZ8hPxlohSzsh7d13L7byOVUmxAxzolzDTvRSH9l1gjMm-RjtdW95eYkgPlRoM3nJ0WCledYAp5NA3kUGNhsc5uNDp3lWIzS7gZGIMfTyg9/https%3A%2F%2Fwww.massgeneralbrigham.org%2Fcomplianceline
>>>>>>>>     <
>>>>>>>>    
>>>>>>>>   https://secure-web.cisco.com/11OmVChs0jRoe-4AH2iRxvEdMN0dxZcFsunG07PJ0sXxdW7tj7-BUiDwEEi3gkqOms_qFRdQbCLHJQW0jD6cG8-griFmte8mXIoPZSDzIE8dHcew9yMCpQxJnYVVs8mK5aB-9o4ospPlPqxo3FA0LN8gpJSrsBKOxr5m7T3Jd7FY04zJnehrYc0FQwfWAPJx523fZTqVTTmwZgdEFZAQtURZ8hPxlohSzsh7d13L7byOVUmxAxzolzDTvRSH9l1gjMm-RjtdW95eYkgPlRoM3nJ0WCledYAp5NA3kUGNhsc5uNDp3lWIzS7gZGIMfTyg9/https%3A%2F%2Fwww.massgeneralbrigham.org%2Fcomplianceline>
>>>>>>>>     .
>>>>>>>>     Please note that this e-mail is not secure (encrypted).  If you
>>>>>>>>   do
>>>>>>>>   not
>>>>>>>>     wish to continue communication over unencrypted e-mail, please
>>>>>>>>   notify
>>>>>>>>     the
>>>>>>>>     sender of this message immediately.  Continuing to send or
>>>>>>>>   respond
>>>>>>>>   to
>>>>>>>>     e-mail after receiving this message means you understand and
>>>>>>>>   accept
>>>>>>>>     this
>>>>>>>>     risk and wish to continue to communicate over unencrypted e-mail.
>>>>>>>> 
>>>>>>>>
>>>>>>     The information in this e-mail is intended only for the person to
>>>>>>   whom
>>>>>>     it
>>>>>>     is addressed.  If you believe this e-mail was sent to you in error
>>>>>>   and
>>>>>>     the
>>>>>>     e-mail contains patient information, please contact the Mass
>>>>>>   General
>>>>>>     Brigham Compliance HelpLine at
>>>>>>    
>>>>>>   https://secure-web.cisco.com/1PmxwaOlMXSsSWGbWWYKPBpt4ErgJCfQ6yIrz0i6wWWrPX1wiWCiBfeZ2xu-pFqQOqhme1wR4J-sGQNCjJYx8LQrRYLlgTpQZSpSKneskIT28Lv2VyVO25H03n8yo9waPPhoudMF-cow9scYJMVCkyJnSN_R1yg51kb6zS1-MjPgru7ZS0BZTzlPeX-7KoHpQahSlFu0vjOWUq4nqdEVwh_g44-YCL1zmSrzRTkg96oS8Bm8Bwo3jZ7AOml-adns9Fr6Q9QVg31f2N9NsGviytLoSWv8s8wFQCwlgVNfPTTwKZxjkIxeWK8HBmc4vgE9D/https%3A%2F%2Fwww.massgeneralbrigham.org%2Fcomplianceline
>>>>>>    
>>>>>>   <https://secure-web.cisco.com/1PmxwaOlMXSsSWGbWWYKPBpt4ErgJCfQ6yIrz0i6wWWrPX1wiWCiBfeZ2xu-pFqQOqhme1wR4J-sGQNCjJYx8LQrRYLlgTpQZSpSKneskIT28Lv2VyVO25H03n8yo9waPPhoudMF-cow9scYJMVCkyJnSN_R1yg51kb6zS1-MjPgru7ZS0BZTzlPeX-7KoHpQahSlFu0vjOWUq4nqdEVwh_g44-YCL1zmSrzRTkg96oS8Bm8Bwo3jZ7AOml-adns9Fr6Q9QVg31f2N9NsGviytLoSWv8s8wFQCwlgVNfPTTwKZxjkIxeWK8HBmc4vgE9D/https%3A%2F%2Fwww.massgeneralbrigham.org%2Fcomplianceline>
>>>>>>   .
>>>>>>     Please note that this e-mail is not secure (encrypted).  If you do
>>>>>>   not
>>>>>>     wish to continue communication over unencrypted e-mail, please
>>>>>>   notify
>>>>>>     the
>>>>>>     sender of this message immediately.  Continuing to send or respond
>>>>>>   to
>>>>>>     e-mail after receiving this message means you understand and accept
>>>>>>   this
>>>>>>     risk and wish to continue to communicate over unencrypted e-mail.
>>>>>> 
>>>>>
>>>>>    --
>>>>>    Dipl.-Inf. Marcus Wagner
>>>>>
>>>>>    IT Center
>>>>>    Gruppe: Server, Storage, HPC
>>>>>    Abteilung: Systeme und Betrieb
>>>>>    RWTH Aachen University
>>>>>    Seffenter Weg 23
>>>>>    52074 Aachen
>>>>>    Tel: +49 241 80-24383
>>>>>    Fax: +49 241 80-624383
>>>>>    wagner at itc.rwth-aachen.de
>>>>>    http://secure-web.cisco.com/1B5jcsbcQ02Q4GJn4-URoKY8HqxxzgGBFYqjthyPxPxdXzmf7UHSU6_5MS9jp4IyQqV_2eP2GOlJADbOsU1JCQektcVysY0wFUzTT6iJXeukZSEwwsS1a9fEa5A5A9V3YXL2ew7-1i2_EbER_b0LzSNZxFTuZZhFFec7CwG5_VBxAPznWJN6V5UeiPae_PAclOALCf9dVkQKsja5wf6gn9opTN6LBBqXFodRzRdf00Tohpr3X0gVJqiS0wkdhNsDi4lAgnTIPn712RdlPYwpWdc_wx4lZyn6UyuSWitp0muk8aauPf_DTotbHJ56Uc8V3/http%3A%2F%2Fwww.itc.rwth-aachen.de
>>>>>
>>>>>    Social Media Kanäle des IT Centers:
>>>>>    https://secure-web.cisco.com/1ez5FTX8pxpjTtZWBVFVRAWafNU7u3KXdZlwv16JCbfnwcYPDxCftSdbp14fmyndNHEYU-Pso33tUA8Vkj4QgeCgq8gbGJkHfuV-kJSqb1Rmg9b_j6fQ2GaXGDs-8zlteOuZi5bK3ePaj36iUxmVD-IfIv9AWM_39cZ6ZA9D2PjfqPV8lgFi5VqOR7fuj5hCH2Sp9Xv0nY_w0RIaLDlKh16HPvmGRZWg5EWNnEkAmgeVhGtDasU77Y292iq3wLJGZ0xkJifdzbOVEHfBG6X9Y98oDRNO_qFPiAsaOcdpk_eCxFl3L2KijPbh5b99AGtiC/https%3A%2F%2Fblog.rwth-aachen.de%2Fitc%2F
>>>>>    https://secure-web.cisco.com/1CW5aaHqTou4LGMrEyU-6QwYallSKgvhhp71rescRZCgVtIRexR6DcQS5NXxhMghQcwnM07RraLcLU2RufHZCj983olA9VGoI063ZUppLuV3aWjKmTYFZglCggrjb_5yB1GdYmb0sMtxBL8JBf_FFuTnOh4JYwUpuopXyH5I1aHPI3Ywr0zYhWCGA3EkxAcpBCpV738s8tpfpMXVOuDJzZfUapgRTtDQf7bv7NE00-10grSjQOCv8QZg22L-c_6O37DhBIQ42goSurJUehTubM5f7acoG6XONaiKufNWhp48f31SeQ6nsxOMmvm_pOVfa/https%3A%2F%2Fwww.facebook.com%2Fitcenterrwth
>>>>>    https://secure-web.cisco.com/1s2yL6gEKoiHrJk0M0CRf9vptCb77ZmQoicVknhVL3O-3EEDFPOdYYlSq_K6pvUubzaOLiKIe9-OA-6xT9i-HPYamEdbBDJNCOGF8Tgr3eZqpDe7GhbdLnh8j1K9SuzIbG-2tRP5E1QgNx4m2zfQBWkK5Gu6vTkQbNDQQSeDZxVHfKYdbYsil21qLauAHloys7KGs7gqqcKnBPN6RZCRC-x5X506txXjkGcqm6xU6U6J85FMIFGhdTukCDbJtZEJIdgZgHDBaIvfDRfk1WgMUlFvTOe7eEpzLkBHBl9j-w4SfNjtdVfN7dUx9w3B-bGtg/https%3A%2F%2Fwww.linkedin.com%2Fcompany%2Fitcenterrwth
>>>>>    https://secure-web.cisco.com/1hzaxwdTNMmSQsMHSOQZXdDlQcfiJ7_MPscUdPKHWZsfpNEsZQHYao5lDIqJajwoH1Vs_U2sH5N-sNQDGOcjzeUVwnjbrPtILmKLpBYi8BMPGePr7LiGM5Ehn_H_e8UvB38Sp-SbZcOOw_4421AARSlSiZibRKYWvIGnEevq8PKSmAIRshzZ1KX63V32VRGhJD6AQLhWUYIVIAlKAaZvZ_kqR-KeriN2cgZgM24guNTJcFw9eKTUleKB10kVs5EU-eh-CT5Yai-M4WsclwyrJUIQ9SzZgxrwWsIuOTlXtuf0szkzHxEXjxRTmjkn4xqi9/https%3A%2F%2Ftwitter.com%2FITCenterRWTH
>>>>>    https://secure-web.cisco.com/1uxcAJMPXZb2B1FJm65QOa0OEFNrH0W_Vo4oDBHKlvJCgxIfjgHhyip1Qlqap_05F4BQzdiRMuaxbQJS1vikwziB80jfQfmq-kgCkBml0pN80U8YzpZCpYSeAc0eoOoHN0RzutppHGkUP2Fzzlgk22qszo5PQZWJxjXW8J7X1FozVJEiYYs38gCHTyoALjPnoGdadThFVBSufbwVsqj2JG29I3M2vSE-IMPidEONSt6klggc-nGCdN-M_BwbmmVf8INbVah-UmPWh7B9UFpd13QVKpuDe_LqBvBTHuqLOxGjj0KVrTn6HCF58c_VioJkx/https%3A%2F%2Fwww.youtube.com%2Fchannel%2FUCKKDJJukeRwO0LP-ac8x8rQ
>>>>> 
>>>>
>>>   The information in this e-mail is intended only for the person to whom
>>>   it
>>>   is addressed.  If you believe this e-mail was sent to you in error and
>>>   the
>>>   e-mail contains patient information, please contact the Mass General
>>>   Brigham Compliance HelpLine at
>>>   https://secure-web.cisco.com/1PmxwaOlMXSsSWGbWWYKPBpt4ErgJCfQ6yIrz0i6wWWrPX1wiWCiBfeZ2xu-pFqQOqhme1wR4J-sGQNCjJYx8LQrRYLlgTpQZSpSKneskIT28Lv2VyVO25H03n8yo9waPPhoudMF-cow9scYJMVCkyJnSN_R1yg51kb6zS1-MjPgru7ZS0BZTzlPeX-7KoHpQahSlFu0vjOWUq4nqdEVwh_g44-YCL1zmSrzRTkg96oS8Bm8Bwo3jZ7AOml-adns9Fr6Q9QVg31f2N9NsGviytLoSWv8s8wFQCwlgVNfPTTwKZxjkIxeWK8HBmc4vgE9D/https%3A%2F%2Fwww.massgeneralbrigham.org%2Fcomplianceline
>>>   <https://secure-web.cisco.com/1PmxwaOlMXSsSWGbWWYKPBpt4ErgJCfQ6yIrz0i6wWWrPX1wiWCiBfeZ2xu-pFqQOqhme1wR4J-sGQNCjJYx8LQrRYLlgTpQZSpSKneskIT28Lv2VyVO25H03n8yo9waPPhoudMF-cow9scYJMVCkyJnSN_R1yg51kb6zS1-MjPgru7ZS0BZTzlPeX-7KoHpQahSlFu0vjOWUq4nqdEVwh_g44-YCL1zmSrzRTkg96oS8Bm8Bwo3jZ7AOml-adns9Fr6Q9QVg31f2N9NsGviytLoSWv8s8wFQCwlgVNfPTTwKZxjkIxeWK8HBmc4vgE9D/https%3A%2F%2Fwww.massgeneralbrigham.org%2Fcomplianceline>
>>>   .
>>>   Please note that this e-mail is not secure (encrypted).  If you do not
>>>   wish to continue communication over unencrypted e-mail, please notify
>>>   the
>>>   sender of this message immediately.  Continuing to send or respond to
>>>   e-mail after receiving this message means you understand and accept this
>>>   risk and wish to continue to communicate over unencrypted e-mail.
>>
>>  --
>>  Dipl.-Inf. Marcus Wagner
>>
>>  IT Center
>>  Gruppe: Server, Storage, HPC
>>  Abteilung: Systeme und Betrieb
>>  RWTH Aachen University
>>  Seffenter Weg 23
>>  52074 Aachen
>>  Tel: +49 241 80-24383
>>  Fax: +49 241 80-624383
>>  wagner at itc.rwth-aachen.de
>>  http://secure-web.cisco.com/1B5jcsbcQ02Q4GJn4-URoKY8HqxxzgGBFYqjthyPxPxdXzmf7UHSU6_5MS9jp4IyQqV_2eP2GOlJADbOsU1JCQektcVysY0wFUzTT6iJXeukZSEwwsS1a9fEa5A5A9V3YXL2ew7-1i2_EbER_b0LzSNZxFTuZZhFFec7CwG5_VBxAPznWJN6V5UeiPae_PAclOALCf9dVkQKsja5wf6gn9opTN6LBBqXFodRzRdf00Tohpr3X0gVJqiS0wkdhNsDi4lAgnTIPn712RdlPYwpWdc_wx4lZyn6UyuSWitp0muk8aauPf_DTotbHJ56Uc8V3/http%3A%2F%2Fwww.itc.rwth-aachen.de
>>
>>  Social Media Kanäle des IT Centers:
>>  https://secure-web.cisco.com/1ez5FTX8pxpjTtZWBVFVRAWafNU7u3KXdZlwv16JCbfnwcYPDxCftSdbp14fmyndNHEYU-Pso33tUA8Vkj4QgeCgq8gbGJkHfuV-kJSqb1Rmg9b_j6fQ2GaXGDs-8zlteOuZi5bK3ePaj36iUxmVD-IfIv9AWM_39cZ6ZA9D2PjfqPV8lgFi5VqOR7fuj5hCH2Sp9Xv0nY_w0RIaLDlKh16HPvmGRZWg5EWNnEkAmgeVhGtDasU77Y292iq3wLJGZ0xkJifdzbOVEHfBG6X9Y98oDRNO_qFPiAsaOcdpk_eCxFl3L2KijPbh5b99AGtiC/https%3A%2F%2Fblog.rwth-aachen.de%2Fitc%2F
>>  https://secure-web.cisco.com/1CW5aaHqTou4LGMrEyU-6QwYallSKgvhhp71rescRZCgVtIRexR6DcQS5NXxhMghQcwnM07RraLcLU2RufHZCj983olA9VGoI063ZUppLuV3aWjKmTYFZglCggrjb_5yB1GdYmb0sMtxBL8JBf_FFuTnOh4JYwUpuopXyH5I1aHPI3Ywr0zYhWCGA3EkxAcpBCpV738s8tpfpMXVOuDJzZfUapgRTtDQf7bv7NE00-10grSjQOCv8QZg22L-c_6O37DhBIQ42goSurJUehTubM5f7acoG6XONaiKufNWhp48f31SeQ6nsxOMmvm_pOVfa/https%3A%2F%2Fwww.facebook.com%2Fitcenterrwth
>>  https://secure-web.cisco.com/1s2yL6gEKoiHrJk0M0CRf9vptCb77ZmQoicVknhVL3O-3EEDFPOdYYlSq_K6pvUubzaOLiKIe9-OA-6xT9i-HPYamEdbBDJNCOGF8Tgr3eZqpDe7GhbdLnh8j1K9SuzIbG-2tRP5E1QgNx4m2zfQBWkK5Gu6vTkQbNDQQSeDZxVHfKYdbYsil21qLauAHloys7KGs7gqqcKnBPN6RZCRC-x5X506txXjkGcqm6xU6U6J85FMIFGhdTukCDbJtZEJIdgZgHDBaIvfDRfk1WgMUlFvTOe7eEpzLkBHBl9j-w4SfNjtdVfN7dUx9w3B-bGtg/https%3A%2F%2Fwww.linkedin.com%2Fcompany%2Fitcenterrwth
>>  https://secure-web.cisco.com/1hzaxwdTNMmSQsMHSOQZXdDlQcfiJ7_MPscUdPKHWZsfpNEsZQHYao5lDIqJajwoH1Vs_U2sH5N-sNQDGOcjzeUVwnjbrPtILmKLpBYi8BMPGePr7LiGM5Ehn_H_e8UvB38Sp-SbZcOOw_4421AARSlSiZibRKYWvIGnEevq8PKSmAIRshzZ1KX63V32VRGhJD6AQLhWUYIVIAlKAaZvZ_kqR-KeriN2cgZgM24guNTJcFw9eKTUleKB10kVs5EU-eh-CT5Yai-M4WsclwyrJUIQ9SzZgxrwWsIuOTlXtuf0szkzHxEXjxRTmjkn4xqi9/https%3A%2F%2Ftwitter.com%2FITCenterRWTH
>>  https://secure-web.cisco.com/1uxcAJMPXZb2B1FJm65QOa0OEFNrH0W_Vo4oDBHKlvJCgxIfjgHhyip1Qlqap_05F4BQzdiRMuaxbQJS1vikwziB80jfQfmq-kgCkBml0pN80U8YzpZCpYSeAc0eoOoHN0RzutppHGkUP2Fzzlgk22qszo5PQZWJxjXW8J7X1FozVJEiYYs38gCHTyoALjPnoGdadThFVBSufbwVsqj2JG29I3M2vSE-IMPidEONSt6klggc-nGCdN-M_BwbmmVf8INbVah-UmPWh7B9UFpd13QVKpuDe_LqBvBTHuqLOxGjj0KVrTn6HCF58c_VioJkx/https%3A%2F%2Fwww.youtube.com%2Fchannel%2FUCKKDJJukeRwO0LP-ac8x8rQ
>> 
>> 
>
The information in this e-mail is intended only for the person to whom it is addressed.  If you believe this e-mail was sent to you in error and the e-mail contains patient information, please contact the Mass General Brigham Compliance HelpLine at https://www.massgeneralbrigham.org/complianceline <https://www.massgeneralbrigham.org/complianceline> .
Please note that this e-mail is not secure (encrypted).  If you do not wish to continue communication over unencrypted e-mail, please notify the sender of this message immediately.  Continuing to send or respond to e-mail after receiving this message means you understand and accept this risk and wish to continue to communicate over unencrypted e-mail. 


More information about the slurm-users mailing list