Thanks Loris and Gareth. here is the job submission script. if you find any errors please let me know.
since i am not the admin but just an user, i think i dont have access to the prolog and epilogue files.

If the jobs are independent, why do you want to run them all on the same
node?
I am running sequential codes. Essentially 50 copies of the same node with a variation in parameter.
Since I am using the Slurm scheduler, the nodes and cores are allocated depending upon the
available resources. So there are instances, when 20 of them goes to 20 free cores located on a particular 
node and the rest 30 goes to the free 30 cores on another node. It turns out that only 1 job out of 20 and 1 job 
out of 30 are completed succesfully with exitcode 0 and the rest gets terminated with exitcode 9.
for information, i run sjobexitmod -l jobid to check the exitcodes.

----------------------------------
the submission script is as follows:



#!/bin/bash
################
# Setting slurm options
################


# lines starting with "#SBATCH" define your jobs parameters
# requesting the type of node on which to run job
##SBATCH --partition <patition name>
#SBATCH --partition=standard

# telling slurm how many instances of this job to spawn (typically 1)

##SBATCH --ntasks <number>
##SBATCH --ntasks=1
#SBATCH --nodes=1
##SBATCH -N 1
##SBATCH --ntasks-per-node=1



# setting number of CPUs per task (1 for serial jobs)

##SBATCH --cpus-per-task <number>

##SBATCH --cpus-per-task=1

# setting memory requirements

##SBATCH --mem-per-cpu <memory in MB>
#SBATCH --mem-per-cpu=1G

# propagating max time for job to run

##SBATCH --time <days-hours:minute:seconds>
##SBATCH --time <hours:minute:seconds>
##SBATCH --time <minutes>
#SBATCH --time 10:0:0
#SBATCH --job-name gstate

#module load compiler/intel/2018_4
module load fftw-3.3.10-intel-2021.6.0-ppbepka
echo "Running on $(hostname)"
echo "We are in $(pwd)"


################
# run the program
################
/home/arkoroy.sps.iitmandi/ferro-detun/input1/a_1.out &