![]() SlurmctldLogFile =/var/log/slurm/slurmctld.log Modify the following parameters in /etc/slurm/nf to match with your cluster: ClusterName =IRD Visit to make a configuration file for Slurm. Using the command lscpu on each node to get processors’ informations. This will populate the slurm_acct_db with tables Configuration file /etc/slurm/nf: Then enable and start the slurmdbd service $ systemctl start slurmdbd Modify the /etc/slurm/nf with the following parameters: AuthType =auth/munge Mysql> create database slurm_acct_db Configure the slurm db backend: * TO 'localhost' identified by 'some_pass' with grant option In the RPMs’folder, launch the following command: $ yum -nogpgcheck localinstall slurm- * Create and configure the slurm_acct_db database: $ mysql -u root -p RPMs are located in /root/rpmbuild/RPMS/x86_64/ Install slurm on master and nodes Slurm installation: Install the following prerequisites: $ yum install openssl openssl-devel pam-devel rpmbuild numactl numactl-devel hwloc hwloc-devel lua lua-devel readline-devel rrdtool-devel ncurses-devel man2html libibmad libibumad -y Retrieve the tarball $ wget Create the RPMs: $ rpmbuild -ta slurm-19.05.0.tar.bz2 To implement this change you have to shut down the database and move/remove logfiles: $ systemctl stop mariadb To do that, create a the file /etc/my.cnf.d/innodb.cnf with the following lines: Setting innodb_lock_wait_timeout,innodb_log_file_size and innodb_buffer_pool_size to larger values than the default is recommended. Launch the following command to set up the root password an secure mariadb: $ mysql_secure_installation Modify the innodb configuration: Systemctl enable mariadb secure the installation: ![]() $ munge -n | ssh unmunge Mariadb installation and configuration Install mariadb with the following command: $ yum install mariadb-server -y Activate and start the mariadb service: $ systemctl start mariadb $ cexec systemctl start munge Test munge from the master node: $ munge -n | unmunge $ cexec chmod 0700 /etc/munge/ /var/log/munge/ /var/lib/munge/ /run/munge/ Enable and Start the munge service with: $ systemctl enable munge $ cexec chown -R munge: /etc/munge/ /var/log/munge/ /var/lib/munge/ /run/munge/ $ chmod 0700 /etc/munge/ /var/log/munge/ /var/lib/munge/ /run/munge/ $ cexec cp /home/munge.key /etc/munge Set the rights: $ chown -R munge: /etc/munge/ /var/log/munge/ /var/lib/munge/ /run/munge/ $ useradd -m -c "SLURM workload manager" -d /var/lib/slurm -u $SLURMUSER -g slurm -s /bin/bash slurm Munge Installation for authentication: $ yum install munge munge-libs munge-devel -y Create a munge authentication key: $ /usr/sbin/create-munge-key Copy the munge authentication key on every node: $ cp /etc/munge/munge.key /home ![]() $ useradd -m -c "MUNGE Uid 'N' Gid Emporium" -d /var/lib/munge -u $MUNGEUSER -g munge -s /sbin/nologin munge Slurm and Munge require consistent UID and GID across every node in the cluster.įor all the nodes, before you install Slurm or Munge: $ export MUNGEUSER =1001 Slurm is an open source, fault-tolerant, and highly scalable cluster management and job scheduling system for large and small Linux clusters.Īuthentication and databases: Create the user for munge and slurm:
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |