Joining the debian HPC team
Hi Debian-HPC Team,
I'd like to join you in order to improve the admin/user experience of Debian
on HPC systems.
My background:
- Debian developer already for quite some time:
+ FAI https://fai-project.org (my main contribution: 'softupdate', i.e.
configuration management after the initial installation)
+ maintainer of PDL (Perl Data Language) for some years
- PhD in computational physics (high-throughput screening for new
superconductors)
- HPC purchase/setup/administration
+ started with a 2-rack Opteron HPC at Freie Universitaet Berlin,
Physics department before 2006 (PBS/Torque)
+ several smaller HPC-Clusters at
* Max-Planck-Institute for Microstructure Physics (Halle/Germany)
* Max-Planck-Institute for the Structure and Dynamics of Matter
(Hamburg/Germany)
- involved in the build/test farm (based on buildbot) of octopus
(https://octopus-code.org)
Ideas:
- CUDA-aware MPI in Debian, in the context of nvlink/sxm2:
+ package gdrcopy kernel drivers (https://github.com/NVIDIA/gdrcopy)
+ UCX support for gdrcopy, and integration into Debian's OpenMPI packages
(https://www.open-mpi.org/faq/?category=buildcuda)
- official Debian packages for https://octopus-code.org
I am a bit confused about the distribution of packages among the debian-hpc,
debian-science and pkg-nvidia-devel teams:
- I imagine gdrcopy in the context of pkg-nvidia-devel, i.e. the team that
maintains also other nvidia cuda kernel drivers
- why is ucx (a fairly low-level HPC lib) maintained by debian-science,
while openmpi is maintained by debian-hpc?
--
Mit freundlichen Grüßen
Henning Glawe
Dr. Henning Glawe
Max-Planck-Institut für Struktur und Dynamik der Materie
Geb. 99 (CFEL), Luruper Chaussee 149, 22761 Hamburg, Germany
http://www.mpsd.mpg.de/, Email: henning.glawe@mpsd.mpg.de
Building/Room: 99/O2.100, Phone: +49-40-8998-88334
Reply to: