[Date Prev][Date Next] [Thread Prev][Thread Next] [Date Index] [Thread Index]

Joining the debian HPC team



Hi Debian-HPC Team,
I'd like to join you in order to improve the admin/user experience of Debian
on HPC systems.

My background:
- Debian developer already for quite some time:
  + FAI https://fai-project.org (my main contribution: 'softupdate', i.e.
    configuration management after the initial installation)
  + maintainer of PDL (Perl Data Language) for some years
- PhD in computational physics (high-throughput screening for new
  superconductors)
- HPC purchase/setup/administration
  + started with a 2-rack Opteron HPC at Freie Universitaet Berlin,
    Physics department before 2006 (PBS/Torque)
  + several smaller HPC-Clusters at
    * Max-Planck-Institute for Microstructure Physics (Halle/Germany)
    * Max-Planck-Institute for the Structure and Dynamics of Matter
      (Hamburg/Germany)
- involved in the build/test farm (based on buildbot) of octopus
  (https://octopus-code.org)

Ideas:
- CUDA-aware MPI in Debian, in the context of nvlink/sxm2:
  + package gdrcopy kernel drivers (https://github.com/NVIDIA/gdrcopy)
  + UCX support for gdrcopy, and integration into Debian's OpenMPI packages
    (https://www.open-mpi.org/faq/?category=buildcuda)
- official Debian packages for https://octopus-code.org

I am a bit confused about the distribution of packages among the debian-hpc, 
debian-science and pkg-nvidia-devel teams:
- I imagine gdrcopy in the context of pkg-nvidia-devel, i.e. the team that
  maintains also other nvidia cuda kernel drivers
- why is ucx (a fairly low-level HPC lib) maintained by debian-science,
  while openmpi is maintained by debian-hpc?

-- 
Mit freundlichen Grüßen
Henning Glawe

Dr. Henning Glawe
Max-Planck-Institut für Struktur und Dynamik der Materie
Geb. 99 (CFEL), Luruper Chaussee 149, 22761 Hamburg, Germany
http://www.mpsd.mpg.de/, Email: henning.glawe@mpsd.mpg.de
Building/Room: 99/O2.100, Phone: +49-40-8998-88334


Reply to: