High Performance Computing Modernization Program

The United States Department of Defense High Performance Computing Modernization Program (HPCMP) was initiated in 1992 in response to Congressional direction to modernize the Department of Defense (DoD) laboratories’ high performance computing capabilities.[1] The HPCMP provides supercomputers, a national research network, high-end software tools, a secure environment, and computational science experts[2] that together enable the Defense laboratories and test centers to conduct research, development, test and technology evaluation activities.[3]

High Performance Computing Modernization Program
AbbreviationHPCMP
Formation1992; 32 years ago (1992)
HeadquartersVicksburg, MS
Director
Kevin Newmeyer (Acting)
Parent organization
ERDC
Websitewww.hpc.mil
ASN

The program was administered by the Office of the Director, Defense Research and Engineering (now called the Assistant Secretary of Defense for Research and Engineering) through FY2011, at which point it was transferred to the office of the United States Assistant Secretary of the Army for Acquisition, Logistics, and Technology, where it is managed by the Deputy Assistant Secretary for Research and Technology.[4]

The program comprises three primary elements: DoD Supercomputing Resource Centers (DSRCs), which provide large scale supercomputers and operations staff; Defense Research and Engineering Network (DREN), a nationwide high speed, low latency, R&D network connecting the centers and major user communities; and a collection of efforts in software applications to develop, modernize, and maintain software to address DoD's science and engineering challenges. Dr. Kevin Newmeyer is currently the Acting Director of HPCMP.

DoD Supercomputing Resource Centers

edit

The HPCMP funds and oversees the operation of five supercomputing centers, called DoD Supercomputing Resource Centers, or DSRCs. The centers are operated by the Engineer Research and Development Center in Vicksburg, MS, the Army Research Laboratory in Aberdeen, MD, the Naval Meteorology and Oceanography Command in Stennis Space Center, MS, the Air Force Research Laboratory in Dayton, OH, and Maui High Performance Computing Center in Maui, HI.[5] The Arctic Region Supercomputing Center (ARSC) in Fairbanks, AK was a sixth DSRC[6] until funding for it was discontinued in 2011.[7]

Each center hosts large-scale supercomputers, high-speed networks, multi-petabyte archival mass storage systems, and computational experts. The centers are managed by the HPCMP Assistant Director for Centers, who also funds program-wide activities in user support (the HPC Help Desk) and scientific visualization (the Data Analysis and Assessment Center, or DAAC).

Defense Research and Engineering Network

edit

The Defense Research and Engineering Network (DREN) — a high-speed national computer network for US Department of Defense (DoD) computational research, engineering, and testing — is a significant component of the DoD High Performance Computing Modernization Program (HPCMP).

DREN is DoD's premier wide area network (WAN) for research, test and engineering missions. DREN is a high-speed, high-capacity, low-latency, low-jitter nationwide computer network in support of the DoD's High Performance Computing, Science and Technology, Test and Evaluation and Acquisition Engineering communities. DREN connects scientists and engineers with the HPCMP's geographically dispersed high performance computing (HPC) sites, including the five DoD Supercomputing Resource Centers (DSRCs). DREN is installed at more than 210 DoD sites including research laboratories, test centers, universities, and industrial locations throughout the United States (including Hawaii and Alaska).[8]

The fourth generation DREN network (DREN 4) is provided under a commercial services contract awarded to Verizon in 2021.[9] DREN has been awarded in the past to Lumen (DREN 3), Verizon (DREN 2), and AT&T (DREN 1). DREN service providers typically build DREN as a virtual private network overlay on their commercial network infrastructure. Capabilities provided by DREN 4 include digital data transfer speeds ranging from 1 Gbit/s through 100 Gbit/s. DREN 4 is an IPv6 network, with support for legacy IPv4. The HPCMP is currently in the process of building out DREN 4 which will ultimately replace DREN 3 when fully tested in 2022. The two networks will run in parallel for about one year while the 210 sites are transitioned from DREN 3 to DREN 4.

In 2003, DREN was designated the Department of Defense's first IPv6 network by the Assistant Secretary of Defense for Networks & Information Integration. Since that time, DREN has been a pioneer within the DoD and US Federal Government for IPv6 deployment, running both IPv4/IPv6 dual stack and IPv6-only networks. DREN provides an extensive IPv6 Knowledge Base containing best practices and lessons learned, available at no charge on the HPCMP website.

Other research networks of interest:

Software Applications

edit

In addition to supercomputers and the national wide area research network, the HPCMP funds software applications that help the DoD achieve its research objectives by ensuring that important DoD applications run effectively on the large-scale supercomputers it deploys.

HPC Software Applications Institutes

edit

HPC Software Applications Institutes, or HSAIs, are cross-disciplinary projects funded by the HPCMP but executed in the DoD labs. The HSAI program both develops tools to solve important computational problems facing the DoD and build organic HPC expertise within the department.[10]

User Productivity Enhancement and Training (PET)

edit

The High Performance Computing Modernization Program's (HPCMP) User Productivity Enhancement and Training (PET) program gives users access to computational experts with experience spanning a wide variety of HPC technical areas. The PET Team is available to help HPC users become more productive using HPCMP resources with the goal of expediting milestones and reducing overall costs related to technological development to DoD users.[11]

In addition to internal expertise, the PET Team also provides virtual training on effective usage of software, hardware, networking, data transfer, data storage, and visualization of data. PET facilitates and promotes the transfer of scientific and technical knowledge between HPCMP users and the broader computational communities, including DoD and other federal agencies. Help in specific Technical Thrust Areas (TTAs) provide a wide range of expertise including, but not limited to, new delivery methods, emerging hardware exploration, High Performance Data Analytics (HPDA), debugging software, performance improvement with user codes, scalability issues, and porting of software to emerging hardware.[11]

PET offers three modes of support depending on the type, duration, and level of user support required:

  • Mission Projects (MPs): Mission Projects are designed to assist users with solving complex issues in mission-critical areas. MPs will last anywhere from 1 to 12 months in duration. PET Team scientists will advise and assist in developing, applying, and/or using algorithms and in maximizing usage of existing software.
  • Special Projects (SPs): Concentrated efforts involving separately funded projects that may include PET Team support staff or other subcontractors for project/user specific challenges.
  • Training: Trains the HPC community in applicable tools and technologies using the latest collaboration, distance learning, and training technologies.[11]

Computational Research and Engineering Acquisition Tools and Environments

edit

CREATE is a federated program approved in the DoD 2006 Program Objective Memorandum (POM) process, with funding starting in 2008. The CREATE Program is part of the Tri-Service Department of Defense (DoD) High Performance Computing Modernization Program (HPCMP). The HPCMP is executed by the US Army Corps of Engineers (USACE) Engineer Research and Development Center (ERDC) Information Technology Laboratory (ITL) in Vicksburg, Mississippi.

The charter of CREATE is to reduce the cost, time, and risks of DoD acquisition programs by developing and deploying multidisciplinary, physics-based software applications for the design and analysis of military aircraft, naval ships, and radio frequency antenna systems (expanded to include ground vehicles in 2012) by DoD engineering organizations.

The CREATE software product portfolio was established to enable physics-based virtual prototyping and testing analysis for major DoD acquisition programs and include the following program areas:

  • Air vehicle design (AV)
  • Foundational technologies (FT)
  • Ground vehicle design (GV)
  • Radio frequency antenna design (RF)
  • Ship design (SH)
  • Educational software (Genesis)

CREATE-AV – Air Vehicle Products

Kestrel – Fixed-wing Design and Analysis

High-fidelity, multi-disciplinary analysis platform supporting a wide range of coupled physics to include: aerodynamics, thermochemistry, structural dynamics, thermodynamics, propulsion and flight controls.

Helios – Rotorcraft Design and Analysis

High-fidelity aero-structural software for coupled CFD/CSD analysis of rotary-wing aircraft: aerodynamics, structural dynamics and trim, maneuver, interactional aerodynamics, and air-launched effects.

ADAPT - Aircraft Design, Analysis, Performance, and Trade-space

Multi-disciplinary Analysis and Optimization (MDAO) environment enabling designers to support DoD fixed-wing aircraft pre-program requirements development.

CREATE-FT – Foundational Technologies Product

Capstone – Geometry, Mesh and Attribution Generation

Software platform to create, modify, and query geometry mesh and attribution information to define a numerical representation (digital model) for physics-based simulation of complex engineering systems.

CREATE-GV – Ground Vehicle Product

Mercury – Ground Vehicle Design and Analysis

Integrates the physics domains of vehicle dynamics, powertrain, tire-soil and track-soil interaction, and driver and control models, enabling multiple performance tests such as ride quality, discrete-obstacle shock, soft-soil mobility, sand slope climbing, maximum speed, lane change stability, and circular turn stability.

CREATE-RF – Radio Frequency Antenna Products

SENTRi – Antenna Modeling, Design and Analysis

Modeling of complex structures, including highly heterogeneous material structures with multi-scaled features to include: dielectric, magnetic, impedance boundary conditions, and complex valued resistive sheets.

FLO-K – Antenna Rapid Design Tool

Rapid design tool for periodic structures such as frequency selective surfaces, phased array antennas and band gap structures, including full three-dimensional structures with dielectric, magnetic, and impedance sheets.

Aurora – Antenna Pattern Prediction Tool

Accurate prediction of antenna patterns for complex platform integration analysis.

CREATE-SH – Ships Products

RSDE – Concept Design

Concept design tool that allows for the assessment of a spectrum of competing performance parameters in ship design to include: range, speed, armament, aviation support, varying hull form size and shape, systems, structures, powering, and payloads.

IHDE – Integration, Design and Analysis

Enables the integration of a suite of hull form design and analysis tools for users to evaluate hydrodynamic performance (including visualization) in areas of resistance, seakeeping, hydrodynamic loads, and operability in the form of percent time operable (PTO).

NavyFOAM – Ship Modeling and Analysis

Enables high-fidelity hydrodynamic analysis and prediction of ship performance, including resistance, propulsion, maneuvering, seakeeping, and seaway loads.

NESM – Multi-physics Toolkit

Extends the Department of Energy Sandia National Laboratory's multi-physics toolkit, Sierra Mechanics, to provide a means to assess ship and component response to external shock and blast.

ISDE – Structural Analysis

A multi-fidelity structural analysis environment to address the inadequacies in current structural design tools by providing a computational data-bridge between structural concept design, first order assessments, and detailed FEA while introducing the design capabilities needed for future Navy requirements.

CREATE-Genesis – Tool Suite for Aerospace Education

Genesis CFD – Fluid Dynamics Analysis

Provides basic computational fluid dynamics analysis capabilities to include: single mesh unstructured Navier-Stokes solver (core count constrained), motion (prescribed and 6-DOF), modal model based aeroelastic structures, and 0-D linear engine model for propulsion boundary conditions.

Capstone – Geometry, Mesh and Attribution Generation

Software platform to create, modify, and query geometry mesh and attribution information to define a numerical representation (digital model) for physics-based simulation of complex engineering systems.

ADAPT - Aircraft Design, Analysis, Performance, and Trade-space

Multi-disciplinary Analysis and Optimization (MDAO) environment enabling designers to support DoD fixed-wing aircraft pre-program requirements development.

Current and Former Program Directors

edit

The following table summarizes the chronology of program directors.[12][citation needed]

Name From To
1 Anthony Pressley[13] 1992 1995
2 Kay Howell [14] 1995 1997
3 Tom Dunn [15] 1997 1998
4 Charles Holland [15] 1998 1999
5 Cray Henry[16] 2000 Sep 2011
6 John West[17] Oct 2011 Dec 2014
7 David Horner[18] Jan 2015 Jan 2019
8 Will McMahon Jan 2019 Jan 2022
9 Jerry Ballard Jan 2022 Jan 2024
10 Kevin Newmeyer (Acting) Jan 2024 Present

The HPCMP in the DoD Budget

edit

Since FY2012, the HPCMP's base funding has been provided on two lines in the Army budget, which is itself a portion of the Department of Defense Budget submitted each year by the President and approved by Congress. PE 0603461A provides for RDT&E funds that operate the centers and DREN, and funds R&D efforts in support of program goals.[19] Line item number B66501 (line 103, BA 02, BSA 92) provides procurement funds for the annual purchase of new supercomputing hardware (both supercomputers and related systems).[20]

Prior to FY2012, the HPCMP's RDT&E funding was provided on PE 0603755D8Z,[21] while procurement was funded on PE 0902198D8Z (P011).[22]

The following table summarizes requested and committee-approved funding amounts for the RDT&E portion of the program for the most recent federal fiscal years (procurement funding, which is supplied on a different line in the federal budget, is not included in this table).

Requested and approved HPCMP funding amounts, RDT&E (in millions of dollars).
Fiscal Year Program Element Line President's Request Congress Approved Difference
2022 0603461A 42 $189.12 $229.12 +$40.0
2021 0603461A 64 $188.02 $228.02 +$40.0
2020 0603461A 60 $184.76 $224.76 +$40.0
2019 0603461A 46 $183.32 $218.32 +$35.0
2018 0603461A 45 $182.33 $221.33 +$39.0
2017[23] 0603461A 46 $177.190 $222.19 +$45.0
2016[24] 0603461A 46 $177.159 $222.16 +$45.0
2015[25] 0603461A 47 $181.609 $221.61 +$40.0
2014 0603461A 47 $180.66 $220.66 +$40.0
2013[26] 0603461A 47 $180.582 $228.18 +$47.6
2012[27] 0603461A 47 $183.150 $228.15 +$45.0
2011[28] 0603755D8Z 52 $200.986 $255.49 +$54.5
2010[29] 0603755D8Z 49 $221.286 $245.19 +$23.9
2009 0603755D8Z 49 $208.079[30] $220.345[31] +$12.266
2008[32] 0603755D8Z 50 $187.587 $208.487 +$20.9
2007[33] 0603755D8Z 43 $175.313 $207.213 +$31.9
2006[34] 0603755D8Z 45 $189.747 $213.247 +$23.5
2005[35] 0603755D8Z 42 $186.666 $236.766 +$50.1
2004 0603755F[36] Project 5093 $185.282[37] $202.492[38] +$17.21
2003[39] 0603755D8Z $188.642 $217.142 +$28.5
2002[40] 0603755D8Z $188.376 $187.200 -$1.2
2001[41] 0603755D8Z $164.027 $177.527 +$13.5
2000[42] 0603755D8Z $159.099 $168.099 +$9.0
1999[43] 0603755D8Z $140.927 $153.927 +$13.0
1998[44] 0603755D8Z $126.211 $149.880 +$23.67

The temporary change in Program Element number for FY2004 reflects a planned transition of the program from management by the Office of the Secretary of Defense to the Air Force; this transition did not ultimately occur.

References

edit
  1. ^ National Defense Authorization Act for Fiscal Years 1992 and 1993, Sec. 215 Supercomputer Modernization Program
  2. ^ The HPCMP provides World-class computational resources, a Nationwide research and engineering network, and vision, funding, and expertise to develop advanced physics-based computational analysis capabilities through the Department’s network of laboratories and warfare centers. "About HPCMP Public". High Performance Computing Modernization Program (HPCMP). The United States Department of Defense. 2013-08-29. Retrieved 2019-07-14.
  3. ^ "HPCMP website, Computational Technology Areas list". Archived from the original on 2015-07-23. Retrieved 2015-07-23.
  4. ^ "Spring 2011 HPCMP HPC Insights, p. 1" (PDF). Archived from the original (PDF) on 2015-07-22. Retrieved 2015-07-22.
  5. ^ "HPCMP Centers website". HPCMP Centers website. High Performance Computing Modernization Program. Retrieved 2019-07-11.
  6. ^ In 2011 the Chugach and Wiseman systems are listed under the Arctic Region Supercomputing Center on the HPCMP Consolidated Customer Assistance Center (CCAC) Hardware page along with systems from the other 5 DSRCs."HPCMP Consolidated Customer Assistance Center (CCAC) Hardware". Internet Archive Wayback Machine. The Internet Archive, a 501(c)(3) non-profit. Archived from the original on 2011-09-11. Retrieved 2019-07-11.
  7. ^ The Arctic Region Supercomputing Center (ARSC), currently one of six HPCMP centers, is slated to lose its Department of Defense (DoD) funding at the end of May 2011. Today ARSC is funded to the tune of $12 to $15 million, and the DoD slice represents around 95 percent of the total. The center’s main production machine is Chugach, a Cray XE6 ‘Baker’ supercomputer, which was part of a recent big procurement under HPCMP. That system has been moved to Vicksburg center and is being run remotely.Michael Feldman (2010-01-22). "Arctic Region Supercomputing Center Gets Cold Shoulder from DoD". HPC Wire. Tabor Communications. Retrieved 2019-07-06.
  8. ^ "HPCMP Networking Overview". Archived from the original on 2015-07-24. Retrieved 2015-07-23.
  9. ^ Eversden, Andrew (2021-06-18). "Verizon wins $495 million contract for DoD research network". C4ISRNet. Retrieved 2022-06-01.
  10. ^ "HPCMP HSAI web page". Archived from the original on 2015-07-24. Retrieved 2015-07-23.
  11. ^ a b c "HPC Centers: User Productivity and Enhancement Training (PET)". centerswww.afrl.hpc.mil. US Department of Defense. Retrieved 11 June 2022.   This article incorporates text from this source, which is in the public domain.
  12. ^ Private Communication, 07/23/2015
  13. ^ 50 Years of Army Computing: from ENIAC to MSRC, p .116
  14. ^ "DOD'S HPCMP director loves the cutting edge"FCW, 02 November 1997. Retrieved on 02 January 2016.
  15. ^ a b "Roster Change", FCW, 21 February 1999. Retrieved on 02 January 2016.
  16. ^ "HPCMP web archive, Cray Henry biography". Archived from the original on 2003-02-24. Retrieved 2008-05-19.
  17. ^ "DOD’s Supercomputing Program Relocates to Miss.", ERDC Press Release, 21 June 2012. Retrieved on 02 January 2016.
  18. ^ "HPCMP chief selected", ERDC Press Release, 12 March 2015. Retrieved on 02 January 2016.
  19. ^ "Department of the Army FY 2016 President's Budget Exhibit R-1, February 2015, p. A-4. Retrieved on 02 January 2016.
  20. ^ "Department of Defense Fiscal Year (FY) 2016 President's Budget Submission, Army Justification Book of Other Procurement, Army Communications and Electronics Equipment, Budget Activity 2", February 2015, p. 563. Retrieved on 02 January 2016.
  21. ^ "Office of the Secretary of Defense Fiscal Year (FY) 2011 Budget Estimates, Volume 3A, p. 359", February 2010. Retrieved on 02 January 2016.
  22. ^ "Office of the Secretary of Defense Fiscal Year (FY) 2011 Budget Estimates, Procurement, Defense-Wide", February 2010. P-1 Item Nomenclature: Major Equipment, OSD High Performance Computing Modernization Program (HPCMP) (P011). Retrieved on 02 January 2016.
  23. ^ FY2017 Omnibus Summary Department of Defense Appropriations, which points to funding tables in the written markup; page 206 in that document
  24. ^ S. Rept. 114-63 - 114th Congress (2015-2016)
  25. ^ S. Rept. 113-211 - 113th Congress (2013-2014)
  26. ^ S. Rept. 112-196 - 112th Congress (2011-2012)
  27. ^ S. Rept. 112-77 - 112th Congress (2011-2012)
  28. ^ S. Rept. 111-295 - 111th Congress (2009-2010)
  29. ^ S. Rept. 111-74 - 111th Congress (2009-2010)
  30. ^ FY2009 OSD RDT&E BUDGET ITEM JUSTIFICATION (R2 Exhibit)
  31. ^ FY2010 OSD RDT&E BUDGET ITEM JUSTIFICATION (R2 Exhibit)
  32. ^ S. Rept. 110-155 - 110th Congress (2007-2008)
  33. ^ S. Rept. 109-292 - 109th Congress (2005-2006)
  34. ^ S. Rept. 109-141 - 109th Congress (2005-2006)
  35. ^ S. Rept. 108-284 - 108th Congress (2003-2004)
  36. ^ FY2004 Exhibit R-2, RDT&E Budget Item Justification
  37. ^ DEPARTMENT OF THE AIR FORCE FISCAL YEAR (FY) 2004/2005 BIENNIAL BUDGET ESTIMATES RESEARCH, DEVELOPMENT, TEST AND EVALUATION (RDT&E) DESCRIPTIVE SUMMARIES, Vol 2
  38. ^ Fiscal Year (FY) 2005 Budget Estimates Exhibit R-2, RDT&E Budget Item Justification
  39. ^ S. Rept. 107-213 - 107th Congress (2001-2002)
  40. ^ S. Rept. 107-109 - 107th Congress (2001-2002)
  41. ^ S. Rept. 106-298 - 106th Congress (1999-2000)
  42. ^ H. Rept. 106-371 - 106th Congress (1999-2000)
  43. ^ H. Rept. 105-746 - 105th Congress (1997-1998)
  44. ^ H. Rept. 105-265 - 105th Congress (1997-1998)
edit

  This article incorporates public domain material from High Performance Computing Modernization Program. United States Department of Defense.