Submission information
Submission Number: 78
Submission ID: 110
Submission UUID: a21e9844-8ffc-4624-b397-865cb8a42151
Submission URI: /form/project
Created: Tue, 11/17/2020 - 14:42
Completed: Tue, 11/17/2020 - 15:24
Changed: Tue, 08/02/2022 - 15:05
Remote IP address: 98.235.167.220
Submitted by: Jeffrey J. Nucciarone
Language: English
Is draft: No
Webform: Project
Project Title | Amsterdam Modeling Suite Software Usage on Penn State’s Roar Computer Clusters |
---|---|
Program | CAREERS |
Project Image |
![]() |
Tags | batch-jobs (76), cluster-support (244), computational-chemistry (81), file-transfer (252), hpc-operations (43), job-sizing (42), management (334), pbs (169), permissions (177), provisioning (13), putty (250), quantum-mechanics (281), schedulers (164), scheduling (52), ssh (78), terminal-emulation-and-window-management (318) |
Status | Complete |
Project Leader | Kofi Adu |
cxa269@psu.edu | |
Mobile Phone | |
Work Phone | 814-940-3335 |
Mentor(s) | Jorge Sofo, Jeffrey J. Nucciarone |
Student-facilitator(s) | Mohamed Mauroof Umar |
Mentee(s) | |
Project Description | Penn State recently acquired licenses for the Amsterdam Modelling Suite (AMS) package. The AMS is a collection of software tools for performing computational chemistry calculations. The AMS-jobs graphical user interface serves as a job management interface. AMS-input is used to model the required structures. The AMS suite has native code to perform density functional theory (DFT) calculations. The ADF and BAND codes work on molecular and periodic systems. The AMS suite can also be used as a portal to create input files, submit jobs, and analyze results for other popular computational chemistry codes. Semi-empirical codes such as DFTB+ and plane-wave DFT codes such as Quantum Espresso (QE) and Vienna Ab-initio Simulation Package (VASP) calculations can be performed using the AMS suite. Additionally, with the setup of a proper secure shell (ssh) portal and defining the necessary queues in the AMS-jobs interface, remote jobs can be submitted directly from the AMS suite. Further AMS-jobs acts as a convenient interface to monitor the remote-jobs, handles the transfer of input files to the remote server and on completion of the job automatically retrieves the relevant output files. Analysis of results can be done conveniently with the various visualization tools included in the AMS suite. These convenient factors make the AMS suite a promising candidate to simplify the workflow of computational chemistry research projects. This project aims to document the necessary steps to connect the AMS suite with Penn State’s Roar supercomputer. Calculations and analysis of results for a current research project on inversion domain boundaries in Wurzite crystal structure will be performed as a proof of concept. |
Project Deliverables | Documentation to assist new users of the AMS suite with Roar Supercomputer. These instructions can serve as a base for other institutions to set up AMS in a similar fashion, providing desktop users the ability to interface directly with central HPC resources. Results for publication on inversion domain boundaries in Wurzite crystal structure. |
Project Deliverables | |
Student Research Computing Facilitator Profile | Graduate student in physics/materials science |
Mentee Research Computing Profile | |
Student Facilitator Programming Skill Level | Practical applications |
Mentee Programming Skill Level | |
Project Institution | Penn State University - Altoona Campus |
Project Address | 3000 Ivyside Park Altoona, Pennsylvania. 16601 |
Anchor Institution | CR-Penn State |
Preferred Start Date | 12/01/2020 |
Start as soon as possible. | No |
Project Urgency | Already behind3Start date is flexible |
Expected Project Duration (in months) | 6 |
Launch Presentation | |
Launch Presentation Date | 03/10/2021 |
Wrap Presentation | |
Wrap Presentation Date | 01/12/2022 |
Project Milestones |
|
Github Contributions | |
Planned Portal Contributions (if any) | |
Planned Publications (if any) | Inversion Domain Boundaries in Wurzite Crystal Structure. |
What will the student learn? | Submitting and managing jobs on the remote server with external tools. A scientific approach to analyzing complex data. |
What will the mentee learn? | |
What will the Cyberteam program learn from this project? | Potential democratization of HPC resources by showing how a desktop application can be seamlessly integrated into a central HPC resource. |
HPC resources needed to complete this project? | Access to Penn State ICDS ROAR HPC system |
Notes | |
What is the impact on the development of the principal discipline(s) of the project? | The project developed a method to leverage computing from a laptop to a remote-based back-end HPC cluster without the end-user needing to run through multiple login steps or manual file transfers. |
What is the impact on other disciplines? | The integration of other scientific computing applications that have a native front end to integrate with back-end HPC systems without relying on other portals. |
Is there an impact physical resources that form infrastructure? | Other projects that use a similar approach to extending local to HPC resources have the potential to reduce the requirements for large front-end interactive systems needed to support classroom or other mass-login uses. The use of the vendors' own front end requires the queuing system for the HPC resources to be set up to support these multiple remote requests. |
Is there an impact on the development of human resources for research computing? | There is a lower entry barrier to getting started as incoming students or other researchers do not have to be excessively fluent in how the back-end resources operate. |
Is there an impact on institutional resources that form infrastructure? | Implementation and system engineers will have to understand how these types of applications will integrate with the HPC systems. The current nature of some of these applications is not as mature as we would like it to be, resulting in a lot of technical assistance to be provided in its current form. This project helped expose some of these weaknesses. |
Is there an impact on information resources that form infrastructure? | See above. |
Is there an impact on technology transfer? | By simplifying and unifying integration from local to remote HPC resources we can make it easier to share results across multiple disciplines and allow greater sharing and collaboration. |
Is there an impact on society beyond science and technology? | This project helped lay the groundwork for the further democratization of high-performance computing resources, bringing it out of its more traditional "cloistered" environment of login nodes and captive portals and directly into end-use, locally run, and managed applications. |
Lessons Learned | The original goals of this project were more ambitious than we realized. The vendor's software suite did not provide for a robust enough interface to the PSU HPC cluster queueing and job management system, as the site used a non-standard way of interfacing the management and queuing systems. These "extra" messages generated by the queue system were enough to confuse the vendor's API. Given more time we could have solved the issue but this part was moving beyond the original scope of the project. The project did provide us with a good feasibility study where we can continue to pursue the original approach. |
Overall results | Despite some of the technical limitations the project encountered, we demonstrated that a local application run on a laptop could be extended to use available HPC resources to run problems more complex than can be handled by a local system. The researcher can set up and test their models locally and then seamlessly run on the back-end HPC system, and then automatically have the results return to the local system for further analysis. |