Skip to main content

Breadcrumb

  1. ACCESS Home
  2. Support
  3. Knowledge Base
  4. Knowledge Base Resources

Knowledge Base Resources

These resources are contributed by researchers, facilitators, engineers, and HPC admins. Please upvote resources you find useful!
Add a Resource

Topics

  • machine-learning (50)
  • ai (45)
  • training (41)
  • data-analysis (40)
  • deep-learning (28)
  • documentation (28)
  • big-data (26)
  • neural-networks (24)
  • workforce-development (21)
  • professional-development (18)
  • visualization (18)
  • parallelization (16)
  • community-outreach (14)
  • programming (14)
  • image-processing (13)
  • cybersecurity (12)
  • gpu (12)
  • r (12)
  • pytorch (11)
  • slurm (10)
  • c (9)
  • cloud-computing (9)
  • compiling (9)
  • mpi (9)
  • plotting (9)
  • administering-hpc (8)

Topics

  • machine-learning (50)
  • ai (45)
  • training (41)
  • data-analysis (40)
  • deep-learning (28)
  • documentation (28)
  • big-data (26)
  • neural-networks (24)
  • workforce-development (21)
  • professional-development (18)
  • visualization (18)
  • parallelization (16)
  • community-outreach (14)
  • programming (14)
  • image-processing (13)
  • cybersecurity (12)
  • gpu (12)
  • r (12)
  • pytorch (11)
  • slurm (10)
  • c (9)
  • cloud-computing (9)
  • compiling (9)
  • mpi (9)
  • plotting (9)
  • administering-hpc (8)

If you'd like to use more filters, please login to view them all.

Educause HEISC-800-171 Community Group
0
  • Educause HEISC-800-171 Community Group
The purpose of this group is to provide a forum to discuss NIST 800-171 compliance. Participants are encouraged to collaborate and share effective practices and resources that help higher education institutions prepare for and comply with the NIST 800-171 standard as it relates to Federal Student Aid (FSA), CMMC, DFARS, NIH, and NSF activities.
cybersecurity
0 Likes

Login to like
Type
website
Level
Beginner, Intermediate, Advanced
DAGMan for orchestrating complex workflows on HTC resources (High Throughput Computing)
0
  • DAGMan
  • DAGMan Workflows
DAGMan (Directed Acyclic Graph Manager) is a meta-scheduler for HTCondor. It manages dependencies between jobs at a higher level than the HTCondor Scheduler. It is a workflow management system developed by the High-Throughput Computing (HTC) community, specifically for managing large-scale scientific computations and data analysis tasks. It enables users to define complex workflows as directed acyclic graphs (DAGs). In a DAG, nodes represent individual computational tasks, and the directed edges represent dependencies between the tasks. DAGMan manages the execution of these tasks and ensures that they are executed in the correct order based on their dependencies. The primary purpose of DAGMan is to simplify the management of large-scale computations that consist of numerous interdependent tasks. By defining the dependencies between tasks in a DAG, users can easily express the order of execution and allow DAGMan to handle the scheduling and coordination of the tasks. This simplifies the development and execution of complex scientific workflows, making it easier to manage and track the progress of computations.
open-science-grid
0 Likes

Login to like
Type
tool
Level
Intermediate, Advanced
Learn Python Online
0
  • Python Courses Online
Learn Python online with these distance learning courses.
professional-developmenttrainingpython
0 Likes

Login to like
Type
website
Level
Beginner
Campus Research Computing Consortium (CaRCC)
0
  • CaRCC
CaRCC – the Campus Research Computing Consortium – is an organization of dedicated professionals developing, advocating for, and advancing campus research computing and data and associated professions. Vision: CaRCC advances the frontiers of research by improving the effectiveness of research computing and data (RCD) professionals, including their career development and visibility, and their ability to deliver services and resources for researchers. CaRCC connects RCD professionals and organizations around common objectives to increase knowledge sharing and enable continuous innovation in research computing and data capabilities.
community-outreachprofessional-developmentresearch-facilitationworkforce-development
0 Likes

Login to like
Type
website
Level
Beginner, Intermediate, Advanced
GIS: Projections and their distortions
0
  • Map Projections
In GIS, projections are helpful to take something plotted on a globe and convert it to a flat map that we can print or show on a screen. Unfortunately it also introduces distortions to the objects and features on the map. This not only distorts the objects visually, but the results for any spatial attribute calculations will also reflect this distortion (such as distance and area ). Below is a link to a quick primer on projections, types of distortions that can occur, and suggestions on how to choose a correct projection for your work.
gis
0 Likes

Login to like
Type
learning
Level
Beginner, Intermediate
Vulkan Support Survey across Systems
0
  • OSF hosted knowledge base submission
It's not uncommon to see beautiful visualizations in HPC center galleries, but the majority of these are either rendered off the HPC or created using programs that run on OpenGL or custom rasterization techniques. To put it simply the next generation of graphics provided by OpenGL's successor Vulkan is strangely absent in the super computing world. The aim of this survey of available resources is to determine the systems that can support Vulkan workflows and programs. This will assist users in getting past some of the first hurdles in using Vulkan in HPC contexts.
big-datacomputer-graphicsworkflow
0 Likes

Login to like
Type
learning
Level
Intermediate
Jetstream Home
0
  • Jetstream Website
Jetstream2 makes cutting-edge high-performance computing and software easy to use for your research regardless of your project’s scale—even if you have limited experience with supercomputing systems.Cloud-based and on-demand, the 24/7 system includes discipline-specific apps. You can even create virtual machines that look and feel like your lab workstation or home machine, with thousands of times the computing power.
jetstream
0 Likes

Login to like
Type
website
Level
Beginner, Intermediate, Advanced
Framework to help in scaling Machine Learning/Deep Learning/AI/NLP Models to Web Application level
0
  • Framework to help in scaling Machine Learning/Deep Learning/AI/NLP Models to Web Application level
This framework will help in scaling Machine Learning/Deep Learning/Artificial Intelligence/Natural Language Processing Models to Web Application level almost without any time.
aideep-learningmachine-learningneural-networks
0 Likes

Login to like
Type
learning
Level
Intermediate
Introduction to Probabilistic Graphical Models
0
  • https://ermongroup.github.io/cs228-notes/
This website summarizes the notes of Stanford's introductory course on probabilistic graphical models. It starts from the very basics and concludes by explaining from first principles the variational auto-encoder, an important probabilistic model that is also one of the most influential recent results in deep learning.
aimachine-learning
0 Likes

Login to like
Type
learning
Level
Beginner, Intermediate
High performance computing 101
0
  • High performance computing 101
An introductory guide to High Performance Computing.
administering-hpc
0 Likes

Login to like
Type
website
Level
Beginner
CyberAmbassadors: Professional Skills for Interdisciplinary Work
0
  • CyberAmbassador Program
The CyberAmbassadors project was funded through a workforce development grant from the National Science Foundation (Award #1730137). Starting in 2017, the initial focus of this project was to develop, test, and refine new curriculum to help CyberInfrastructure (CI) Professionals strengthen their communications, teamwork and leadership skills. With support and collaboration from a number of academic and professional organizations, the CyberAmbassadors project was expanded to offer professional skills training to college students and professionals working across STEM (science, technology, engineering, math) disciplines.
mentorshipprofessional-developmentresearch-facilitationtrainingworkforce-development
0 Likes

Login to like
Type
learning
Level
Beginner, Intermediate, Advanced
NCSA HPC-Moodle
0
  • NCSA HPC-Moodle
Self-paced tutorials on high-end computing topics such as parallel computing, multi-core performance, and performance tools. Some of the tutorials also offer digital badges.
trainingworkforce-development
0 Likes

Login to like
Type
learning
Level
Beginner, Intermediate, Advanced
ACCESS Support Portal
0
  • ACCESS Support Portal
affinity-grouppegasusACCESS-websiteopen-ondemand
0 Likes

Login to like
Type
website
Level
Beginner, Intermediate, Advanced
Introduction to MP
0
  • A “Hands-on” Introduction to OpenMP*
Open Multi-Processing, is an API designed to simplify the integration of parallelism in software development, particularly for applications running on multi-core processors and shared-memory systems. It is an important resource as it goes over what openMP and ways to work with it. It is especially important because it provides a straightforward way to express parallelism in code through pragma directives, making it easier to create parallel regions, parallelize loops, and define critical sections. The key benefit of OpenMP lies in its ease of use, automatic thread management, and portability across various compilers and platforms. For app development, especially in the context of mobile or desktop applications, OpenMP can enhance performance by leveraging the capabilities of modern multi-core processors. By parallelizing computationally intensive tasks, such as image processing, data analysis, or simulations, apps can run faster and more efficiently, providing a smoother user experience and taking full advantage of the available hardware resources. OpenMP's scalability allows apps to adapt to different hardware configurations, making it a valuable tool for developers aiming to optimize their software for a range of devices and platforms.
expansefastercc++compilingopenmpprogramming
0 Likes

Login to like
Type
presentation
Level
Intermediate
ACCESS KB Guide - DELTA
0
  • ACCESS KB Guide - DELTA
NCSA is the home of Delta, a computing and data resource that balances cutting-edge graphics processor and CPU architectures with a non-POSIX file system with a POSIX-like interface. Delta allows applications to reap the benefits of modern file systems without rewriting code.
delta
0 Likes

Login to like
Type
documentation
Level
Beginner, Intermediate, Advanced
Metadata Systems
0
  • Metadata Systems
Metadata is a vital topic in libraries and librarianship, encompassing structured information used for accessing digital resources. The definition of metadata varies but is essentially data about data. It has evolved beyond simply describing metadata schemas and now focuses on topics like interoperability, non-descriptive metadata (administrative and preservation metadata), and the effective application of metadata schemas for user discovery. Interoperability, the ability to seamlessly exchange metadata between systems, is a major concern. Different levels of interoperability are examined, including schema-level, record-level, and repository-level. Challenges to interoperability include variations in standards, collaboration barriers, and costs.Metadata management is discussed in terms of the holistic management of metadata across an entire library. Steps include analyzing metadata requirements, adopting schema, creating metadata content, delivery/access, evaluation, and maintenance. Administrative metadata, which encompasses ownership and production information, is becoming more critical, particularly for electronic resource licensing. Preservation metadata is also gaining importance in ensuring the long-term viability of digital objects.
metadata
0 Likes

Login to like
Type
learning
Level
Intermediate
Probabilistic Semantic Data Association for Collaborative Human-Robot Sensing
0
  • Probabilistic Semantic Data Association for Collaborative Human-Robot Sensing
Humans cannot always be treated as oracles for collaborative sensing. Robots thus need to maintain beliefs over unknown world states when receiving semantic data from humans, as well as account for possible discrepancies between human-provided data and these beliefs. To this end, this paper introduces the problem of semantic data association (SDA) in relation to conventional data association problems for sensor fusion. It then, develops a novel probabilistic semantic data association (PSDA) algorithm to rigorously address SDA in general settings. Simulations of a multi-object search task show that PSDA enables robust collaborative state estimation under a wide range of conditions.
aimachine-learning
0 Likes

Login to like
Type
documentation
Level
Advanced
Docker Tutorial for Beginners
0
  • Docker Tutorial for Beginners
A Docker tutorial for beginners is a course that teaches the basics of Docker, a containerization platform that allows you to package your application and its dependencies into a standardized unit for development, shipment, and deployment.
docker
0 Likes

Login to like
Type
video_link
Level
Beginner, Intermediate, Advanced
Jetstream2 Status
0
  • Jetstream2 Status
Jetstream2 makes cutting-edge high-performance computing and software easy to use for your research regardless of your project’s scale—even if you have limited experience with supercomputing systems.Cloud-based and on-demand, the 24/7 system includes discipline-specific apps. You can even create virtual machines that look and feel like your lab workstation or home machine, with thousands of times the computing power.
jetstream
0 Likes

Login to like
Type
website
Level
Beginner, Intermediate, Advanced
CUDA Toolkit Documentation
0
  • CUDA Toolkit Documentation
NVIDIA CUDA Toolkit Documentation: If you are working with GPUs in HPC, the NVIDIA CUDA Toolkit is essential. You can access the CUDA Toolkit documentation, including programming guides and API references, at this provided website
documentationcc++fortranpython
0 Likes

Login to like
Type
documentation
Level
Intermediate, Advanced
What are LSTMs?
0
  • Introduction to LSTMs
This reading will explain what a long short-term memory neural network is. LSTMs are a type of neural networks that rely on both past and present data to make decisions about future data. It relies on loops back to previous data to make such decisions. This makes LSTMs very good for predicting time-dependent behavior.
aideep-learningmachine-learningneural-networks
0 Likes

Login to like
Type
learning
Level
Intermediate, Advanced
Benchmarking with a cross-platform open-source flow solver, PyFR
0
  • PyFR documentation
  • PyFR source code from Github
  • Discourse channel for discussions and help
What is PyFR and how does it solve fluid flow problems? PyFR is an open-source Computational Fluid Dynamics (CFD) solver that is based on Python and employs the high-order Flux Reconstruction technique. It effectively solves fluid flow problems by utilizing streaming architectures, making it suitable for complex fluid dynamics simulations. How does PyFR achieve scalability on clusters with CPUs and GPUs? PyFR achieves scalability by leveraging distributed memory parallelism through the Message Passing Interface (MPI). It implements persistent, non-blocking MPI requests using point-to-point (P2P) communication and organizes kernel calls to enable local computations while exchanging ghost states. This design approach allows PyFR to efficiently operate on clusters with heterogeneous architectures, combining CPUs and GPUs. Why is PyFR valuable for benchmarking clusters? PyFR's exceptional performance has been recognized by its selection as a finalist in the ACM Gordon Bell Prize for High-Performance Computing. It demonstrates strong-scaling capabilities by effectively utilizing low-latency inter-GPU communication and achieving strong-scaling on unstructured grids. PyFR has been successfully benchmarked with up to 18,000 NVIDIA K20X GPUs on Titan, showcasing its efficiency in handling large-scale simulations.
finite-element-analysisbenchmarkingparallelizationgithubfluid-dynamicsopenmpic++cudampi
0 Likes

Login to like
Type
tool
Level
Intermediate
Astronomy data analysis with astropy
0
  • astropy
Astropy is a community-driven package that offers core functionalities needed for astrophysical computations and data analysis. From coordinate transformations to time and date handling, unit conversions, and cosmological calculations, Astropy ensures that astronomers can focus on their research without getting bogged down by the intricacies of programming. This guide walks you through practical usage of astropy from CCD data reduction to computing galactic orbits of stars.
visualizationimage-processingastrophysics
0 Likes

Login to like
Type
learning
Level
Intermediate, Advanced

Pagination

  • First page « First
  • Previous page ‹‹
  • …
  • Page 3
  • Page 4
  • Page 5
  • Page 6
  • Current page 7
  • Page 8
  • Page 9
  • Page 10
  • Page 11
  • …
  • Next page ››
  • Last page Last »