Agentic Tools and Coding with AI/LLMs
Explore how to integrate NRP-managed LMs and AI agents into JupyterHub, Coder and your own namespace.
Explore how to integrate NRP-managed LMs and AI agents into JupyterHub, Coder and your own namespace. Presented on September 16, 2025
This event will help users to make the most effective use of Expanse’s AMD EPYC processors. Topics include an introduction to the EPYC architecture, AMD compilers and math libraries, strategies for mapping processes and tasks to compute cores, Slurm, application tuning and profiling tools. Presented on April 21, 2021
Foundations of containers and orchestration: Pods, Deployments, PVCs, and S3 storage. Presented on September 9, 2025
A brief introduction to the Linux scheduler, how to interact with it, and run your research workloads on your personal computer, a shared workstation, or even a high-performance computing system. Presented on February 6, 2025
How to schedule your batch jobs on high-performance computing systems using the Slurm Workload Manager. Presented on August 21, 2025
This webinar will cover practices that projects at all scales can adopt to promote using the FAIR data principles in their work. I will demonstrate how to apply these practices by deploying a data portal using the Serverless Research Data Repository (SRDR) model, building on free and allocated resources like Globus collections and GitHub repositories. During the webinar, participants will learn the basic components of a research data repository, FAIR data guidelines, dataset elements including files, metadata, identifiers, and landing pages, the role of machine-readable metadata in data discovery, and how to apply access control policies to datasets. Presented on May 13, 2025
Introduction to porting codes and workflows to HPC resources. Presented on June 12, 2025
A brief introduction on how to schedule your batch jobs on high-performance computing systems using the Slurm Workload Manager. Presented on March 21, 2024
Introduction to porting codes and workflows to HPC resources. Presented on October 3, 2024
Introduction to porting codes and workflows to HPC resources. Presented on March 7, 2024
How to use the data storage and file systems you’ll find mounted on high-performance computing systems. Presented on December 12, 2024
How to get the data you need for your research to and from high-performance computing systems. Presented on March 6, 2025
How to get the data you need for your research to and from high-performance computing systems. Presented on June 6, 2024
How to build and run your high-throughput and many-task computing workflows on high-performance computing systems using the Slurm Workload Manager. Presented on October 17, 2024
A brief introduction into what makes up a HPC system and how users should use this information. No programming required. Presented on November 7, 2024
A brief introduction into what makes up a HPC system and how users should use this information. No programming required. Presented on April 4, 2024
Discussion on best practices for using HPC systems and getting support Presented on August 15, 2024
Discussion on best practices for using HPC systems and getting support Presented on February 15, 2024
Interactive high-performance computing (HPC) involves real-time user inputs that result in actions being performed on HPC compute nodes. This session presents an overview of interactive computing tools and methods. Presented on December 5, 2024
Interactive high-performance computing (HPC) involves real-time user inputs that result in actions being performed on HPC compute nodes. This session presents an overview of interactive computing tools and methods. Presented on April 18, 2024
A survey of intermediate Linux skills for effectively using advanced cyberinfrastructure. Presented on September 19, 2024
A survey of intermediate Linux skills for effectively using advanced cyberinfrastructure. Presented on May 16, 2024
A survey of intermediate Linux skills for effectively using advanced cyberinfrastructure. Presented on January 18, 2024
An overview of commonly used Linux tools for searching and manipulating text. Presented on July 18, 2024
An overview of commonly used Linux tools for searching and manipulating text. Presented on February 1, 2024
A brief introduction to fundamental concepts in parallel computing. No programming experience needed. Presented on January 9, 2025
A brief introduction to fundamental concepts in parallel computing. No programming experience needed. Presented on September 5, 2024
A brief introduction to fundamental concepts in parallel computing. No programming experience needed. Presented on May 2, 2024
A brief introduction to fundamental concepts in parallel computing for anyone who uses HPC resources. Presented on January 4, 2024
The CIML Summer Institute will involve introducing ML researchers, developers and educators to the techniques and methods needed to migrate their ML applications from smaller, locally run resources, such as laptops and workstations, to large-scale HPC systems, such as the SDSC Expanse supercomputer. Presented on June 27, 2022
The CIML Summer Institute will involve introducing ML researchers, developers and educators to the techniques and methods needed to migrate their ML applications from smaller, locally run resources, such as laptops and workstations, to large-scale HPC systems, such as the SDSC Expanse supercomputer. Presented on June 22, 2021
Managing data efficiently on a supercomputer is important from both users' and system's perspectives. In this webinar, we will cover a few basic data management techniques and I/O best practices in the context of the Expanse system at SDSC. Presented on March 16, 2023
How to use the data storage and file systems you’ll find mounted on high-performance computing systems. Presented on May 15, 2025
How to get the data you need for your research to and from high-performance computing systems. Presented on September 18, 2025
Learn how to deploy and manage JupyterHub and Coder environments for groups or courses. Presented on September 23, 2025
This webinar covers the basics of accessing the SDSC Expanse supercomputer, managing the user environment, compiling and running jobs on Expanse. Presented on October 8, 2020
This webinar covers the basics of accessing SDSC's Expanse supercomputer, managing the user environment, compiling and running jobs using Slurm, where to run them, and how to run batch jobs. We will also cover interactive computing using applications such as Jupyter Notebooks and how to run them via the command line or from the Expanse portal. It is assumed that you have mastered the basic skills of logging onto HPC systems using SSH and running basic Unix commands on these systems. Presented on February 17, 2022
This webinar covers the basics of accessing SDSC's Expanse supercomputer, managing the user environment, compiling and running jobs using Slurm, where to run them, and how to run batch jobs. Presented on September 16, 2021
Come learn about Singularity containers and how you might use them in your own work. Presented on January 21, 2021
This webinar will present the approach and the architecture of the composable systems component of Expanse. We will also summarize scientific case studies that demonstrate the application of this new infrastructure and its federation with Nautilus, a Kubernetes-based GPU geo-distributed cluster. Presented on April 15, 2021
Managing data efficiently on a supercomputer is very important from both users' and system's perspectives. In this webinar, we will cover some of the basic data management techniques, I/O best practices in the context of the Expanse system at SDSC. Presented on October 21, 2021
The first in a recurring webinar series on using Expanse and other SDSC HPC resources securely. This webinar will cover security and security-related topics relevant to researchers and the trustworthiness of their work produced on these resources. Presented on April 21, 2022
This webinar will give a brief introduction to GPU computing and programming on Expanse. We will cover the GPU architecture, programming with the NVIDIA HPC SDK via libraries, OpenACC compiler directives, CUDA, profiling and debugging, and submitting GPU enabled jobs on Expanse. Presented on May 20, 2021
Come learn how to launch your Jupyter notebook sessions on Expanse in a simple, secure way. Presented on December 14, 2021
This webinar will be a quick introduction and overview of neural networks, convolution networks, and deep learning on Expanse. Presented on May 19, 2022
In this webinar we cover supercomputer architectures, the differences between threads and processes, implementations of parallelism (e.g., OpenMP and MPI), strong and weak scaling, limitations on scalability (Amdahl’s and Gustafson’s Laws) and benchmarking. Presented on September 14, 2022
In this webinar we cover supercomputer architectures, the differences between threads and processes, implementations of parallelism (e.g., OpenMP and MPI), strong and weak scaling, limitations on scalability (Amdahl’s and Gustafson’s Laws) and benchmarking. Presented on January 20, 2022
Presentation will cover cache-level optimizations and other techniques for achieving optimal software performance. We will also cover AMD specific compiler options, libraries and performance tools. Presented on February 18, 2021
In this webinar, we will present SDSC’s multitiered approach to running notebooks more securely: hosting Jupyter services on Expanse using SSH Tunneling or using the SDSC Jupyter Reverse Proxy Service (JRPS), which connects the user over an HTTPS connection. The JRPS will launch a batch script that creates a securely hosted HTTPS access point for the user, resulting in a safer, more secure notebook environment. Presented on December 10, 2020
In this webinar we demonstrate how to transition your Jupyter Notebooks from a local machine to the Expanse HPC system using command-line tools and the Expanse Portal. We cover creating transferable software environments, scaling up calculations to large datasets, parallel processing, and running Jupyter Notebooks in batch mode. Presented on June 16, 2022
In this webinar we provide recipes for transitioning scientific workloads that currently run on traditional batch systems to Kubernetes systems. Kubernetes is batch-like in nature, but there are some differences that science users should be aware of. We will also briefly describe capabilities that are not found in traditional batch systems that can improve the effectiveness of scientific computing. Presented on October 20, 2022
Come learn all about Singularity containers. In this webinar, we'll provide an overview of Singularity and how you might incorporate the use of containers in your own research. We'll also show you how to access and use some of the containerized applications that we make available to users on Expanse at SDSC. Presented on March 17, 2022
Learn how to write your first batch job script and submit it to a Slurm batch job scheduler. We discuss best practices on how to structure your batch job scripts, teach you how to leverage Slurm environment variables, and provide you with some tips on how to request resources from the scheduler to get your work done faster. We also introduce you to some advanced features like Slurm job arrays and job dependencies for more structured computational workflows. Presented on February 16, 2023
Use GitLab pipelines to automate image building and deploy custom JupyterHub setups. Presented on September 30, 2025
This webinar gives a brief introduction to GPU computing and programming on Expanse. We will cover the GPU architecture, programming with the Nvidia CUDA Toolkit and HPC SDK via libraries, OpenACC compiler directives, and CUDA, and submitting GPU enabled jobs on Expanse. Presented on October 26, 2023
This year’s Summer Institute continues SDSC’s strategy of bringing HPC to the “long tail of science,” i.e., providing resources to a larger number of modest-sized computational research projects that represent, in aggregate, a tremendous amount of scientific research and discovery. Presented on August 2, 2021
The HPC and Data Science Summer Institute is a week-long workshop focusing on a broad spectrum of introductory-to-intermediate topics in High Performance Computing and Data Science. The program is aimed at researchers in academia and industry, especially in domains not traditionally engaged in supercomputing, who have problems that cannot typically be solved using local computing resources. Presented on August 1, 2022
How to build and run your high-throughput and many-task computing workflows on high-performance computing systems using the Slurm Workload Manager. Presented on December 4, 2025
SDSC’s High Performance Computing (HPC)/ Cyberinfrastructure (CI) Training Series was developed to support UC San Diego undergraduates and graduates interested in furthering their knowledge of HPC concepts and hands-on training, as well as, building a team interested in competing in the Student Cluster Competition held at the annual International Conference for High Performance Computing, Networking, Storage, and Analysis (SC). This program is available to any who are interested in advancing their knowledge and experience on HPC systems and concepts. Presented on January 14, 2022
The mini-workshop is designed to provide the UCSD research community with a streamlined pathway to swiftly engage with the Expanse cluster for their scientific endeavors. Collaboratively organized by Research IT and SDSC, this workshop series offers participants the opportunity to start using the Expanse cluster through the Campus Champions allocation, while benefiting from comprehensive training resources and expert guidance provided by SDSC. Presented on October 3, 2023
Implement a practical and well supported data management plan for your research lab, project or grant with SeedMeLab. Presented on April 28, 2021
Interactive high-performance computing (HPC) involves real-time user inputs that result in actions being performed on HPC compute nodes. This session presents an overview of interactive computing tools and methods. Presented on March 20, 2025
Interactive high-performance computing (HPC) involves real-time user inputs that result in actions being performed on HPC compute nodes. This session presents an overview of interactive computing tools and methods. Presented on October 9, 2025
Interactive computing includes commonly used programs, such as word processors or spreadsheet applications running user devices (mobile phones, laptops). Interactive high-performance computing (HPC) involves real-time user inputs that result in actions being performed on HPC compute nodes. In this session we’ll present an overview of interactive computing tools and methods. Presented on December 7, 2023
Scripting with kubectl, exposing services, using taints/tolerations, node affinity, and Python APIs. Presented on September 16, 2025
A survey of intermediate Linux skills for effectively using advanced cyberinfrastructure. Presented on January 23, 2025
A survey of intermediate Linux skills for effectively using advanced cyberinfrastructure. Presented on July 10, 2025
This webinar will be a quick introduction and overview of neural networks, convolution networks, and demonstration of executing deep learning models in an HPC environment. Presented on April 20, 2023
In this webinar, Yuwu Chen from TSCC User Services will show how to build Singularity images and then run them on the SDSC supercomputer clusters such as TSCC. Yuwu will also be sharing his insider knowledge of best practices along with pitfalls to avoid while working with Singularity. Presented on October 13, 2022
Overview of NRP's mission, capabilities, how to gain access, and policy guidelines. Presented on September 2, 2025
This training will cover everything users need to know about the new TSCC 2.0 system that will be launched in phases starting late spring. Topics will include changes to the TSCC system, scheduler, queues, software stack, accounting, and policies for using TSCC. Presented on April 13, 2023
Several new scientific compute resources are becoming available only through Kubernetes and their users will have to adapt their workloads to interface to it. This tutorial provides the basic Kubernetes notions any science user will need, paired with extensive hands-on exercises on a production-quality system to better explore the details. Presented on May 4, 2022
An introduction to shell scripting Presented on October 23, 2025
An introduction to shell scripting. Presented on April 17, 2025
An overview of commonly used Linux tools for searching and manipulating text. Presented on February 20, 2025
An overview of commonly used Linux tools for searching and manipulating text. Presented on September 4, 2025
In this session you will learn how to solve and accelerate computationally and data-intensive problems that are becoming common in the areas of machine learning and deep learning using multicore processors, GPUs, and computer clusters. Presented on April 27, 2022
In this webinar we cover supercomputer architectures, the differences between threads and processes, implementations of parallelism (e.g., OpenMP and MPI), strong and weak scaling, limitations on scalability (Amdahl’s and Gustafson’s Laws) and benchmarking. Presented on January 19, 2023
A PEARC25 Workshop aimed at bringing together resource providers, researchers, software engineers and coding science users that are interested in using and supporting applications that benefit from the concurrent use of CPU and GPU resources, like those on the SDSC Cosmos and TACC Vista systems. The event consisted of both short talks and panel discussions, which made the event truly engaging. Presented on July 21, 2025
Slides by Mahidhar Tatineni, SDSCSlides by Amit Ruhela, TACCSlides by Toshihiro Hanawa, JCAHPCSlides by Dan Stanzione, TACC
This webinar is a condensed version of the Performance Tuning session, targeting code developers who want to speed up their calculations. It will cover various optimization techniques, including cache usage, loop optimization, and compiler optimization, with exercises primarily in C but applicable to any language. Presented on February 11, 2025
This session is intended for attendees who do their own code development and need their calculations to finish as quickly as possible. We cover effective use of cache, loop-level optimizations, and other topics for writing and building optimal code. Presented on September 21, 2023
SYCL is an open standard for programming heterogeneous architectures in ISO C++. This webinar will give a high-level overview of the SYCL programming language and the software ecosystem to write and tune SYCL code for different accelerator architectures. We will focus on GPUs and discuss how we can provide SYCL performance portability across hardware from different vendors, including Intel, Nvidia, and AMD GPUs, by employing a single-source model based on a modern C++ standard. As an example of complex scientific software, we will demonstrate briefly how the Amber molecular dynamics software was ported from CUDA to SYCL using Intel oneAPI software development tools and Intel Xe architecture GPUs. We will discuss numerical results and benchmark data that demonstrate the accuracy and performance of the SYCL implementation on data center and consumer-grade GPU hardware. Presented on November 4, 2025
Voyager offers a unique architecture to optimize and scale deep learning applications. In this tutorial we will introduce Voyager and its Gaudi architecture. Then we will take a Pytorch application and port it to the Voyager system. Finally, how to run any Huggingface models to Voyager will be also discussed. Presented on April 8, 2025
In this session, we will explore two transformative Python technologies—Numba and Dask—that empower researchers to bridge the gap between Python’s flexibility and the performance demands of supercomputing environments. These tools unlock new possibilities for accelerating computationally intensive tasks and scaling workflows across clusters. Presented on September 9, 2025
We will focus on (1) R package management and user environment configuration on the HPC cluster and (2) understanding the R parallelism in HPC, such as using the "parallel" package in R and a few related packages to parallelize and enhance the performance of R programs. Presented on December 9, 2025
This free webinar will introduce HPCShare, a web-based resource for users of SDSC’s high-performance computing resources, including Expanse, to easily share small-to medium-scale datasets in an efficient and organized manner. Attendees will learn about using HPCShare and SDSC’s SeedMeLab scientific data management system. Hosted by SDSC Visualization Group Lead Amit Chourasia. Presented on April 22, 2021
In this webinar, we demonstrate how to transition your Jupyter Notebooks from a local machine to the Expanse HPC system using command-line tools and the Expanse Portal. We cover creating transferable software environments, scaling up calculations to large datasets, parallel processing, and running Jupyter Notebooks in batch mode. Presented on June 15, 2023
This webinar is an introduction to performing machine learning at scale. An overview of approaches for parallelizing R code on HPC will be provided. We will also cover the essentials of Spark and demonstrate how to use Spark for large-scale data analytics and machine learning. Demonstrations will allow participants to gain practical guidance for building and scaling machine learning workflows. Presented on January 14, 2025
In this webinar we provide recipes for transitioning scientific workloads that currently run on traditional batch systems to Kubernetes systems. Kubernetes is batch-like in nature, but there are some differences that science users should be aware of. We will also briefly describe capabilities that are not found in traditional batch systems that can improve the effectiveness of scientific computing. Presented on July 20, 2023
This webinar will briefly introduce how to build Singularity images and how to run them on the SDSC supercomputer clusters. We will also share some insider knowledge of best practices and pitfalls to avoid while working with Singularity. Presented on May 18, 2023
SDSC's newest Supercomputer, Expanse, supports SDSC's vision of 'Computing without Boundaries' by increasing the capacity and performance for thousands of users of batch-oriented and science gateway computing, and by providing new capabilities that will enable research increasingly dependent upon heterogeneous and distributed resources composed into integrated and highly usable cyberinfrastructure. It also implements new technical capabilities such as Direct Liquid Cooling. SDSC has acquired additional capacity for Expanse specifically to support industrial research and collaborations. Presented on April 15, 2021
Navigate the NRP portal, manage namespaces/groups, and explore the integrated LiteLLM interface. Presented on September 2, 2025
The talk will highlight the advances in the TAU Performance System(R). It will describe the different instrumentation, measurement and analysis options that are available and how TAU is integrated in the Extreme-scale Scientific Software Stack (E4S). A hands-on demo on AWS with E4S will be shown. Presented on March 11, 2025
This training will cover everything new users need to know about using the TSCC system. Topics will include: an overview of condo/hotel program; how to apply; accounts and allocation usage monitoring; environment and software modules; overview of various queues, building PBS job scripts, job submission and monitoring; data transfers; and file systems. Presented on March 3, 2022
During this workshop, we will provide an overview of TSCC 2.0 including the new authentication method, new allocation system, new filesystems, shared data transfer options from the current TSCC to TSCC 2.0, software stack, new partition characteristics, and provide examples of SLURM job scripts. Presented on November 6, 2023
During this workshop, we will provide an overview of TSCC, including authentication, allocation, filesystems, software stack, partition characteristics, and job submission, with examples of SLURM job scripts. Presented on September 12, 2024
Launch and use interactive development environments. Ideal for researchers, students, and educators. Presented on September 9, 2025
This workshop will focus on providing guidelines for setting up customized Python environments, how to install and manage packages using Miniconda/pip, and how to run secure Jupyter notebooks on Triton Shared Computing Cluster (TSCC) HPC system. Presented on September 2, 2021
Essentials of using regular expressions with the Linux tools grep, awk and sed. Presented on November 6, 2025
Essentials of using regular expressions with the Linux tools grep, awk and sed Presented on May 29, 2025
Launch and connect to VMs (Linux/Windows), configure networking, and explore advanced edge scenarios. Presented on September 30, 2025
Voyager offers a unique architecture to optimize and scale AI applications. In this tutorial we will introduce Voyager and its Gaudi architecture. We will show how to port a Pytorch application and deploy it into the Voyager system using Kubernetes. Some examples including Jupyter Notebooks and Huggingface models (Diffusers, LLMs) will be presented. Presented on October 14, 2025
This is the first of a two-part Voyager training session. Voyager is based on Intel’s Habana Lab AI processors and provides a unique opportunity to use AI focused hardware for exploring AI in science and engineering. Voyager features Habana’s Gaudi processors optimized for training, Goya processors optimized for inference, 100 GbE all-to-all connection within Gaudi nodes, 24 x 100GbE RDMA RoCE for scale-out across Gaudi nodes, and a Ceph file system. Presented on October 18, 2022
This is the second of a two-part Voyager training session. Voyager is based on Intel’s Habana Lab AI processors and provides a unique opportunity to use AI focused hardware for exploring AI in science and engineering. Voyager features Habana’s Gaudi processors optimized for training, Goya processors optimized for inference, 100 GbE all-to-all connection within Gaudi nodes, 24 x 100GbE RDMA RoCE for scale-out across Gaudi nodes, and a Ceph file system. Presented on November 8, 2022