The mission of the Center for Scientific Computation at Southern Methodist University is to stimulate interdisciplinary education and research in simulation-based engineering and science.
We are motivated by the fact that computer simulation has become an essential component of research in most disciplines in engineering and science, and that advances in computing, networking, and data storage technologies are likely to accelerate this trend. At the same time, it is recognized that our nation's position of dominance in this area can no longer be taken for granted. Thus we believe that the education of the next generation of computational scientists and engineers is an urgent challenge which SMU must accept.
Goals of the CSC
- Education and training focused on high-performance computing algorithms, software, and hardware.
- Formation of interdisciplinary research teams addressing cutting-edge applications.
- Rapid communication of faculty and student research accomplishments.
ManeFrame II (M2)
SMU’s new high-performance compute cluster will dramatically increase the computational capacity and performance that SMU provides to its researchers. The new cluster features state of the art CPUs, accelerators, and networking technologies, significantly more memory per node, and advanced interactive GPU-accelerated remote desktop experiences. Also, the cluster is much more energy efficient making it more economical to run and more environmentally friendly!
The new cluster will provide a similar interactive experience for researchers currently using ManeFrame. Similarities include the CentOS 7 operating system (replacing Scientific Linux 6; both are Red Hat Enterprise Linux derivatives), the SLURM resource scheduler, and the Lmod environment module system. Additionally, updated, but familiar, development tool chains will be available including the GCC, Intel, and PGI compiler suites. Optimized high-level programming environments such as MATLAB, Python, and R will also be installed in addition to the domain specific software packages that SMU researchers depend on for their work.
More information on ManeFrame II and use can be found here. If you currently use ManeFrame I, you can find more information about transitioning to ManeFrame II here.
|Number of Nodes
|Total CPU Cores
||29.2 TB (29,856 GB)
||116.5 TB (119,336 GB)
|Total Accelerator Cores
|Node Interconnect Bandwidth
||1.4 PB (1,229 TB)
||2.8 PB (2,867 TB)
||Scientific Linux 6
Request an Account
Faculty can request an account by sending an email to SMU HPC Admins with the following details. If you are a student or postdoc please ask your supervisor, sponsor, or adviser to send an email to SMU HPC Admins with the following details.
- Full Name
- A valid SMU email for SMU student/post-doc/faculty/staff.
- Duration for which the account should remain active
If you are an external research collaborator please have your sponsor request a sponsored account at firstname.lastname@example.org. Once your account has been created please request HPC access at SMU HPC Admins.
More information on account management can be found here.
Here you can can find information on how to effectively use SMU’s HPC resources. Topics covered in the documentation include:
For questions about using resources or setting up accounts please email the SMU HPC Admins with "HPC" in the subject line.
ManeFrame I Documentation
The documentation for ManeFrame I can be found here.
Summer 2017 HPC Workshop
This year the summer 2017 workshop will be given twice and focus exclusively on using ManeFrame II. The workshops will be largely identical except that second workshop will cover additional features of ManeFrame as they become available.
- August 7-11, 9:00 AM to 4:00 PM, Fondren Science Building Room 306A
- September 11-15, 9:00 AM to 4:00 PM, Location to be determined
Monday - Using ManeFrame II
- Introduction to High-Performance Computing and ManeFrame II
- Accessing ManeFrame II and Introduction to UNIX
- Using ManeFrame II
Tuesday - Scripting Introduction
- Programming Concepts
- Scripting with Bash and Python
- Jupyter Notebooks
Wednesday - Programming Introduction
Thursday - Parallel Programming Introduction
- Parallel Programming Concepts
- Programming with OpenMP and MPI
- Debugging and Profiling Parallel Programs
Friday - Heterogeneous Programming Introduction
- Heterogeneous Programming Concepts
- Programming Accelerators using Libraries, OpenACC, CUDA, and OpenCL
- Debugging and Profiling Heterogeneous Programs
Fall 2017 Workshop Series
The Fall 2017 workshop series will be announced shortly.
Report HPC Usage
Faculty are encouraged to report their usage of SMU’s HPC facilities here
. The data will be used to assess and document usage of these resources.