Search Documentation

Search Documentation

Owens
As of Monday, Jan. 6, 2025, the Owens high performance computing (HPC) cluster has been partially decommissioned. OSC has moved two-thirds of the regular compute nodes and one-half of the GPU nodes (a total of about 60% of the cluster cores) on the Owens cluster offline. The remainder of the Owens nodes will power down on Monday, Feb. 3, 2025.
TIP: Remember to check the menu to the right of the page for related pages with more information about Owens' specifics.

OSC's Owens cluster being installed in 2016 is a Dell-built, Intel® Xeon® processor-based supercomputer.

2024_0903 Owens Cluster Graphic Update.png


Pitzer

Linaro MAP is a full scale profiler for HPC programs. We recommend using Linaro MAP after reviewing reports from Linaro Performance Reports. MAP supports pthreads, OpenMP, and MPI software on CPU, GPU, and MIC based architectures.

Ascend, Pitzer

Linaro Performance Reports is a simple tool used to generate a single-page HTML or plain text report that presents the overall performance characteristics of HPC applications. It supports pthreads, OpenMP, or MPI code on CPU, GPU, and MIC based architectures.

Many software packages require a license.  These licenses are usually made available via a license server, which allows software to check out necessary licenses.  In this document external refers to a license server that is not hosted inside OSC.

If you have such a software license server set up using a license manager, such as FlexNet, this guide will instruct you on the necessary steps to connect to and use the licenses at OSC.

Users who wish to host their software licenses inside OSC should consult OSC Help.

 

Cardinal

The Schrodinger molecular modeling software suite includes a number of popular programs focused on drug design and materials science but of general applicability, for example Glide, Jaguar, and MacroModel.  Maestro is the graphical user interface for the suite.  It allows the user to construct and graphically manipulate both simple and complex chemical structures, to apply molecular mechanics and dynamics techniques to evaluate the energies and geometries of

Ascend, Cardinal, Pitzer

MPI is a standard library for performing parallel processing using a distributed memory model. The Pitzer, Ascend, and Cardinal clusters at OSC can use the OpenMPI implementation of the Message Passing Interface (MPI).

Cardinal

ANSYS Workbench platform is the backbone for delivering a comprehensive and integrated simulation system to users. See ANSYS Workbench platform for more information. 

Cardinal

ANSYS Mechanical is a finite element analysis (FEA) tool that enables you to analyze complex product architectures and solve difficult mechanical problems. You can use ANSYS Mechanical to simulate real world behavior of components and sub-systems, and customize it to test design variations quickly and accurately.

Wednesday, December 2nd

Gateway

Connections to OSC

In order for users to access OSC resources through the web your firewall rules should allow for connections to the following publicly-facing IP ranges.  Otherwise, users may be blocked or denied access to our services.

  • 192.148.248.0/24
  • 192.148.247.0/24
  • 192.157.5.0/25

The followingg TCP ports should be opened:

Pages