Here are the queues available on Owens. Please note that you will be routed to the appropriate queue based on your walltime and job size request.
The following are technical specifications for Owens. We hope these may be of use to the advanced user.
C, C++ and Fortran are supported on the Owens cluster. Intel, PGI and GNU compiler suites are available. The Intel development tool chain is loaded by default. Compiler commands and recommended options for serial programs are listed in the table below. See also our compilation guide.
On July 12th, 2016 OSC migrated its old GPFS and Lustre filesystems to new Project and Scratch services, respectively. We've moved 1.22 PB of data, and the new capacities are 3.4 PB for Project, and 1.1 PB for Scratch. If you store data on these services, there are a few important details to note.
For more information about citations of OSC, visit https://www.osc.edu/citation.
To cite Owens, please use the following Archival Resource Key:
Please adjust this citation to fit the citation style guidelines required.
Ohio Supercomputer Center. 2016. Owens Supercomputer. Columbus, OH: Ohio Supercomputer Center. http://osc.edu/ark:19495/hpc6h5b1
Here is the citation in BibTeX format:
"Condo model" refers to an arrangment when a client leases one or more compute nodes for the shared cluster while OSC provides all infrastructure, as well as maintenance and services. BMI's Condo on the Oakley Cluster is leased by the Biomedical Informatics Institute at The Ohio State University. This condo is a temporary "bridge" condo to server the community needs until the Owens Cluster is available.
OSCusage is command developed at OSC for use on OSC's systems. It allows for a user to see information on their project's current RU balance, including which users and jobs incurred what charges.
We have been adding some output from
qsub that should aid you in creating better job scripts. We've documented the various messages here.
A "NOTE" message is informational; your job has been submitted, but
qsub made some assumptions about your job that you may not have intended.
This page includes a summary of differences to keep in mind when migrating jobs from Glenn to one of our other clusters.
Most Oakley nodes have 12 cores and 48GB memory. There are eight large-memory nodes with 12 cores and 192GB memory, and one huge-memory node with 32 cores and 1TB of memory. Most Ruby nodes have 20 cores and 64GB of memory. There is one huge-memory node with 32 cores and 1TB of memory. By contrast, most Glenn nodes have 8 cores and 24GB memory, with eight nodes having 16 cores and 64GB memory.
OSC's Owens cluster being installed in 2016 is a Dell-built, Intel® Xeon® processor-based supercomputer. More details will be forthcoming as we finalize facilities changes and deployment schedules.