Work Description

Title: Simple Physics CAM6 Codebase for Training Machine Learning Algorithms Open Access Deposited

h
Attribute Value
Methodology
  • This contains the codebase for developing machine learning emulators for simplified physical parameterizations within the Community Atmosphere Model, version 6 (CAM6).
Description
  • The work guides the processing of CAM6 data for use in machine learning applications. We also provide workflow scripts for training both random forests and neural networks to emulate physic s schemes from the data, as well as analysis scripts written in both Python and NCL in order to process our results.
Creator
Depositor
  • glimon@umich.edu
Contact information
Discipline
Funding agency
  • National Science Foundation (NSF)
Keyword
Citations to related material
  • Limon, G. C., Jablonowski, C. (2022) Probing the Skill of Random Forest Emulators for Physical Parameterizations via a Hierarchy of Simple CAM6 Configurations [Pre Print]. ESSOAr. https://10.1002/essoar.10512353.1
Related items in Deep Blue Documents
  • Limon, G. C. (2022). Simple Physics CAM6 Dataset for Training Machine Learning Algorithms [Data set], University of Michigan - Deep Blue Data. https://doi.org/10.7302/r6v3-s064
Resource type
Last modified
  • 05/05/2023
Published
  • 05/05/2023
Language
DOI
  • https://doi.org/10.7302/kxrz-9k87
License
To Cite this Work:
Limon, G. C. (2023). Simple Physics CAM6 Codebase for Training Machine Learning Algorithms [Data set], University of Michigan - Deep Blue Data. https://doi.org/10.7302/kxrz-9k87

Files (Count: 2; Size: 57.3 KB)

# Simple Physics CAM6 Codebase for Training Machine Learning Algorithms

## Creator

Garrett Limon (glimon@umich.edu)

## Funding

NSF GRFP

## Research and Description

This is a collection of Python and NCL scripts intended to process data, train
machine learning algorithms to emulate physical parameterization schemes,
as well as to analyze the results. See Associated Work for detailed information
and findings.

## Format

Directory is organized as follows.
simplePhysicsML/data_process/
- contains data processing scripts and functions.
We use input data from the Community Atmosphere
Model (CAM) for these tasks; see Associated Work(2)
for data initially used for this work.
simplePhysicsML/19/
- contains directories for each case studied (HS,TJ,TJBM), each case
containing the fields (PTTEND, PTEQ, PRECL, etc) that were emulated
with machine learning, and each field directory containing a sub-
directory for each machine learning method used (random forests and
neural networks). This final directory in this sub-structure is
associated with the training of our machine leraning emulators using
Python scripts.
simplePhysicsML/analysis
- contains python and NCL scripts that analyze and plot relevant fields
and metrics for us to understand the skillfullness of our emulators.

## Data Citation

Limon, G. C. (2023) Simple Physics CAM6 Codebase for Training Machine Learning Algorithms [Data set], University of Michigan - Deep Blue Data.

## Associated Work

(1) Limon, G. C., Jablonowski, C. (2022) Probing the Skill of Random Forest Emulators for Physical Parameterizations via a Hierarchy of Simple CAM6 Configurations [Preprint], ESSOAr. https://10.1002/essoar.10512353.1

(2) Limon, G. C. (2022) Simple Physics CAM6 Dataset for Training Machine Learning Algorithms [Data set], University of Michigan - Deep Blue Data. https://doi.org/10.7302/r6v3-s064

Download All Files (To download individual files, select them in the “Files” panel above)

Best for data sets < 3 GB. Downloads all files plus metadata into a zip file.



Best for data sets > 3 GB. Globus is the platform Deep Blue Data uses to make large data sets available.   More about Globus

Remediation of Harmful Language

The University of Michigan Library aims to describe library materials in a way that respects the people and communities who create, use, and are represented in our collections. Report harmful or offensive language in catalog records, finding aids, or elsewhere in our collections anonymously through our metadata feedback form. More information at Remediation of Harmful Language.