Using Nvidia GPUs (G4DN instances with CUDA)
Using PyTorch
Off the shelf models come from Hugging Face and GitHub
We are going to go in reverse order.
Go into Coder and create a new workspace from the ESDS GPU PyTorch
template.
Name the workspace your-initials-esds-dl
.
Choose at least 200GB of storage.
Keep the defaults for the rest. (4 vCPU, 16 GiB RAM, 1 NVIDIA T4 GPU; US East)
Let your workspace spin up.
Once your workspace has deployed and you can access JupyterLab:
Go to the terminal and set up your AWS Credentials:
In your home directory:
Copy and paste your credentials from the SSO Login Page
Make sure that the first line is [default]
.
Save out the file
As admin, move the file to the .aws
directory:
Clone the Module 4 GitLab Repo to your home directory.
You should see a
day-3
directory.
We are going to go in reverse order and start with:
Taken and lightly adapted from PyTorch Examples.
Navigate to the 03-super-resolution
directory.
Run uv sync
to load the environment.
Activate the environment source .venv/bin/activate
.
Run the training:
python super_resolve.py --input_image dataset/BSDS300/images/test/16077.jpg --model model_epoch_30.pth --output_filename test_out.png --cuda
From the abstract:
Aardvark Weather, an end-to-end data-driven weather prediction system, ingests observations and produces global gridded forecasts and local station forecasts. The global forecasts outperform an operational NWP baseline for multiple variables and lead times. The local station forecasts are skillful up to ten days lead time, competing with a post-processed global NWP baseline and a state-of-the-art end-to-end forecasting system with input from human forecasters. End-to-end tuning further improves the accuracy of local forecasts. Our results show that skillful forecasting is possible without relying on NWP at deployment time, which will enable the full speed and accuracy benefits of data-driven models to be realised.
Materials taken and adapted from Aardvark Zenodo.
Navigate to the 02-aardvark
directory.
Copy the necessary files from S3:
Run uv sync
to set up environment.
Activate the environment source .venv/bin/activate
.
Set up notebook to talk to environment:
Open notebooks/forecast_demo.ipynb
.
Ensure the Aardvark Environment Kernel is selected.
Have fun!
Using Segment Anything and Self Supervised Learning to understand insect diversity.
Navigate to the 01-insect-images
directory.
Pull insect images from AWS s3:
Run uv sync
to load environment
Activate environment: source .venv/bin/activate
Run python scripts in order and monitor outputs.
Earth System Data Science in the Cloud