A Snakemake storage plugin for downloading files via HTTP with local caching, checksum verification, and adaptive rate limiting.
Supported sources:
- zenodo.org - Zenodo data repository (checksum from API)
- data.pypsa.org - PyPSA data repository (checksum from manifest.yaml)
- storage.googleapis.com - Google Cloud Storage (checksum from GCS JSON API)
- any http(s) URL - Generic fallback with size/mtime from HTTP headers
- Local caching: Downloads are cached to avoid redundant transfers (can be disabled)
- Checksum verification: Automatically verifies checksums (from Zenodo API, data.pypsa.org manifests, or GCS object metadata)
- Rate limit handling: Automatically respects Zenodo's rate limits using
X-RateLimit-*headers with exponential backoff retry - Concurrent download control: Limits simultaneous downloads to prevent overwhelming servers
- Resumable downloads: Interrupted transfers resume from where they left off using HTTP range requests
- Progress bars: Shows download progress with tqdm
- Immutable URLs: Returns mtime=0 for Zenodo and data.pypsa.org (persistent URLs); uses actual mtime for GCS and generic HTTP
- Environment variable support: Configure via environment variables for CI/CD workflows
From the pypsa-eur repository root:
pip install -e plugins/snakemake-storage-plugin-cached-httpThe Zenodo storage plugin works alongside other storage providers (like HTTP). Snakemake automatically routes URLs to the correct provider based on the URL pattern.
Register additional settings in your Snakefile if you want to customize the defaults:
# Optional: Configure cached HTTP storage with custom settings
# This extends the existing storage configuration (e.g., for HTTP)
storage cached_http:
provider="cached-http",
cache="~/.cache/snakemake-pypsa-eur", # Default location
max_concurrent_downloads=3, # Download max 3 files at onceIf you don't explicitly configure it, the plugin will use default settings automatically.
-
cache (optional): Cache directory for downloaded files
- Default: Platform-dependent user cache directory (via
platformdirs.user_cache_dir("snakemake-pypsa-eur")) - Set to
""(empty string) to disable caching - Files are cached here to avoid re-downloading
- Environment variable:
SNAKEMAKE_STORAGE_CACHED_HTTP_CACHE
- Default: Platform-dependent user cache directory (via
-
skip_remote_checks (optional): Skip metadata checking with remote API
- Default:
False(perform checks) - Set to
Trueor"1"to skip remote existence/size checks (useful for CI/CD) - Environment variable:
SNAKEMAKE_STORAGE_CACHED_HTTP_SKIP_REMOTE_CHECKS
- Default:
-
max_concurrent_downloads (optional): Maximum concurrent downloads
- Default:
3 - Controls how many files can be downloaded simultaneously
- No environment variable support
- Default:
Use any HTTP(S) URL directly in your rules. Snakemake automatically routes all HTTP(S) URLs to this plugin:
rule download_zenodo:
input:
storage("https://zenodo.org/records/3520874/files/natura.tiff"),
output:
"resources/natura.tiff"
shell:
"cp {input} {output}"
rule download_pypsa:
input:
storage("https://data.pypsa.org/workflows/eur/eez/v12_20231025/World_EEZ_v12_20231025_LR.zip"),
output:
"resources/eez.zip"
shell:
"cp {input} {output}"
rule download_gcs:
input:
storage("https://storage.googleapis.com/open-tyndp-data-store/CBA_projects.zip"),
output:
"resources/cba_projects.zip"
shell:
"cp {input} {output}"
rule download_generic:
input:
storage("https://example.com/data/dataset.csv"),
output:
"resources/dataset.csv"
shell:
"cp {input} {output}"Or if you configured a tagged storage entity:
rule download_data:
input:
storage.cached_http(
"https://zenodo.org/records/3520874/files/natura.tiff"
),
output:
"resources/natura.tiff"
shell:
"cp {input} {output}"The plugin will:
- Check if the file exists in the cache (if caching is enabled)
- If cached, copy from cache (fast)
- If not cached, download with:
- Progress bar showing download status
- Automatic rate limit handling with exponential backoff retry
- Concurrent download limiting
- Checksum verification where available (Zenodo API, data.pypsa.org manifest, GCS metadata)
- Store in cache for future use (if caching is enabled)
For continuous integration environments where you want to skip caching and remote checks:
# GitHub Actions example
- name: Run snakemake workflows
env:
SNAKEMAKE_STORAGE_CACHED_HTTP_CACHE: ""
SNAKEMAKE_STORAGE_CACHED_HTTP_SKIP_REMOTE_CHECKS: "1"
run: |
snakemake --cores allZenodo API limits:
- Guest users: 60 requests/minute
- Authenticated users: 100 requests/minute
The plugin automatically:
- Monitors
X-RateLimit-Remainingheader - Waits when rate limit is reached
- Uses
X-RateLimit-Resetto calculate wait time - Retries failed requests with exponential backoff (up to 5 attempts)
- Handles transient errors: HTTP errors, timeouts, checksum mismatches, and network issues
- Resumes interrupted downloads using
Rangerequests where supported by the server
This plugin accepts all HTTP(S) URLs and replaces snakemake-storage-plugin-http. It provides
enhanced support for specific sources:
| Source | Checksum | mtime | Immutable |
|---|---|---|---|
zenodo.org, sandbox.zenodo.org |
✓ (from API) | — | ✓ |
data.pypsa.org |
✓ (from manifest.yaml) | — | ✓ |
storage.googleapis.com |
✓ (from GCS API) | ✓ | — |
| any other HTTP(S) | — | ✓ (Last-Modified) | — |
Generic HTTP URLs are treated as mutable: size and mtime are read from Content-Length and
Last-Modified response headers. Servers that do not support HEAD requests are handled
gracefully (size and mtime default to 0).
MIT License