galaxy-automation

Galaxy Workflow Automation with BioBlend and Planemo

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "galaxy-automation" with this command: npx skills add delphine-l/claude_global/delphine-l-claude-global-galaxy-automation

Galaxy Workflow Automation with BioBlend and Planemo

Purpose

This skill provides expert knowledge for automating Galaxy workflows using BioBlend (Python Galaxy API library) and Planemo (Galaxy workflow testing and execution tool).

When to Use This Skill

Use this skill when:

  • Automating Galaxy workflow execution via API

  • Building batch processing systems for Galaxy

  • Using BioBlend to interact with Galaxy

  • Testing workflows with Planemo

  • Managing Galaxy histories, datasets, and collections programmatically

  • Polling workflow invocation status

  • Implementing error handling and retry logic for Galaxy operations

  • Creating Galaxy automation pipelines

  • Integrating Galaxy into larger bioinformatics workflows

This skill is NOT project-specific - it's useful for ANY Galaxy automation project.

Supporting Files

Detailed reference material is split into separate files:

  • bioblend-reference.md -- BioBlend API patterns: connection, history management, workflow invocation, status checking, error handling, rerun API, dataset operations, and collections

  • planemo-reference.md -- Planemo command structure, job YAML format, programmatic command generation, output parsing, and Galaxy API curl/authentication patterns

  • automation-patterns.md -- Thread-safe operations, batch processing, resume capability, and debugging (history inspection, invocation step analysis)

Security Best Practices

  1. API Key Management

Store in environment variables:

import os

api_key = os.environ.get('GALAXY_API_KEY') if not api_key: raise ValueError("GALAXY_API_KEY environment variable not set")

gi = GalaxyInstance(url, api_key)

Mask in logs:

def mask_api_key(key): """Mask API key for display""" if len(key) <= 8: return '' * len(key) return f"{key[:4]}{'' * (len(key) - 8)}{key[-4:]}"

masked_key = mask_api_key(api_key) print(f"Using API key: {masked_key}")

  1. Path Handling

Always quote paths in shell commands:

Good - handles spaces

command = f'planemo run "{workflow_path}" "{job_yaml}"'

Bad - breaks with spaces

command = f'planemo run {workflow_path} {job_yaml}'

Common Pitfalls

Planemo failures vs Galaxy failures

  • Planemo return code != 0: Workflow was NOT launched, no invocation exists

  • Invocation state = 'failed': Workflow was launched but Galaxy job failed

  • Don't confuse these two failure modes

Concurrent uploads

  • Too many simultaneous uploads can overwhelm Galaxy

  • Use max_concurrent limits (typically 3-5)

  • Consider --simultaneous_uploads vs sequential

Dataset state checking

  • Don't invoke workflows before uploads complete

  • Always wait for dataset state = 'ok'

History name conflicts

  • Use unique history names (add timestamps or suffixes)

  • Check for existing histories before creating

Return code interpretation

  • os.system() shifts exit codes (exit 1 -> return 256)

  • Use return_code >> 8 to get actual exit code

Invocation ID recovery

  • Terminal disconnection loses invocation ID

  • Always save invocation IDs to file immediately

  • Use --test_output_json with planemo

Best Practices Summary

  • Use environment variables for API keys

  • Mask API keys in logs and output

  • Quote all file paths in shell commands

  • Implement thread-safety for concurrent operations

  • Save state frequently for resume capability

  • Wait for dataset uploads before invoking workflows

  • Poll invocation status with reasonable intervals (30-60s)

  • Distinguish planemo failures from Galaxy failures

  • Implement proper error handling and retry logic

  • Use unique history names to avoid conflicts

Galaxy MCP Connection

When using the Galaxy MCP tools (mcp__Galaxy__*), connect at the start of each session.

Connection Pattern

MCP tools cannot read shell environment variables directly. Resolve them via Bash first:

Resolve env vars

echo "$GXYVGP" # Galaxy instance URL echo "$MAINKEY" # API key

Then pass the resolved values:

mcp__Galaxy__connect(url="<resolved_url>", api_key="<resolved_key>")

Known Instances

Env Var Instance Notes

$GXYVGP

https://vgp.usegalaxy.org VGP production, user: delphinel (admin)

$MAINKEY

API key for VGP instance Used with $GXYVGP

Related Skills

  • galaxy-tool-wrapping: For creating Galaxy tool wrappers

  • galaxy-workflow-development: For creating Galaxy workflows

  • vgp-pipeline: VGP-specific orchestration (uses this skill as dependency)

Resources

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

token-efficiency

No summary provided by upstream source.

Repository SourceNeeds Review
General

bioinformatics-fundamentals

No summary provided by upstream source.

Repository SourceNeeds Review
General

folder-organization

No summary provided by upstream source.

Repository SourceNeeds Review
General

obsidian

No summary provided by upstream source.

Repository SourceNeeds Review