Training & EvalMachine Learning & Research AIK-Dense-AI/claude-scientific-skillsModel Training & Evaluation
PU

PufferLib

Maintainer K-Dense Inc. · Last updated April 1, 2026

PufferLib is a high-performance reinforcement learning library designed for fast parallel environment simulation and training. It achieves training at millions of steps per second through optimi.

Claude CodeTrainingEvaluationpufferlibmachine-learningpackagemachine learning & deep learning

Original source

K-Dense-AI/claude-scientific-skills

https://github.com/K-Dense-AI/claude-scientific-skills/tree/main/scientific-skills/pufferlib

Maintainer
K-Dense Inc.
License
MIT license
Last updated
April 1, 2026

Skill Snapshot

Key Details From SKILL.md

2 min

Key Notes

  • PufferLib is a high-performance reinforcement learning library designed for fast parallel environment simulation and training. It achieves training at millions of steps per second through optimized vectorization, native multi-agent support, and efficient PPO implementation (PuffeRL). The library provides the Ocean suite of 20+ environments and seamless integration with Gymnasium, PettingZoo, and specialized RL frameworks.
  • puffer train procgen-coinrun --train.device cuda --train.learning-rate 3e-4.

Source Doc

Excerpt From SKILL.md

When to Use This Skill

Use this skill when:

  • Training RL agents with PPO on any environment (single or multi-agent)
  • Creating custom environments using the PufferEnv API
  • Optimizing performance for parallel environment simulation (vectorization)
  • Integrating existing environments from Gymnasium, PettingZoo, Atari, Procgen, etc.
  • Developing policies with CNN, LSTM, or custom architectures
  • Scaling RL to millions of steps per second for faster experimentation
  • Multi-agent RL with native multi-agent environment support

1. High-Performance Training (PuffeRL)

PuffeRL is PufferLib's optimized PPO+LSTM training algorithm achieving 1M-4M steps/second.

Quick start training:


## Distributed training

torchrun --nproc_per_node=4 train.py
python
import pufferlib
from pufferlib import PuffeRL

Use cases

  • **Training RL agents** with PPO on any environment (single or multi-agent).
  • **Creating custom environments** using the PufferEnv API.

Not for

  • Do not rely on this catalog entry alone for installation or maintenance details.

Related skills

Related skills

Back to directory
BI
Training & EvalMachine Learning & Research AI

bio-epitranscriptomics-m6anet-analysis

Nanopore direct RNA m6A detection with m6Anet deep learning.

OpenClawNanoClawTraining
FreedomIntelligence/OpenClaw-Medical-SkillsView
BI
Training & EvalMachine Learning & Research AI

bio-imaging-mass-cytometry-interactive-annotation

Interactive cell type annotation for IMC data. Covers napari-based annotation, marker-guided labeling, training data generation, and annotat…

OpenClawNanoClawTraining
FreedomIntelligence/OpenClaw-Medical-SkillsView
BI
Training & EvalMachine Learning & Research AI

bio-immunoinformatics-tcr-epitope-binding

Predict TCR-epitope specificity using ERGO-II and deep learning models for T-cell receptor antigen recognition. Match TCRs to their cognate…

OpenClawNanoClawTraining
FreedomIntelligence/OpenClaw-Medical-SkillsView
CI
Training & EvalMachine Learning & Research AI

cirq

Cirq is Google Quantum AI's open-source framework for designing, simulating, and running quantum circuits on quantum computers and simulator…

Claude CodeTraining
K-Dense-AI/claude-scientific-skillsView