modal-serverless-gpu by davila7/claude-code-templates
npx skills add https://github.com/davila7/claude-code-templates --skill modal-serverless-gpu在 Modal 的无服务器 GPU 云平台上运行机器学习工作负载的全面指南。
在以下情况使用 Modal:
主要特性:
请改用以下替代方案:
pip install modal
modal setup # 打开浏览器进行身份验证
广告位招租
在这里展示您的产品或服务
触达数万 AI 开发者,精准高效
import modal
app = modal.App("hello-gpu")
@app.function(gpu="T4")
def gpu_info():
import subprocess
return subprocess.run(["nvidia-smi"], capture_output=True, text=True).stdout
@app.local_entrypoint()
def main():
print(gpu_info.remote())
运行:modal run hello_gpu.py
import modal
app = modal.App("text-generation")
image = modal.Image.debian_slim().pip_install("transformers", "torch", "accelerate")
@app.cls(gpu="A10G", image=image)
class TextGenerator:
@modal.enter()
def load_model(self):
from transformers import pipeline
self.pipe = pipeline("text-generation", model="gpt2", device=0)
@modal.method()
def generate(self, prompt: str) -> str:
return self.pipe(prompt, max_length=100)[0]["generated_text"]
@app.local_entrypoint()
def main():
print(TextGenerator().generate.remote("Hello, world"))
| 组件 | 用途 |
|---|---|
App | 函数和资源的容器 |
Function | 具有计算规格的无服务器函数 |
Cls | 具有生命周期钩子的基于类的函数 |
Image | 容器镜像定义 |
Volume | 用于模型/数据的持久化存储 |
Secret | 安全凭证存储 |
| 命令 | 描述 |
|---|---|
modal run script.py | 执行并退出 |
modal serve script.py | 开发模式,支持热重载 |
modal deploy script.py | 持久化云部署 |
| GPU | 显存 | 最佳适用场景 |
|---|---|---|
T4 | 16GB | 预算型推理,小型模型 |
L4 | 24GB | 推理,Ada Lovelace 架构 |
A10G | 24GB | 训练/推理,比 T4 快 3.3 倍 |
L40S | 48GB | 推荐用于推理(最佳性价比) |
A100-40GB | 40GB | 大模型训练 |
A100-80GB | 80GB | 超大模型 |
H100 | 80GB | 最快,支持 FP8 + Transformer Engine |
H200 | 141GB | 可从 H100 自动升级,4.8TB/s 带宽 |
B200 | 最新 | Blackwell 架构 |
# 单个 GPU
@app.function(gpu="A100")
# 特定显存版本
@app.function(gpu="A100-80GB")
# 多个 GPU(最多 8 个)
@app.function(gpu="H100:4")
# GPU 备选方案
@app.function(gpu=["H100", "A100", "L40S"])
# 任意可用 GPU
@app.function(gpu="any")
# 带 pip 的基础镜像
image = modal.Image.debian_slim(python_version="3.11").pip_install(
"torch==2.1.0", "transformers==4.36.0", "accelerate"
)
# 从 CUDA 基础镜像构建
image = modal.Image.from_registry(
"nvidia/cuda:12.1.0-cudnn8-devel-ubuntu22.04",
add_python="3.11"
).pip_install("torch", "transformers")
# 包含系统包
image = modal.Image.debian_slim().apt_install("git", "ffmpeg").pip_install("whisper")
volume = modal.Volume.from_name("model-cache", create_if_missing=True)
@app.function(gpu="A10G", volumes={"/models": volume})
def load_model():
import os
model_path = "/models/llama-7b"
if not os.path.exists(model_path):
model = download_model()
model.save_pretrained(model_path)
volume.commit() # 持久化更改
return load_from_path(model_path)
@app.function()
@modal.fastapi_endpoint(method="POST")
def predict(text: str) -> dict:
return {"result": model.predict(text)}
from fastapi import FastAPI
web_app = FastAPI()
@web_app.post("/predict")
async def predict(text: str):
return {"result": await model.predict.remote.aio(text)}
@app.function()
@modal.asgi_app()
def fastapi_app():
return web_app
| 装饰器 | 使用场景 |
|---|---|
@modal.fastapi_endpoint() | 简单函数 → API |
@modal.asgi_app() | 完整 FastAPI/Starlette 应用 |
@modal.wsgi_app() | Django/Flask 应用 |
@modal.web_server(port) | 任意 HTTP 服务器 |
@app.function()
@modal.batched(max_batch_size=32, wait_ms=100)
async def batch_predict(inputs: list[str]) -> list[dict]:
# 输入自动批处理
return model.batch_predict(inputs)
# 创建密钥
modal secret create huggingface HF_TOKEN=hf_xxx
@app.function(secrets=[modal.Secret.from_name("huggingface")])
def download_model():
import os
token = os.environ["HF_TOKEN"]
@app.function(schedule=modal.Cron("0 0 * * *")) # 每日午夜
def daily_job():
pass
@app.function(schedule=modal.Period(hours=1))
def hourly_job():
pass
@app.function(
container_idle_timeout=300, # 保持预热 5 分钟
allow_concurrent_inputs=10, # 处理并发请求
)
def inference():
pass
@app.cls(gpu="A100")
class Model:
@modal.enter() # 容器启动时运行一次
def load(self):
self.model = load_model() # 在预热期间加载
@modal.method()
def predict(self, x):
return self.model(x)
@app.function()
def process_item(item):
return expensive_computation(item)
@app.function()
def run_parallel():
items = list(range(1000))
# 分发到并行容器
results = list(process_item.map(items))
return results
@app.function(
gpu="A100",
memory=32768, # 32GB 内存
cpu=4, # 4 个 CPU 核心
timeout=3600, # 最长 1 小时
container_idle_timeout=120,# 保持预热 2 分钟
retries=3, # 失败时重试
concurrency_limit=10, # 最大并发容器数
)
def my_function():
pass
# 本地测试
if __name__ == "__main__":
result = my_function.local()
# 查看日志
# modal app logs my-app
| 问题 | 解决方案 |
|---|---|
| 冷启动延迟 | 增加 container_idle_timeout,使用 @modal.enter() |
| GPU 内存不足 | 使用更大的 GPU(A100-80GB),启用梯度检查点 |
| 镜像构建失败 | 固定依赖版本,检查 CUDA 兼容性 |
| 超时错误 | 增加 timeout,添加检查点 |
每周安装次数
241
代码仓库
GitHub 星标数
23.4K
首次出现
2026年1月21日
安全审计
安装于
opencode203
gemini-cli194
codex188
claude-code187
cursor184
github-copilot172
Comprehensive guide to running ML workloads on Modal's serverless GPU cloud platform.
Use Modal when:
Key features:
Use alternatives instead:
pip install modal
modal setup # Opens browser for authentication
import modal
app = modal.App("hello-gpu")
@app.function(gpu="T4")
def gpu_info():
import subprocess
return subprocess.run(["nvidia-smi"], capture_output=True, text=True).stdout
@app.local_entrypoint()
def main():
print(gpu_info.remote())
Run: modal run hello_gpu.py
import modal
app = modal.App("text-generation")
image = modal.Image.debian_slim().pip_install("transformers", "torch", "accelerate")
@app.cls(gpu="A10G", image=image)
class TextGenerator:
@modal.enter()
def load_model(self):
from transformers import pipeline
self.pipe = pipeline("text-generation", model="gpt2", device=0)
@modal.method()
def generate(self, prompt: str) -> str:
return self.pipe(prompt, max_length=100)[0]["generated_text"]
@app.local_entrypoint()
def main():
print(TextGenerator().generate.remote("Hello, world"))
| Component | Purpose |
|---|---|
App | Container for functions and resources |
Function | Serverless function with compute specs |
Cls | Class-based functions with lifecycle hooks |
Image | Container image definition |
Volume | Persistent storage for models/data |
Secret | Secure credential storage |
| Command | Description |
|---|---|
modal run script.py | Execute and exit |
modal serve script.py | Development with live reload |
modal deploy script.py | Persistent cloud deployment |
| GPU | VRAM | Best For |
|---|---|---|
T4 | 16GB | Budget inference, small models |
L4 | 24GB | Inference, Ada Lovelace arch |
A10G | 24GB | Training/inference, 3.3x faster than T4 |
L40S | 48GB | Recommended for inference (best cost/perf) |
A100-40GB | 40GB | Large model training |
# Single GPU
@app.function(gpu="A100")
# Specific memory variant
@app.function(gpu="A100-80GB")
# Multiple GPUs (up to 8)
@app.function(gpu="H100:4")
# GPU with fallbacks
@app.function(gpu=["H100", "A100", "L40S"])
# Any available GPU
@app.function(gpu="any")
# Basic image with pip
image = modal.Image.debian_slim(python_version="3.11").pip_install(
"torch==2.1.0", "transformers==4.36.0", "accelerate"
)
# From CUDA base
image = modal.Image.from_registry(
"nvidia/cuda:12.1.0-cudnn8-devel-ubuntu22.04",
add_python="3.11"
).pip_install("torch", "transformers")
# With system packages
image = modal.Image.debian_slim().apt_install("git", "ffmpeg").pip_install("whisper")
volume = modal.Volume.from_name("model-cache", create_if_missing=True)
@app.function(gpu="A10G", volumes={"/models": volume})
def load_model():
import os
model_path = "/models/llama-7b"
if not os.path.exists(model_path):
model = download_model()
model.save_pretrained(model_path)
volume.commit() # Persist changes
return load_from_path(model_path)
@app.function()
@modal.fastapi_endpoint(method="POST")
def predict(text: str) -> dict:
return {"result": model.predict(text)}
from fastapi import FastAPI
web_app = FastAPI()
@web_app.post("/predict")
async def predict(text: str):
return {"result": await model.predict.remote.aio(text)}
@app.function()
@modal.asgi_app()
def fastapi_app():
return web_app
| Decorator | Use Case |
|---|---|
@modal.fastapi_endpoint() | Simple function → API |
@modal.asgi_app() | Full FastAPI/Starlette apps |
@modal.wsgi_app() | Django/Flask apps |
@modal.web_server(port) | Arbitrary HTTP servers |
@app.function()
@modal.batched(max_batch_size=32, wait_ms=100)
async def batch_predict(inputs: list[str]) -> list[dict]:
# Inputs automatically batched
return model.batch_predict(inputs)
# Create secret
modal secret create huggingface HF_TOKEN=hf_xxx
@app.function(secrets=[modal.Secret.from_name("huggingface")])
def download_model():
import os
token = os.environ["HF_TOKEN"]
@app.function(schedule=modal.Cron("0 0 * * *")) # Daily midnight
def daily_job():
pass
@app.function(schedule=modal.Period(hours=1))
def hourly_job():
pass
@app.function(
container_idle_timeout=300, # Keep warm 5 min
allow_concurrent_inputs=10, # Handle concurrent requests
)
def inference():
pass
@app.cls(gpu="A100")
class Model:
@modal.enter() # Run once at container start
def load(self):
self.model = load_model() # Load during warm-up
@modal.method()
def predict(self, x):
return self.model(x)
@app.function()
def process_item(item):
return expensive_computation(item)
@app.function()
def run_parallel():
items = list(range(1000))
# Fan out to parallel containers
results = list(process_item.map(items))
return results
@app.function(
gpu="A100",
memory=32768, # 32GB RAM
cpu=4, # 4 CPU cores
timeout=3600, # 1 hour max
container_idle_timeout=120,# Keep warm 2 min
retries=3, # Retry on failure
concurrency_limit=10, # Max concurrent containers
)
def my_function():
pass
# Test locally
if __name__ == "__main__":
result = my_function.local()
# View logs
# modal app logs my-app
| Issue | Solution |
|---|---|
| Cold start latency | Increase container_idle_timeout, use @modal.enter() |
| GPU OOM | Use larger GPU (A100-80GB), enable gradient checkpointing |
| Image build fails | Pin dependency versions, check CUDA compatibility |
| Timeout errors | Increase timeout, add checkpointing |
Weekly Installs
241
Repository
GitHub Stars
23.4K
First Seen
Jan 21, 2026
Security Audits
Gen Agent Trust HubPassSocketPassSnykWarn
Installed on
opencode203
gemini-cli194
codex188
claude-code187
cursor184
github-copilot172
Azure RBAC 权限管理工具:查找最小角色、创建自定义角色与自动化分配
104,600 周安装
use-agently CLI:去中心化AI智能体市场命令行工具,支持A2A/MCP协议与链上支付
237 周安装
Docker容器化最佳实践指南:生产就绪容器构建、安全优化与CI/CD部署
237 周安装
Excel/XLSX文件编程操作指南:Python openpyxl/pandas与JavaScript xlsx库教程
237 周安装
DaleStudy技能创建指南:skill-creator使用教程与SEO优化
237 周安装
Every 风格编辑器 - AI 文案审阅与校对工具,遵循专业风格指南
237 周安装
Claude Code 集群编排指南:掌握多智能体协同与任务管理系统
237 周安装
A100-80GB | 80GB | Very large models |
H100 | 80GB | Fastest, FP8 + Transformer Engine |
H200 | 141GB | Auto-upgrade from H100, 4.8TB/s bandwidth |
B200 | Latest | Blackwell architecture |