🔧

持續整合/部署

13574 skills in DevOps > 持續整合/部署

condition-based-waiting

Use when tests have race conditions, timing dependencies, or inconsistent pass/fail behavior - replaces arbitrary timeouts with condition polling to wait for actual state changes, eliminating flaky tests from timing guesses

mneves75/dnschat
62
6
更新於 1w ago

behavioral-modes

Defines different operational modes for AI behavior. Each mode optimizes for specific scenarios like brainstorming, implementation, or debugging.

xenitV1/claude-code-maestro
62
15
更新於 1w ago

nemo-guardrails

NVIDIA's runtime safety framework for LLM applications. Features jailbreak detection, input/output validation, fact-checking, hallucination detection, PII filtering, toxicity detection. Uses Colang 2.0 DSL for programmable rails. Production-ready, runs on T4 GPU.

zechenzhangAGI/AI-research-SKILLs
62
2
更新於 1w ago

serving-llms-vllm

Serves LLMs with high throughput using vLLM's PagedAttention and continuous batching. Use when deploying production LLM APIs, optimizing inference latency/throughput, or serving models with limited GPU memory. Supports OpenAI-compatible endpoints, quantization (GPTQ/AWQ/FP8), and tensor parallelism.

zechenzhangAGI/AI-research-SKILLs
62
2
更新於 1w ago

dispatching-parallel-agents

Use when facing 3+ independent failures that can be investigated without shared state or dependencies - dispatches multiple Claude agents to investigate and fix independent problems concurrently

mneves75/dnschat
62
6
更新於 1w ago

plan-writing

Structured task planning with clear breakdowns, dependencies, and verification criteria. Use when implementing features, refactoring, or any multi-step work.

xenitV1/claude-code-maestro
62
15
更新於 1w ago

game-design

Game design principles. GDD structure, balancing, player psychology, progression.

xenitV1/claude-code-maestro
62
15
更新於 1w ago

tensorrt-llm

Optimizes LLM inference with NVIDIA TensorRT for maximum throughput and lowest latency. Use for production deployment on NVIDIA GPUs (A100/H100), when you need 10-100x faster inference than PyTorch, or for serving models with quantization (FP8/INT4), in-flight batching, and multi-GPU scaling.

zechenzhangAGI/AI-research-SKILLs
62
2
更新於 1w ago

writing-skills

Use when creating new skills, editing existing skills, or verifying skills work before deployment - applies TDD to process documentation by testing with subagents before writing, iterating until bulletproof against rationalization

mneves75/dnschat
62
6
更新於 1w ago

faiss

Facebook's library for efficient similarity search and clustering of dense vectors. Supports billions of vectors, GPU acceleration, and various index types (Flat, IVF, HNSW). Use for fast k-NN search, large-scale vector retrieval, or when you need pure similarity search without metadata. Best for high-performance applications.

zechenzhangAGI/AI-research-SKILLs
62
2
更新於 1w ago

gptq

Post-training 4-bit quantization for LLMs with minimal accuracy loss. Use for deploying large models (70B, 405B) on consumer GPUs, when you need 4× memory reduction with <2% perplexity degradation, or for faster inference (3-4× speedup) vs FP16. Integrates with transformers and PEFT for QLoRA fine-tuning.

zechenzhangAGI/AI-research-SKILLs
62
2
更新於 1w ago

root-cause-tracing

Use when errors occur deep in execution and you need to trace back to find the original trigger - systematically traces bugs backward through call stack, adding instrumentation when needed, to identify source of invalid data or incorrect behavior

mneves75/dnschat
62
6
更新於 1w ago

sentence-transformers

Framework for state-of-the-art sentence, text, and image embeddings. Provides 5000+ pre-trained models for semantic similarity, clustering, and retrieval. Supports multilingual, domain-specific, and multimodal models. Use for generating embeddings for RAG, semantic search, or similarity tasks. Best for production embedding generation.

zechenzhangAGI/AI-research-SKILLs
62
2
更新於 1w ago

vulnerability-scanner

Vulnerability scanning principles. DAST, SAST, SCA selection and integration.

xenitV1/claude-code-maestro
62
15
更新於 1w ago

constitutional-ai

Anthropic's method for training harmless AI through self-improvement. Two-phase approach - supervised learning with self-critique/revision, then RLAIF (RL from AI Feedback). Use for safety alignment, reducing harmful outputs without human labels. Powers Claude's safety system.

zechenzhangAGI/AI-research-SKILLs
62
2
更新於 1w ago

pc-games

PC and console game development principles. Engine selection, platform features, optimization strategies.

xenitV1/claude-code-maestro
62
15
更新於 1w ago

testing-skills-with-subagents

Use when creating or editing skills, before deployment, to verify they work under pressure and resist rationalization - applies RED-GREEN-REFACTOR cycle to process documentation by running baseline without skill, writing to address failures, iterating to close loopholes

mneves75/dnschat
62
6
更新於 1w ago

mcp-builder

MCP (Model Context Protocol) server building principles. Tool design, resource patterns, best practices.

xenitV1/claude-code-maestro
62
15
更新於 1w ago

huggingface-accelerate

Simplest distributed training API. 4 lines to add distributed support to any PyTorch script. Unified API for DeepSpeed/FSDP/Megatron/DDP. Automatic device placement, mixed precision (FP16/BF16/FP8). Interactive config, single launch command. HuggingFace ecosystem standard.

zechenzhangAGI/AI-research-SKILLs
62
2
更新於 1w ago

llama-cpp

Runs LLM inference on CPU, Apple Silicon, and consumer GPUs without NVIDIA hardware. Use for edge deployment, M1/M2/M3 Macs, AMD/Intel GPUs, or when CUDA is unavailable. Supports GGUF quantization (1.5-8 bit) for reduced memory and 4-10× speedup vs PyTorch on CPU.

zechenzhangAGI/AI-research-SKILLs
62
2
更新於 1w ago