Back to Skills

nemo-guardrails

davila7
Updated 4 days ago
204 views
18,478
1,685
18,478
View on GitHub
TestingSafety AlignmentNeMo GuardrailsNVIDIAJailbreak DetectionGuardrailsColangRuntime SafetyHallucination DetectionPII FilteringProduction

About

NeMo Guardrails is a runtime safety framework for LLM applications that adds programmable guardrails. It provides key safety features like jailbreak detection, input/output validation, and hallucination detection using the Colang 2.0 DSL. Use it to enforce safety and compliance rules in production LLM deployments.

Quick Install

Claude Code

Recommended
Primary
npx skills add davila7/claude-code-templates -a claude-code
Plugin CommandAlternative
/plugin add https://github.com/davila7/claude-code-templates
Git CloneAlternative
git clone https://github.com/davila7/claude-code-templates.git ~/.claude/skills/nemo-guardrails

Copy and paste this command in Claude Code to install this skill

GitHub Repository

davila7/claude-code-templates
Path: cli-tool/components/skills/ai-research/safety-alignment-nemo-guardrails
0
anthropicanthropic-claudeclaudeclaude-code

Related Skills

huggingface-tokenizers

Documents

This skill provides high-performance tokenization using HuggingFace's Rust-based library, processing 1GB of text in under 20 seconds. It supports BPE, WordPiece, and Unigram algorithms while enabling custom tokenizer training and alignment tracking. Use it when you need production-fast tokenization or to build custom tokenizers integrated with the transformers ecosystem.

View skill

qdrant-vector-search

Meta

The qdrant-vector-search skill provides a high-performance vector similarity search engine for building production RAG systems. It enables fast nearest neighbor search, hybrid search with filtering, and scalable vector storage powered by Rust. Use it when you need low-latency semantic search with horizontal scaling capabilities and full data control.

View skill

crewai-multi-agent

Meta

CrewAI is a lightweight multi-agent orchestration framework for building teams of specialized AI agents that collaborate autonomously on complex tasks. It enables role-based agent collaboration with memory and supports sequential or hierarchical workflows for production use. The framework is built without LangChain dependencies for lean, fast execution.

View skill

training-llms-megatron

Design

This skill trains massive LLMs (2B-462B parameters) using NVIDIA's Megatron-Core framework for maximum GPU efficiency. Use it when training models over 1B parameters and needing advanced parallelism like tensor, pipeline, or expert parallelism. It's a production-ready framework proven on models like Nemotron and LLaMA.

View skill