tracking-application-response-times
About
This skill tracks application response times across APIs, databases, and services using a plugin that calculates key percentiles and metrics. Use it to identify performance bottlenecks, monitor SLOs, and receive alerts for degradation. Trigger it with phrases like "track response times" or "optimize latency" when you need to analyze and improve application performance.
Quick Install
Claude Code
Recommended/plugin add https://github.com/jeremylongshore/claude-code-plugins-plusgit clone https://github.com/jeremylongshore/claude-code-plugins-plus.git ~/.claude/skills/tracking-application-response-timesCopy and paste this command in Claude Code to install this skill
Documentation
Overview
This skill empowers Claude to proactively monitor and improve application performance by tracking response times across various layers. It provides detailed metrics and insights to identify and resolve performance bottlenecks.
How It Works
- Initiate Tracking: The user requests response time tracking.
- Configure Monitoring: The plugin automatically begins monitoring API endpoints, database queries, external service calls, frontend rendering, and background jobs.
- Report Metrics: The plugin generates reports including P50, P95, P99 percentiles, average, and maximum response times.
When to Use This Skill
This skill activates when you need to:
- Identify performance bottlenecks in your application.
- Monitor service level objectives (SLOs) related to response times.
- Receive alerts about performance degradation.
Examples
Example 1: Diagnosing Slow API Endpoint
User request: "Track response times for the user authentication API endpoint."
The skill will:
- Activate the response-time-tracker plugin.
- Monitor the specified API endpoint and report response time metrics, highlighting potential bottlenecks.
Example 2: Monitoring Database Query Performance
User request: "Monitor database query performance for the product catalog."
The skill will:
- Activate the response-time-tracker plugin.
- Track the execution time of database queries related to the product catalog and provide performance insights.
Best Practices
- Granularity: Track response times at a granular level (e.g., individual API endpoints, specific database queries) for more precise insights.
- Alerting: Configure alerts for significant deviations from baseline performance to proactively address potential issues.
- Contextualization: Correlate response time data with other metrics (e.g., CPU usage, memory consumption) to identify root causes.
Integration
This skill can be integrated with other monitoring and alerting tools to provide a comprehensive view of application performance. It can also be used in conjunction with optimization tools to automatically address identified bottlenecks.
GitHub Repository
Related Skills
content-collections
MetaThis skill provides a production-tested setup for Content Collections, a TypeScript-first tool that transforms Markdown/MDX files into type-safe data collections with Zod validation. Use it when building blogs, documentation sites, or content-heavy Vite + React applications to ensure type safety and automatic content validation. It covers everything from Vite plugin configuration and MDX compilation to deployment optimization and schema validation.
creating-opencode-plugins
MetaThis skill provides the structure and API specifications for creating OpenCode plugins that hook into 25+ event types like commands, files, and LSP operations. It offers implementation patterns for JavaScript/TypeScript modules that intercept and extend the AI assistant's lifecycle. Use it when you need to build event-driven plugins for monitoring, custom handling, or extending OpenCode's capabilities.
sglang
MetaSGLang is a high-performance LLM serving framework that specializes in fast, structured generation for JSON, regex, and agentic workflows using its RadixAttention prefix caching. It delivers significantly faster inference, especially for tasks with repeated prefixes, making it ideal for complex, structured outputs and multi-turn conversations. Choose SGLang over alternatives like vLLM when you need constrained decoding or are building applications with extensive prefix sharing.
evaluating-llms-harness
TestingThis Claude Skill runs the lm-evaluation-harness to benchmark LLMs across 60+ standardized academic tasks like MMLU and GSM8K. It's designed for developers to compare model quality, track training progress, or report academic results. The tool supports various backends including HuggingFace and vLLM models.
