Skip to content

Benchmarks for JSON Schema Integration #13

@jenny-s51

Description

@jenny-s51

#12 introduces the component-schemas tool, which provides structured JSON Schema data for PatternFly React components.

We need to benchmark the updated MCP server using actual AI agents in Cursor (or similar AI coding assistants).

Implement a benchmark that:

  1. Uses AI Agent: Integrate with Cursor's AI agent or use an LLM API (e.g., Claude, GPT-4) to generate actual AI responses
  2. Tests on a "Golden Set": Create a curated set of prompts that represent common developer queries about PatternFly components
  3. Compares Two Scenarios:
    • Before: AI agent using only the use-patternfly-docs and fetch-docs tools (parsing unstructured documentation)
    • After: AI agent using all tools including component-schemas (structured JSON Schema data)
  4. Measures Metrics:
    • Accuracy: Did the agent provide the correct prop names / types?
    • Completeness: Did it find all relevant props?
    • Response Time: How long did it take to generate response?

Success Criteria

  • Golden set of prompts
  • Benchmark script that runs prompts against the MCP server (with and without component-schemas)
  • LLM responses captured and evaluated for accuracy
  • Document results and share findings

Metadata

Metadata

Assignees

Labels

No labels
No labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions