-
Notifications
You must be signed in to change notification settings - Fork 386
Closed
Labels
enhancementNew feature or requestNew feature or requesttriageIssues / Features awaiting triageIssues / Features awaiting triage
Milestone
Description
🧭 Type of Feature
Please select the most appropriate category:
- Enhancement to existing functionality
- New feature or capability
- New MCP-compliant server
- New component or integration
- Developer tooling or test improvement
- Packaging, automation and deployment (ex: pypi, docker, quay.io, kubernetes, terraform)
- Other (please describe below)
🧭 Epic
Title: Add IBM Watsonx.ai LLM provider for LLM Chat
Goal: Integrate IBM Watsonx.ai as a supported LLM provider to enhance AI-driven chat capabilities for enterprise users.
Why now: So users can choose their own LLM provider out of all available.
🙋♂️ User Story 1
As a: business enterprise using IBM Watsonx.ai,
I want: integrate Watsonx.ai as an LLM provider for agentic llm chat in context forge,
So that: I can leverage its advanced AI capabilities, ensuring secure, context-aware, and industry-specific conversational intelligence for my customer support and internal operations.
🔗 MCP Standards Check
- Change adheres to current MCP specifications
- No breaking changes to existing MCP-compliant integrations
- If deviations exist, please describe them below:
Metadata
Metadata
Assignees
Labels
enhancementNew feature or requestNew feature or requesttriageIssues / Features awaiting triageIssues / Features awaiting triage