mirror of
https://github.com/k8sgpt-ai/k8sgpt.git
synced 2026-05-03 18:02:43 +00:00
- Add .github/ISSUE_TEMPLATE/ (bug_report, feature_request, config) - Add MAINTAINERS.md with maintainer list and roles - Add GOVERNANCE.md with decision-making, lifecycle, vendor neutrality - Add ADOPTERS.md with adopter collection template - Add ROADMAP.md with current focus areas and planned initiatives - Add INTEGRATIONS.md with CNCF, AI provider, and tool integrations - Add RELEASE.md documenting automated release process Part of CNCF incubation preparation tracking issue #1641 Signed-off-by: Alex Jones <axjns@example.com> Co-authored-by: Alex Jones <axjns@example.com>
63 lines
4.2 KiB
Markdown
63 lines
4.2 KiB
Markdown
# Integrations
|
|
|
|
k8sgpt integrates with a variety of cloud native tools, platforms, and services.
|
|
|
|
## CNCF Project Integrations
|
|
|
|
| Project | Integration Type | Description |
|
|
|---------|-----------------|-------------|
|
|
| [Prometheus](https://prometheus.io/) | Exporter / Metrics | k8sgpt operator can export analysis results to Prometheus for monitoring |
|
|
| [Prometheus Operator](https://github.com/prometheus-operator/prometheus-operator) | Operator Integration | Integration with Prometheus Operator for service discovery and alerting |
|
|
| [Alertmanager](https://prometheus.io/docs/alerting/alertmanager/) | Alert Integration | Send k8sgpt analysis alerts to Alertmanager |
|
|
| [OpenTelemetry](https://opentelemetry.io/) | Observability | Export analysis metrics and traces via OpenTelemetry |
|
|
| [Grafana](https://grafana.com/) | Dashboard | Visualize k8sgpt analysis results in Grafana dashboards |
|
|
| [Kubernetes](https://kubernetes.io/) | Core Platform | Native Kubernetes resource analysis and diagnostics |
|
|
| [Helm](https://helm.sh/) | Packaging | k8sgpt available as a Helm chart in the [charts](https://github.com/k8sgpt-ai/charts) repository |
|
|
| [Krew](https://krew.dev/) | Plugin Distribution | k8sgpt distributed as a Krew kubectl plugin via [.krew.yaml](https://github.com/k8sgpt-ai/k8sgpt/blob/main/.krew.yaml) |
|
|
|
|
## AI/LLM Provider Integrations
|
|
|
|
| Provider | Backend Name | Description |
|
|
|----------|-------------|-------------|
|
|
| [OpenAI](https://openai.com/) | `openai` | Default provider - supports GPT-3.5, GPT-4, and other OpenAI models |
|
|
| [Azure OpenAI](https://azure.microsoft.com/services/openai/) | `azureopenai` | Azure-hosted OpenAI models |
|
|
| [Cohere](https://cohere.com/) | `cohere` | Cohere's command models |
|
|
| [Amazon Bedrock](https://aws.amazon.com/bedrock/) | `amazonbedrock` | AWS Bedrock - supports Claude, Llama, Titan, and more |
|
|
| [Amazon SageMaker](https://aws.amazon.com/sagemaker/) | `amazonsagemaker` | AWS SageMaker JumpStart models |
|
|
| [Google Gemini](https://ai.google/gemini) | `google` | Google's Gemini models |
|
|
| [Google Vertex AI](https://cloud.google.com/vertex-ai) | `googlevertexai` | Google Cloud Vertex AI models |
|
|
| [Ollama](https://ollama.com/) | `ollama` | Local LLM inference with Ollama |
|
|
| [LocalAI](https://localai.io/) | `localai` | Self-hosted OpenAI-compatible API |
|
|
| [Hugging Face](https://huggingface.co/) | `huggingface` | Hugging Face Inference API |
|
|
| [IBM WatsonX](https://www.ibm.com/watsonx) | `watsonxai` | IBM WatsonX AI models |
|
|
| [IBM WatsonxAI](https://www.ibm.com/watsonx) | `ibmwatsonxai` | IBM WatsonxAI specific integration |
|
|
| [Custom REST](https://) | `customrest` | Any REST API that follows the OpenAI chat completion format |
|
|
|
|
## Other Tool Integrations
|
|
|
|
| Tool | Integration Type | Description |
|
|
|------|-----------------|-------------|
|
|
| [Claude Desktop](https://claude.ai/) | MCP Server | k8sgpt MCP server integrates with Claude Desktop for AI-assisted cluster analysis |
|
|
| [Docker](https://www.docker.com/) | Container | Container image available on GitHub Container Registry |
|
|
| [Minikube](https://minikube.sigs.k8s.io/) | Development | Works with Minikube clusters for development and testing |
|
|
| [Kubeblocks](https://kubeblocks.io/) | Database Analysis | Analyzer support for KubeBlocks-managed databases |
|
|
| [OpenShift](https://www.openshift.com/) | Platform | Analysis support for OpenShift-specific resources (CatalogSource, ClusterCatalog, etc.) |
|
|
| [FluxCD](https://fluxcd.io/) | GitOps | Compatible with GitOps workflows using FluxCD |
|
|
| [ArgoCD](https://argoproj.github.io/cd/) | GitOps | Compatible with GitOps workflows using ArgoCD |
|
|
|
|
## Remote Caching
|
|
|
|
k8sgpt supports remote caching of analysis results:
|
|
|
|
| Provider | Type | Description |
|
|
|----------|------|-------------|
|
|
| AWS S3 | Object Storage | Store analysis cache in AWS S3 buckets |
|
|
| Azure Blob | Object Storage | Store analysis cache in Azure Blob Storage |
|
|
| Google Cloud Storage | Object Storage | Store analysis cache in GCS buckets |
|
|
|
|
## How to Integrate
|
|
|
|
For custom analyzer integrations, see the [Custom Analyzers documentation](https://docs.k8sgpt.ai/tutorials/custom-analyzers/) and the [custom analyzer schema](https://github.com/k8sgpt-ai/schemas/blob/main/protobuf/schema/v1/custom_analyzer.proto).
|
|
|
|
For MCP server integration, see [MCP.md](MCP.md).
|