* chore(deps): update cohere client implementation to v2 and to use chat endpoint
Signed-off-by: Miguel Varela Ramos <miguel@cohere.ai>
* chore: remove renovate rule for cohere-go
Signed-off-by: Miguel Varela Ramos <miguel@cohere.ai>
* style: remove unused attribute
Signed-off-by: Miguel Varela Ramos <miguel@cohere.ai>
* fix: go mod
Signed-off-by: Miguel Varela Ramos <miguel@cohere.ai>
---------
Signed-off-by: Miguel Varela Ramos <miguel@cohere.ai>
Signed-off-by: Miguel Varela Ramos <miguelvramos92@gmail.com>
Co-authored-by: Alex Jones <alexsimonjones@gmail.com>
* refactor: Simplified IAI; made caching and processing consisent.
Signed-off-by: bwplotka <bwplotka@gmail.com>
* feat: Added Google AI API e.g. for Gemini models.
Signed-off-by: bwplotka <bwplotka@gmail.com>
---------
Signed-off-by: bwplotka <bwplotka@gmail.com>
Co-authored-by: Alex Jones <alexsimonjones@gmail.com>
Co-authored-by: Thomas Schuetz <38893055+thschue@users.noreply.github.com>
* feat: openAI explicit value for maxToken and temp
Because when k8sgpt talks with vLLM, the default MaxToken is 16,
which is so small.
Given the most model supports 2048 token(like Llama1 ..etc), so
put here for a safe value.
Signed-off-by: Peter Pan <Peter.Pan@daocloud.io>
* feat: make temperature a flag
Signed-off-by: Peter Pan <Peter.Pan@daocloud.io>
---------
Signed-off-by: Peter Pan <Peter.Pan@daocloud.io>