Could not install TabbyML hosted in K8s & integrated with Cloud Based APIs like OpenAI #3109
Replies: 3 comments 1 reply
-
Hello, it appears that your configuration file is not in the |
Beta Was this translation helpful? Give feedback.
-
Hi @wsxiaoys, Thanks for reverting back. Just to give a little background. I'm trying to self-host aforesaid open-source docker image version of tabbyml onto my K8s cluster. As a part of self-hosting I have used following K8s resources:
K8s installation of any product never have file with name config.toml. Pls guide how to make tabbyml work while self-hosting. Just FYI, my pods are up and running but still instances are not working as it expects the llama-server model to be present in my k8s infra instead of picking up from openAI model. Many thanks |
Beta Was this translation helpful? Give feedback.
-
@siddharthgaur2590 how do you mount the configMap into your pod? looks like its not formatted in maybe your should create a configMap by a file containing toml configuration described in https://tabby.tabbyml.com/docs/administration/model/, and then mount the configMap into the tabby pod under |
Beta Was this translation helpful? Give feedback.
-
Describe the bug
Trying to self host the TabbyML on K8s using Cloud Based API like openAI ones which have the GPUs capability but facing below error.
WARN llama_cpp_server::supervisor: crates/llama-cpp-server/src/supervisor.rs:98: llama-server exited with status code 127, args:
Command { std: "/opt/tabby/bin/llama-server" "-m" "/data/models/TabbyML/Nomic-Embed-Text/ggml/model.gguf" "--cont-batching" "--port" "30888" "-np" "1" "--log-disable" "--ctx-size" "4096" "-ngl" "9999" "--embedding" "--ubatch-size" "4096", kill_on_drop: true }
2024-09-05T08:16:26.589Z | 2024-09-05T08:16:26.589532Z WARN llama_cpp_server::supervisor: crates/llama-cpp-server/src/supervisor.rs:110: : /opt/tabby/bin/llama-server: error while loading shared libraries: libcuda.so.1: cannot open shared object file: No such file or directory
Information about your version
tabbyml/tabby:20240826
Information about your GPU
Using the config mentioned herein: https://tabby.tabbyml.com/docs/references/models-http-api/openai/
Additional context
Following config are used to direct the tabbyML to use GPU based models present in OpenAI platform.
Beta Was this translation helpful? Give feedback.
All reactions