Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

sharpa/triton/whisper下面的readme文件里的操作能不能合并到docker file中? #592

Open
taorui-plus opened this issue May 17, 2024 · 1 comment

Comments

@taorui-plus
Copy link

taorui-plus commented May 17, 2024

这是sharpa/triton/whisper下面的readme文件,我想把这里的操作写到dockerfile里面,将server一键run起来,但是标记出来的编译操作在dockerfile中用RUN执行会报很多环境错误,直接在docker中执行没有这些问题,想知道这个操作是不是不适合写在dockerfile中。

执行的操作:


# Build the large-v3 model using a single GPU with plugins.
python3 build.py --output_dir whisper_large_v3 --use_gpt_attention_plugin --use_gemm_plugin  --use_bert_attention_plugin --enable_context_fmha

报错:


OSError: libcuda.so.1: cannot open shared object file: No such file or directory

解决:


export LD_LIBRARY_PATH=/usr/local/cuda-12.4/compat/lib.real:$LD_LIBRARY_PATH

继续报错:


TypeError: pybind11::init(): factory function returned nullptr

@yuekaizhang

@taorui-plus
Copy link
Author

在 Dockerfile 中使用 RUN 命令执行 Python 脚本可能会遇到一些问题,因为 Dockerfile 中的每个 RUN 命令都在一个新的层中执行,这意味着它们不会共享环境变量或其他状态。
解决方案是使用一个 shell 脚本来执行所有的命令,然后在 Dockerfile 中运行这个脚本

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant