-
-
Notifications
You must be signed in to change notification settings - Fork 7.7k
Streaming broken in OpenAI server in v0.2.3 (0.2.2 works) #1967
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Ok great, I am not crazy. Here is my Dockerfile that I've been using: # FROM nvidia/cuda:12.1.0-devel-ubuntu22.04 as base
FROM nvcr.io/nvidia/pytorch:23.04-py3 as base
WORKDIR /workspace
RUN apt update && \
apt install -y python3-pip python3-packaging \
git ninja-build && \
pip3 install -U pip
# Tweak this list to reduce build time
# https://developer.nvidia.com/cuda-gpus
ENV TORCH_CUDA_ARCH_LIST "8.6"
RUN pip3 install "torch>=2.0.0"
RUN pip3 install "xformers>=0.0.22.post7" "transformers>=4.34.0" "fschat[model_worker]>=0.2.30" "numpy"
RUN pip3 install https://github.com/vllm-project/vllm/archive/main.zip Note that neither of the base images above have worked and that installing vllm from pip also did not work. Update: |
@wanmok I don't believe this was my code changes...but I haven't bisected yet to double check. |
Same problem, with OpenAI request. Seem coming when |
Reproduce :
|
Hmm... this was a bug we encountered during the development. But we then fixed it before merging. Will take a look later. |
I'm having the same issue here, happens only when |
Here's the fix: #1992 Feel free to checkout my branch to address production issue. We will be merging this in by EOD. |
Uh oh!
There was an error while loading. Please reload this page.
After upgrading to the new 0.2.3, I get the following error on a Mistral 7B finetune. I am not really sure what the cause is of the
output.logprobs
beingNone
. I suspect the error is being thrown after one of these PRs: #1504 #1756 (probably first one)Python Code:
Traceback:
The text was updated successfully, but these errors were encountered: