File tree 2 files changed +15
-18
lines changed
2 files changed +15
-18
lines changed Original file line number Diff line number Diff line change @@ -41,24 +41,16 @@ jobs:
41
41
sanitizer : " "
42
42
fail-fast : false # While -DLLAMA_SANITIZE_THREAD=ON is broken
43
43
44
- container :
45
- image : ubuntu:latest
46
- ports :
47
- - 8888
48
- options : --cpus 4
49
-
50
44
steps :
51
45
- name : Dependencies
52
46
id : depends
53
47
run : |
54
- apt-get update
55
- apt-get -y install \
48
+ sudo apt-get update
49
+ sudo apt-get -y install \
56
50
build-essential \
57
51
xxd \
58
52
git \
59
53
cmake \
60
- python3-pip \
61
- python3-venv \
62
54
curl \
63
55
wget \
64
56
language-pack-en \
71
63
fetch-depth : 0
72
64
ref : ${{ github.event.inputs.sha || github.event.pull_request.head.sha || github.sha || github.head_ref || github.ref_name }}
73
65
66
+ - name : Python setup
67
+ id : setup_python
68
+ uses : actions/setup-python@v5
69
+ with :
70
+ python-version : ' 3.11'
71
+
72
+ - name : Tests dependencies
73
+ id : test_dependencies
74
+ run : |
75
+ pip install -r examples/server/tests/requirements.txt
76
+
74
77
- name : Verify server deps
75
78
id : verify_server_deps
76
79
run : |
@@ -101,13 +104,6 @@ jobs:
101
104
-DLLAMA_SANITIZE_${{ matrix.sanitizer }}=ON ;
102
105
cmake --build . --config ${{ matrix.build_type }} -j $(nproc) --target server
103
106
104
- - name : Setup python env
105
- id : pipenv
106
- run : |
107
- cd examples/server/tests
108
- python3 -m venv venv
109
- . venv/bin/activate
110
- pip install -r requirements.txt
111
107
112
108
- name : Tests
113
109
id : server_integration_tests
@@ -133,6 +129,7 @@ jobs:
133
129
uses : actions/checkout@v4
134
130
with :
135
131
fetch-depth : 0
132
+ ref : ${{ github.event.inputs.sha || github.event.pull_request.head.sha || github.sha || github.head_ref || github.ref_name }}
136
133
137
134
- name : libCURL
138
135
id : get_libcurl
Original file line number Diff line number Diff line change @@ -1208,7 +1208,7 @@ struct server_context {
1208
1208
}
1209
1209
1210
1210
auto n_ctx_train = llama_n_ctx_train (model);
1211
- if (slot.params .n_predict < 1 && slot.ga_n == 1
1211
+ if (slot.params .n_predict < 1 && slot.n_predict < 1 && slot. ga_n == 1
1212
1212
&& slot.n_prompt_tokens + slot.n_decoded >= n_ctx_train) {
1213
1213
LOG_WARNING (" n_predict is not set and self-context extend is disabled."
1214
1214
" Limiting generated tokens to n_ctx_train to avoid EOS-less generation infinite loop" , {
You can’t perform that action at this time.
0 commit comments