File tree 1 file changed +2
-2
lines changed
1 file changed +2
-2
lines changed Original file line number Diff line number Diff line change @@ -1151,7 +1151,7 @@ def chain_speculative_sampling(
1151
1151
Shape: ``(batch_size, num_speculate_tokens, vocab_size)``
1152
1152
draft_token_ids: torch.Tensor
1153
1153
The draft model's generated token indices.
1154
- Shape: ``(batch_size, num_specutate_tokens )``
1154
+ Shape: ``(batch_size, num_speculate_tokens )``
1155
1155
target_probs: torch.Tensor
1156
1156
Expected to be uniformly distributed in ``[0, 1)``.
1157
1157
target_probs: torch.Tensor
@@ -1183,7 +1183,7 @@ def chain_speculative_sampling(
1183
1183
Compared to input :attr:`draft_token_ids`, the output tensor has an additional
1184
1184
token index at the end for the final token, if all previous tokens are accepted,
1185
1185
another "bonus" token will be sampled from the target model's probability.
1186
- Shape: (batch_size, num_specutate_tokens + 1)
1186
+ Shape: (batch_size, num_speculate_tokens + 1)
1187
1187
output_accepted_token_num: torch.Tensor
1188
1188
The number of tokens that can be accepted if each token is considered independently for each request.
1189
1189
This metric does not consider the fact that rejection sampling will stop at the first token that does not
You can’t perform that action at this time.
0 commit comments