Skip to content

Commit 0f3f883

Browse files
committed
fix(embed): mark all tokens for output to suppress llama.cpp 'overriding' warning (#2208)
1 parent f774690 commit 0f3f883

1 file changed

Lines changed: 2 additions & 1 deletion

File tree

llama_cpp/llama.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1040,7 +1040,8 @@ def embed(
10401040

10411041
# get pooling information
10421042
pooling_type = self.pooling_type()
1043-
logits_all = pooling_type == llama_cpp.LLAMA_POOLING_TYPE_NONE
1043+
# All tokens need outputs for embeddings; llama.cpp otherwise logs an "overriding" warning per input.
1044+
logits_all = True
10441045

10451046
if self.context_params.embeddings is False:
10461047
raise RuntimeError(

0 commit comments

Comments
 (0)