Skip to content

Commit 634fbac

Browse files
authored
test: fix tests (#497)
1 parent db4a243 commit 634fbac

File tree

2 files changed

+7
-3
lines changed

2 files changed

+7
-3
lines changed

src/bindings/Llama.ts

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -465,7 +465,7 @@ export class Llama {
465465
}
466466

467467
try {
468-
const transformedLogLevel = getTransformedLogLevel(level, message);
468+
const transformedLogLevel = getTransformedLogLevel(level, message, this.gpu);
469469
if (LlamaLogLevelGreaterThanOrEqual(transformedLogLevel, this._logLevel))
470470
this._logger(transformedLogLevel, message);
471471
} catch (err) {
@@ -665,7 +665,7 @@ function logMessageIsOnlyDots(message: string | null) {
665665
return true;
666666
}
667667

668-
function getTransformedLogLevel(level: LlamaLogLevel, message: string): LlamaLogLevel {
668+
function getTransformedLogLevel(level: LlamaLogLevel, message: string, gpu: BuildGpu): LlamaLogLevel {
669669
if (level === LlamaLogLevel.warn && message.endsWith("the full capacity of the model will not be utilized"))
670670
return LlamaLogLevel.info;
671671
else if (level === LlamaLogLevel.warn && message.startsWith("ggml_metal_init: skipping kernel_") && message.endsWith("(not supported)"))
@@ -684,6 +684,8 @@ function getTransformedLogLevel(level: LlamaLogLevel, message: string): LlamaLog
684684
return LlamaLogLevel.info;
685685
else if (level === LlamaLogLevel.warn && message.startsWith("load: special_eog_ids contains both '<|return|>' and '<|call|>' tokens, removing '<|end|>' token from EOG list"))
686686
return LlamaLogLevel.info;
687+
else if (gpu === false && level === LlamaLogLevel.warn && message.startsWith("llama_adapter_lora_init_impl: lora for '") && message.endsWith("' cannot use buft 'CPU_REPACK', fallback to CPU"))
688+
return LlamaLogLevel.info;
687689

688690
return level;
689691
}

test/modelDependent/llama3.1/tokenBias.test.ts

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,9 +25,11 @@ describe("llama 3.1", () => {
2525
const text = model.detokenize([token]);
2626

2727
if (text.toLowerCase().includes("hello"))
28-
customBias.set(token, -0.99);
28+
customBias.set(token, -1);
2929
else if (text.toLowerCase().includes("hi"))
3030
customBias.set(token, "never");
31+
else if (text.toLowerCase().includes("well"))
32+
customBias.set(token, -0.99);
3133
}
3234

3335
const res = await chatSession.prompt('Greet me by saying "hello" to me', {

0 commit comments

Comments
 (0)