Skip to content

Commit

Permalink
Merge pull request #56 from jehna/fix/nogputypo
Browse files Browse the repository at this point in the history
Fix disabling GPU use
  • Loading branch information
jehna authored Aug 24, 2024
2 parents 85d17e7 + 7fb58d4 commit 14e4ae5
Show file tree
Hide file tree
Showing 2 changed files with 18 additions and 7 deletions.
4 changes: 3 additions & 1 deletion src/commands/local.ts
Original file line number Diff line number Diff line change
Expand Up @@ -25,9 +25,11 @@ export const local = cli()
verbose.enabled = true;
}

verbose.log("Starting local inference with options: ", opts);

const prompt = await llama({
model: opts.model,
disableGPU: opts.disableGPU,
disableGpu: opts.disableGpu,
seed: opts.seed ? parseInt(opts.seed) : undefined
});
await unminify(filename, opts.outputDir, [
Expand Down
21 changes: 15 additions & 6 deletions src/plugins/local-llm-rename/llama.ts
Original file line number Diff line number Diff line change
@@ -1,6 +1,12 @@
import { getLlama, LlamaChatSession, LlamaGrammar } from "node-llama-cpp";
import {
getLlama,
LlamaChatSession,
LlamaGrammar,
LlamaModelOptions
} from "node-llama-cpp";
import { Gbnf } from "./gbnf.js";
import { getModelPath, getModelWrapper } from "../../local-models.js";
import { verbose } from "../../verbose.js";

export type Prompt = (
systemPrompt: string,
Expand All @@ -13,13 +19,16 @@ const IS_CI = process.env["CI"] === "true";
export async function llama(opts: {
seed?: number;
model: string;
disableGPU?: boolean;
disableGpu?: boolean;
}): Promise<Prompt> {
const llama = await getLlama();
const model = await llama.loadModel({
const disableGpu = opts.disableGpu ?? IS_CI;
const llama = await getLlama({ gpu: disableGpu ? false : "auto" });
const modelOpts: LlamaModelOptions = {
modelPath: getModelPath(opts?.model),
gpuLayers: (opts?.disableGPU ?? IS_CI) ? 0 : undefined
});
gpuLayers: disableGpu ? 0 : undefined
};
verbose.log("Loading model with options", modelOpts);
const model = await llama.loadModel(modelOpts);

const context = await model.createContext({ seed: opts?.seed });

Expand Down

0 comments on commit 14e4ae5

Please sign in to comment.