@huggingface/tasks
Advanced tools
Comparing version 0.10.10 to 0.10.11
{ | ||
"name": "@huggingface/tasks", | ||
"packageManager": "pnpm@8.10.5", | ||
"version": "0.10.10", | ||
"version": "0.10.11", | ||
"description": "List of ML tasks for huggingface.co/tasks", | ||
@@ -6,0 +6,0 @@ "repository": "https://github.com/huggingface/huggingface.js.git", |
@@ -51,12 +51,6 @@ import type { ModelData } from "./model-data"; | ||
return [ | ||
`## Install llama.cpp via brew | ||
`# Option 1: use llama.cpp with brew | ||
brew install llama.cpp | ||
## or from source with curl support | ||
## see llama.cpp README for compilation flags to optimize for your hardware | ||
git clone https://github.com/ggerganov/llama.cpp | ||
cd llama.cpp | ||
LLAMA_CURL=1 make | ||
`, | ||
`## Load and run the model | ||
# Load and run the model | ||
llama \\ | ||
@@ -67,2 +61,13 @@ --hf-repo "${model.id}" \\ | ||
-n 128`, | ||
`# Option 2: build llama.cpp from source with curl support | ||
git clone https://github.com/ggerganov/llama.cpp.git | ||
cd llama.cpp | ||
LLAMA_CURL=1 make | ||
# Load and run the model | ||
./main \\ | ||
--hf-repo "${model.id}" \\ | ||
-m file.gguf \\ | ||
-p "I believe the meaning of life is" \\ | ||
-n 128`, | ||
]; | ||
@@ -69,0 +74,0 @@ }; |
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is too big to display
Sorry, the diff of this file is not supported yet
1159856
28369