Skip to content

Commit ea66027

Browse files
authored
Merge branch 'main' into functionary-fixes
2 parents cbe9604 + e6bbfb8 commit ea66027

File tree

5 files changed

+20
-8
lines changed

5 files changed

+20
-8
lines changed

.github/dependabot.yml

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -9,3 +9,7 @@ updates:
99
directory: "/" # Location of package manifests
1010
schedule:
1111
interval: "weekly"
12+
- package-ecosystem: "github-actions"
13+
directory: "/"
14+
schedule:
15+
interval: "weekly"

.github/workflows/build-wheels-metal.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,7 @@ jobs:
4141
with:
4242
submodules: "recursive"
4343

44-
- uses: actions/setup-python@v4
44+
- uses: actions/setup-python@v5
4545
with:
4646
python-version: ${{ matrix.pyver }}
4747

@@ -78,7 +78,7 @@ jobs:
7878
VERBOSE=1 python -m build --wheel
7979
fi
8080
81-
- uses: softprops/action-gh-release@v1
81+
- uses: softprops/action-gh-release@v2
8282
with:
8383
files: dist/*
8484
# set release name to <tag>-metal

CHANGELOG.md

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,11 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
77

88
## [Unreleased]
99

10+
## [0.2.65]
11+
12+
- feat: Update llama.cpp to ggerganov/llama.cpp@46e12c4692a37bdd31a0432fc5153d7d22bc7f72
13+
- feat: Allow for possibly non-pooled embeddings by @iamlemec in #1380
14+
1015
## [0.2.64]
1116

1217
- feat: Update llama.cpp to ggerganov/llama.cpp@4e96a812b3ce7322a29a3008db2ed73d9087b176

examples/low_level_api/quantize.py

Lines changed: 8 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -4,14 +4,16 @@
44

55

66
def main(args):
7+
fname_inp = args.fname_inp.encode("utf-8")
8+
fname_out = args.fname_out.encode("utf-8")
79
if not os.path.exists(fname_inp):
810
raise RuntimeError(f"Input file does not exist ({fname_inp})")
911
if os.path.exists(fname_out):
1012
raise RuntimeError(f"Output file already exists ({fname_out})")
11-
fname_inp = args.fname_inp.encode("utf-8")
12-
fname_out = args.fname_out.encode("utf-8")
13-
itype = args.itype
14-
return_code = llama_cpp.llama_model_quantize(fname_inp, fname_out, itype)
13+
ftype = args.type
14+
args = llama_cpp.llama_model_quantize_default_params()
15+
args.ftype = ftype
16+
return_code = llama_cpp.llama_model_quantize(fname_inp, fname_out, args)
1517
if return_code != 0:
1618
raise RuntimeError("Failed to quantize model")
1719

@@ -20,6 +22,7 @@ def main(args):
2022
parser = argparse.ArgumentParser()
2123
parser.add_argument("fname_inp", type=str, help="Path to input model")
2224
parser.add_argument("fname_out", type=str, help="Path to output model")
23-
parser.add_argument("type", type=int, help="Type of quantization (2: q4_0, 3: q4_1)")
25+
parser.add_argument("type", type=int, help="Type of quantization (2: q4_0, 3: q4_1), see llama_cpp.py for enum")
2426
args = parser.parse_args()
2527
main(args)
28+

llama_cpp/__init__.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
11
from .llama_cpp import *
22
from .llama import *
33

4-
__version__ = "0.2.64"
4+
__version__ = "0.2.65"

0 commit comments

Comments
 (0)