Skip to content

Commit be84bb9

Browse files
committed
clean up
1 parent 893b40a commit be84bb9

File tree

5 files changed

+3
-6
lines changed

5 files changed

+3
-6
lines changed

ggml/src/ggml-quants.c

+1-1
Original file line numberDiff line numberDiff line change
@@ -15744,7 +15744,7 @@ bool ggml_validate_row_data(enum ggml_type type, const void * data, size_t nbyte
1574415744
if (!validate_float(q[i].d, i)) {
1574515745
return false;
1574615746
}
15747-
// NAN?
15747+
// NAN
1574815748
for (size_t k = 0; k < QK_K; ++k) {
1574915749
nans += (q[i].qs[k] & 0x7f) == 0x7f;
1575015750
}

ggml/src/ggml.c

-1
Original file line numberDiff line numberDiff line change
@@ -588,7 +588,6 @@ void ggml_bf16_to_fp32_row(const ggml_bf16_t * x, float * y, int64_t n) {
588588
int64_t i = 0;
589589
#if defined(__AVX512BF16__)
590590
for (; i + 16 <= n; i += 16) {
591-
// __m512 _mm512_cvtpbh_ps (__m256bh a)
592591
_mm512_storeu_ps(y + i,
593592
_mm512_cvtpbh_ps((__m256bh)
594593
_mm256_loadu_si256(

include/llama.h

-1
Original file line numberDiff line numberDiff line change
@@ -175,7 +175,6 @@ extern "C" {
175175
LLAMA_FTYPE_MOSTLY_Q4_0_8_8 = 35, // except 1d tensors
176176
LLAMA_FTYPE_MOSTLY_TQ1_0 = 36, // except 1d tensors
177177
LLAMA_FTYPE_MOSTLY_TQ2_0 = 37, // except 1d tensors
178-
179178
LLAMA_FTYPE_MOSTLY_E5M2 = 38, // except 1d tensors
180179
LLAMA_FTYPE_MOSTLY_E4M3 = 39, // except 1d tensors
181180
LLAMA_FTYPE_MOSTLY_E4M3_Q = 40, // except 1d tensors

scripts/sync-ggml.sh

+2-2
Original file line numberDiff line numberDiff line change
@@ -15,14 +15,14 @@ cp -rpv ../ggml/src/ggml-cann.cpp ./ggml/src/ggml-cann.cpp
1515
cp -rpv ../ggml/src/ggml-common.h ./ggml/src/ggml-common.h
1616
cp -rpv ../ggml/src/ggml-cuda/* ./ggml/src/ggml-cuda/
1717
cp -rpv ../ggml/src/ggml-cuda.cu ./ggml/src/ggml-cuda.cu
18+
cp -rpv ../ggml/src/ggml-fp8.cpp ./ggml/src/ggml-fp8.cpp
19+
cp -rpv ../ggml/src/ggml-fp8.h ./ggml/src/ggml-fp8.h
1820
cp -rpv ../ggml/src/ggml-impl.h ./ggml/src/ggml-impl.h
1921
cp -rpv ../ggml/src/ggml-kompute.cpp ./ggml/src/ggml-kompute.cpp
2022
cp -rpv ../ggml/src/ggml-metal.m ./ggml/src/ggml-metal.m
2123
cp -rpv ../ggml/src/ggml-metal.metal ./ggml/src/ggml-metal.metal
2224
cp -rpv ../ggml/src/ggml-quants.c ./ggml/src/ggml-quants.c
2325
cp -rpv ../ggml/src/ggml-quants.h ./ggml/src/ggml-quants.h
24-
cp -rpv ../ggml/src/ggml-fp8.h ./ggml/src/ggml-fp8.h
25-
cp -rpv ../ggml/src/ggml-fp8.cpp ./ggml/src/ggml-fp8.cpp
2626
cp -rpv ../ggml/src/ggml-rpc.cpp ./ggml/src/ggml-rpc.cpp
2727
cp -rpv ../ggml/src/ggml-sycl/* ./ggml/src/ggml-sycl/
2828
cp -rpv ../ggml/src/ggml-sycl.cpp ./ggml/src/ggml-sycl.cpp

src/llama.cpp

-1
Original file line numberDiff line numberDiff line change
@@ -5298,7 +5298,6 @@ static std::string llama_model_ftype_name(llama_ftype ftype) {
52985298
case LLAMA_FTYPE_MOSTLY_Q4_0_4_4: return "Q4_0_4_4";
52995299
case LLAMA_FTYPE_MOSTLY_Q4_0_4_8: return "Q4_0_4_8";
53005300
case LLAMA_FTYPE_MOSTLY_Q4_0_8_8: return "Q4_0_8_8";
5301-
53025301
case LLAMA_FTYPE_MOSTLY_E5M2: return "E5M2";
53035302
case LLAMA_FTYPE_MOSTLY_E4M3: return "E4M3";
53045303
case LLAMA_FTYPE_MOSTLY_E4M3_Q: return "E4M3_Q";

0 commit comments

Comments
 (0)