Skip to content

Commit 9e27222

Browse files
committed
feat: add llama_model_n_head_kv module function
1 parent cc85f68 commit 9e27222

File tree

1 file changed

+17
-0
lines changed

1 file changed

+17
-0
lines changed

ext/llama_cpp/llama_cpp.c

Lines changed: 17 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1452,6 +1452,20 @@ static VALUE rb_llama_model_n_head(VALUE self, VALUE model) {
14521452
return INT2NUM(llama_model_n_head(model_wrapper->model));
14531453
}
14541454

1455+
/**
1456+
* @overload llama_model_n_head_kv(model)
1457+
* @param [LlamaModel] model
1458+
* @return [Integer]
1459+
*/
1460+
static VALUE rb_llama_model_n_head_kv(VALUE self, VALUE model) {
1461+
if (!rb_obj_is_kind_of(model, rb_cLlamaModel)) {
1462+
rb_raise(rb_eArgError, "model must be a LlamaModel");
1463+
return Qnil;
1464+
}
1465+
llama_model_wrapper* model_wrapper = get_llama_model_wrapper(model);
1466+
return INT2NUM(llama_model_n_head_kv(model_wrapper->model));
1467+
}
1468+
14551469
/**
14561470
* @overload llama_model_rope_freq_scale_train(model)
14571471
* @param [LlamaModel] model
@@ -4714,6 +4728,9 @@ void Init_llama_cpp(void) {
47144728
/* llama_model_n_head */
47154729
rb_define_module_function(rb_mLlamaCpp, "llama_model_n_head", rb_llama_model_n_head, 1);
47164730

4731+
/* llama_model_n_head_kv */
4732+
rb_define_module_function(rb_mLlamaCpp, "llama_model_n_head_kv", rb_llama_model_n_head_kv, 1);
4733+
47174734
/* llama_model_rope_freq_scale_train */
47184735
rb_define_module_function(rb_mLlamaCpp, "llama_model_rope_freq_scale_train", rb_llama_model_rope_freq_scale_train, 1);
47194736

0 commit comments

Comments
 (0)