Skip to content

Commit b095f78

Browse files
committed
fix llama2 test.
1 parent 4e1b509 commit b095f78

File tree

1 file changed

+8
-1
lines changed

1 file changed

+8
-1
lines changed

src/llama2-tasks-test.cpp

Lines changed: 8 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -524,6 +524,8 @@ float expectedOutput[4096] = {
524524
1.00493455, 1.00216055, 1.02500832, 1.01412213, 0.997673035, 1.01922369, 1.01705575, 1.01369667,
525525
};
526526

527+
void nop(TASK_ARGS) {}
528+
527529
int main() {
528530
TransformerSpec spec;
529531
spec.headerSize = sizeof(TransformerFileOldHeader) + sizeof(int);
@@ -542,6 +544,8 @@ int main() {
542544
spec.weightsFloatType = F32;
543545
spec.bufferFloatType = F32;
544546
spec.nSlices = 1;
547+
spec.hiddenAct = GELU;
548+
spec.ropeTheta = 10000.0f;
545549

546550
size_t beforeBlockBytes = /* embedding */ 524288000;
547551
size_t blockBytes = 809533440;
@@ -567,8 +571,11 @@ int main() {
567571
for (int i = 0; i < spec.dim; i++) x[i] = randomF32(&state) / 120.0;
568572

569573
TransformerArch arch = buildLlama2Arch(&spec);
574+
arch.inference.tasks[arch.inference.nTasks - 3].handler = &nop;
575+
arch.inference.tasks[arch.inference.nTasks - 2].handler = &nop;
576+
arch.inference.tasks[arch.inference.nTasks - 1].handler = &nop;
570577

571-
int nThreads = 4;
578+
int nThreads = 1;
572579
TransformerContext context;
573580
context.transformer = &transformer;
574581
context.currentBlockIndex = 0;

0 commit comments

Comments
 (0)