From 114df388fec8324386167048992997bc3b02e0f1 Mon Sep 17 00:00:00 2001 From: Georgi Gerganov Date: Mon, 10 Apr 2023 23:09:15 +0300 Subject: [PATCH] talk-llama : increase context to 2048 --- examples/talk-llama/talk-llama.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/examples/talk-llama/talk-llama.cpp b/examples/talk-llama/talk-llama.cpp index de915a6b8c9..65fd20ddf35 100644 --- a/examples/talk-llama/talk-llama.cpp +++ b/examples/talk-llama/talk-llama.cpp @@ -239,7 +239,7 @@ int main(int argc, char ** argv) { auto lparams = llama_context_default_params(); // tune these to your liking - lparams.n_ctx = 512; + lparams.n_ctx = 2048; lparams.seed = 1; lparams.f16_kv = true; lparams.n_parts = params.n_parts_llama;