Skip to content

Conversation

@ekcasey
Copy link
Contributor

@ekcasey ekcasey commented Sep 15, 2025

When the model runtime bundle contains a chat template file, use it by passing it to llama.cpp with the --chat-template flag. See CLI PR for usage example.

Related PRs:

@ericcurtin ericcurtin requested a review from Copilot September 15, 2025 14:25
Copy link
Contributor

Copilot AI left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Pull Request Overview

This pull request adds support for custom chat templates in the llamacpp backend by enabling the use of chat template files from model bundles.

Key changes:

  • Adds chat template file support to the llamacpp backend configuration
  • Updates the argument generation logic to include --chat-template-file parameter when available
  • Updates model-distribution dependency to a newer version

Reviewed Changes

Copilot reviewed 3 out of 4 changed files in this pull request and generated 1 comment.

File Description
pkg/inference/backends/llamacpp/llamacpp_config.go Implements chat template file argument handling in completion mode
pkg/inference/backends/llamacpp/llamacpp_config_test.go Adds test coverage for chat template functionality and updates test bundle
go.mod Updates model-distribution dependency version

Tip: Customize your code reviews with copilot-instructions.md. Create the file or learn how to get started.

@ekcasey ekcasey merged commit d6567bf into main Sep 18, 2025
4 checks passed
@ekcasey ekcasey deleted the template branch September 18, 2025 17:14
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants