Skip to content

server: support add_generation_prompt query param #12062

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Feb 25, 2025

Conversation

ochafik
Copy link
Collaborator

@ochafik ochafik commented Feb 25, 2025

While this isn't part of OpenAI's API, it is part of vLLM and very useful w/ the /apply-template endpoint

(I'm using it in this PR for Cline's support of llama.cpp to compute the tool call examples used in the system prompt)

@ochafik ochafik marked this pull request as ready for review February 25, 2025 10:30
@ochafik ochafik requested a review from ngxson as a code owner February 25, 2025 10:30
@ochafik ochafik merged commit 0b52745 into ggml-org:master Feb 25, 2025
47 checks passed
orca-zhang pushed a commit to orca-zhang/llama.cpp that referenced this pull request Feb 26, 2025
mglambda pushed a commit to mglambda/llama.cpp that referenced this pull request Mar 8, 2025
arthw pushed a commit to arthw/llama.cpp that referenced this pull request Mar 19, 2025
mostlyuseful pushed a commit to mostlyuseful/llama.cpp that referenced this pull request May 12, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants