Skip to content

readme : add link to Autopen under UIs #11684

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Feb 6, 2025
Merged

Conversation

blackhole89
Copy link
Contributor

Sorry that after all the months of AWOL this triviality is all I come back with, but I'd be grateful if my little project could be added to the README list.

Autopen is a graphical text editor that uses llama.cpp to tokenize the buffer on the fly, score the buffer, visualise token logits and allow you to switch back and forth between different possible completions at any point. There's a demo video here. I hope the criteria for inclusion are met, as I'm stating the dependency prominently.

(I think I might also have found some subtle issues (of numerical stability, snapshot save/restore, and/or the CUDA kernels) in the process of working on it, where different sequences of decode calls that should give identical results don't, but I'm still working on isolating the exact conditions in a way that is actionable.)

Autopen (https://github.com/blackhole89/autopen) is a graphical text editor that uses llama.cpp to tokenize the buffer on the fly, score the buffer, visualise token logits and allow you to switch back and forth between different possible completions at any point. It hopefully meets the criteria for inclusion, as the dependency on llama.cpp is stated prominently.
@slaren slaren merged commit c3db048 into ggml-org:master Feb 6, 2025
2 checks passed
@ggerganov
Copy link
Member

(I think I might also have found some subtle issues (of numerical stability, snapshot save/restore, and/or the CUDA kernels) in the process of working on it, where different sequences of decode calls that should give identical results don't, but I'm still working on isolating the exact conditions in a way that is actionable.)

Indeed there are some subtle effects that lead to differences when processing the same input with different batch sizes (see for example #7745). Not clear how to make this fully deterministic. There are at least 2 points in the computation that are problematic in this regard:

  • The max computation over the sequence length in the softmax operator of the attention
  • The KQ*V reduction over the sequence length at the end of the attention

@blackhole89
Copy link
Contributor Author

blackhole89 commented Feb 7, 2025

Ah, good to know that it's a known problem. I've observed discrepancies on the order of 0.5 logit units (usually less), though my use case may be a bit pathological since I frequently reevaluate the same run of ~10 tokens (the window between two snapshots, with default params, in my program) batched up differently, in a context where the ordering of two choices suddenly flipping is quite conspicuous.

Thanks for putting in my link!

tinglou pushed a commit to tinglou/llama.cpp that referenced this pull request Feb 13, 2025
Autopen (https://github.com/blackhole89/autopen) is a graphical text editor that uses llama.cpp to tokenize the buffer on the fly, score the buffer, visualise token logits and allow you to switch back and forth between different possible completions at any point. It hopefully meets the criteria for inclusion, as the dependency on llama.cpp is stated prominently.
orca-zhang pushed a commit to orca-zhang/llama.cpp that referenced this pull request Feb 26, 2025
Autopen (https://github.com/blackhole89/autopen) is a graphical text editor that uses llama.cpp to tokenize the buffer on the fly, score the buffer, visualise token logits and allow you to switch back and forth between different possible completions at any point. It hopefully meets the criteria for inclusion, as the dependency on llama.cpp is stated prominently.
arthw pushed a commit to arthw/llama.cpp that referenced this pull request Feb 26, 2025
Autopen (https://github.com/blackhole89/autopen) is a graphical text editor that uses llama.cpp to tokenize the buffer on the fly, score the buffer, visualise token logits and allow you to switch back and forth between different possible completions at any point. It hopefully meets the criteria for inclusion, as the dependency on llama.cpp is stated prominently.
mglambda pushed a commit to mglambda/llama.cpp that referenced this pull request Mar 8, 2025
Autopen (https://github.com/blackhole89/autopen) is a graphical text editor that uses llama.cpp to tokenize the buffer on the fly, score the buffer, visualise token logits and allow you to switch back and forth between different possible completions at any point. It hopefully meets the criteria for inclusion, as the dependency on llama.cpp is stated prominently.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants