Skip to content

Releases: xiaobing318/llama.cpp

b6179

15 Aug 16:44
13d4335

Choose a tag to compare

Merge branch 'ggml-org:master' into master

b5077

08 Apr 11:43
8ca6e1c

Choose a tag to compare

server : webui : Improve Chat Input with Auto-Sizing Textarea (#12785)

* Update ChatScreen.tsx

* useAutosizeTextarea.ts

useAutosizeTextarea to encapsulate the logic.

* Implement responsive auto-sizing chat textarea

Replaces the manual textarea resizing with an automatic height adjustment based on content.

- `useChatTextarea` hook to manage textarea state and auto-sizing logic via refs, preserving the optimization
- Textarea now grows vertically up to a maximum height (`lg:max-h-48`) on large screens (lg breakpoint and up).
- Disables auto-sizing and enables manual vertical resizing (`resize-vertical`) on smaller screens for better mobile usability.
- Aligns the "Send" button to the bottom of the textarea (`items-end`) for consistent positioning during resize.

* -update compressed index.html.gz after npm run build
-refactor: replace OptimizedTextareaValue with AutosizeTextareaApi in VSCode context hook

* chore: normalize line endings to LF
refactor: AutosizeTextareaApi -> chatTextareaApi

* refactor: Rename interface to PascalCase

---------

Co-authored-by: Xuan Son Nguyen <son@huggingface.co>

b4923

19 Mar 10:16
108e53c

Choose a tag to compare

llama : add support for GPT2, Bloom and CodeShell tied word embedding…

b4735

17 Feb 23:08
73e2ed3

Choose a tag to compare

CUDA: use async data loading for FlashAttention (#11894)

* CUDA: use async data loading for FlashAttention

---------

Co-authored-by: Diego Devesa <slarengh@gmail.com>

b4695

12 Feb 16:17
fef0cbe

Choose a tag to compare

cleanup: fix compile warnings associated with gnu_printf (#11811)