Skip to content

Releases: ggerganov/llama.cpp

b3883

04 Oct 17:02
1788077
Compare
Choose a tag to compare
sync : ggml

b3880

04 Oct 11:17
f3fdcfa
Compare
Choose a tag to compare
ci : fine-grant permission (#9710)

b3879

04 Oct 10:54
133c7b4
Compare
Choose a tag to compare
Fixed RNG seed docs (#9723)

* Update README.md

fixed RNG seed info

* changed print format to unsigned

b3878

03 Oct 19:42
d5ed2b9
Compare
Choose a tag to compare
metal : remove abort (skip) (ggml/0)

b3874

03 Oct 19:41
5d5ab1e
Compare
Choose a tag to compare
metal : fix compute pass descriptor autorelease crash (#9718)

b3873

03 Oct 17:06
a7ad553
Compare
Choose a tag to compare
ggml-backend : add device description to CPU backend (#9720)

b3872

03 Oct 16:50
d6fe7ab
Compare
Choose a tag to compare
ggml: unify backend logging mechanism (#9709)

* Add scaffolding for ggml logging macros

* Metal backend now uses GGML logging

* Cuda backend now uses GGML logging

* Cann backend now uses GGML logging

* Add enum tag to parameters

* Use C memory allocation funcs

* Fix compile error

* Use GGML_LOG instead of GGML_PRINT

* Rename llama_state to llama_logger_state

* Prevent null format string

* Fix whitespace

* Remove log callbacks from ggml backends

* Remove cuda log statement

b3870

03 Oct 10:53
841713e
Compare
Choose a tag to compare
rpc : enable vulkan (#9714)

closes #8536

b3869

03 Oct 08:00
5639971
Compare
Choose a tag to compare
Fixed dequant precision issues in Q4_1 and Q5_1 (#9711)

b3868

03 Oct 00:59
c83ad6d
Compare
Choose a tag to compare
ggml-backend : add device and backend reg interfaces (#9707)

Co-authored-by: Johannes Gäßler <[email protected]>