Skip to content

Releases: mgonzs13/llama_ros

4.0.9

02 Nov 08:54
Compare
Choose a tag to compare
  • github action for jazzy
  • SmolLM2 model added
  • requirements removed (lark, packaging)
  • llama.cpp b4011

4.0.8

31 Oct 08:39
Compare
Choose a tag to compare
  • fixing llama_rag demo
  • Tail-Free sampling removed
  • signal removed from mains
  • llama.cpp b3995

4.0.7

26 Oct 20:21
Compare
Choose a tag to compare
  • DRY sampling
  • llama.cpp b3982

4.0.6

25 Oct 08:26
Compare
Choose a tag to compare
  • Dockerfile created
  • CI github actions for formatter and docker build
  • python formatted with black
  • llama.cpp b3974

4.0.5

23 Oct 08:29
Compare
Choose a tag to compare
  • lark added to requirements
  • normalization types for embeddings
  • llama.cpp b3962

4.0.4

17 Oct 09:03
Compare
Choose a tag to compare
  • vendor CMakeLists fixed
  • llama.cpp b3933

4.0.3

16 Oct 09:21
Compare
Choose a tag to compare
  • new XTC sampling added
  • new system_prompt param
  • llama.cpp b3923

This version does not compile due to errors in the vendor CMakeLists

4.0.2

11 Oct 11:36
Compare
Choose a tag to compare
  • common prefix added for llama.cpp commons
  • llama.cpp b3906

This version does not compile due to errors in the vendor CMakeLists

4.0.1

07 Oct 07:24
Compare
Choose a tag to compare
  • llama_rag_demo fixed
  • llama.cpp b3889

4.0.0

03 Oct 11:48
Compare
Choose a tag to compare
  • reranking added
  • separate LLM, embedding models and reranking models
  • new services (reranking and detokenize)
  • models for reranking and embeddings added
  • vicuna promopt added
  • llama namespace removed from LlamaClientNode
  • full demo with LLM + chat template + RAG + reranking + stream
  • README:
    • model shards example added
    • reranking langchain and demo added
    • embedding demo added
    • minor fixes
  • langchain reranking added
  • langchain upgraded to 0.3
  • llama.cpp b3870