Releases: mgonzs13/llama_ros
Releases · mgonzs13/llama_ros
4.0.3
- new XTC sampling added
- new system_prompt param
- llama.cpp b3923
This version does not compile due to errors in the vendor CMakeLists
4.0.2
- common prefix added for llama.cpp commons
- llama.cpp b3906
This version does not compile due to errors in the vendor CMakeLists
4.0.1
- llama_rag_demo fixed
- llama.cpp b3889
4.0.0
- reranking added
- separate LLM, embedding models and reranking models
- new services (reranking and detokenize)
- models for reranking and embeddings added
- vicuna promopt added
- llama namespace removed from LlamaClientNode
- full demo with LLM + chat template + RAG + reranking + stream
- README:
- model shards example added
- reranking langchain and demo added
- embedding demo added
- minor fixes
- langchain reranking added
- langchain upgraded to 0.3
- llama.cpp b3870
3.9.2
- chat_llama_ros added to README
- model shard files download added
- llama.cpp b3827
3.9.1
- qwen2 updated to qwen2.5
- llama.cpp b3799
3.9.0
- new sampling from llama.cpp
- grammar functions removed
- n_remain removed
- threadpool added
- llama.cpp b3756
3.8.3
- fixed stop when n_remain is 0
- llama.cpp updated
3.8.2
- ChatLlamaROS stream fix
- ChatLlamaROS demo video added
- Fix passing image as data
3.8.1
- llama.cpp updated
- new cpuparams