ext_server
llm: allow gemma 2 to context shift ( #5534 )
2024-07-07 13:41:51 -04:00
generate
Bump ROCm on windows to 6.1.2
2024-07-10 11:01:22 -07:00
patches
Update llama.cpp submodule to a8db2a9c ( #5530 )
2024-07-07 13:03:09 -04:00
ggla.go
llm: speed up gguf decoding by a lot ( #5246 )
2024-06-24 21:47:52 -07:00
ggml.go
gemma2 graph
2024-06-27 13:34:52 -07:00
gguf.go
llm: speed up gguf decoding by a lot ( #5246 )
2024-06-24 21:47:52 -07:00
llm_linux.go
Switch back to subprocessing for llama.cpp
2024-04-01 16:48:18 -07:00
llm.go
Statically link c++ and thread lib
2024-07-09 11:34:30 -07:00
memory.go
handle asymmetric embedding KVs
2024-06-20 09:57:27 -07:00
payload.go
Fix corner cases on tmp cleaner on mac
2024-07-03 13:10:14 -07:00
server.go
Merge pull request #5126 from ollama/mxyng/messages
2024-07-09 09:20:44 -07:00