ext_server
llm: use json.hpp from common ( #6642 )
2024-09-04 19:34:42 -04:00
generate
llm: update llama.cpp commit to 8962422 ( #6618 )
2024-09-03 21:12:39 -04:00
patches
llm: update llama.cpp commit to 8962422 ( #6618 )
2024-09-03 21:12:39 -04:00
ggla.go
update convert test to check result data
2024-07-31 10:59:38 -07:00
ggml.go
Merge pull request #6260 from ollama/mxyng/mem
2024-09-05 13:22:08 -07:00
llm.go
lint
2024-08-01 17:06:06 -07:00
memory_test.go
llama3.1
2024-08-21 11:49:31 -07:00
memory.go
Introduce GPU Overhead env var ( #5922 )
2024-09-05 13:46:35 -07:00
payload.go
Add Jetson cuda variants for arm
2024-08-19 09:38:53 -07:00
server.go
Log system memory at info ( #6617 )
2024-09-03 14:55:20 -07:00
status.go
Catch one more error log
2024-08-05 09:28:07 -07:00