ollama/README.md

91 lines
1.4 KiB
Markdown
Raw Normal View History

2023-06-28 00:08:52 +08:00
# Ollama
2023-06-23 00:45:31 +08:00
2023-06-28 05:26:27 +08:00
The easiest way to run ai models.
2023-06-28 05:13:07 +08:00
## Download
2023-06-28 10:05:41 +08:00
- [macOS](https://ollama.ai/download/darwin_arm64) (Apple Silicon)
2023-06-28 05:13:07 +08:00
- macOS (Intel Coming soon)
- Windows (Coming soon)
- Linux (Coming soon)
## Python SDK
2023-06-23 00:45:31 +08:00
```
2023-06-28 00:08:52 +08:00
pip install ollama
2023-06-23 00:45:31 +08:00
```
2023-06-28 05:13:07 +08:00
### Python SDK quickstart
2023-06-26 01:08:03 +08:00
2023-06-28 00:08:52 +08:00
```python
import ollama
2023-06-28 00:48:50 +08:00
ollama.generate("./llama-7b-ggml.bin", "hi")
2023-06-26 01:08:03 +08:00
```
2023-06-28 00:51:36 +08:00
### `ollama.generate(model, message)`
2023-06-28 00:08:52 +08:00
2023-06-28 00:51:36 +08:00
Generate a completion
2023-06-28 00:08:52 +08:00
```python
2023-06-28 00:51:36 +08:00
ollama.generate("./llama-7b-ggml.bin", "hi")
2023-06-26 01:08:03 +08:00
```
2023-06-28 00:51:36 +08:00
### `ollama.load(model)`
2023-06-26 01:08:03 +08:00
2023-06-28 00:51:36 +08:00
Load a model for generation
2023-06-26 01:08:03 +08:00
2023-06-28 00:08:52 +08:00
```python
2023-06-28 00:56:53 +08:00
ollama.load("model")
2023-06-26 01:08:03 +08:00
```
2023-06-28 00:51:36 +08:00
### `ollama.models()`
2023-06-26 01:08:03 +08:00
2023-06-28 00:44:12 +08:00
List available local models
2023-06-28 00:08:52 +08:00
```
models = ollama.models()
2023-06-26 01:08:03 +08:00
```
2023-06-28 00:51:36 +08:00
### `ollama.serve()`
2023-06-26 01:10:15 +08:00
2023-06-28 00:08:52 +08:00
Serve the ollama http server
2023-06-26 01:08:03 +08:00
2023-06-28 05:36:02 +08:00
### `ollama.add(filepath)`
2023-06-26 01:08:03 +08:00
2023-06-28 05:36:02 +08:00
Add a model by importing from a file
2023-06-28 00:08:52 +08:00
```python
2023-06-28 05:36:02 +08:00
ollama.add("./path/to/model")
2023-06-26 01:08:03 +08:00
```
2023-06-28 05:36:02 +08:00
## Cooming Soon
2023-06-28 00:08:52 +08:00
2023-06-28 05:36:02 +08:00
### `ollama.pull(model)`
Download a model
2023-06-28 00:08:52 +08:00
```python
2023-06-28 05:36:02 +08:00
ollama.pull("huggingface.co/thebloke/llama-7b-ggml")
2023-06-28 00:08:52 +08:00
```
2023-06-26 01:08:03 +08:00
2023-06-28 00:51:36 +08:00
### `ollama.search("query")`
2023-06-26 02:29:26 +08:00
2023-06-28 00:08:52 +08:00
Search for compatible models that Ollama can run
2023-06-26 02:29:26 +08:00
2023-06-28 00:08:52 +08:00
```python
ollama.search("llama-7b")
```
2023-06-26 01:08:03 +08:00
2023-06-28 00:08:52 +08:00
## Future CLI
2023-06-26 02:29:26 +08:00
2023-06-28 01:51:20 +08:00
In the future, there will be an `ollama` CLI for running models on servers, in containers or for local development environments.
2023-06-28 00:44:12 +08:00
2023-06-28 00:08:52 +08:00
```
2023-06-28 01:52:32 +08:00
ollama generate huggingface.co/thebloke/llama-7b-ggml "hi"
2023-06-28 00:44:12 +08:00
> Downloading [================> ] 66.67% (2/3) 30.2MB/s
2023-06-28 00:08:52 +08:00
```
2023-06-28 01:46:46 +08:00
## Documentation
- [Development](docs/development.md)