llama-desktop 2.0.2

Desktop interface for Ollama
[MIT]: https://github.com/cacilhas/llama-desktop/blob/master/COPYING
[Llama]: https://raw.githubusercontent.com/cacilhas/llama-desktop/master/ui/icon.png
[Ollama]: https://ollama.ai/

# Llamma Desktop

![Llama][]

Desktop app to connect to [Ollama][] and send queries.

Llama Desktop reads the Ollama service URI from the environment variable
`OLLAMA_HOST`, defaults to `http://localhost:11434`.

## Installation

```sh
cargo install llama-desktop
```

## License

- [MIT][]