From 2d33aa547ff6a516c90ca2b47b13e2add200583a Mon Sep 17 00:00:00 2001 From: Ben Sima Date: Tue, 14 May 2024 09:35:45 -0400 Subject: Add simonw/llm as cli/library client for running LLMs This is basically exactly the client library that I would write myself. Some parts of it are still beta quality, but it's the sort of thing that I would contribute to anyway. Unfortunately I couldn't get the llm-llama-cpp plugin to work because it depends on llama-cpp-python which is not packaged for nix and is hard to package because the upstream project vendors a patched version of llama.cpp. So I'm stuck with ollama for now, but that's fine because it actually works. --- Biz/Bild/Sources.json | 13 +++++++++++++ 1 file changed, 13 insertions(+) (limited to 'Biz/Bild/Sources.json') diff --git a/Biz/Bild/Sources.json b/Biz/Bild/Sources.json index 40bcdfb..3b1e4fd 100644 --- a/Biz/Bild/Sources.json +++ b/Biz/Bild/Sources.json @@ -77,6 +77,19 @@ "url": "https://github.com/MegaIng/interegular/archive/v0.2.1.tar.gz", "url_template": "https://github.com///archive/.tar.gz" }, + "llm": { + "branch": "main", + "description": "Access large language models from the command-line", + "homepage": "https://llm.datasette.io", + "owner": "simonw", + "repo": "llm", + "rev": "0.13.1", + "sha256": "0305xpmigk219i2n1slgpz3jwvpx5pdp5s8dkjz85w75xivakbin", + "type": "tarball", + "url": "https://github.com/simonw/llm/archive/0.13.1.tar.gz", + "url_template": "https://github.com///archive/.tar.gz", + "version": "0.13.1" + }, "niv": { "branch": "master", "description": "Easy dependency management for Nix projects", -- cgit v1.2.3