(ql:quickload :cl-llama) (cl-llama:load-model "path/to/llama-2-7b.Q4_K_M.gguf") (cl-llama:generate "Once upon a time") If a Lisp library expects local weights:
(ql:quickload :cl-markov) (e.g., Shakespeare):
wget https://www.gutenberg.org/files/100/100-0.txt -O shakespeare.txt
| Model Source | Command / Link | |--------------|----------------| | | wget https://huggingface.co/gpt2/resolve/main/model.safetensors | | BERT | wget https://huggingface.co/bert-base-uncased/resolve/main/pytorch_model.bin | | CodeLlama (7B) | Request from Meta, then download .gguf from Hugging Face |
For a modern LLM generator in Lisp, use (easy) or cl-llama + llama.cpp (more powerful). Avoid implementing transformers from scratch unless educational.
(ql:quickload :cl-gpt2) It automatically downloads a (124M parameter GPT-2) from Hugging Face (~500 MB) on first use.