Inference package for Guido language models — model code, HuggingFace loader, cartridge support, CLI.
Project description
guido-inference
Run Guido models locally. Model code with Titans memory update during inference, HuggingFace loader, cartridge support, and CLI.
Guido's memory evolves as you talk to it. Save the memory state as a .cart cartridge file and reload it later — the model picks up where it left off.
Installation
pip install guido-inference
For Flash Linear Attention kernel support (GPU):
pip install guido-inference[fla]
Quick Start
CLI
guido run northsea-ai/Guido-3B --prompt "What is the capital of the Netherlands?"
guido run northsea-ai/Guido-3B # Interactive mode
guido run northsea-ai/Guido-3B --effort high # Adaptive compute
guido run northsea-ai/Guido-3B --cartridge ./john.cart # Resume with saved memory
guido info northsea-ai/Guido-3B # Model info (no GPU)
Python
from guido_inference import load_model
model, tokenizer = load_model("northsea-ai/Guido-3B")
input_ids = tokenizer.encode("Hello!", return_tensors="pt").to(model.device)
output = model.generate(input_ids)
print(tokenizer.decode(output[0], skip_special_tokens=True))
Cartridges — Create, Save, Reload
Memory updates happen automatically during inference (Titans architecture). Save and reload the evolved state:
# Conversation evolves the model's memory
for msg in ["I'm a researcher at TU Delft.", "My focus is renewable energy."]:
ids = tokenizer.encode(msg, return_tensors="pt").to(model.device)
model.generate(ids)
# Save evolved memory as portable cartridge (~250KB-1MB)
model.save_memory("./researcher.cart")
# Later: reload and continue where you left off
model.load_memory("./researcher.cart")
output = model.generate(tokenizer.encode("What papers should I read?", return_tensors="pt").to(model.device))
# Reset memory to blank slate
model.reset_memory()
Model Downloads
| Model | Params (Total) | Params (Active) | HuggingFace |
|---|---|---|---|
| Guido-300M | 0.42B | 0.29B | northsea-ai/Guido-300M |
| Guido-3B | 3.33B | 0.84B | northsea-ai/Guido-3B |
| Guido-7B | 7.36B | 1.32B | northsea-ai/Guido-7B |
For fine-tuning, use third-party training harnesses like Axolotl or LLaMA-Factory with a LoRA/QLoRA adapter.
See the monorepo README for architecture details and full documentation.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file guido_inference-0.1.0.tar.gz.
File metadata
- Download URL: guido_inference-0.1.0.tar.gz
- Upload date:
- Size: 6.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.10.2 {"installer":{"name":"uv","version":"0.10.2","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"macOS","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
654ab300d2a025bb19a442bd6d001901f9d0f2a06e3b8083e2407d7d12ae935f
|
|
| MD5 |
e9d1f24f721f65911271fe0834cd6ab6
|
|
| BLAKE2b-256 |
5d3b72dfefa40d4716512a2b7310146997ae3ba3b8686fc953c36b5c018ed206
|
File details
Details for the file guido_inference-0.1.0-py3-none-any.whl.
File metadata
- Download URL: guido_inference-0.1.0-py3-none-any.whl
- Upload date:
- Size: 6.2 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.10.2 {"installer":{"name":"uv","version":"0.10.2","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"macOS","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
665f2fe24e51a6c0725265526ab734a0476cab29b26f870e81743599d6f7214f
|
|
| MD5 |
03ac86c108b656d5bf450f9dcc295ab8
|
|
| BLAKE2b-256 |
67ee11f55472162eb72d79271ab600508e55ad94a06e26d641b8f9a548483fba
|