Skip to main content

Inference package for Guido language models — model code, HuggingFace loader, cartridge support, CLI.

Project description

guido-inference

Run Guido models locally. Model code with Titans memory update during inference, HuggingFace loader, cartridge support, and CLI.

Guido's memory evolves as you talk to it. Save the memory state as a .cart cartridge file and reload it later — the model picks up where it left off.

Installation

pip install guido-inference

For Flash Linear Attention kernel support (GPU):

pip install guido-inference[fla]

Quick Start

CLI

guido run northsea-ai/Guido-3B --prompt "What is the capital of the Netherlands?"
guido run northsea-ai/Guido-3B                          # Interactive mode
guido run northsea-ai/Guido-3B --effort high             # Adaptive compute
guido run northsea-ai/Guido-3B --cartridge ./john.cart   # Resume with saved memory
guido info northsea-ai/Guido-3B                          # Model info (no GPU)

Python

from guido_inference import load_model

model, tokenizer = load_model("northsea-ai/Guido-3B")
input_ids = tokenizer.encode("Hello!", return_tensors="pt").to(model.device)
output = model.generate(input_ids)
print(tokenizer.decode(output[0], skip_special_tokens=True))

Cartridges — Create, Save, Reload

Memory updates happen automatically during inference (Titans architecture). Save and reload the evolved state:

# Conversation evolves the model's memory
for msg in ["I'm a researcher at TU Delft.", "My focus is renewable energy."]:
    ids = tokenizer.encode(msg, return_tensors="pt").to(model.device)
    model.generate(ids)

# Save evolved memory as portable cartridge (~250KB-1MB)
model.save_memory("./researcher.cart")

# Later: reload and continue where you left off
model.load_memory("./researcher.cart")
output = model.generate(tokenizer.encode("What papers should I read?", return_tensors="pt").to(model.device))

# Reset memory to blank slate
model.reset_memory()

Model Downloads

Model Params (Total) Params (Active) HuggingFace
Guido-300M 0.42B 0.29B northsea-ai/Guido-300M
Guido-3B 3.33B 0.84B northsea-ai/Guido-3B
Guido-7B 7.36B 1.32B northsea-ai/Guido-7B

For fine-tuning, use third-party training harnesses like Axolotl or LLaMA-Factory with a LoRA/QLoRA adapter.

See the monorepo README for architecture details and full documentation.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

guido_inference-0.1.0.tar.gz (6.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

guido_inference-0.1.0-py3-none-any.whl (6.2 kB view details)

Uploaded Python 3

File details

Details for the file guido_inference-0.1.0.tar.gz.

File metadata

  • Download URL: guido_inference-0.1.0.tar.gz
  • Upload date:
  • Size: 6.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.10.2 {"installer":{"name":"uv","version":"0.10.2","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"macOS","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}

File hashes

Hashes for guido_inference-0.1.0.tar.gz
Algorithm Hash digest
SHA256 654ab300d2a025bb19a442bd6d001901f9d0f2a06e3b8083e2407d7d12ae935f
MD5 e9d1f24f721f65911271fe0834cd6ab6
BLAKE2b-256 5d3b72dfefa40d4716512a2b7310146997ae3ba3b8686fc953c36b5c018ed206

See more details on using hashes here.

File details

Details for the file guido_inference-0.1.0-py3-none-any.whl.

File metadata

  • Download URL: guido_inference-0.1.0-py3-none-any.whl
  • Upload date:
  • Size: 6.2 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.10.2 {"installer":{"name":"uv","version":"0.10.2","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"macOS","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}

File hashes

Hashes for guido_inference-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 665f2fe24e51a6c0725265526ab734a0476cab29b26f870e81743599d6f7214f
MD5 03ac86c108b656d5bf450f9dcc295ab8
BLAKE2b-256 67ee11f55472162eb72d79271ab600508e55ad94a06e26d641b8f9a548483fba

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page