Thoth AI

THOTH AI BLOG

BLOG POST

ALiBi: The Simple Bias Adjustment That Enables Transformers to Handle Long Contexts Without Losing Track

Consider a scenario in extending a model’s context window for tasks like analyzing extensive legal transcripts that could reach 100,000 tokens. Various approaches might be tried, such as scaling rotary positional embeddings (RoPE), fine-tuning with methods like LongLoRA, or increasing computational resources. However, when testing on longer sequences not seen

Read More »

The Future of Innovation
Starts Here.

The Future
of Innovation
Starts Here.

a close-up of a molecule

Expertise

A purple and blue cube on a white background.

Resources