← Volver a los artículos
April 19, 2026
5 min de lectura

Kronos: A Foundation Model That Teaches Candlestick Charts to Speak Transformer Language

Kronos: A Foundation Model That Teaches Candlestick Charts to Speak Transformer Language
#kronos
#machine learning
#transformer
#foundation model
#qlib
#review
#open-source
#tokenization

Kronos — candlestick tokenization

If the market is noise, then any attempt to predict the next candle is like decoding radio static. The authors of Kronos propose a radical approach: convert exchange candles into discrete tokens (like words in a language) and train a transformer to predict the "next word" — that is, the next candle.

This isn't a universal time-series transformer "for all occasions" but a specialized model for multidimensional K-lines (OHLCV) with calendar features. The repository shiyu-coder/Kronos is open under MIT, weights on Hugging Face, paper accepted at AAAI 2026.

Core Idea: Candles = Language

Just as ChatGPT learns to predict the next word in text, Kronos learns to predict the next candle in a price series. But first, you need to solve a key problem: a candle is a continuous vector (open, high, low, close, volume), while transformers work with discrete tokens.

The solution — a two-level architecture:

  1. Tokenizer (BSQ) — compresses the continuous candle into a discrete code
  2. Decoder (transformer) — predicts the next code autoregressively

BSQ Tokenizer: How to Compress a Candle into Two Numbers

BSQ tokenizer Kronos

The key module — Binary Spherical Quantizer (BSQ):

  • The continuous candle vector is projected onto a sphere (F.normalize)
  • Then quantized into a binary code
  • The code is split into two levels: coarse (S1) and detailed (S2)

This is vocabulary factorization: instead of one massive dictionary of size S1 × S2, the model uses two small embedding tables — much more efficient.

Hierarchical Decoder: First the Scenario, Then Details

The dual head (DualHead) works in sequence:

  1. First the model predicts S1 — the "rough scenario" (going up? down? approximately how much?)
  2. Then with S1 fixed, refines S2 — details (exact prices, volume)

Links

Conclusion

Kronos isn't "yet another transformer for time series" but a coherent pipeline: BSQ tokenizer with codebook control, hierarchical language model with two discretization levels, autoregression with two-stage sampling, local window normalization. It's an open implementation of the "candles = language" idea.

blog.disclaimer

MarketMaker.cc Team

Investigación Cuantitativa y Estrategia

Discutir en Telegram
Newsletter

Mantente a la vanguardia

Suscríbete a nuestro boletín para recibir información exclusiva sobre trading con IA, análisis de mercado y actualizaciones de la plataforma.

Respetamos tu privacidad. Puedes darte de baja en cualquier momento.