Standard Large Language Models (LLMs) are trained on a simple objective: Next-Token Prediction (NTP). By maximizing the probability of the immediate subsequent token xt+1, given the previous context, models have achieved remarkable fluency and reasoning capabilities.
However, this approach is really inefficient as the model has to spend the same amount of compute in predicting filler words (eg, “the”, “and”, “have”) as information-carrying words (eg, “red”, “apple”, “lazy”). This is exacerbated by the fact that more than 50% of the words you see in the English language are filler (Nordquist, 2024)3. This raises a practical question: Do all words need a full inference cycle to be predicted, or do models already have the filler words in their hidden states long before they are predicted?
Motivation For MTP
The idea that transformers are capable of processing more than just the immediate next step is supported by recent empirical research. (Pal et al., 2023)1 demonstrated that the internal representations of transformer models often encode trajectories of future text long before they are generated.
To illustrate, the researchers performed a “transplantation” experiment. They extracted the hidden states from a model processing the sentence “Madison Square Garden is located in…”— just before it was about to predict the next word as “New.” They then placed this vector into a model processing a completely unrelated context, such as “Tell me something about…” Despite the unrelated prompt, the model autoregressively completed the sentence as “Tell me something about New York City.” This confirmed that the model did not just encode solely for the next token, but for the entire future sequence.
To capitalize on this latent capacity of LLMs, researchers at Meta FAIR (Gloeckle et al., 2024)2 propose a novel approach. Instead of treating this foresight as an emergent byproduct, they explicitly use it as a training objective. By tasking the model with predicting “n” future tokens simultaneously at each position instead of just one, they were effectively able to make the model look ahead. The authors demonstrate that the Multi-Token Prediction (MTP) paradigm yields significantly stronger performance on various benchmarks while boosting inference speeds to up to 3 times faster than the baseline.
The MTP Architecture: Parallelizing Prediction
If the information for the next few tokens is already embedded in the current hidden states of LLMs, the question then becomes architectural: How do we extract this information in advance, without increasing the compute requirements compared to standard NTP?
The architecture proposed by the authors aims to modify the existing transformer backbone to predict n future tokens simultaneously. Unlike the standard NTP paradigm, where the cross-entropy loss is minimized for the immediate next token (xt+1) only, Multi-Token Prediction (MTP) minimizes the average loss over n different output heads:

xt+i: Represents future “i” tokens
x1:t: Represents the prompt context
P
To implement this, the authors divide the model into two components:
- A Shared Trunk (fs): The bulk of the model is a standard transformer backbone, whose job is to process the prompted context x1:t into an information-dense global representation zt, which will be used for all subsequent predictions.
- Independent Heads (fh_i): The output of the trunk is fed to n independent heads. Each head has its own transformer layer and is responsible for predicting a future offset token (e.g., head 1 predicts t+1, head 2 predicts t+2, etc.).
Ultimately, the output of each individual head is passed to the shared un-embedding layer, which is implemented as a simple linear projection from the model’s hidden dimension to the length of the vocabulary. The diagram below serves to sum up the most important aspects of the MTP architecture:

The model processes the shared trunk only once. Then, it activates each head sequentially. For steps 4-6, it activates the first head, calculates its logits, and then backpropagates the changes in steps 6-8. Head 2 is activated in a similar fashion, followed by heads 3 and 4.
Overcoming the Memory Bottleneck
The architecture described above presents a significant engineering hurdle: GPU memory utilization.
The vocabulary size (V) of Large Language Models is typically in the realm of 32k-256k, which is astronomically big. This makes the raw prediction scores for every word in the vocabulary, aka the output logits, also very big. In a standard NTP setup, the model needs to materialize these logits only once per step, making it tractable. However, in the MTP setup, n different sets of these massive logits are produced simultaneously, which can easily overwhelm the GPU memory. This makes the MTP method impractical for researchers, unless they drastically reduce batch sizes, slowing down the entire training process.
The authors circumvent this bottleneck with a sequential forward/backward pass strategy. Rather than computing the loss for all n heads at once, the training loop iterates through them sequentially:
- The shared trunk computes the latent state zt.
- The model computes the logits for head 1, calculates the loss, backpropagates gradients throughout the entire model, and immediately discards the logits from memory.
- It then repeats this process for head 2, head 3, and so on.
By deleting these massive logit vectors from memory after each head computation, the peak memory usage of the training process remains O(V) instead of O(nV). This allows the MTP models to be trained in similar batch sizes as the standard models.
Critical Design Choices
Beyond memory optimization, the authors also made two specific design decisions that are important to understand the performance metrics and scientific validity of MTP.
1. The Parameter Parity Constraint
In an MTP model with n=4 heads, the four additional head layers with transformer backbones lead to an increase in parameters. To compensate for this increase, the authors removed an equivalent number of layers from the model’s trunk, making it shallower. This is done so that any performance changes in the MTP with respect to the baseline can be solely credited to the MTP architecture itself, and not to the increase in parameters of the model.
The fact that MTP still outperforms standard NTP-based models despite having a shallower trunk only goes on to show the merits of the architecture.
2. Head Topology: Parallel vs. Causal
The authors also experimented with the arrangement of the heads themselves, specifically comparing two approaches:
- Parallel Heads: This is the standard MTP design described above. In this design, every head predicts its specific future token based only on the shared state zt, without seeing the predictions of other heads.
- Causal Heads: In this setup, head 2 (predicting t+2) would receive the output of head 1 as input. This creates a “mini-autoregressive” chain at the end of the model, which allows each head to look at the state of the previous head. The architecture of MTP with n=4 causal heads is given below:

In the causal design, heads are arranged in a sequential order. This is done so that each head knows what the head preceding it predicted.
Surprisingly, the Parallel design performed better. The authors hypothesize that in the design with causal heads, the shared trunk “got lazy,” relying on the heads to figure out the sequential information. But by forcing the heads to act independently, the trunk was effectively coerced into learning a global representation, which could satisfy all heads at once. This is the exact property that also manifests itself as the model’s ability to plan into the future, which is essential in reasoning tasks.
Experimental Results: The Scale of Improvement
The authors conducted extensive evaluations comparing MTP models against standard Next-Token Prediction (NTP) baselines across model sizes ranging from 300M to 13B parameters.
1. The “Scaling Law” of Multi-Token Prediction
Arguably, the most interesting finding is that the model’s performance scales with its size. For smaller models from 300M-1.3B parameters, the difference between MTP and NTP is negligible (oftentimes MTP performs worse). But as the size increases, MTP starts to perform significantly better than the baseline. As illustrated below, MTP outperforms NTP by 17% on the MBPP benchmark and 12% on the HumanEval benchmark.

Note: These graphs depict the absolute point changes compared to the baseline. For example, in the top left graph, the 13B NTP model scored 26% on the MBPP benchmark while MTP scored 30.5%, which is a 4.5% point increase in absolute terms and 17% increase in relative terms.
A possible reason behind this disparity could stem from the fact that larger models, with their larger parameter counts, can afford to allocate more capacity to future planning than smaller models can. This allows the bigger models to take advantage of the multi-token objective to develop superior reasoning.
2. Three-Fold Inference Speedup via Self-Speculation
Apart from performance metrics, MTP also solves one of the most persistent bottlenecks in LLM operations: inference latency.
To fully appreciate this contribution, we must first understand what Speculative Decoding is. In standard inference, the model has to iteratively generate tokens. It has to wait for xt to be generated before computing xt+1. Speculative decoding speeds this process up by using a smaller, faster draft model (usually of the same family as the main model but with many fewer parameters), which takes in the hidden state from the main model and predicts the next few tokens. The main model is then tasked to verify all of these tokens in a single forward pass, ensuring it agrees with the predictions of the smaller model. Since a single forward pass is faster than generating tokens through numerous iterations, this results in a net speedup. (Read more about Speculative Decoding)
Speculative decoding generally requires a smaller model to be loaded into memory, which can be memory-intensive. However, the authors propose that the extra MTP heads—usually discarded after training—can be used to serve the role of a built-in draft model. As these heads share the same trunk, these heads are highly accurate drafters. By using up to four heads to draft a subsequence and then verifying it in parallel, MTP achieves a 3x speedup in inference with zero loss in performance accuracy.
4. Faster Formation of “Induction Heads”
The authors also analyze the emergence of induction capabilities in MTP. Induction heads are circuits in transformers that are mainly responsible for pattern-matching abilities (e.g., recognizing that [A]…[B]…[A] is likely followed by [B]). The graph below shows that for smaller model sizes, MTP shows a greater induction ability than similarly sized NTP models. This indicates that by forcing the model to predict the consequences of the immediate next token, it creates a gradient signal that is conducive to the emergence of pattern recognition and in-context learning.

The authors took 100 children’s stories and replaced the names of characters with names that span two tokens. The induction success plotted on the y-axis is the accuracy with which the model correctly predicts the second token of the two-token names, given that the name has been shown to the model at least once before.
5. Unlocking Byte-Level Training
In a more radical experiment, the authors applied MTP to byte-level models, which predict a sequence of bytes instead of token representations. Historically, byte-level models have always performed poorly because contextual information among bytes is weak, and byte sequences tend to become very large. However, as demonstrated in the table below, with n=8 heads (predicting 8 bytes at once), the MTP model significantly outperforms the baseline NTP with n=1 head, consistently across all three benchmarks. This suggests that the MTP model can efficiently navigate the byte-realm, allowing models to process raw data natively without any compromises in performance.

This table presents the Pass@k accuracies of the MTP and NTP models on different benchmarks. For example, the column @10 measures the probability that at least one of the top 10 solutions generated by the model is correct.
The Price of Foresight: Shortcomings and Trade-offs
While Multi-Token Prediction offers a compelling alternative to the standard paradigm, the paper’s results clarify that it is not a universal “silver bullet.” The architecture introduces specific trade-offs that engineers must consider.
1. Regression on Knowledge-Intensive Task
While MTP improves reasoning (how to structure an answer), it appears to hurt retrieval (knowing a specific fact).
As shown below, MTP models dominate in code generation and reasoning benchmarks, but actually underperform the baseline on standard NLP tasks, including benchmarks like MMLU, TriviaQA, and ARC Challenge (which test fact retrieval and world knowledge).

The average accuracy across 7 benchmarks, namely arc challenge, copa, hellaswag, nq, piqa, siqa, and tqa, is plotted on the y-axis against the training steps on the x-axis.
A possible explanation can be that answering recall-based questions like “What is the capital of France?” requires a precise focus on the word “Paris”. By forcing the model to predict multiple tokens at once, as in “Paris is a city in…,” it might dilute the overall signal from the most critical token, tanking the model’s performance on the overall benchmark. If your aim is to build a RAG (Retrieval Augmented Generation) system or a Trivia bot, MTP might actually be detrimental.
2. The “Goldilocks” Sensitivity of n
There is no “more is better” rule here. The authors found that performance is highly sensitive to the number of heads (n).
The authors also concluded that the number of heads (n) doesn’t scale linearly with MTP performance. There exists a “sweet spot” where the model can most efficiently exploit the MTP paradigm:
- Too few (n=2): Negligible gain, as the model does not receive enough incentive to develop any foresight.
- Too many (n=8): Performance degrades rapidly, as the information for all 8 heads starts to overcrowd the hidden state of the shared trunk.
- Just right (n=4): Best performance
This introduces a new hyperparameter that must be tuned. Unlike Next-Token Prediction, which just “works,” MTP requires finding the specific horizon that matches the complexity of your data.
Conclusion
With its demonstrated ability to improve coding performance and inference speedups, one obvious question remains: If MTP is so revolutionary, why haven’t any major AI labs used it yet?
The answer to it is actually DeepSeek-V3.
In their technical report (Liu et al., 2024)4, the DeepSeek team revealed that MTP was a core component during training of the model. Similar to Meta, they performed vigorous ablation studies comparing standard NTP models against MTP at both the 15.7B and 228.7B parameter scales. Using a configuration of n=2 during training (predicting one extra future token), they found that MTP-trained models consistently outperformed their NTP counterparts across all datasets, like MMLU, PILE-test, HumanEval, MBPP, etc. Moreover, by keeping that second prediction head during inference for speculative decoding as described earlier, DeepSeek achieved an inference speedup of up to 1.8x.
This successful deployment by DeepSeek serves as practical validation for MTP to be widely used as a training objective in Large Language Models, as it demonstrates a clear path to improving the reasoning capabilities and inference efficiency of the model with minimal associated drawbacks.
If you like these kinds of breakdowns, I share more insights, notes, and explainers here: https://steadysurfdom.substack.com/
References
[1] Pal, Koyena, et al. “Future lens: Anticipating subsequent tokens from a single hidden state.” arXiv preprint arXiv:2311.04897 (2023).
[2] Gloeckle, Fabian, et al. “Better & faster large language models via multi-token prediction.” arXiv preprint arXiv:2404.19737 (2024).
[3] Nordquist, R. (2024, July 20). Definition and examples of function words in English. ThoughtCo.
[4] Liu, Aixin, et al. “Deepseek-v3 technical report.” arXiv preprint arXiv:2412.19437 (2024).
Source link
#Weve #Optimizing #Wrong #LLMs #Years









