mirror of
https://github.com/jackyzha0/quartz.git
synced 2025-12-20 03:14:06 -06:00
First draft
This commit is contained in:
parent
2de928c576
commit
1d8c387366
BIN
content/assets/VSR_results.png
Normal file
BIN
content/assets/VSR_results.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 100 KiB |
129
content/research/Can AI Models Predict What Youll Say Next.md
Normal file
129
content/research/Can AI Models Predict What Youll Say Next.md
Normal file
@ -0,0 +1,129 @@
|
|||||||
|
---
|
||||||
|
title: Can AI Models Predict What You'll Say Next? Developing Verifiable Social Rewards
|
||||||
|
author: Dani Balcells
|
||||||
|
date: 02.28.25
|
||||||
|
tags:
|
||||||
|
- research
|
||||||
|
- ml
|
||||||
|
---
|
||||||
|
|
||||||
|
|
||||||
|
## TL;DR
|
||||||
|
|
||||||
|
We developed a benchmark to evaluate how well language models can predict social interactions in conversation settings. By testing various models on the task of predicting the next message in real Discord conversations, with and without different types of context, we found that Claude 3.7 Sonnet significantly outperforms other models. We discovered that generating context summaries with a smaller model (Llama 3.3 70B) and injecting these into inference yields comparable or better results than providing raw conversation history. On one hand, this validates key aspects of the thesis behind our product Honcho. On the other hand, we discovered that models highly optimized for technical reasoning often underperform on social cognition tasks.
|
||||||
|
|
||||||
|
## The Challenge: Finding Verifiable Social Rewards
|
||||||
|
|
||||||
|
The machine learning community has made significant progress optimizing language models for tasks with clear, verifiable answers—like math, coding, and factual reasoning. These domains offer what are called "verifiable rewards"—objective measures that can be used for reinforcement learning without relying on human preferences or subjective judgments.
|
||||||
|
|
||||||
|
While this approach has yielded impressive results for technical reasoning, at Plastic Labs we've become increasingly curious about whether similar verifiable reward structures could be developed for social intelligence—an area that has largely relied on more subjective evaluation metrics.
|
||||||
|
|
||||||
|
To address this gap, we developed a multiple-choice next-message prediction task using real conversations from our team's Discord. The premise is straightforward: given a snippet of conversation between two people and four possible options for what came next (with only one being the actual message), can a model identify the genuine response?
|
||||||
|
|
||||||
|
This creates a clear, verifiable reward signal for social understanding: either the model correctly identifies the real message or it doesn't. Yet unlike many technical tasks, success requires the model to understand conversational dynamics, recognize individual communication patterns, track context across multiple turns, and model how specific people behave in specific social contexts.
|
||||||
|
|
||||||
|
This benchmark also allows us to test whether models specifically optimized for technical reasoning excel at social understanding, and to get a granular, quantifiable understanding of models' social reasoning abilities.
|
||||||
|
|
||||||
|
## Methodology
|
||||||
|
|
||||||
|
### Dataset Creation
|
||||||
|
|
||||||
|
We created our dataset by extracting conversation snippets from our internal team Discord channels (accessible only to our core team of 5-10 people). Each snippet contained:
|
||||||
|
|
||||||
|
- 6-10 messages between exactly two participants
|
||||||
|
- At least two messages from each participant, ensuring each person had at least two conversational turns
|
||||||
|
- A target message (the message that was sent immediately after the snippet)
|
||||||
|
|
||||||
|
For each genuine target message, we generated three convincing decoys using Claude 3.7 Sonnet. These decoys were designed to be plausible alternatives that matched the style and substance of the real message while being distinctly different.
|
||||||
|
|
||||||
|
We specifically chose to use our internal Discord data because it represents authentic, messy communication—complete with inconsistent formatting, shorthand, emojis, and other real-world conversational artifacts that make this a particularly challenging test.
|
||||||
|
|
||||||
|
### Context Modes
|
||||||
|
|
||||||
|
Upon visual inspection of the resulting dataset, we found that the decoys were remarkably similar to the real messages, making it difficult even for us to consistently identify the genuine response. We wondered if providing additional context about the users might help determine the correct answer, which led us to explore different context modes:
|
||||||
|
|
||||||
|
1. **No Context**: Models only received the immediate conversation snippet and the four options.
|
||||||
|
2. **Raw Context**: Models received the conversation snippet plus the previous 50 or 100 messages from the Discord history (unfiltered and unprocessed).
|
||||||
|
3. **Summary Context**: Models received the conversation snippet plus a generated personality profile of the target user, created by processing the previous 50 or 100 messages through Llama 3.3 70B.
|
||||||
|
|
||||||
|
This design allowed us to compare whether any context provides useful signals for predicting social behavior, and whether a summary can provide results comparable to the full context.
|
||||||
|
|
||||||
|
### Experimental Setup
|
||||||
|
|
||||||
|
We tested a wide range of models including:
|
||||||
|
- Claude 3.7 Sonnet, Claude 3.5 Sonnet, Claude 3.5 Haiku
|
||||||
|
- GPT-4.5, GPT-4o, GPT-4o Mini, O-1, O-3 Mini
|
||||||
|
- Google's Gemini 2.0 Flash
|
||||||
|
- Meta's Llama 3.3 70B Instruct
|
||||||
|
- Nous Research's Hermes 3 (405B and 70B variants)
|
||||||
|
- DeepSeek models (Chat and R1)
|
||||||
|
|
||||||
|
For each model and context mode combination, we ran three trials with different random seeds to control for position bias in option selection. Ideally we would have run more trials, but we wanted to constrain the compute needed for this experiment.
|
||||||
|
|
||||||
|
## Results and Discussion
|
||||||
|
|
||||||
|

|
||||||
|
|
||||||
|
Our evaluation produced several notable results:
|
||||||
|
|
||||||
|
### Context Helps Regardless of Form
|
||||||
|
|
||||||
|
Additional context helps models predict social behavior, whether that context is provided as raw conversation history or as a processed summary. Moving from no context to either raw or summary context yielded substantial improvements for virtually all models tested. This confirms what might seem intuitive: knowing more about someone helps predict what they might say next.
|
||||||
|
|
||||||
|
### Efficient Context Processing Works
|
||||||
|
|
||||||
|
What's particularly significant is that injecting pre-processed summaries of user context works as well as or better than providing raw context for most models. This has important implications for system design:
|
||||||
|
|
||||||
|
1. The summaries contain far fewer tokens than raw context (approximately one paragraph versus potentially thousands of tokens)
|
||||||
|
2. The summarization can be done once with a smaller, cheaper model
|
||||||
|
3. The resulting performance gains are substantial compared to no-context baselines, and in some cases even better than providing the full context
|
||||||
|
|
||||||
|
This supports a core thesis behind Honcho: ambient processing of user context to generate compressed representations can improve model performance while keeping inference costs manageable. Rather than injecting massive amounts of data into the context window, models can achieve better results with distilled personality profiles.
|
||||||
|
|
||||||
|
We didn't observe significant performance differences between 50-message and 100-message contexts, suggesting there may be diminishing returns beyond a certain point. This is likely dependent on factors like user count and conversation density.
|
||||||
|
|
||||||
|
### Newest Models Lead the Way
|
||||||
|
|
||||||
|
Only the newest models perform well on this task. Claude 3.7 Sonnet (released last week) and GPT-4.5 (released yesterday) were the only models to achieve accuracy significantly above 40% in any context mode, with Claude 3.7 reaching nearly 60% accuracy with summary context—more than doubling the 25% random baseline.
|
||||||
|
|
||||||
|
This is particularly interesting because tasks that would have seemed impossible for models that existed just months ago are now becoming tractable. This rapid progress also informs how we should think about designing evaluations—creating hard tasks that aren't saturated from the start rather than ones where models already perform at ceiling.
|
||||||
|
|
||||||
|
### Different Models Benefit from Different Contexts
|
||||||
|
|
||||||
|
While summary context generally outperformed raw context, this pattern wasn't universal. Some models (notably Claude 3.5 Sonnet) performed better with raw context than with summaries. This suggests different architectures may vary in their ability to extract relevant information from different types of context.
|
||||||
|
|
||||||
|
### Reasoning and Social Understanding May Involve Trade-offs
|
||||||
|
|
||||||
|
The relatively poor performance of models optimized for technical reasoning, like DeepSeek R1, O-1, and O-3 Mini, raises interesting questions. Despite their strong results on math and coding benchmarks, these models achieved well below random performance on our social prediction task.
|
||||||
|
|
||||||
|
This suggests potential trade-offs in model optimization. The reinforcement learning or supervised fine-tuning techniques used to enhance reasoning abilities might come at the expense of social cognition capabilities, possibly through:
|
||||||
|
|
||||||
|
1. Training data composition that emphasizes technical content over social interactions
|
||||||
|
2. Alignment techniques that optimize for step-by-step reasoning rather than social intuition
|
||||||
|
3. Reinforcement learning objectives that prioritize factual accuracy over social understanding
|
||||||
|
|
||||||
|
### Caveat: Decoy Generation
|
||||||
|
|
||||||
|
We should note that our decoys were generated using Claude 3.7 Sonnet, which was also the best-performing model on the task. It's possible that Claude 3.7 is better at recognizing the subtleties in its own generations. However, this almost creates a generative adversarial setup—Claude 3.7 is both generating challenging decoys and trying to identify them—which makes its strong performance even more notable.
|
||||||
|
|
||||||
|
## Future Directions
|
||||||
|
|
||||||
|
### Verifiable Social Rewards for RL
|
||||||
|
|
||||||
|
So far, we've used this task purely as an evaluation metric, but with a large enough dataset, it could potentially serve as a reward signal for reinforcement learning. This would allow for optimization of social cognition abilities with objective metrics, similar to how technical reasoning has been enhanced. Expanding our toolkit of objective social evaluation metrics could help bridge the gap between technical and social intelligence.
|
||||||
|
|
||||||
|
### Social-Reasoning Balance
|
||||||
|
|
||||||
|
Can we develop training techniques that enhance reasoning capabilities without sacrificing social cognition? This might involve carefully designed datasets that balance technical and social tasks, or novel fine-tuning approaches that preserve multiple types of capabilities. Understanding the apparent trade-off between these abilities could be crucial for developing more well-rounded AI systems.
|
||||||
|
|
||||||
|
### Context Optimization and Alternative Approaches
|
||||||
|
|
||||||
|
We're also interested in exploring several technical improvements to the methodology: finding the minimum effective context window size across different environments; testing different prompting techniques and models for generating personality summaries; experimenting with combinations of raw and summary contexts; and trying different models for decoy generation to address potential advantages Claude 3.7 might have in recognizing its own outputs.
|
||||||
|
|
||||||
|
## Conclusion
|
||||||
|
|
||||||
|
We were excited to find that this social prediction task was genuinely challenging for most current models, with only the very latest releases showing strong performance. The fact that models optimized for reasoning performed poorly suggests interesting trade-offs in current training approaches. Meanwhile, the effectiveness of pre-processed context summaries supports a key principle behind Honcho: ambient processing of user context can significantly improve personalization while managing compute costs.
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
*If you're interested in discussing this research or exploring how improved social modeling could benefit your AI applications, [join our Discord](https://discord.gg/plasticlabs) or reach out to us at [hello@plasticlabs.ai](mailto:hello@plasticlabs.ai).*
|
||||||
Loading…
Reference in New Issue
Block a user