As we enter the new year, I wanted to share some of my (tentative) predictions about AI development in 2025. Below, I predict three changes that I believe will have profound impact on AI research and development. This post is inherently an incomplete list, many other predictions could be added but are omitted (e.g. around agent use-cases). For simplicity, I refer to the largest multi-modal generative AI models as “AI” here — my predictions don’t necessarily apply to other AI models. Disclaimer: These predictions were made at the beginning of 2025 and may very well turn out to be wrong.


Description of image

Interacting with AI models via text-based chat interfaces can be tedious. Voice-based interfaces promise a more natural interaction. Yet, traditional voice-based interfaces (such as Google Assistant, Alexa, or Siri) lack the latency and reasoning capabilities to serve as a satisfying voice-based alternative to ChatGPT. Traditionally, such interfaces rely on multiple separate components to solve speech recognition, reasoning and text-to-speech tasks. In 2024, we saw significant progress towards (probably) fully end-to-end models like OpenAI’s GPT-4o via Advanced Voice Mode or the Realtime API. Yet, these methods are still prohibitively expensive for many use-cases and interactions still leave something to be desired, e.g. they could feel more natural. I believe that in 2025 we will see further progress in this front as other developers release their Advanced Voice Mode competitors. This development will transfer users from text-based to voice-based AI interfaces and enable AI use in scenarios where text-only interfaces prevented use. Use-cases for which a call with a person is perceived as more helpful or easier than text chat will likely be most affected. Whilst claiming an increase of voice-based AI interactions is not controversial, I believe people generally under-appreciate the extent of the incoming shift towards voice. In my opinion, voice may eventually become the dominant (but not only) interface with AI models.


Description of image

Prediction 2: AI will become more personalized

There is no single person that everyone likes to talk to. People differ in who they enjoy talking with. Similarly, it is unlikely that we want to keep to talking to effectively the same vanilla AI models. Many interfaces already provide some form of personalization (e.g. ChatGPT’s memory), though the overall level of personalization remains limited. I expect the level of personalization to increase beyond what is currently offered. In particular, I expect that, as for many chat-based tasks models solve tasks functionally reliably well, the focus in 2025 will be more on the style and personality of models. A key bottleneck will be trust: full personalisation requires access to extensive private data. Each model provider’s credibility in protecting such data will be critical for such use-cases. Personally, there is a lot of useful data that I am unwilling to send to most providers (e.g. health data). On-device models may be able to address these privacy concerns for some use-cases.


Description of image

Prediction 3: Performance will scale better with inference cost

Prior to the o3 model generation, it was difficult to improve an AI model’s response quality for a given task beyond a certain point by generating more tokens. For many tasks, prompting GPT-4o with some form of chain-of-thought was the best you could do. The marginal utility of spending another dollar on token inference diminished quickly beyond this point. For many tasks, this meant that someone willing to spend 1 million USD on generating a response could not get more utility than someone spending 100 USD. AI responses were a form of satiated good, similar to the iPhone. With the new generation of models that perform some form of search and are fine-tuned for long multi-step reasoning (e.g. OpenAI’s o3 model), the marginal utility of additional spending potentially becomes non-zero even at large scales - spending more money on exploring the solution-space pays off better. I predict that in 2025 AI will increasingly become less of a satiated good: for a given task, the marginal utility of additional AI inference will diminish less quickly. Performance will scale better with inference cost. This change will have a profound impact on the economics of serving AI models.


Conclusion

I look forward to finding out whether and how correct these predictions will have been. Reach out if you have any feedback or thoughts to share!

Acknowledgement

Big thanks to Timo Kaufmann, Benjamin Minixhofer, Victor Bourgin for their helpful feedback on earlier versions of this post. All errors are my own of course.

Citation

If you found this post useful for your work, please consider citing it as:

Findeis, Arduin. (Jan 2025). Some predictions for AI in 2025. Retrieved from https://arduin.io/blog/ai-predictions-2025/.

or
 @article{Findeis2023SomepredictionsforAIin2025,
        title = "Some predictions for AI in 2025",
        author = "Findeis, Arduin",
        journal = "arduin.io",
        year = "2025",
        month = "January",
        url = "https://arduin.io/blog/ai-predictions-2025/"
 }