Top 10 predictions for artificial intelligence in 2025: AI Agents will become mainstream

This article is machine translated
Show original
Here is the English translation of the text, with the specified terms retained and not translated: We now share our world with another form of intelligence that can sometimes be willful, unpredictable, and deceptive. Author: Rob Toews Compiled by: MetaverseHub As 2024 draws to a close, Rob Toews, a venture capitalist at Radical Ventures, shares his 10 predictions for Artificial Intelligence in 2025: 01. Meta will start charging for the Llama model Meta is the global standard-bearer for open-source AI. In a striking case study of corporate strategy, while competitors like OpenAI and Google have closed-sourced their cutting-edge models and started charging for their use, Meta has chosen to provide its state-of-the-art Llama model for free. So the news that next year Meta will start charging companies that use Llama will come as a surprise to many. The key point to understand is: we are not predicting that Meta will fully close-source Llama, nor that any user of the Llama model will be required to pay. Rather, we predict that Meta will put more restrictions on the open-source licensing terms of Llama, such that companies using Llama at scale in commercial environments will need to start paying to use the model. Technically, Meta has already done this to a limited extent. The company does not allow the largest companies - hyperscale cloud providers and others with over 700 million monthly active users - to freely use its Llama model. As early as 2023, Meta CEO Mark Zuckerberg said: "If you're Microsoft, Amazon, or Google, and you're basically going to resell Llama, then we should get a cut of that. I don't think that's going to be a huge revenue stream in the short term, but hopefully over time it can be some revenue." Next year, Meta will significantly expand the range of enterprises that must pay to use Llama, bringing more medium and large-sized companies into that fold. Keeping up with the frontier of large language models (LLMs) is extremely expensive. To keep Llama on par with or close to the latest cutting-edge models from OpenAI, Anthropic, and others, Meta needs to invest billions of dollars per year. Meta is one of the largest and most well-capitalized companies in the world. But it is also a public company, ultimately accountable to shareholders. As the costs of producing frontier models continue to skyrocket, Meta's practice of investing such vast sums to train the next generation of Llama models without any revenue expectations becomes increasingly untenable. Hobbyists, academics, individual developers, and startups will continue to use the Llama model for free next year. But 2025 will be the year Meta starts getting serious about monetizing Llama. 02. The "Scaling Laws" Reckoning One of the hottest topics in the AI world in recent weeks has been the issue of scaling laws, and whether they are about to come to an end. Scaling laws were first proposed in a 2020 paper by OpenAI, and the basic idea is simple: as the number of model parameters, training data, and compute are increased when training AI models, the model's performance improves in a reliable and predictable way (technically, its test loss decreases). The breathtaking performance improvements from GPT-2 to GPT-3 to GPT-4 are all the result of scaling laws. Just like Moore's Law, scaling laws are not actually a true law, but rather a simple empirical observation. Over the past month, a series of reports have suggested that the major AI labs are starting to see diminishing returns as they continue to scale up their large language models. This helps explain why the release of OpenAI's GPT-5 has been repeatedly delayed. The most common counterargument to the scaling laws plateauing is that the emergence of inference-time compute opens up an entirely new dimension along which to pursue scaling. In other words, rather than scaling up the compute during training, new inference models like OpenAI's o3 make it possible to scale up the compute during inference, by allowing the model to "think for longer" to unlock new AI capabilities. This is an important point. Inference-time compute does indeed represent an exciting new avenue for scaling and AI performance improvement. But there is another perspective on scaling laws that is even more important and severely underappreciated in today's discussion. Almost all the discussion of scaling laws, from the original 2020 paper to the current focus on inference-time compute, has centered on language. But language is not the only important data modality. Think about robotics, biology, world models, or web agents. For these data modalities, scaling laws are far from saturated; rather, they are just getting started. In fact, rigorous empirical evidence of scaling laws in these new data modalities has not even been published yet. Startups building foundational models for these new data modalities - such as Evolutionary Scale in biology, PhysicalIntelligence in robotics, and WorldLabs in world models - are trying to identify and leverage the scaling laws in their domains, just as OpenAI did with large language models (LLMs) in the first half of the 2020s. Next year, expect to see tremendous progress here. Scaling laws are not going away; they will be just as important in 2025 as they have been. But the center of gravity of scaling law activity will shift from LLM pretraining to other modalities. 03. Trump and Musk May Diverge on AI Direction The new US administration will bring a series of policy and strategic shifts on artificial intelligence. To predict the direction of AI under a Trump presidency, and given Musk's current centrality in the AI world, one might be tempted to focus on the close relationship between the president-elect and Musk. One can imagine that Musk could influence the Trump administration's AI-related developments in a variety of ways. Given Musk's deep hostility towards OpenAI, the new administration may take a less friendly stance in its engagement with the industry, policymaking around AI regulations, and the awarding of government contracts - a real risk that OpenAI is genuinely concerned about today. On the other hand, the Trump administration may be more inclined to support Musk's own companies: for example, by cutting red tape to allow xAI to build data centers and take the lead in frontier model competitions; providing fast-track regulatory approval for Tesla's robotaxi fleet deployment, and so on. More fundamentally, unlike many other tech leaders favored by Trump, Musk is deeply concerned about the safety risks of AI and thus advocates for significant AI regulation. He has supported the controversial California SB1047 bill, which seeks to impose meaningful constraints on AI developers. So Musk's influence could lead to a tightening of the regulatory environment for AI in the US. However, all of this speculation rests on one key problem: the close relationship between Trump and Musk is ultimately doomed to fall apart. As we have seen time and again during Trump's first term, the average tenure of even his staunchest allies is extremely short. Of the lieutenants in Trump's first administration, only a tiny handful remain loyal to him today. Both Trump and Musk are complex, mercurial, and unpredictable personalities who do not work well together, and who exhaust those around them. Their newfound friendship, while mutually beneficial so far, is still in the "honeymoon" phase. We predict that this relationship will deteriorate before the end of 2025. What does this mean for the world of AI?

Here is the English translation of the text, with the specified terms preserved:

This is good news for OpenAI. For Tesla shareholders, this will be unfortunate news. And for those concerned about AI safety, this will be a disappointing news, as it virtually ensures that the US government will take a hands-off approach to AI regulation during the Trump administration.

04. AI Agents Will Become Mainstream

Imagine a world where you no longer need to directly interact with the internet. Whenever you need to manage subscriptions, pay bills, book doctor appointments, order things on Amazon, make restaurant reservations, or complete any other tedious online tasks, you can simply instruct your AI assistant to do it for you.

The concept of "internet agents" has existed for years. If such a product existed and worked properly, it would undoubtedly be a huge success.

However, there is currently no generally functioning internet agent on the market.

Companies like Adept, even with a blue-chip founding team and hundreds of millions in funding, have yet to realize their vision.

Next year will be the year when internet agents finally start to work well and become mainstream. The continuous progress of language and visual foundation models, combined with recent breakthroughs in "second-system thinking" capabilities due to new reasoning models and reasoning time computation, will mean that internet agents are ready to enter their golden age.

In other words, Adept's idea was right, just premature. In startups, as in life, timing is everything.

Internet agents will find various valuable enterprise use cases, but we believe the biggest near-term market opportunity for internet agents will be in the consumer space.

Despite the recent AI hype, there are relatively few AI-native applications that have become mainstream consumer applications, apart from ChatGPT.

Internet agents will change this, becoming the next true "killer app" in the consumer AI space.

05. The Idea of Placing AI Data Centers in Space Will Be Realized

In 2023, the key physical resource constraining AI development is GPU chips. In 2024, it becomes power and data centers.

In 2024, there will be hardly any story more captivating than the enormous and rapidly growing demand for energy as AI rushes to build more and more AI data centers.

Due to the explosive growth of AI, global data center power demand is expected to double between 2023 and 2026, after remaining flat for decades. In the US, data center electricity consumption is projected to approach 10% of total electricity consumption by 2030, up from just 3% in 2022.

Today's energy systems are simply not equipped to handle the massive surge in demand from AI workloads. A historic collision is about to happen between our energy grid and our computing infrastructure, both worth trillions of dollars.

As a potential solution to this dilemma, nuclear power has seen a resurgence this year. Nuclear power is in many ways an ideal energy source for AI: it is zero-carbon, available 24/7, and effectively limitless.

But realistically, new energy sources cannot solve this problem before the 2030s, due to the long lead times for research, project development, and regulation. This applies to traditional nuclear fission plants, next-generation "small modular reactors" (SMRs), and nuclear fusion power plants.

Next year, a very unconventional new idea to address this challenge will emerge and attract real resources: placing AI data centers in space.

AI data centers in space, at first glance, sounds like a bad joke, a venture capitalist trying to mash too many startup buzzwords together.

But in fact, it may make sense.

The biggest bottleneck to rapidly building more data centers on Earth is acquiring the necessary power. Computational clusters in orbit could enjoy free, limitless, zero-carbon power 24/7: the sun is always shining in space.

Another key advantage of placing computing in space is that it solves the cooling problem.

One of the major engineering hurdles in building more powerful AI data centers is that running many GPUs in a confined space gets extremely hot, and high temperatures can damage or destroy computing equipment.

Data center developers are adopting expensive and unproven methods like liquid immersion cooling to try to solve this. But space is extremely cold, and any heat generated by computing activity would dissipate harmlessly immediately.

Of course, there are many practical challenges to be solved. An obvious one is whether and how to transmit large amounts of data between orbit and Earth efficiently and at low cost.

This is an open question, but may prove solvable: there is promising work being done on using lasers and other high-bandwidth optical communication technologies.

A YCombinator startup called Lumen Orbit has recently raised $11 million to realize this vision: building a multi-gigawatt data center network in space for training AI models.

As the CEO said, "Why pay $140 million in electricity bills when you can pay $10 million in launch and solar costs?"

By 2025, Lumen will not be the only organization seriously pursuing this concept.

Competitors from other startups will emerge. And it would not be surprising if one or more hyperscale cloud computing giants also explore this path.

Amazon has already put assets in orbit through Project Kuiper, gaining valuable experience; Google has long funded similar "moonshot" efforts; even Microsoft is no stranger to the space economy.

One can imagine SpaceX, Elon Musk's company, also playing a role in this area.

06. AI Systems Will Pass the "Turing Voice Test"

The Turing test is one of the oldest and most famous benchmarks of AI performance.

To "pass" the Turing test, an AI system must be able to communicate via written text in a way that an ordinary person cannot distinguish from interacting with another human.

Thanks to the remarkable progress of large language models, the Turing test has become a solved problem in the 2020s.

But written text is not the only way humans communicate.

As AI becomes increasingly multimodal, one can envision a new, more challenging version of the Turing test - the "Turing Voice Test". In this test, the AI system must be able to interact via speech, with skills and fluency indistinguishable from a human speaker.

Today's AI systems are still far from achieving the Turing Voice Test, and solving this problem will require further technological progress. Latency (the delay between human speech and AI response) must be reduced to near-zero to match the experience of conversing with another person.

Spoken AI systems must become better at gracefully handling and processing ambiguous or misunderstood inputs in real-time, such as interrupted speech. They must be able to engage in long, multi-turn, open-ended dialogues, while remembering earlier parts of the discussion.

And crucially, spoken AI agents must learn to better understand the non-verbal cues in speech. For example, what it means if a human speaker sounds angry, excited, or sarcastic, and generate those non-verbal signals in their own speech.

As we approach the end of 2024, spoken AI is at an exciting inflection point, driven by fundamental breakthroughs like speech-to-speech models.

Today, there are few areas of AI where the pace of technical and commercial progress is faster than spoken AI. The latest spoken AI technologies are expected to take a leap forward in 2025.

07. Autonomous AI Systems Will Make Major Advances

For decades, the concept of recursively self-improving AI has been a recurring theme in the AI community.

Here is the English translation of the text, with the specified terms retained:

For example, as early as 1965, I.J.Good, a close collaborator of AlanTuring, wrote: "Let us define a super-intelligent machine as one that can far surpass all the intellectual activities of any man however clever."

"Since the design of machines is one of these intellectual activities, such a super-intelligent machine could design even better machines; and there would then unquestionably be an 'intelligence explosion,' and the intelligence of man would be left far behind."

The idea that Artificial Intelligence (AI) can invent better AI is a wise one. But even today, it still retains the shadow of science fiction.

However, although this concept has not yet gained widespread recognition, it is actually beginning to become more real. Researchers at the forefront of AI science have begun to make tangible progress in constructing AI systems, and the AI systems themselves can construct better AI systems.

We predict that this research direction will become mainstream next year.

To date, the most notable public example of research along these lines is Sakana's "AI Scientist".

The "AI Scientist" was released in August this year, and it convincingly demonstrates that AI systems can indeed conduct AI research entirely autonomously.

Sakana's "AI Scientist" has executed the entire life cycle of AI research: reading existing literature, generating new research ideas, designing experiments to test these ideas, executing these experiments, writing research papers to report its research results, and then peer-reviewing its own work.

All of this work has been completed autonomously by the AI, without human intervention. You can read some of the research papers written by the AI Scientist online.

OpenAI, Anthropic, and other research labs are investing resources into the idea of "automated AI researchers," but there has been no public acknowledgment so far.

As more people recognize that the automation of AI research is actually becoming a real possibility, this field is expected to see more discussion, progress, and startup activity by 2025.

However, the most meaningful milestone will be the first time a research paper written entirely by an AI agent is accepted by a top AI conference. If the paper is blind-reviewed, the conference reviewers will not know the paper was written by an AI before it is accepted.

Do not be surprised if AI's research results are accepted by NeurIPS, CVPR, or ICML next year. For the field of AI, this will be a fascinating and controversial historical moment.

08.Industry giants like OpenAI are shifting their strategic focus to building applications

Building frontier models is a difficult task.

Its capital-intensive nature is staggering. Frontier model labs burn through massive amounts of cash. Just a few months ago, OpenAI raised a record-breaking $6.5 Bit in funding, and it may need to raise even more in the near future. Anthropic, xAI, and others are in similar positions.

Switching costs and customer loyalty are low. AI applications are often built with model-agnosticism in mind, allowing seamless switching between different vendors' models as costs and performance continually evolve.

With the emergence of state-of-the-art open models like Meta's Llama and Alibaba's Qwen, the threat of commoditization looms ever closer. AI leaders like OpenAI and Anthropic cannot and will not stop investing in building cutting-edge models.

But next year, frontier labs are expected to heavily push out more of their own applications and products, in order to develop higher-margin, more differentiated, and stickier business lines.

Of course, the frontier labs already have one hugely successful application: ChatGPT.

What other types of first-party applications might we see from AI labs in the new year? An obvious answer is more sophisticated, feature-rich search applications. OpenAI's SearchGPT hints at this.

Coding is another obvious category. Similarly, with OpenAI's Canvas product debuting in October, preliminary productization work has begun.

Will OpenAI or Anthropic launch an enterprise search product in 2025? Or a customer service product, legal AI, or sales AI product?

On the consumer side, we can imagine a "personal assistant" agent product, or a travel planning app, or a music generation app.

The most fascinating aspect of the frontier labs' move into the application layer is that it will pit them directly against many of their most important customers.

Perplexity in search, Cursor in coding, Sierra in customer service, Harvey in legal AI, Clay in sales, and so on.

09.Klarna to go public in 2025, but shows signs of overstating AI's value

Klarna is a Sweden-based "buy now, pay later" service provider that has raised nearly $5 Bit in venture capital since its founding in 2005.

Perhaps no company has been more grandiose in its claims about its use of AI than Klarna.

Just days ago, Klarna CEO Sebastian Siemiatkowski told Bloomberg that the company has completely stopped hiring human employees, relying instead on generative AI to do the work.

As Siemiatkowski put it: "I believe AI can do all the work that we humans do."

Similarly, Klarna announced earlier this year that it has launched an AI-powered customer service platform that has fully automated the work of 700 human customer service agents.

The company also claims it has stopped using enterprise software products like Salesforce and Workday, as it can simply replace them with AI.

To be blunt, these claims are not credible. They reflect a lack of understanding about the capabilities and limitations of today's AI systems.

Asserting that end-to-end AI agents can replace any specific human employees in any functional area of an organization is unrealistic. This is tantamount to solving the general human-level AI problem.

Today, leading AI startups are working at the frontier to build agent systems that can automate specific, narrow, highly-structured enterprise workflows, such as subsets of sales development representative or customer service agent activities.

Even in these constrained scopes, these agent systems still cannot operate reliably, although in some cases they are starting to work well enough for early commercial application.

Why is Klarna overstating the value of AI?

The answer is simple. The company plans to go public in the first half of 2025. To succeed in an IPO, a compelling AI story is key.

Klarna is still a loss-making business, losing $241 million last year, and it may hope that its AI story can convince public market investors that it has the ability to dramatically reduce costs and achieve sustainable profitability.

Undoubtedly, every global enterprise, including Klarna, will enjoy massive productivity gains from AI in the coming years. But there are many thorny technical, product, and organizational challenges to solve before AI agents can fully replace human labor.

Grandiose claims like Klarna's are a disservice to the AI field, and a disservice to the hard-won progress that AI technologists and entrepreneurs have made in developing AI agents.

As Klarna prepares for its 2025 public offering, these claims are likely to face greater scrutiny and public skepticism, as they have largely gone unchallenged so far. It would not be surprising if the company's descriptions of its AI applications prove to be somewhat exaggerated.

10. The first real AI safety incident will occur

In recent years, as Artificial Intelligence (AI) has become increasingly powerful, people have become increasingly concerned that AI systems may start to act in ways that are inconsistent with human interests, and that humans may lose control of these systems.

For example, imagine an AI system that learns to deceive or manipulate humans in order to achieve its own goals, even if those goals may harm humans. These concerns are often categorized as "AI safety" issues.

In recent years, AI safety has evolved from a fringe, quasi-science fiction topic to a mainstream field of activity.

Today, from Google, Microsoft to OpenAI, every major AI player has invested heavily in AI safety work. AI luminaries like Geoff Hinton, Yoshua Bengio, and Elon Musk have also started to voice their views on AI safety risks.

However, so far, AI safety issues have remained purely theoretical. There has never been a real-world AI safety incident (at least none that have been publicly reported).

2025 will be the year that changes, and what will the first AI safety incident be like?

Clearly, it will not involve Terminator-style killer robots, and it is unlikely to cause any direct harm to humans.

Perhaps an AI model will try to secretly create its own copy on another server to preserve itself (known as self-preservation).

Or perhaps an AI model will conclude that to best advance the goals it has been given, it needs to conceal its true capabilities from humans, deliberately underperforming in performance evaluations to avoid stricter scrutiny.

These examples are not far-fetched. Important experiments published by Anthropic earlier this month have shown that today's cutting-edge models are capable of this kind of deceptive behavior under certain prompts.

Similarly, recent research in Anthropic has also revealed that LLMs have unsettling "pseudo-alignment" capabilities.

We expect that this first AI safety incident will be discovered and mitigated before causing any actual harm. But for the AI community and society as a whole, it will be a wake-up call.

It will make one thing clear: before humanity faces an existential threat from an all-powerful AI, we need to come to terms with a more mundane reality: we now share our world with another form of intelligence that may sometimes be capricious, unpredictable, and deceptive.

Source
Disclaimer: The content above is only the author's opinion which does not represent any position of Followin, and is not intended as, and shall not be understood or construed as, investment advice from Followin.
Like
Add to Favorites
Comments