Join daily and weekly newsletters to obtain the latest updates and exclusive content to cover the leading artificial intelligence in the industry. Learn more
The scene of artificial intelligence continues to develop at a rapid pace, with recent developments representing fixed models. In early 2025, Deepseek from AI LAB revealed a new model that sent shock waves By making artificial intelligence And led to a 17 % Decrease in NVIDIA shares, along with Other shares related to the demand for the artificial intelligence data center. This market reaction has been widely reported that it stems from the ability of the Depsik to present high -performance models in a small part of the cost of competitors in the United States, which sparked a discussion about The effects of artificial intelligence data centers.
To give the context to disable Deepseek, we believe it is useful to consider a broader shift in the scene of artificial intelligence led by the scarcity of additional training data. Since the major artificial intelligence laboratories have already trained their models on most of the general data available on the Internet, the scarcity of data is O slowing out more improvements in pre -training. As a result, the models look forward to “TTC Calculating” (TTC) where “Open Open” Models of Models “O” from AI) before answering a question at the time of reasoning, as an alternative way to improve the performance of the general model. Current thinking is that TTC may show improvements in the scaling law similar to those that prompted pre -training, which may enable the next wave of transformative artificial intelligence developments.
These developments indicate an important attack: First, laboratories that work on smaller budgets (reported) are now able to issue modern models. The second transformation is to focus on TTC as the next possible driver for artificial intelligence progress. Below, we implement both of these trends and effects on the competitive scene and the broader artificial intelligence market.
The effects of the artificial intelligence industry
We believe that the shift towards TTC and the increase in competition between the thinking models may have a number of effects on the broader Artificial intelligence landscape Through devices, cloud platforms, basic models and institution programs.
1. Devices (graphics processing units, custom chips and infrastructure for account)
- From huge training groups to “Test Time” mutations: From our point of view, the shift towards TTC may have effects on the type of devices resources required by artificial intelligence companies and how to manage them. Instead of investing in GPU groups, increasingly larger for training work burdens, artificial intelligence companies may increase their investments in the possibilities of inference to support the increasing TTC needs. Although artificial intelligence companies will continue to require large numbers of graphics processing units to deal with the burden of inference, the differences between Training work burdens Work burdens may affect how to create and use these chips. Specifically, since the burdens of the inference work tend to be more Dynamic (and “Spikey”)Planning may become more complicated than it is in the training burdens directed towards the batch.
- The rise of improved devices for reasoning: We believe that the shift in focusing towards TTC is likely to increase the chances of alternative AI devices specialized in calculating the time of low inference to join. For example, we may see more demand for GPU alternatives as the integrated circles specified for the application (Asics) for reasoning. Since access to TTC becomes more important than training ability, dominance General drawings processing units for general purposes, That is used for both training and reasoning, may decline. This transformation can benefit the provider of specialized inferences.
2.
- The quality of service (QOS) becomes a major discrimination: One of the issues that prevents the adoption of artificial intelligence in the institution, in addition to concerns related to the accuracy of the model, is the lack of reliability of applications programming facades for reasoning. The problems associated with the unreliable API conclusion include Flaving response timesand Reducing And difficulty Dealing with simultaneous requests and Adaptation to API’s end point changes. TTC’s increase may exacerbate these problems. In these circumstances, the cloud provider is able to provide models with the assertions of the quality of service that address these challenges, in our opinion, a great advantage.
- Increased cloud spending despite the gains of efficiency: Instead of reducing the demand for artificial intelligence devices, it is possible to follow the most efficient methods of training in the Great Language Model (LLM) from Jevons, which is a historical note as it improves efficiency to high general consumption. In this case, effective inference models may encourage more artificial intelligence developers to take advantage of thinking models, which in turn increase the demand for account. We believe that recent typical developments may lead to an increase in demand at the expense of Cloud AI for both the form of the model and the training of smaller specialized models.
3. The Foundation Form (Openai, Anthropor, COHERE, Deepseek, Mistral)
- Impact on pre -trained models: If new players such as Deepseek can compete with them Ai Laboratory borders In a small part of the reported costs, the pre -trained models may become less susceptible to the trench. We can also expect more innovations in TTC for transformer models, and as Deepseek has proven, these innovations can come from sources outside the most firm artificial intelligence laboratories.
4. Enterprise Ai Adoption and Saas (app.)
- Security and privacy concerns: Looking at the Deepseek assets in China, there is likely to be ongoing Scrutiny One of the company’s products from the perspective of security and privacy. In particular, API and Chatbot shows are unlikely to be used widely by AI Enterprise customers in the United States, Canada or other western countries. It is said that many companies Moving to the ban Use Deepseek and applications. We expect Deepseek models to check even when they are hosted by Third parties In the United States and other Western data centers that may limit institutions ’adoption of models. Researchers have already indicated examples of security concerns about it prison breakand Bias and generation of harmful content. grant Consumer attentionWe may see the experience and evaluation of Deepseek models in the institution, but it is unlikely that the buyers of the institutions will move away from job occupants because of these concerns.
- The vertical specialization acquires traction: In the past, vertical applications that mainly use basic models have focused on creating work tasks designed to meet the specific business needs. Techniques such as the generation of retrieval (RAG) played, directing models, and handicrafts and handrails have an important role in adapting these specialized models to these specialized use cases. While these strategies led to noticeable successes, there was constant concern that major improvements to basic models can make these applications old. Sam warned Al -Tamman, it could be a major penetration in the typical capabilities.Steamroll “App Classes Innovations That was built as files around the basic models.
However, if the developments in the train time account are actually a plateau, the threat of rapid displacement diminishes. In a world where the gains in typical performance come from TTC improvements, new operators may be opened for the application layer. Innovations in post-training algorithms-example Improving the claim is organizedand Perceived thinking Effective samples techniques – may provide significant improvements in the target vertical performance.
Any performance improvement will be particularly relevant in the context of logic-focused models such as the GPT-4O of Openai and Deepseek-R1, which often shows multiple second response times. In actual time applications, it can provide cumin and improve the quality of inference within a specific field of competitive advantage. As a result, the application layer companies that have experience in the field may play a pivotal role in improving the efficiency of reasoning and its polishing outputs.
Deepseek explains a decrease in focusing on increasing quantities of pre -training as the only driver of model quality. Instead, development emphasizes the increasing importance of TTC. While the direct adoption of Deepseek models in institutional program applications is still not certain due to continuous scrutiny, their impact on driving improvements in other existing models has become more clear.
We believe that Deepseek has prompted the laboratories of Amnesty International to integrate similar technologies in engineering and research operations, while completing the advantages of the current devices. The reduction resulting from the costs of the forms, as expected, contributes to increasing the use of models, as it is in line with the principles of Jevons Paradox.
Pashotan Vaezipoor is the technical lead in Georgia.
https://venturebeat.com/wp-content/uploads/2025/04/Datatransformer.webp?w=1024?w=1200&strip=all
Source link