This boot wants to attract a US DeepSeek moment


Since since since Deepseek burst at the scene in January, and momentum grew around Open Source Chinese artificial intelligence models. Some researchers insist on building an even more open approach to building AI whereby modeling can be spread throughout the world.

Prime Intellect, a startup specializing in decentralized AI, is currently training a great language model called IntelleT-3, using a new type of scattered reinforcement learning for finals. The model will demonstrate a new way to build competitive open AI models using a series of hardware in different places in a way that does not rely on large technical companies, says Vincent Weisser, CEO of the company.

Weisser says that the AI ​​world is currently divided between those who rely on closed American models and those who use open Chinese offers. The Technology Prime Intellation is democratizing AI by building and changing more people advanced AI for themselves.

Improving AI models is no longer just to set up and calculate training data. The contemporary boundary models use reinforcement learning to improve after the pre -training process is completed. Do you want your model to perform in math, answer legitimate questions or play Sudoku? Let it improve itself by practicing in an environment where you can measure success and failure.

“These learning environments for reinforcement are now the bottleneck to scale the abilities,” Weisser says.

Prime Intellect has created a framework that allows someone to create a reinforcement learning environment that is adapted for a specific task. The company combines the best environments created by its own team and the community to set intellect-3.

I tried to execute an environment to solve Wordle Puzzles, created by Prime Intellector, Will Brown, and looked as a small model that solved Wordle -Lega cards (it was more methodical than me, to be honest). If I were an AI researcher trying to improve a model, I would tighten a lot of GPUs and have the model exercising over and over again, while a reinforcement -learning algorithm has changed its weights, thereby turning the model into a Wordle Master.

Leave a Reply

Your email address will not be published. Required fields are marked *