Physical Address

304 North Cardinal St.
Dorchester Center, MA 02124

These Startups Are Building Advanced AI Models Without Data Centers


Researchers have trained a new kind of Great Language Model (LLM) use Gypus A dotted and special fed, as well as public data – a movement that has a dominant road of dominant construction artificial intelligence may be broken.

Flower ai and OldTwo starting on non-traditional approaches to the EU’s construction worked together to create a new model called collective-1.

Cicek has created methods that allow training to spread the trainings on the Internet on the computer. The company’s technology is used to train AI models that do not need a pool account or data pool or data pool or data. Vana, X, Reddit and Telegram presented information sources, including personal messages.

Collective-1 is smaller for modern standards, as today’s power programs are 7 billion parameter-value to give their abilities compared to the most developed models today Chatgpt, Claudand Twins.

A computer scientist Nic Lane and Flower Cofounderi in Cambridge University promises to extend a distributed approach to the size of the team. The strip adds that flower prepares a model with 30 billion parameters using the usual data of the EU, and this year later plans to grow another model with another model of 100 billion parameters by industry leaders. “It can really change everyone thinking about AI, so we are following this very difficult.” He says that the beginning also includes pictures and sounds to create multimodal models.

The distributed model building can also disturb the dynamics that form the AI ​​industry.

The AI ​​companies currently have a large amount of learning information, combining large amounts of training information with a large number of computed GPU-filled with a large number of computed GPU, which is well-charged using super fast fiber-optical cables. Sometimes, although copyrighted by copyrights, including copyrights, they are very confident in attractive databases on the public.

The approach means that only the richest companies and the most powerful pounds of the most powerful punishment are able to develop the strongest and valuable models. Open source models, as well as Meta calls and R1 from DeepSEEKIt is built by companies with access to large data centers. Distributed approaches can make smaller companies and universities, pooling the defective sources of defective and establish advanced AI. Or may allow countries to create a more powerful model to create a certain amount of information centers.

Lane believes that the AI ​​industry will look into new methods that allow them to get out of the individual information centers. The distributed approach “allows you to calculate more elegantly from the data center model,” he says.

Helen Toner, AI management specialist on AI management in security and emerging technology, Flower EU’s approach to the AI ​​competition and management is “interesting and potentially very relevant”. “Probably, it will continue to fight to keep up with the border, but it can be an interesting approach in an interesting speed,” he said.

To divide and conquer

Distributed AI education is divided into calculations similar to calculations used to build strong AI systems. Creating an LLM is to feed a model that regulates the parameters to give a large amount of text from a desire to give useful answers. Inside the data center, the training process may work in different GPUs, and then you can interfere in a master model from time to time.

The new approach allows you to be combined in a normal data center in a normal data center and can be combined in a relatively slow or variable internet connection.



Source link

Leave a Reply

Your email address will not be published. Required fields are marked *