Databricks, a company that helps large businesses build customizations AI Models have developed a machine learning technique that improves the performance of AI models without cleaning tagged data.
Databricks Chief AI Scientist Jonathan Frankle discussed with clients over the past year the key challenges they face in making AI reliable.
Frank said the problem was dirty data.
“Everyone has some data and knows what they want to do,” Frank said. But the lack of clean data makes it challenging to fine-tune a model to perform a specific task. [application programming interface]” for the model.
Databricks’ model can allow companies to eventually deploy their own agents to perform tasks without hindering data quality.
The technology rarely looks at some of the key tips engineers are using now to improve the capabilities of advanced AI models, especially when it is difficult to get good data. This approach leverages the idea of helping to generate advanced inference models by combining reinforcement learning, a way in which AI models improve, “synthesize” or AI-generated training data.
From Openai,,,,, Googleand DeepSeek All of this relies heavily on reinforcement learning as well as synthetic training data. The connection reveals this NVIDIA plans to get Gretela company specializing in synthetic data. “We’re all browsing the space,” Frank said.
The Databricks method takes advantage of the fact that given enough attempts, even weak models can score well on a given task or benchmark. The researchers call this method to improve model performance “optimal N.” Databricks trained a model to predict which best results human testers prefer based on examples. Datamap reward model or DBRM can then be used to improve the performance of other models without further labeling of data.
Then use DBRM to select the best output from the given model. This creates synthetic training data to further fine-tune the model, resulting in better output for the first time. Databricks says its new method tests time adaptive optimization or TAO. “The approach we’re talking about uses some relatively light reinforcement learning learning, from baking the benefits of the best-N into the model itself,” Frank said.
He added that the research done by Databricks shows that the TAO method is extended to a larger, more powerful model. Reinforcement learning and synthetic data have been widely used, but combining them to improve language models is a relatively new and technically challenging technique.
Databricks is extremely open to how it develops AI because it wants to show customers that it has the skills it needs to create a powerful custom model for them. The company previously revealed to Wired How it develops DBX, a cutting-edge open source large language model (LLM) Start from scratch.