Blog Post

Prmagazine > News > News > The open-source AI debate: Why selective transparency poses a serious risk
The open-source AI debate: Why selective transparency poses a serious risk

The open-source AI debate: Why selective transparency poses a serious risk


Join our daily and weekly newsletter for the latest updates and exclusive content on industry-leading AI coverage. learn more


As the tech giant announced AI release Open – even write its name – the word “open source” that once internally has entered the modern spirit of the times. During this unstable period, a company’s mistakes may make public comfort for AI make public comfort over a decade or more, and the concepts of openness and transparency are waving casually, sometimes dishonestly to reproduce trust.

Meanwhile, as the new White House administration takes a more open technical regulation approach, the front has been proposed – to prevent regulatory innovation, if the “wrong” aspect prevails, the dire consequences can be predicted.

However, there is a third method that has been tested and proven by other waves Technological Change. True open source collaboration is based on the principles of openness and transparency that can make innovation faster, even if the industry has the ability to develop socially unbiased, ethical and beneficial technologies.

Understand the power of true open source collaboration

In short, open source software has freely available source code that can be viewed, modified, anatomized, adopted and shared for commercial and non-commercial purposes – and historically, it has been huge in terms of breeding innovation. Open source products Linux, Apache, MySQL and PHP, for example, what we know is freeing the Internet.

Now, by democratizing access to AI models, data, parameters, and open source AI tools, communities can once again release faster innovations instead of constantly recreating the steering wheel, which is why IBM’s recent research has been conducted 2,400 IT decision makers Revealing an increasing interest in using open source AI tools to drive ROI. Although faster development and innovation rank first in determining the return on investment in AI, the research also confirmed that adopting open solutions may be related to greater financial viability.

Instead of gaining less short-term gains from companies, open source AI invites to create more diverse and tailored applications across industries and fields that may not have other resources that may not have proprietary models.

Perhaps importantly, open source transparency allows for independent review and review of the behavior and ethics of AI systems – when we take advantage of the existing interests and driving forces of the masses, they find problems and errors. LAION 5B dataset A crushing defeat.

In that case, the crowd is rooted in 1,000 URLs Contains validated child sexual abuse material, hidden in the data, which provides generated AI models such as Stable Diffusion and Midjourney, which generate images from text and image cues and are the basis in many online video generation tools and applications.

Although the discovery caused an uproar, if the dataset was closed, like Openai’s Sora or Google’s Gemini, the consequences could be worse. It’s hard to imagine that if AI’s most exciting video creation tools start picking out disturbing content, it would cause a rebound.

Thankfully, the open nature of the LAION 5B dataset enables the community to inspire its creators to work with industry regulators to find fixes and release Re-Laion 5B, which illustrates why the transparency of truly open source AI benefits not only users and creators from industries and creators who build trust with consumers and the general public.

Dangers of AI Open Source

Although source code is relatively easy to share, AI systems are more complex Not software. They rely on system source code as well as model parameters, datasets, hyperparameters, training source code, random number generation and software frameworks – each of these components must work together to make the AI ​​system work properly.

Among concerns about AI security, it is common to declare that releases are open or open source. But to be accurate, innovators must share various parts of all puzzles so that other players can fully understand, analyze and evaluate the properties of the AI ​​system to ultimately copy, modify and expand its functionality.

For example, Touted llama 3.1 405b As the “first cutting-edge open source AI model”, it can only expose the pre-training parameters or weights of the shared system, as well as some software. While this allows users to download and use the model at will, key components such as source code and datasets are still closed – after which, it will become more disturbing Yuan Announcement Even if it stops reviewing content to ensure accuracy, it injects the AI ​​robot profile into the ether.

To be fair, the shared content will surely contribute to the community. Open weight models offer levels of flexibility, accessibility, innovation and transparency. DeepSeek decides the weight of open source, publishes its technical reports for R1 and makes it free to use, for example, the AI ​​community can research and validate its methods and weave them into their work.

This is MisleadingBut when no one can really look, try and understand every part of the puzzle that creates it, call the AI ​​system open source.

This misleading is not just a threat to public trust. Rather than empowering everyone in the community to collaborate to build and advance models like Llama X, it forces innovators to use such AI systems to blindly trust unshared components.

Embrace the challenge before us

As self-driving cars promote surgeons in operating rooms on the streets of major cities and AI systems, we are only in getting this technology to start using well-known wheels. Promise is huge, and so is the potential for errors – which is why we need new measures that are trustworthy in the AI ​​world.

Even Anka Reuel of Stanford and his colleagues Recently tried To establish a new framework for evaluating the performance of the model, for example, the review practices of industry and public dependencies are not enough. The benchmarks cannot explain the fact that the dataset at the core of the learning system is constantly changing, and the appropriate metrics vary by use case. The field still lacks a wealth of mathematical language to describe the capabilities and limitations of contemporary AI.

By sharing the entire AI system for openness and transparency, rather than relying on under-reporting comments and paying for verbal services to buzzwords, we can foster greater collaboration and foster innovation through secure and ethical AI.

Although true open source AI provides a validation framework for achieving these goals, there is a lack of a transparent framework in the industry. Without the bold leadership and cooperation of tech companies with autonomy, this information gap could undermine public trust and acceptance. Embracing openness, transparency and open source is not only a powerful business model, but also a choice between the future of AI that benefits everyone rather than a few.

Jason Corso is a professor and co-founder at the University of Michigan Voxel51.


Source link

Leave a comment

Your email address will not be published. Required fields are marked *

star360feedback