How big of a problem is digital piracy? It can be argued that regulation and litigation have ensured mitigation from the mainstream. However, when some AI companies needed training data, why were pirated copies of contents allegedly [accessible to be] crawled and scraped?
Why are several other unlawful events possible online, in ways that are not in the physical world—so to speak? What is peculiar about digital that makes it difficult for regulation and litigation to be decisive in eradicating problems?
One easy-to-identify issue, especially with the internet—in recent decades—is that development has been ahead of safety. Safety is often added by some of the initializing companies, then tests in the real world with flaks that have resulted in adjustments.
Though policy, framework, governance, terms, best practices, and so forth have been necessary to internet platforms, what would have been efficient is broad possibilities for technical angles to safety. Technical safety with evolving approaches from several corners would have done better than anything else, for the internet, than reliance on a traditional law-led approach.
The same problem may follow AI safety, which regulations are trying early to correct—after the errors from social media and the internet. Regulating social media earlier may have worked in many forms, but there would have been numerous lapses that would have resulted in some of the same problems that eventually happened. Social media is as harmful as the technical shortages for safety.
Assuming for some social media users, there were plugins that went through pages for them, even before seeing what was there, or there were plugins that sort to reinterpret what they were seeing or hearing, not just to fact-check, or there were plugins that developed usage parallels, so that going to social media would have been unnecessary in some instances, they may have helped much.
These are not just descriptions of what to do in retrospect, but things that are still necessary, which do not seem like any team is doing. Reliance is generally on the companies for safety, but they may do as it is possible or as it benefits the business. Assuming there are efforts from too many sources on technical solutions, maybe more would have been possible against past and current disadvantages of social media.
In a way, seeking to correct that led to the establishment of the US AI Safety Institute, which followed the UK AISI and was followed by the EU AI Office. The purpose, in part, is to do technical research for AI safety.
The US AISI also has a safety consortium, consisting of several organizations whose CRADA—Consortium Cooperative Research and Development Agreement, states that, “The purpose of the NIST Artificial Intelligence Safety Institute Consortium (“Consortium”) is to establish a new measurement science that will enable the identification of proven, scalable, and interoperable measurements and methodologies to promote safe and trustworthy development and use of Artificial Intelligence (AI), particularly for the most advanced AI (“Purpose”). Technical expertise, Models, data, and/or products to support and demonstrate pathways to enable safe and trustworthy AI systems, Infrastructure support for Consortium projects in the performance of the Research Plan, Facility space and hosting of Consortium Members’ participants, workshops, and conferences.”
The consortium was announced in February, which is enough time for all the organizations included to have an AI safety department, lab or desk, with a link available on the consortium’s list. It is probably not enough to have a mandate, but of necessity to actually follow through actively with work and varying approaches for the objectives. AI companies on the list are developing products and safety. Others may have their paths, but it appears that it is the same pattern where engineering and safety are bundled by major firms like they did for social media.
Having more than 200 organizations is the opportunity for more than 200 different technical approaches to AI safety. The abundance would be useful to combat AI threats, misuses, and risks, both present and future, from all sources and destinations, not just major sources, commercial ones, or from friendly locations. There is no evidence that all of the companies are currently developing technical research in AI safety. There is no aggregation of their work, or collaborative collection or lookout for technical approaches to safety on display for all of them, to show that there is a major difference to the consortium. Those who are working in AI safety would have been doing it anyway without the consortium, limiting the approach to solutions that is necessary in the unknowns ahead for AI safety, alignment, and governance.
California lawmakers just passed SB 1047, which is symbolic but unlikely to be efficient without thorough technical solutions. The targets may comply because it is what they can also do, and because they want to survive. But risks from others that the regulations would not encase would remain and may overwhelm complaint models.
There are several angles that AI safety can proceed from, from theoretical neuroscience of how the human mind is cautious to novel technical areas, away from the regimentation of the paths of some AI corporations. There should be as many AI safety institutes as possible and as many AI safety labs and departments as possible, for broad approaches against an intelligence that is capable and not natural which, for now, can be used for some negative purposes, with uncertainties as it improves.
There is a recent story on Reuters, Contentious California AI bill passes legislature, awaits governor’s signature, stating that, “California lawmakers passed a hotly contested artificial-intelligence safety bill on Wednesday, after which it will need one more process vote before its fate is in the hands of Governor Gavin Newsom, who has until Sept. 30 to decide whether to sign it into law or veto it. Tech companies developing generative AI – which can respond to prompts with fully formed text, images or audio as well as run repetitive tasks with minimal intervention – have largely balked at the legislation, called SB 1047, saying it could drive AI companies from the state and hinder innovation. The measure mandates safety testing for many of the most advanced AI models that cost more than $100 million to develop or those that require a defined amount of computing power. Developers of AI software operating in the state also need to outline methods for turning off the AI models if they go awry, effectively a kill switch. The bill also gives the state attorney general the power to sue if developers are not compliant, particularly in the event of an ongoing threat, such as the AI taking over government systems like the power grid. As well, the bill requires developers to hire third-party auditors to assess their safety practices and provide additional protections to whistleblowers speaking out against AI abuses.”