Late last week, the European Commission, the Member States, and the European Parliament reached a deal on the AI Act. The current compromise is a combination of tiered obligations and a limited open source exemption which creates a situation where open source AI models can get away with being less transparent and less well-documented than proprietary GPAI models.
Representatives of the European Parliament, EU member states, and the European Commission reached a provisional agreement on the proposed AI Act. The copyright provisions in the AI Act are a step in the right direction. They further consolidate the existing balanced legislative approach adopted by the EU in the 2019 CDSM Directive.
The blog post argues that with increasing convergence on creator/rightholder opt-outs as an essential mechanism in the governance of generative AI models, there is an urgent need for standardization of machine readable opt outs.
As the debate over how the AI Act should deal with foundation models reaches a climax, we argue that it would be a mistake to leave this crucial aspect of AI regulation to the AI companies themselves, and that there is indeed a need for mandatory transparency and documentation requirements.
The Center for Policy Research at United Nations University (UNU-CPR) published our policy note, which builds on our work on Public Data Commons and informs a recommendation from the report of the High-Level Advisory Board on Effective Multilateralism.
Experts agree that artificial intelligence is the next frontier of market concentration in the Internet economy. The democratization of AI development hinges on public investment in AI infrastructure.
Analysis of the European Commission’s proposal for a Digital Euro that argues that in order to meet its objectives, the Digital Euro must be built on top of public digital infrastructure.
This opinion takes a closer look at how the Falcon 180B model is licensed and is a part of our exploration of the emergent standards for the sharing of AI models.
Some experts believe that open-sourcing AI increases the risk of malicious use. In this opinion, we argue that calls for regulators to intervene and limit the possibility of open-sourcing AI models must consider the impact on freedom of expression.
We agree with Widder, West, and Whittaker that openness alone will not democratize AI. However, it is clear to us that any alternative to current Big Tech-driven AI must be, among other things, open.