New York Times Experiments with Blockchain to Combat Fake News

The New York Times made an announcement regarding their new project by their research and development team focused on experiments using blockchain. They are finding a solution to combat misinformation, adulterated media issues that harm publishers. This solution can be used by the media industry globally in the future.

Their News Provenance Project plans to use Hyperledger Fabric’s permissioned blockchain, an open-source blockchain framework to authenticate news and photographs. Their team consists of technologists and journalists who are also collaborating with IBM Garage on the development of the proof of concept. The project will consist of storing the details of the news item including the information regarding the where a photo/ video was taken, the person who took it as well as the publishing and editing contextual details. The research and development team believes that adding context to images or videos may bring a positive and clarifying effect for the online media ecosystem.

“All hype aside, blockchain offers mechanisms for sharing information between entities in ways we think are essential for establishing and maintaining the provenance of digital files,” stated the News Provenance Project.

As of their current stage, the research and development team are conducting user-centered research by using a blockchain-based system, which the team can benefit from its tamper-proof feature. The participation of the majority of the media industry will affect the success of the project.

Image via Shutterstock

OpenAI Negotiates Licensing with Major Media Amid NY Times Copyright Dispute

OpenAI, known for its groundbreaking AI technology including ChatGPT, is currently engaged in strategic negotiations with several major media companies, including CNN, Fox Corp., and Time. These talks revolve around licensing articles, videos, and images for OpenAI’s use, a move that gains significance amidst the company’s ongoing legal battle with The New York Times over alleged copyright infringement.

The backdrop of these negotiations is OpenAI’s recent lawsuit filed by The New York Times. The newspaper alleges that OpenAI used its content without authorization to train its AI models, including the widely popular ChatGPT. In response, OpenAI has attributed this to a “rare bug” in their system and disputes the accusations, highlighting the complexities surrounding AI ethics and copyright laws.

As part of its strategy to navigate these legal and ethical challenges, OpenAI is seeking to secure formal content licenses with media giants. The discussions with CNN, Fox Corp., and Time are aimed at acquiring a diverse range of content to enhance AI training while mitigating legal risks. Time’s CEO, Jessica Sibley, has confirmed ongoing discussions with OpenAI and expressed optimism about reaching a fair agreement.

However, the situation is not uniformly positive across the media landscape. While some media companies like Guardian News & Media and Axel Springer SE, the parent company of Politico, show willingness to enter negotiations, others like The Washington Post remain hesitant. The concerns mainly revolve around compensation and the implications of AI technology on copyright law. For instance, there’s skepticism among media executives about reaching productive agreements until legal clarity is established.

A significant aspect of these negotiations is the compensation offered to publishers. Reports suggest that OpenAI’s offers range from $1 million to $5 million annually, a figure that some top publishers find insufficient. This is juxtaposed against the backdrop of a multi-year deal with Axel Springer SE for tens of millions of dollars, setting a precedent for higher expectations among other publishers.

Additionally, during a recent Senate Judiciary subcommittee hearing, the need for clearer regulations was voiced, with calls for Congress to step in and establish rules for copyrighted content’s use in commercial generative AI.

OpenAI’s situation illustrates the delicate balance between technological innovation and the protection of intellectual property. As AI continues to evolve, the dynamics between AI developers and content creators are rapidly changing, underscoring the need for legal and ethical frameworks that cater to the complexities of this emerging field.

Exit mobile version