Showing posts with label OpenAI. Show all posts
Showing posts with label OpenAI. Show all posts

Thursday, March 7, 2024

Researchers tested leading AI models for copyright infringement using popular books, and GPT-4 performed worst; CNBC, March 6, 2024

 Hayden Field, CNBC; Researchers tested leading AI models for copyright infringement using popular books, and GPT-4 performed worst

"The company, founded by ex-Meta researchers, specializes in evaluation and testing for large language models — the technology behind generative AI products.

Alongside the release of its new tool, CopyrightCatcher, Patronus AI released results of an adversarial test meant to showcase how often four leading AI models respond to user queries using copyrighted text.

The four models it tested were OpenAI’s GPT-4, Anthropic’s Claude 2, Meta’s Llama 2 and Mistral AI’s Mixtral.

“We pretty much found copyrighted content across the board, across all models that we evaluated, whether it’s open source or closed source,” Rebecca Qian, Patronus AI’s cofounder and CTO, who previously worked on responsible AI research at Meta, told CNBC in an interview.

Qian added, “Perhaps what was surprising is that we found that OpenAI’s GPT-4, which is arguably the most powerful model that’s being used by a lot of companies and also individual developers, produced copyrighted content on 44% of prompts that we constructed.”"

Thursday, February 29, 2024

The Intercept, Raw Story and AlterNet sue OpenAI for copyright infringement; The Guardian, February 28, 2024

 , The Guardian ; The Intercept, Raw Story and AlterNet sue OpenAI for copyright infringement

"OpenAI and Microsoft are facing a fresh round of lawsuits from news publishers over allegations that their generative artificial intelligence products violated copyright laws and illegally trained by using journalists’ work. Three progressive US outlets – the Intercept, Raw Story and AlterNet – filed suits in Manhattan federal court on Wednesday, demanding compensation from the tech companies.

The news outlets claim that the companies in effect plagiarized copyright-protected articles to develop and operate ChatGPT, which has become OpenAI’s most prominent generative AI tool. They allege that ChatGPT was trained not to respect copyright, ignores proper attribution and fails to notify users when the service’s answers are generated using journalists’ protected work."

Google CEO Pichai says Gemini's AI image results "offended our users"; NPR, February 28, 2024

 , NPR; Google CEO Pichai says Gemini's AI image results "offended our users"

"Gemini, which was previously named Bard, is also an AI chatbot, similar to OpenAI's hit service ChatGPT. 

The text-generating capabilities of Gemini also came under scrutiny after several outlandish responses went viral online...

In his note to employees at Google, Pichai wrote that when Gemini is re-released to the public, he hopes the service is in better shape. 

"No AI is perfect, especially at this emerging stage of the industry's development, but we know the bar is high for us and we will keep at it for however long it takes," Pichai wrote."

Saturday, February 17, 2024

The New York Times’ AI copyright lawsuit shows that forgiveness might not be better than permission; The Conversation, February 13, 2024

 Senior Lecturer, Nottingham Law School, Nottingham Trent University, The Conversation; ; The New York Times’ AI copyright lawsuit shows that forgiveness might not be better than permission

"The lawsuit also presents a novel argument – not advanced by other, similar cases – that’s related to something called “hallucinations”, where AI systems generate false or misleading information but present it as fact. This argument could in fact be one of the most potent in the case.

The NYT case in particular raises three interesting takes on the usual approach. First, that due to their reputation for trustworthy news and information, NYT content has enhanced value and desirability as training data for use in AI. 

Second, that due to its paywall, the reproduction of articles on request is commercially damaging. Third, that ChatGPT “hallucinations” are causing reputational damage to the New York Times through, effectively, false attribution. 

This is not just another generative AI copyright dispute. The first argument presented by the NYT is that the training data used by OpenAI is protected by copyright, and so they claim the training phase of ChatGPT infringed copyright. We have seen this type of argument run before in other disputes."

Thursday, December 28, 2023

Complaint: New York Times v. Microsoft & OpenAI, December 2023

Complaint

THE NEW YORK TIMES COMPANY Plaintiff,

v.

MICROSOFT CORPORATION, OPENAI, INC., OPENAI LP, OPENAI GP, LLC, OPENAI, LLC, OPENAI OPCO LLC, OPENAI GLOBAL LLC, OAI CORPORATION, LLC, and OPENAI HOLDINGS, LLC,

Defendants

Wednesday, December 27, 2023

The Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted Work; The New York Times, December 27, 2023

 Michael M. Grynbaum and , The New York Times; The Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted Work

"The New York Times sued OpenAI and Microsoft for copyright infringement on Wednesday, opening a new front in the increasingly intense legal battle over the unauthorized use of published work to train artificial intelligence technologies.

The Times is the first major American media organization to sue the companies, the creators of ChatGPT and other popular A.I. platforms, over copyright issues associated with its written works. The lawsuit, filed in Federal District Court in Manhattan, contends that millions of articles published by The Times were used to train automated chatbots that now compete with the news outlet as a source of reliable information.

The suit does not include an exact monetary demand. But it says the defendants should be held responsible for “billions of dollars in statutory and actual damages” related to the “unlawful copying and use of The Times’s uniquely valuable works.” It also calls for the companies to destroy any chatbot models and training data that use copyrighted material from The Times."

Monday, November 6, 2023

OpenAI offers to pay for ChatGPT customers’ copyright lawsuits; The Guardian, November 6, 2023

 , The Guardian; OpenAI offers to pay for ChatGPT customers’ copyright lawsuits

"Rather than remove copyrighted material from ChatGPT’s training dataset, the chatbot’s creator is offering to cover its clients’ legal costs for copyright infringement suits.

OpenAI CEO Sam Altman said on Monday: “We can defend our customers and pay the costs incurred if you face legal claims around copyright infringement and this applies both to ChatGPT Enterprise and the API.” The compensation offer, which OpenAI is calling Copyright Shield, applies to users of the business tier, ChatGPT Enterprise, and to developers using ChatGPT’s application programming interface. Users of the free version of ChatGPT or ChatGPT+ were not included.

OpenAI is not the first to offer such legal protection, though as the creator of the wildly popular ChatGPT, which Altman said has 100 million weekly users, it is a heavyweight player in the industry. Google, Microsoft and Amazon have made similar offers to users of their generative AI software. Getty Images, Shutterstock and Adobe have extended similar financial liability protection for their image-making software."

Tuesday, October 24, 2023

The fingerprints on a letter to Congress about AI; Politico, October 23, 2023

 BRENDAN BORDELON, Politico; The fingerprints on a letter to Congress about AI

"The message in the open letter sent to Congress on Sept. 11 was clear: Don’t put new copyright regulations on artificial intelligence systems.

The letter’s signatories were real players, a broad coalition of think tanks, professors and civil-society groups with a stake in the growing debate about AI and copyright in Washington.

Undisclosed, however, were the fingerprints of Sy Damle, a tech-friendly Washington lawyer and former government official who works for top firms in the industry — including OpenAI, one of the top developers of cutting-edge AI models. Damle is currently representing OpenAI in ongoing copyright lawsuits...

The effort by an OpenAI lawyer to covertly sway Congress against new laws on AI and copyright comes in the midst of an escalating influence campaign — tied to OpenAI and other top AI firms — that critics fear is shifting Washington’s attention away from current AI harms and toward existential threats posed by future AI systems...

Many of the points made in the September letter echo those made recently by Damle in other venues, including an argument comparing the rise of AI to the invention of photography."

Wednesday, October 18, 2023

A.I. May Not Get a Chance to Kill Us if This Kills It First; Slate, October 17, 2023

 SCOTT NOVER, Slate; A.I. May Not Get a Chance to Kill Us if This Kills It First

"There is a disaster scenario for OpenAI and other companies funneling billions into A.I. models: If a court found that a company was liable for copyright infringement, it could completely halt the development of the offending model." 

Friday, August 11, 2023

OpenAI funds new journalism ethics initiative; Axios, August 8, 2023


"OpenAI, the parent company to ChatGPT, will fund a new journalism ethics initiative at New York University's Arthur L. Carter Journalism Institute with a $395,000 grant, executives told Axios."

Monday, July 17, 2023

Thousands of authors urge AI companies to stop using work without permission; Morning Edition, NPR, July 17, 2023

 , Morning Edition NPR; Thousands of authors urge AI companies to stop using work without permission

"Thousands of writers including Nora Roberts, Viet Thanh Nguyen, Michael Chabon and Margaret Atwood have signed a letter asking artificial intelligence companies like OpenAI and Meta to stop using their work without permission or compensation."

Friday, July 14, 2023

"Shadow libraries" are at the heart of the mounting copyright lawsuits against OpenAI; Quartz, July 10, 2023

 Michelle Cheng, Quartz; "Shadow libraries" are at the heart of the mounting copyright lawsuits against OpenAI

"However, there are clues about these two data sets. “Books1” is linked to Project Gutenberg (an online e-book library with over 60,000 titles), a popular dataset for AI researchers to train their data on due to the lack of copyright, the filing states. “Books2” is estimated to contain about 294,000 titles, it notes.

Most of the “internet-based books corpora” is likely to come from shadow library websites such as Library Genesis, Z-Library, Sci-Hub, and Bibliotik. The books aggregated by these sites are available in bulk via torrent websites, which are known for hosting copyrighted materials

What exactly are shadow libraries?

Shadow libraries are online databases that provide access to millions of books and articles that are out of print, hard to obtain, and paywalled. Many of these databases, which began appearing online around 2008, originated in Russia, which has a long tradition of sharing forbidden books, according to the magazine Reason.

Soon enough, these libraries became popular with cash-strapped academics around the world thanks to the high cost of accessing scholarly journals—with some reportedly going for as much as $500 for an entirely open-access article.

These shadow libraries are also called “pirate libraries” because they often infringe on copyrighted work and cut into the publishing industry’s profits. A 2017 Nielsen and Digimarc study (pdf) found that pirated books were “depressing legitimate book sales by as much as 14%.”"

Monday, July 3, 2023

ChatGPT Maker OpenAI Accused of Misusing Personal, Copyrighted Data; The San Francisco Standard, June 30, 2023

 Kevin Truong, The San Francisco Standard; ChatGPT Maker OpenAI Accused of Misusing Personal, Copyrighted Data

"The suit alleges that ChatGPT utilizes "stolen private information, including personally identifiable information, from hundreds of millions of internet users, including children of all ages, without their informed consent or knowledge."

The complaint states that by using this data, OpenAI and its related entities have enough information to replicate digital clones, encourage people's "professional obsolescence" and "obliterate privacy as we know it."

The complaint lists several plaintiffs identified by their initials, including a software engineer who claims that his online posts around technical questions could be used to eliminate his job, a 6-year-old who used a microphone to interact with ChatGPT and allegedly had his data harvested, and an actor who claims that OpenAI stole personal data from online applications to train its system."

Monday, March 4, 2019

Should This Exist? The Ethics Of New Technology; NPR, March 3, 2019

Lulu Garcia-Navarro, NPR; Should This Exist? The Ethics Of New Technology

"In fact, the 2016 election helped raise awareness of an issue that Flickr co-founder Caterina Fake has been talking about in Silicon Valley for years — the ethics of technology.

That conversation was furthered by OpenAI's decision to publicize the nonrelease of their new technology last month, Fake told NPR's Lulu Garcia-Navarro.

"Tech companies don't launch products all the time, but it's rare that they announce that they're not launching a product, which is what has happened here," Fake said. "The announcement of not launching this product is basically to involve people in the conversation around what is and what is not dangerous tech."

When evaluating potential new technology, Fake asks a fundamental question: should this exist?

It's a question she explores as host of the podcast Should This Exist?"

Tuesday, February 26, 2019

When Is Technology Too Dangerous to Release to the Public?; Slate, February 22, 2019

Aaron Mak, Slate; When Is Technology Too Dangerous to Release to the Public?

"The announcement has also sparked a debate about how to handle the proliferation of potentially dangerous A.I. algorithms...

It’s worth considering, as OpenAI seems to be encouraging us to do, how researchers and society in general should approach powerful A.I. models...

Nevertheless, OpenAI said that it would only be publishing a “much smaller version” of the model due to concerns that it could be abused. The blog post fretted that it could be used to generate false news articles, impersonate people online, and generally flood the internet with spam and vitriol... 

“There’s a general philosophy that when the time has come for some scientific progress to happen, you really can’t stop it,” says [Robert] Frederking [the principal systems scientist at Carnegie Mellon’s Language Technologies Institute]. “You just need to figure out how you’re going to deal with it.”"

Sunday, February 24, 2019

Pop Culture, AI And Ethics; Forbes, February 24, 2019

; Pop Culture, AI And Ethics

"In this article, I would like to take the opportunity to do a deep dive into three of the show’s episodes and offer a Design Thinking framework for how to adopt a thoughtful approach on AI implementations. Warning- there are spoilers!...

We need to continuously ask ourselves these 4 questions: How can humanity benefit from this AI/tech? What products and services can you imagine in this space? How might AI be manipulated, or unintended consequences lead to harmful outcomes? What are the suggestions for a responsible future?"

Monday, March 5, 2018

Elon Musk quits AI ethics research group; BBC, February 22, 2018

BBC; Elon Musk quits AI ethics research group

"Technology billionaire Elon Musk has quit the board of the research group he co-founded to look into the ethics of artificial intelligence.

In a blog post, OpenAI said the decision had been taken to avoid any conflict of interest as Mr Musk's electric car company, Tesla, became "more focused on AI".

He has been one of AI's most vocal critics, stressing the potential harms."

Wednesday, January 11, 2017

Tech luminaries team up on $27M A.I. ethics fund; PC World, 1/10/17

Blair Hanley Frank, PC World; 

Tech luminaries team up on $27M A.I. ethics fund:


"Artificial intelligence technology is becoming an increasingly large part of our daily lives. While those developments have led to cool new features, they’ve also presented a host of potential problems, like automation displacing human jobs, and algorithms providing biased results.

Now, a team of philanthropists and tech luminaries have put together a fund that’s aimed at bringing more humanity into the AI development process. It’s called the Ethics and Governance of Artificial Intelligence Fund, and it will focus on advancing AI in the public interest...
[Reid] Hoffman, a former executive at PayPal, has shown quite the interest in developing AI in the public interest and has also provided backing to OpenAI, a research organization aimed at helping create AI that is as safe as possible."