OpenAI, the creator of ChatGPT, is now legally obligated to preserve all ChatGPT user conversations, including those that have been deleted. This mandate stems from an ongoing copyright lawsuit filed against OpenAI by The New York Times, along with other news organizations. The plaintiffs argue that OpenAI's ChatGPT may be reproducing copyrighted content without permission, and that users might be using the AI to bypass paywalls by requesting summaries of articles.
The court order, issued by a New York magistrate judge, requires OpenAI to "preserve and segregate all output log data" from ChatGPT, regardless of whether users have deleted those conversations. This ruling has sparked widespread privacy concerns, as it means that personal conversations, including potentially sensitive information shared with the AI, will be retained and could be accessed during the litigation process. These conversations may include medical queries, relationship issues, and other private matters shared in confidence.
Prior to this legal pressure, OpenAI had a policy of deleting conversations from user accounts immediately and permanently erasing them from their systems within 30 days. Temporary chats were also purged after 30 days. However, this policy is now on hold for the duration of the case, except for Enterprise and Education accounts with special data agreements. This change means that "delete" doesn't mean permanent erasure at the moment.
OpenAI has expressed strong opposition to the court order, arguing that it violates user privacy and imposes significant technical and financial burdens. CEO Sam Altman has advocated for "AI privilege," suggesting that AI chats should have the same privacy as lawyer-client or doctor-patient conversations. The company has filed a reconsideration motion, which was denied, and has since appealed to the District Judge. Despite these efforts, the court's decision stands for now, and all logs must be kept indefinitely unless the order is changed.
To comply with the ruling, OpenAI says that the retained data will be "stored separately in a secure system" and that access will be limited to "a small, audited OpenAI legal and security team". The company insists that this data is not automatically shared with The New York Times or anyone else and is locked under a separate legal hold, meaning it can only be accessed under strict legal protocols.
This situation has raised several concerns and questions:
The OpenAI case serves as a wake-up call for data privacy in the AI era. It highlights the vulnerability of data when users surrender control to third-party AI providers and raises troubling questions about the power dynamics between AI companies and users. This legal battle could redefine data retention norms for AI technologies and reshape how user privacy is treated in tech litigation for years to come.