OpenAI was ordered by a federal court to preserve all ChatGPT conversation logs indefinitely as part of a copyright lawsuit brought by The New York Times, but that sweeping mandate has now been curtailed: a new October ruling ends the requirement to preserve every output log going forward, though data already preserved and records tied to accounts flagged by the Times must still be retained. (Ars Technica) OpenAI had strongly opposed the original order, arguing it conflicted with its privacy commitments. (Engadget) Under the new ruling, OpenAI can resume its normal data-deletion practices for most users after September 26, but the company is still obligated to hold onto logs deemed relevant to the litigation. (LiveMint)
Key Takeaways
– OpenAI is no longer forced to preserve all ChatGPT logs indefinitely; the court has narrowed the scope of data retention.
– Data already preserved (under the old order) and logs tied to NYT-flagged accounts remain subject to retention.
– The change reduces privacy and operational burdens on OpenAI, though litigation access to key evidence persists.
In-Depth
Last May, a federal court issued a sweeping preservation order requiring OpenAI to retain all ChatGPT conversation logs—even ones that users deleted or that would normally be purged—so that The New York Times (and related plaintiffs) could inspect the logs for evidence of how the AI handled their copyrighted material. Under that mandate, routine deletions were effectively suspended, placing OpenAI in a position of storing masses of new conversational data for no reason other than discovery. OpenAI pushed back hard, arguing that such a broad demand ran counter to user privacy norms and imposed exorbitant operational burdens.
On October 9, the court (via Magistrate Judge Ona T. Wang) struck a more targeted balance. The broad requirement was lifted for new logs going forward: OpenAI no longer must “preserve and segregate all output log data that would otherwise be deleted.” The cutoff date of September 26 marks the point beyond which the blanket retention no longer applies. But it’s not a full rollback: logs already collected remain in the discovery pool, and OpenAI must continue preserving chats tied to accounts that The Times designates as relevant (flagged accounts). In short, the ruling retreats from a dragnet data grab while preserving the plaintiff’s access to material it deems critical.
The implications are meaningful. For users, the risk that every casual ChatGPT session becomes legal fodder is reduced. OpenAI can more confidently re-enable its standard deletion practices, alleviating storage loads and exposure risk. For the lawsuit, it ensures that key evidence remains available without forcing OpenAI into perpetual hoarding of user data. The Times retains recourse: it may flag additional accounts, compelling further retention, as the litigation proceeds.
This ruling signals judicial skepticism toward overbroad data preservation in AI litigation. As cases involving generative models grow, courts may demand more narrowly tailored discovery orders—ones tied to specific accounts, time windows, or content types—rather than blanket obligations over entire platforms. That approach better protects user privacy and operational sanity while preserving the integrity of legal process.

