TR

OpenAI Accuses China’s DeepSeek of Distilling U.S. AI Models to Train R1 Chatbot

OpenAI has alerted U.S. lawmakers that Chinese AI firm DeepSeek allegedly used obfuscated servers to extract knowledge from proprietary U.S. models like GPT-4, training its R1 chatbot through model distillation—a technique that bypasses access controls and raises concerns over intellectual property and global AI competition.

calendar_today🇹🇷Türkçe versiyonu
OpenAI Accuses China’s DeepSeek of Distilling U.S. AI Models to Train R1 Chatbot

OpenAI has formally accused Chinese artificial intelligence startup DeepSeek of systematically distilling knowledge from advanced U.S.-developed language models, including GPT-4, to train its own high-performing R1 chatbot, according to a confidential memo shared with the U.S. House Select Committee on Artificial Intelligence. The allegations, first reported by Bloomberg and corroborated by Reuters, suggest DeepSeek employed sophisticated, obfuscated server infrastructure to circumvent access restrictions and licensing protocols, effectively free-riding on billions of dollars in American AI research and development.

Model distillation—a well-documented machine learning technique—typically involves training a smaller, more efficient model to mimic the outputs of a larger, more complex one. While the method itself is not inherently illicit, OpenAI contends that DeepSeek’s implementation violates ethical and legal boundaries by accessing proprietary models through indirect, unauthorized channels. According to Bloomberg, internal OpenAI analyses indicate DeepSeek’s R1 model exhibits near-identical reasoning patterns, response structures, and even idiosyncratic behavioral quirks found in GPT-4, despite lacking equivalent training data or computational resources.

Reuters reports that OpenAI’s memo details how DeepSeek researchers allegedly used third-party cloud providers and anonymized API gateways to query GPT-4 millions of times, harvesting outputs that were then used to fine-tune R1. This approach, known as "black-box distillation," avoids direct access to model weights or training data, making it harder to detect and legally challenge. OpenAI’s legal team argues this constitutes a form of intellectual property theft, as it exploits the public-facing interface of a commercially licensed model to replicate its intelligence without payment or permission.

The implications extend beyond corporate rivalry. U.S. lawmakers are now weighing whether to expand export controls on AI model access, tighten API usage agreements, or mandate transparency requirements for foreign entities training on U.S.-developed systems. The Department of Commerce’s Bureau of Industry and Security has reportedly initiated a preliminary review into whether DeepSeek’s methods violate the Export Administration Regulations (EAR), particularly under provisions governing "technology transfer" and "deemed exports."

DeepSeek has not publicly responded to the allegations. However, industry analysts note that the company has previously emphasized its commitment to open-source development and independent training. Its R1 model, released in late 2025, achieved top-tier performance on multiple international benchmarks—including MMLU, GSM8K, and HumanEval—without disclosing its training methodology. Some experts suggest that while distillation is plausible, the extent of replication may be overstated by OpenAI as part of a broader strategic effort to influence U.S. policy and restrict Chinese AI advancement.

Meanwhile, academic researchers warn that such accusations could trigger a dangerous escalation in the global AI arms race. "If every major AI developer begins treating model outputs as proprietary secrets, we risk fragmenting the open scientific ecosystem that propelled modern AI forward," said Dr. Lena Zhao, a machine learning ethicist at Stanford University. "The real challenge is distinguishing between legitimate innovation and exploitation—and crafting policy that protects innovation without stifling global collaboration."

As the U.S. Congress prepares hearings on AI model governance, the DeepSeek case may become a landmark test of how intellectual property law adapts to the era of generative AI. With China’s state-backed AI ambitions intensifying and U.S. firms racing to maintain dominance, the line between competitive advantage and unethical extraction is blurring—and the world’s most powerful technology may be the one most vulnerable to theft by proxy.

AI-Powered Content

recommendRelated Articles