Openai has reportedly claimed that Deepseek might have disturbed its artificial intelligence (ai) models to build the r1 model. As per the report, the San Francisco-Based AI Firm Stated It Has Evidence that some users were used its ai models' outputs for a competitor, which is suspected to before. Notable, the chinese company released the open-source Deepsek-R1 ai model last week and hosted it on github and hugging face. The reasoning-focused model surpassed the capability of the chatgpt-maker's o1 ai models in Several Benchmarks.

Openai says it has evidence of foulplay

According to a Financial Times ReportOpenai Claimed that its proprietary ai models was used to train Deepsek's models. The company told the publication that it had seen evidence of distillation from Several Accounts Using The Openai Application Programming Interface (API). The AI ​​Firm and Its Cloud Partner Microsoft Investigated the issue and blocked their access.

In a statement to the financial time, Openai said, “We know [China]-Based Companies -And Others -Are Constantly Trying to Distil The Models of Leading Us Ai Companies. ” The chatgpt-maker also highlighted that it is working closely with the US government to protect its frontier models from from competitors and adversaries.

Notably, AI Model Distillation is a Technique Used to Transfer Knowledge from a large model to a smaller and more efficient model. The goal here is to brings the smaller model on par or ahead of the larger model while reduction computational requirements. Notably, Openai's GPT-4 has roughly 1.8 trillion parameters while Deepsek-R1 has 1.5 billion parameters, which would fit the description.

The knowledge transfer typically takes place by using the relevant dataset from the larger model to train the smaller model, when a company is creating more efficient versions of its model in-hues. For instance, meta used the llama 3 ai model to create several coding-focused llama models.

However, this is not possible when a competitor, which does not have access to the datasets of a proprietary model, wants to distil a model. If Openai's Allegations are true, this could have been done by adding prompt injections to its apis to generate a large number of outputs. This natural language data is then converted to code and fed to a base model.

Notably, Openai has not publicly issued a statement Regarding this. Recently, the company ceo sam altman praised deepsek for creating

Source link

Leave a Reply

Your email address will not be published. Required fields are marked *