Normal view

There are new articles available, click to refresh the page.
Today — 30 January 2025Main stream

Report: DeepSeek’s chat histories and internal data were publicly exposed

30 January 2025 at 11:49

A cloud security firm found a publicly accessible, fully controllable database belonging to DeepSeek, the Chinese firm that has recently shaken up the AI world, "within minutes" of examining DeepSeek's security, according to a blog post by Wiz.

An analytical ClickHouse database tied to DeepSeek, "completely open and unauthenticated," contained more than 1 million instances of "chat history, backend data, and sensitive information, including log streams, API secrets, and operational details," according to Wiz. An open web interface also allowed for full database control and privilege escalation, with internal API endpoints and keys available through the interface and common URL parameters.

"While much of the attention around AI security is focused on futuristic threats, the real dangers often come from basic risks—like accidental external exposure of databases," writes Gal Nagli at Wiz's blog. "As organizations rush to adopt AI tools and services from a growing number of startups and providers, it’s essential to remember that by doing so, we’re entrusting these companies with sensitive data. The rapid pace of adoption often leads to overlooking security, but protecting customer data must remain the top priority."

Read full article

Comments

© Getty Images

I agree with OpenAI: You shouldn’t use other peoples’ work without permission

ChatGPT developer OpenAI and other players in the generative AI business were caught unawares this week by a Chinese company named DeepSeek, whose open source R1 simulated reasoning model provides results similar to OpenAI's best paid models (with some notable exceptions) despite being created using just a fraction of the computing power.

Since ChatGPT, Stable Diffusion, and other generative AI models first became publicly available in late 2022 and 2023, the US AI industry has been undergirded by the assumption that you'd need ever-greater amounts of training data and compute power to continue improving their models and get—eventually, maybe—to a functioning version of artificial general intelligence, or AGI.

Those assumptions were reflected in everything from Nvidia's stock price to energy investments and data center plans. Whether DeepSeek fundamentally upends those plans remains to be seen. But at a bare minimum, it has shaken investors who have poured money into OpenAI, a company that reportedly believes it won't turn a profit until the end of the decade.

Read full article

Comments

© Benj Edwards / OpenAI

Microsoft now hosts AI model accused of copying OpenAI data

Fresh on the heels of a controversy in which ChatGPT-maker OpenAI accused the Chinese company behind DeepSeek R1 of using its AI model outputs against its terms of service, OpenAI's largest investor, Microsoft, announced on Wednesday that it will now host DeepSeek R1 on its Azure cloud service.

DeepSeek R1 has been the talk of the AI world for the past week because it is a freely available simulated reasoning model that reportedly matches OpenAI's o1 in performance—while allegedly being trained for a fraction of the cost.

Azure allows software developers to rent computing muscle from machines hosted in Microsoft-owned data centers, as well as rent access to software that runs on them.

Read full article

Comments

© Wong Yu Liang via Getty Images

Yesterday — 29 January 2025Main stream

Microsoft brings a DeepSeek model to its cloud

29 January 2025 at 13:08

Microsoft’s close partner and collaborator, OpenAI, might be suggesting that DeepSeek stole its IP and violated its terms of service. But Microsoft still wants DeepSeek’s shiny new models on its cloud platform. Microsoft today announced that R1, DeepSeek’s so-called reasoning model, is available on Azure AI Foundry service, Microsoft’s platform that brings together a number […]

© 2024 TechCrunch. All rights reserved. For personal use only.

What DeepSeek, China, and Trump mean for Apple ahead of its Q1 earnings

29 January 2025 at 12:26

These days, nothing is certain about the tech market or the world at large. Even Nvidia’s seemingly bulletproof stock took a hammering on Monday, enduring a $589 billion market cap decline after China-based DeepSeek raised questions for investors about more efficient AI models. But Apple’s stock has remained steady ahead of its first-quarter earnings report, […]

© 2024 TechCrunch. All rights reserved. For personal use only.

Before yesterdayMain stream

DeepSeek panic triggers tech stock sell-off as Chinese AI tops App Store

On Monday, Nvidia stock dove 17 percent amid worries over the rise of Chinese AI company DeepSeek, whose R1 reasoning model stunned industry observers last week by challenging American AI supremacy with a low-cost, freely available AI model, and whose AI assistant app jumped to the top of the iPhone App Store's "Free Apps" category over the weekend, overtaking ChatGPT.

What’s the big deal about DeepSeek?

The drama started around January 20 when Chinese AI startup DeepSeek announced R1, a new simulated reasoning (SR) model that it claimed could match OpenAI's o1 in reasoning benchmarks. Like o1, R1 is trained to work through a simulated chain of thought process before providing an answer, which can potentially improve the accuracy or usefulness of the AI models' outputs for some types of questions posed by the user.

That first part wasn't too surprising since other AI companies like Google are hot on the heels of OpenAI with their own simulated reasoning models. In addition, OpenAI itself has announced an upcoming SR model (dubbed "o3") that can surpass o1 in performance.

Read full article

Comments

© Luis Diaz Devesa via Getty Images

Cutting-edge Chinese “reasoning” model rivals OpenAI o1—and it’s free to download

On Monday, Chinese AI lab DeepSeek released its new R1 model family under an open MIT license, with its largest version containing 671 billion parameters. The company claims the model performs at levels comparable to OpenAI's o1 simulated reasoning (SR) model on several math and coding benchmarks.

Alongside the release of the main DeepSeek-R1-Zero and DeepSeek-R1 models, DeepSeek published six smaller "DeepSeek-R1-Distill" versions ranging from 1.5 billion to 70 billion parameters. These distilled models are based on existing open source architectures like Qwen and Llama, trained using data generated from the full R1 model. The smallest version can run on a laptop, while the full model requires far more substantial computing resources.

The releases immediately caught the attention of the AI community because most existing open-weights models—which can often be run and fine-tuned on local hardware—have lagged behind proprietary models like OpenAI's o1 in so-called reasoning benchmarks. Having these capabilities available in an MIT-licensed model that anyone can study, modify, or use commercially potentially marks a shift in what's possible with publicly available AI models.

Read full article

Comments

© Wong Yu Liang

❌
❌