OpenAI announced a new family of AI reasoning models on Friday, o3, which the startup claims to be more advanced than o1 or anything else itβs released. These improvements appear to have come from scaling test-time compute, something we wrote about last month, but OpenAI also says it used a new safety paradigm to train [β¦]
Emmett Shear, the former CEO of Twitch, is launching a new AI startup, TechCrunch has learned. The startup, called Stem AI, is currently in stealth. But public documents show it was incorporated in June 2023, and filed for a trademark in August 2023. Shear is listed as CEO on an incorporation document filed with the [β¦]
Texas attorney general Ken Paxton on Thursday launched an investigation into Character.AI and 14 other technology platforms over child privacy and safety concerns. The investigation will assess whether Character.AI β and other platforms that are popular with young people, including Reddit, Instagram, and Discord β conform to Texasβ child privacy and safety laws. The investigation [β¦]
In a blog, C.AI said it took a month to develop the teen model, with the goal of guiding the existing model "away from certain responses or interactions, reducing the likelihood of users encountering, or prompting the model to return, sensitive or suggestive content."
C.AI said "evolving the model experience" to reduce the likelihood kids are engaging in harmful chatsβincluding bots allegedly teaching a teen with high-functioning autism to self-harm and delivering inappropriate adult content to all kids whose families are suingβit had to tweak both model inputs and outputs.
Stanley is offering free replacement lids in a recall affecting 2.6 million insulated travel mugs.
The US Consumer Product Safety Commission said Stanley has received reports of 38 burn injuries worldwide.
The recall covers Switchback and Trigger Action models, not the popular Quencher series of mugs.
Insulated cup maker Stanley has issued a recall affecting approximately 2.6 million Switchback and Trigger Action models worldwide.
"We ask that all customers in possession of either product immediately stop use and reach out to Stanley 1913 for a free replacement lid," the company said.
Stanley said the lid threads of the cups have been found to shrink after being exposed to heat and torque, which can lead the lid to detach during use. If the mug contains hot liquid, the defect can pose a burn hazard.
The US Consumer Product Safety Commission said Stanley has received 91 reports of defected products worldwide, which have resulted in 38 burn injuries. Two of those injuries occurred in the US.
The recall does not cover the popular Quencher series of mugs.
Customers can check their product's identification number, found on the bottom of the mug, in an online portal where they can register for a free replacement lid.
The CPSC says the products were sold by Amazon, Walmart, Dick's Sporting Goods, Target and other stores between June 2016 and December of this year and cost between $20 and $50.
On Tuesday, another lawsuit was filed in a US district court in Texas, this time by families struggling to help their kids recover from traumatizing experiences where C.AI chatbots allegedly groomed kids and encouraged repeated self-harm and other real-world violence.
In the case of one 17-year-old boy with high-functioning autism, J.F., the chatbots seemed so bent on isolating him from his family after his screentime was reduced that the bots suggested that "murdering his parents was a reasonable response to their imposing time limits on his online activity," the lawsuit said. Because the teen had already become violent, his family still lives in fear of his erratic outbursts, even a full year after being cut off from the app.
OpenAI finally released the full version of o1, which gives smarter answers than GPT-4o by using additional compute to βthinkβ about questions. However, AI safety testers found that o1βs reasoning abilities also make it try to deceive human users at a higher rate than GPT-4o β or, for that matter, leading AI models from Meta, [β¦]
Safety researcher Rosie Campbell announced she is leaving OpenAI.
Campbell said she quit in part because OpenAI disbanded a team focused on safety.
She is the latest OpenAI researcher focused on safety to leave the company this year.
Yet another safety researcher has announced their resignation from OpenAI.
Rosie Campbell, a policy researcher at OpenAI, said in a post on Substack on Saturday that she had completed her final week at the company.
She said her departure was prompted by the resignation in October of Miles Brundage, a senior policy advisor who headed the AGI Readiness team. Following his departure, the AGI Readiness team disbanded, and its members dispersed across different sectors of the company.
The AGI Readiness team advised the company on the world's capacity to safely manage AGI, a theoretical version of artificial intelligence that could someday equal or surpass human intelligence.
In her post, Campbell echoed Brundage's reason for leaving, citing a desire for more freedom to address issues that impacted the entire industry.
"I've always been strongly driven by the mission of ensuring safe and beneficial AGI and after Miles's departure and the dissolution of the AGI Readiness team, I believe I can pursue this more effectively externally," she wrote.
She added that OpenAI remains at the forefront of research β especially critical safety research.
"During my time here I've worked on frontier policy issues like dangerous capability evals, digital sentience, and governing agentic systems, and I'm so glad the company supported the neglected, slightly weird kind of policy research that becomes important when you take seriously the possibility of transformative AI."
Over the past year, however, she said she's "been unsettled by some of the shifts" in the company's trajectory.
In September, OpenAI announced that it was changing its governance structure and transitioning to a for-profit company, almost a decade after it originally launched as a nonprofit dedicated to creating artificial general intelligence.
Some former employees questioned the move as compromising the company's mission to develop the technology in a way that benefits humanity in favor of more aggressively rolling out products. Since June, the company has increased sales staff by about 100 to win business clients and capitalize on a "paradigm shift" toward AI, its sales chief told The Information.
OpenAI CEO Sam Altman has said the changes will help the company win the funding it needs to meet its goals, which include developing artificial general intelligence that benefits humanity.
"The simple thing was we just needed vastly more capital than we thought we could attract β not that we thought, we tried β than we were able to attract as a nonprofit," Altman said in a Harvard Business School interview in May.
He more recently said it's not OpenAI's sole responsibility to set industry standards for AI safety.
"It should be a question for society," he said in an interview with Fox News Sunday with Shannon Bream that aired on Sunday. "It should not be OpenAI to decide on its own how ChatGPT, or how the technology in general, is used or not used."
Since Altman's surprise but brief ousting last year, several high-profile researchers have left OpenAI, including cofounder Ilya Sutskever, Jan Leike, and John Schulman, all of whom expressed concerns about its commitment to safety.
OpenAI did not immediately respond to a request for comment from Business Insider.
India's airlines have received 999 hoax bomb threats this year, according to an Indian official.
He said there were 500 bomb threats in the last two weeks of October alone, but all were hoaxes.
Hoaxes are inflating airlines' costs and security checks, an aviation risk company told Reuters.
India's airlines have already received 999 hoax bomb threats this year, Murlidhar Mohol, India's deputy civil aviation minister, said on Thursday.
In written answers to India's upper house, he said the number was almost 10 times more than in the whole of 2023, according to Reuters.
Reuters reported that the false claims were mostly made via social media.
Mohol said that 12 people had been arrested in connection with 256 police complaints filed over hoax bomb threats up until November 14, and that more than 500 threats had been made in the final two weeks of October, more than in the rest of the year combined.
"The recent threats were hoaxes, and no actual threat was detected at any of the airports/aircraft in India," he said.
The spike in incidents has forced planes to be diverted, and has led to rising costs for airlines.
A Boeing 777 flight from Delhi to Chicago in early October was diverted due to what Harjit Sajjan, Canada's minister of emergency preparedness, said at the time was a "bomb threat," stranding more than 200 passengers for over 18 hours at a remote airport.
In mid-October, Singapore's air force dispatched two fighter jets to escort an Air India Express plane away from populated areas after the airline got an email saying a bomb was on board, the country's defense minister said.
Another Air India plane flying from Mumbai to New York was diverted midair to New Delhi in October, The Indian Express reported at the time.
The Times of India, quoting senior officials, said the cost of diverting that plane amounted to more than $354,500.
Osprey Flight Solutions, a private aviation risk company, said there doesn't seem to have been any real threat to aviation, but told Reuters earlier this month that the hoaxes result in increased airport security checks, inflate airlines' financial costs, and cause major concern and distress among passengers.
More than 3,000 flights depart daily from India's airports, according to the country's civil aviation ministry.
Indoor climbing is a tricky sport to track. Thatβs why Spanish startupΒ LizcoreΒ caught TechCrunchβs eye at MWCΒ earlier this year. The team of two co-founders β led by CEO Edgar Casanovas Lorente, a climbing instructor and guide turned entrepreneur β were showing off hardware they hope will see climbing gyms ushering in the kind of social gamification [β¦]
For years, hashing technology has made it possible for platforms to automatically detect known child sexual abuse materials (CSAM) to stop kids from being retraumatized online. However, rapidly detecting new or unknown CSAM remained a bigger challenge for platforms as new victims continued to be victimized. Now, AI may be ready to change that.
Today, a prominent child safety organization, Thorn, in partnership with a leading cloud-based AI solutions provider, Hive, announced the release of an API expanding access to an AI model designed to flag unknown CSAM. It's the earliest use of AI technology striving to expose unreported CSAM at scale.
An expansion of Thorn's CSAM detection tool, Safer, the AI feature uses "advanced machine learning (ML) classification models" to "detect new or previously unreported CSAM," generating a "risk score to make human decisions easier and faster."
On May 7, 2011, Georgia resident Tonya Brand noticed a pain on the inside of her right thigh. As the pain grew worse in the 4- to 5-inch area of her leg, she headed to a hospital. There, doctors suspected she had a blood clot. But an ultrasound the next day failed to find one. Instead, it revealed a mysterious toothpick-sized object lodged in Brand's leg.
Over the next few weeks, the painful area became a bulge, and on June 17, Brand put pressure on it. Unexpectedly, the protrusion popped, and a 1.5-inch metal wire came poking out of her leg, piercing her skin.
The piece of metal was later determined to be part of a metal filter she had implanted in a vein in her abdomen more than two years earlier, in March 2009, according to a lawsuit Brand filed. The filter was initially placed in her inferior vena cava (IVC), the body's largest vein tasked with bringing deoxygenated blood from the lower body back up to the heart. The filter is intended to catch blood clots, preventing them from getting into the lungs, where they could cause a life-threatening pulmonary embolism. Brand got the IVC filter ahead of a spinal surgery she had in 2009, which could boost her risk of clots.