Alexis Ohanian predicts AI will drive demand for more raw human experiences.
In 10 years, live theater will be more popular than ever, the Reddit cofounder contends.
He says no matter what jobs are replaced by AI, humans will always have an advantage in empathy.
Alexis Ohanian predicted that in a future oversaturated with artificial intelligence, people will seek out more raw, emotive human experiences.
And in 10 years, he said, live theater will be more popular than ever.
The 41-year-old, who co-founded social media platform Reddit in 2005, told the "On Purpose with Jay Shetty" podcast this week that AI will soon have an undeniable impact on nearly every aspect of society, including the entertainment sector.
Ohanian, who also founded venture capital firm Seven Seven Six in 2020, said that the industry will see a big shift when AI makes on-screen entertainment better, faster, cheaper, and more dynamic β which he said is happening.
Every screen we look at will become so programmed to show us "what we want, when we want it, how we want it," he said, that "a part of our humanity will miss, you know, thousands of years ago when we were sitting around a campfire and that great storyteller was doing the voices and the impressions.'"
"That's ingrained in our species," he said.
And that kind of raw, in-person magic will feel novel, he suggested.
"I actually bet 10 years from now live theater will be more popular than ever," Ohanian said. "Because, again, we'll look at all these screens with all these AI-polished images, and we'll actually want to sit in a room with other humans to be captivated for a couple hours in a dark room to feel the goosebumps of seeing live human performances."
The same is true for sports, he told Shetty. "We need humans doing that. We need to feel their pain and their success and their triumphs," he said. "Those are the areasthat get me most hopeful."
AI can't replace genuine human empathy, Ohanian suggested.
No matter what jobs robots take over from us in the future, fields of work where empathy is a core component of the job will have an advantage, he said. And that's why one of the most important, marketable skills he's teaching his kids is empathy, he said.
OpenAI's o1 model was hailed as a breakthrough in September.
By November, a Chinese AI lab had released a similar model called DeepSeek.
On Thursday, Google came out with a challenger called Gemini 2.0 Flash Thinking.
In September, OpenAI unveiled a radically new type of AI model called o1. In a matter of months, rivals introduced similar offerings.
On Thursday, Google released Gemini 2.0 Flash Thinking, which uses reasoning techniques that look a lot like o1.
Even before that, in November, a Chinese company announced DeepSeek, an AI model that breaks challenging questions down into more manageable tasks like OpenAI's o1 does.
This is the latest example of a crowded AI frontier where pricey innovations are swiftly matched, making it harder to stand out.
"It's amazing how quickly AI model improvements get commoditized," Rahul Sonwalkar, CEO of the startup Julius AI, said. "Companies spend massive amounts building these new models, and within a few months they become a commodity."
The proliferation of multiple AI models with similar capabilities could make it difficult to justify charging high prices to use these tools. The price of accessing AI models has indeed plunged in the past year or so.
That, in turn, could raise questions about whether it's worth spending hundreds of millions of dollars, or even billions, to build the next top AI model.
September is a lifetime ago in the AI industry
When OpenAI previewed its o1 model in September, the product was hailed as a breakthrough. It uses a new approach called inference-time compute to answer more challenging questions.
It does this by slicing queries into more digestible tasks and turning each of these stages into a new prompt that the model tackles. Each step requires running a new request, which is known as the inference stage in AI.
This produces a chain of thought or chain of reasoning in which each part of the problem is answered, and the model doesn't move on to the next stage until it ultimately comes up with a full response.
The model can even backtrack and check its prior steps and correct errors, or try solutions and fail before trying something else. This is akin to how humans spend longer working through complex tasks.
DeepSeek rises
In a mere two months, o1 had a rival. On November 20, a Chinese AI company released DeepSeek.
"They were probably the first ones to reproduce o1," said Charlie Snell, an AI researcher at UC Berkeley who coauthored a Google DeepMind paper this year on inference-time compute.
He's tried DeepSeek's AI model and says it performs well on complex math problems that must be solved by thinking for longer and in stages.Β
He noted that in DeepSeek's DeepThink mode, the model shows users every step of its thought process. With o1, these intermediate steps are hidden from users.Β
"I've asked people at OpenAI what they think of it," Snell told BI. "They say it looks like the same thing, but they don't how DeepSeek did this so fast."
OpenAI didn't respond to a request for comment.Β On Friday, the startup previewed an o1 successor, called o3. Francois Chollet, a respected AI expert, called the update a "significant breakthrough."
Andrej Karpathy, an OpenAI cofounder, praised Google's new "Thinking" model for the same reasoning feature.
"The prominent and pleasant surprise here is that unlike o1 the reasoning traces of the model are shown," he wrote on X. "As a user I personally really like this because the reasoning itself is interesting to see and read β the models actively think through different possibilities, ideas, debate themselves, etc., it's part of the value add."
A DeepSeek demo
Snell shared a multistep math problem with Business Insider, which we used to test DeepSeek for ourselves:Β
"Find a sequence of +, -, /, * which can be applied to the numbers 7, 3, 11, 5 to get to 24, where each of the given numbers is used exactly once."
BI put that prompt in DeepSeek's chat window on its website. The model responded initially by laying out the challenge ahead.
"Alright, so I've got this problem here: I need to use the numbers 7, 3, 11, and 5, and combine them with the operations of addition, subtraction, multiplication, and division, using each number exactly once, to get to 24," it replied. "At first glance, this seems a bit tricky, but I think with some systematic thinking, I can figure it out."
It then proceeded through multiple steps over roughly 16 pages of discussion that included mathematical calculationsΒ and equations. The model sometimes got it wrong, but it spotted this and didn't give up. Instead, it swiftly moved on to another possible solution.Β
"Almost got close there with 33 / 7 * 5 β 23.57, but not quite 24. Maybe I need to try a different approach," it wrote at one point.Β
After a few minutes, it found the correct solution.Β
"You can see it try different ideas and backtrack," Snell said in an interview on Wednesday.Β He highlighted this part of DeepSeek's chain of thought as particularly noteworthy:
"This is getting really time-consuming. Maybe I need to consider a different strategy," the AI model wrote. "Instead of combining two numbers at a time, perhaps I should look for a way to group them differently or use operations in a nested manner."
Then Google appears
Snell said other companies are likely working on AI models that use the same inference-time compute approach as OpenAI.
"DeepSeek does this already, so I assume others are working on this," he added on Wednesday.
The following day, Google releasedΒ Gemini 2.0 Flash Thinking. Like DeepSeek, this new model shows users each step of its thought process while tackling problems.Β
Jeff Dean, a Google AI veteran, shared a demo on X that showed this new model solving a physics problem and explained its reasoning steps.Β
"This model is trained to use thoughts to strengthen its reasoning," Dean wrote. "We see promising results when we increase inference time computation!"
Suchir Balaji helped OpenAI collect data from the internet for AI model training, the NYT reported.
He was found dead in an apartment in San Francisco in late November, according to police.
About a month before, Balaji published an essay criticizing how AI models use data.
The recent death of former OpenAI researcherΒ Suchir Balaji has brought an under-discussed AI debate back into the limelight.
AI models are trained on information from the internet. These tools answer user questions directly, so fewer people visit the websites that created and verified the original data. This drains resources from content creators, which could lead to a less accurate and rich internet.
Elon Musk calls this "Death by LLM." Stack Overflow, a coding Q&A website, has already been damaged by this phenomenon. And Balaji was concerned about this.
Balaji was found dead in late November. The San Francisco Police Department said it found "no evidence of foul play" during the initial investigation. TheΒ city's chief medical examiner determined the death to be suicide.
Balaji's concerns
About a month before Balaji died, he published an essay on his personal website that addressed how AI models are created and how this may be bad for the internet.Β
He citedΒ researchΒ that studied the impact of AI models using online data for free to answer questions directly while sucking traffic away from the original sources.
The study analyzed Stack Overflow and found that traffic to this site declined by about 12% after the release of ChatGPT. Instead of going to Stack Overflow to ask coding questions and do research, some developers were just asking ChatGPT for the answers.Β
Other findings from the research Balaji cited:Β
There was a decline in the number of questions posted on Stack Overflow after the release of ChatGPT.
The average account age of the question-askers rose after ChatGPT came out, suggesting fewer people signed up to Stack Overflow or that more users left the online community.
This suggests that AI models could undermine some of the incentives that created the information-rich internet as we know it today.
If people can get their answers directly from AI models, there's no need to go to the original sources of the information. If people don't visit websites as much, advertising and subscription revenue may fall, and there would be less money to fund the creation and verification of high-quality online data.
MKBHD wants to opt out
It's even more galling to imagine that AI models might be doing this based partly on your own work.Β
Tech reviewer Marques Brownlee experienced this recently when he reviewed OpenAI's Sora video model and found that it created a clip with a plant that looked a lot like a plant from his own videos posted on YouTube.Β
"Are my videos in that source material? Is this exact plant part of the source material? Is it just a coincidence?" said Brownlee, who's known as MKBHD.
Naturally, he also wanted to know if he could opt out and prevent his videos from being used to train AI models.Β "We don't know if it's too late to opt out," Brownlee said.
'Not a sustainable model'
In an interview with The New York Times published in October, Balaji said AI chatbots like ChatGPT are stripping away the commercial value of people's work and services.
The publication reported that while working at OpenAI, Balaji was part of a team that collected data from the internet for AI model training. He joined the startup with high hopes for how AI could help society, but became disillusioned, NYT wrote.Β
"This is not a sustainable model for the internet ecosystem," he told the publication.
In a statement to the Times about Balaji's comments, OpenAI said the way it builds AI models is protected by fair use copyright principles and supported by legal precedents. "We view this principle as fair to creators, necessary for innovators, and critical for US competitiveness," it added.
In his essay, Balaji disagreed.
One of the four tests for copyright infringement is whether a new work impacts theΒ potential market for, or value of, the original copyrighted work. If it does this type of damage, then it's notΒ "fairΒ use" and is not allowed.Β
Balaji concluded that ChatGPT and other AI models don't quality for fair use copyright protection.Β
"None of the four factors seem to weigh in favor of ChatGPT being a fair use of its training data," he wrote. "That being said, none of the arguments here are fundamentally specific to ChatGPT either, and similar arguments could be made for many generative AI products in a wide variety of domains."
Talking about data
Tech companies producing these powerful AI models don'tΒ like to talk about the value of training data. They've even stopped disclosing where they get the data from, which was a common practice until a few years ago.Β
"They always highlight their clever algorithms, not the underlying data," Nick Vincent, an AI researcher, told BI last year.
Balaji's death may finally give this debate the attention it deserves.Β
"We are devastated to learn of this incredibly sad news today and our hearts go out to Suchir's loved ones during this difficult time," an OpenAI spokesperson told BI recently.Β
If you or someone you know is experiencing depression or has had thoughts of harming themself or taking their own life, get help. In the US, call or text 988 to reach the Suicide & Crisis Lifeline, which provides 24/7, free, confidential support for people in distress, as well as best practices for professionals and resources to aid in prevention and crisis situations. Help is also available through the Crisis Text Line β just text "HOME" to 741741. The International Association for Suicide Prevention offers resources for those outside the US.
Congress's bipartisan task force on artificial intelligence (AI) released its long-anticipated report this week, detailing strategies for how the U.S. can protect itself against emerging AI-related threats while ensuring the nation remains a leader in innovation within this rapidly evolving sector.
Responses to the report, which sought to strike a "flexible sectoral regulatory framework," were positive and with mixed concerns.Β
"The Task Force report offers a thoughtful and forward-thinking framework that balances AI's transformative economic potential with the imperative to address legitimate safety concerns," said Dr. Vahid Behzadan, a professor in the computer science department at the University of New Haven. "That said, there's still work to be done."
He pointed to the importance of developing an "international collaboration strategy," especially with U.S. allies, the need to establish "clearer priorities among the many recommendations provided" and the need for more guidance on market competition and consolidation.Β
The Center for AI Policy, a nonpartisan research organization based in the nation's capital, issued a press release that commended lawmakers for their work on the report. But the group echoed Behzadan's remarks about the need for more detail.
"The body of the report does not contain enough detail about how or when these frameworks will be created," the group said after the report's release. It also expressed concern over the report's lack of emphasis on "catastrophic risks" posed by AI.
"Congress has deliberated on AI for two years now, and it is time to start moving forward with decisive action," the press release stated.Β
Yaron Litwin is the chief marketing officer for Canopy, a digital parenting app and an expert in how AI technology is revolutionizing parental control and internet safety. He said "we need faster" and "stronger" protections than what was laid out in the report. "To me, the report appears more business-friendly than not."
The report pointed out that it would be "unreasonable to expect Congress to enact legislation this year that could serve as its last word on AI policy." But while Congress may be slow to act, some states have already moved the ball forward on regulating AI, and experts who spoke to Fox News Digital said the report could serve to bolster those efforts.
Lawmakers in Colorado enacted the first comprehensive piece of AI legislation this year, which placed certain obligations on developers of "high-risk artificial intelligence systems." Meanwhile, in California, lawmakers passed a bill this month aiming to regulate AI in health care.
"These federal soft law standards could work alongside state efforts to protect consumers and give businesses clear, consistent, and science-based federal guidelines," said Tatiana Rice, Deputy Director for U.S. Legislation at the Future of Privacy Forum, a nonprofit that explores challenges posed by technological innovation. Rice pointed out that an increasing number of state AI laws "include carveouts or assumptions of compliance if businesses adhere to federally recognized standards," and she noted that Congress's approach will likely "make it easier for businesses to meet legal requirements, incentivize consumerΒ trust and safety, and reduce regulatory complexity."
Craig Albright, Senior Vice President of U.S. Government Relations for the Business Software Alliance, posited that the report could likely encourage states "to be more aggressively [sic] next year than what we are expecting to see in Congress."
On the issue of whether the 25-page report strikes the balance that lawmakers were hoping for in terms of balancing regulation with the need to foster innovation, experts who spoke to Fox News Digital expressed optimism.
"While there will always be debate over regulation versus not enough government oversight, the report is a step in the right direction," said Griffiths. "With the development of any new technology, regulation requires a nuanced and flexible approach. My recommendation going forward will be for Congress to pick and choose to legislate on specific aspects of AI policy."
Griffiths' reaction to the report was echoed by others who warned that in such a rapidly evolving industry, it will be critical not to get trigger-happy with regulations that could soon become obsolete.
"It is encouraging that the report suggests taking an incremental approach to AI policy,"Β said JD Harriman, a partner at Foundation Law Group who has worked as outside patent council at technology corporations like Apple and Pixar. "Many areas of technology have been stifled by over-regulation before a complete understanding of the technology was undertaken."
"The task forceβs honesty β βWe donβt know what we donβt knowβ β is both refreshing and daunting," added Cassidy Reid, the founder of Women in Automation, a nonprofit group that supports women in the tech sector. "It acknowledges the speed of AIβs evolution but raises a bigger question: Are we ready to govern something so inherently unpredictable?"
Early testers of Google's new AI video generator compared it with OpenAI's Sora.
So far, Google's results are blowing OpenAI's out of the water.
Google has tapped YouTube to train its AI models but says other companies can't do the same.
Not to let OpenAI have all the fun with its 12 days of 'Shipmas,'Google on Monday revealed its new AI video generator, Veo 2. Early testers found it's blowing OpenAI's Sora out of the water.
OpenAI made its Sora AI video generator available for general use earlier this month, while Google's is still in early preview. Still, people are sharing comparisons of the two models running the same prompts, and so far, Veo has proved more impressive.
Why is Google's Veo 2 doing better than Sora so far? The answer may be YouTube, which Google owns and has used to train these models.
TED host and former Googler Bilawal Sidhu shared some comparisons of Google's Veo 2 and OpenAI's Sora on X.
He said he used the prompt, "Eating soup like they do in Europe, the old fashioned way," which generated a terrifying result in Sora and something more impressive in Google's Veo 2.
Tried MKBHD's prompt: "A side scrolling shot of a rhinoceros walking through a dry field of low grass plants"
Took the 1st video I got out of Veo, and it's not even close. Prompt adherence & physics modeling? Leagues apart. Sora nails the look, but⦠pic.twitter.com/mus9MdRsWo
This one from EasyGen founder Ruben Hassid included a prompt for someone cutting a tomato with a knife. In the video he shared, we see how Google's Veo 2 had the knife going cleanly through the tomato and avoiding fingers, while the knife in Sora's video cut through the hand.
Granted, these are cherry-picked examples, but the consensus among AI enthusiasts is that Google has outperformed.
Andreessen Horowitz partner Justine Moore wrote on X that she had spent a few hours testing Veo against Sora and believed Sora's "biases towards more motion" while Veo focuses on accuracy and physics.
Sora vs. Veo 2:
I spent a few hours running prompts on both models and wanted to share some comparisons β¬οΈ.
IMO - Sora biases towards more motion, whereas Veo focuses more on accuracy / physics. And a larger % of clips from Veo are usable.
Google has been open about tapping YouTube data for its AI, but it does not permit others to do the same.Β The New York TimesΒ previously reported that OpenAI had trained its models using some YouTube data anyway. YouTube CEO Neal Mohan said OpenAI doing this would violate Google's policies.
Google did not immediately respond to a request for comment.
Are you a Google or OpenAI employee? Got more insight to share? You can reach the reporter Hugh Langley via the encrypted messaging app Signal (+1-628-228-1836) or email ([email protected]).
Generative AI tools have made it easier to create fake images, videos, and audio.
That sparked concern that this busy election year would be disrupted by realistic disinformation.
The barrage of AI deepfakes didn't happen. An AI researcher explains why and what's to come.
Oren Etzioni has studied artificial intelligence and worked on the technology for well over a decade, so when he saw the huge election cycle of 2024 coming, he got ready.
India, Indonesia, and the US were just some of the populous nations sending citizens to the ballot box. Generative AI had been unleashed upon the world about a year earlier, and there were major concerns about a potential wave of AI-powered disinformation disrupting the democratic process.
"We're going into the jungle without bug spray," Etzioni recalled thinking at the time.
He responded by starting TrueMedia.org, a nonprofit that uses AI-detection technologies to help people determine whether online videos, images, and audio are real or fake.
The group launched an early beta version of its service in April, so it was ready for a barrage of realistic AI deepfakes and other misleading online content.
In the end, the barrage never came.
"It really wasn't nearly as bad as we thought," Etzioni said. "That was good news, period."
He's still slightly mystified by this, although he has theories.
First, you don't need AI to lie during elections.
"Out-and-out lies and conspiracy theories were prevalent, but they weren't always accompanied by synthetic media," Etzioni said.
Second, he suspects that generative AI technology is not quite there yet, particularly when it comes to deepfake videos.Β
"Some of the most egregious videos that are truly realistic β those are still pretty hard to create," Etzioni said. "There's another lap to go before people can generate what they want easily and have it look the way they want. Awareness of how to do this may not have penetrated the dark corners of the internet yet."
One thing he's sure of: High-end AI video-generation capabilities will come. This might happen during the next major election cycle or the one after that, but it's coming.
With that in mind, Etzioni shared learnings from TrueMedia's first go-round this year:
Democracies are still not prepared for the worst-case scenario when it comes to AI deepfakes.
There's no purely technical solution for this looming problem, and AI will need regulation.Β
Social media has an important role to play.Β
TrueMedia achieves roughly 90% accuracy, although people asked for more. It will be impossible to be 100% accurate, so there's room for human analysts.
It's not always scalable to have humans at the end checking every decision, so humans only get involved in edge cases, such as when users question a decision made by TrueMedia's technology.Β
The group plans to publishΒ research on its AI deepfake detection efforts, and it's working on potential licensing deals.Β
"There's a lot of interest in our AI models that have been tuned based on the flurry of uploads and deepfakes," Etzioni said. "We hope to license those to entities that are mission-oriented."
Broadcom's stock surged in recent weeks, pushing the company's market value over $1 trillion.
Broadcom is crucial for companies seeking alternatives to Nvidia's AI chip dominance.
Custom AI chips are gaining traction, enhancing tech firms' bargaining power, analysts say.
The rise of AI, and the computing power it requires, is bringing all kinds of previously under-the-radar companies into the limelight. This week it's Broadcom.
Broadcom's stock has soared since late last week, catapulting the company into the $1 trillion market cap club. The boost came from a blockbuster earnings report in which custom AI chip revenue grew 220% compared to last year.
In addition to selling lots of parts and components for data centers, Broadcom designs and sells ASICs, or application-specific integrated circuits β an industry acronym meaning custom chips.
Designers of custom AI chips, chief among them Broadcom and Marvell, are headed into a growth phase, according to Morgan Stanley.
Custom chips are picking up speed
The biggest players in AI buy a lot of chips from Nvidia, the $3 trillion giant with an estimated 90% of market share of advanced AI chips.
Heavily relying on one supplier isn't a comfortable position for any company, though, and many large Nvidia customers are also developing their own chips. Most tech companies don't have large teams of silicon and hardware experts in house. Of the companies they might turn to design them a custom chip, Broadcom is the leader.
Though multi-purpose chips like Nvidia's and AMD's graphics processing units are likely to maintain the largest share of the AI chip market in the long-term, custom chips are growing fast.
Morgan Stanley analysts this week forecast the market for ASICs to nearly double to $22 billion next year.
Much of that growth is attributable to Amazon Web Services' Trainium AI chip, according to Morgan Stanley analysts. Then there are Google's in-house AI chips, known as TPUs, which Broadcom helps make.
In terms of actual value of chips in use, Amazon and Google dominate. But OpenAI, Apple, and TikTok parent company ByteDance are all reportedly developing chips with Broadcom, too.
ASICs bring bargaining power
Custom chips can offer more value, in terms of the performance you get for the cost, according to Morgan Stanley's research.
ASICs can also be designed to perfectly match unique internal workloads for tech companies, accord to the bank's analysts. The better these custom chips get, the more bargaining power they may provide when tech companies are negotiating with Nvidia over buying GPUs. But this will take time, the analysts wrote.
In addition to Broadcom, Silicon Valley neighbor Marvell is making gains in the ASICs market, along with Asia-based players Alchip Technologies and Mediatek, they added in a note to investors.
Analysts don't expect custom chips to ever fully replace Nvidia GPUs, but without them, cloud service providers like AWS, Microsoft, and Google would have much less bargaining power against Nvidia.
"Over the long term, if they execute well, cloud service providers may enjoy greater bargaining power in AI semi procurement with their own custom silicon," the Morgan Stanley analysts explained.
Nvidia's big R&D budget
This may not be all bad news for Nvidia. A $22 billion ASICs market is smaller than Nvidia's revenue for just one quarter.
Nvidia's R&D budget is massive, and many analysts are confident in its ability to stay at the bleeding edge of AI computing.
And as Nvidia rolls out new, more advanced GPUs, its older offerings get cheaper and potentially more competitive with ASICs.
"We believe the cadence of ASICs needs to accelerate to stay competitive to GPUs," the Morgan Stanley analysts wrote.
Still, Broadcom and chip manufacturers on the supply chain rung beneath, such as TSMC, are likely to get a boost every time a giant cloud company orders up another custom AI chip.
Vercel said it added Steffan Tomlinson to its board.
Tomlinson is the CFO of Stripe and has experience taking tech startups public.
He used to be CFO at several other tech companies, including Palo Alto Networks and Confluent.
Vercel, an AI startup valued at more than $3 billion, just bulked up its board with the addition of a finance executive who has experience taking tech companies public.
Stripe Chief Financial Officer Steffan Tomlinson will serve as a director on Vercel's board, the startup said on Tuesday.
Tomlinson was previously CFO at several other tech startups, guiding Palo Alto Networks, Confluent, and Aruba Networks through the IPO process.
Stripe, one of the world's most valuable startups, has long been mentioned as an IPO candidate. Vercel is earlier in its lifecycle, however the AI startup has been putting some of the early pieces in place to potentially go public someday.
"Steffan's experience leading developer-focused companies from startup to public markets makes him an ideal addition to Vercel's Board of Directors as we continue to put our products in the hands of every developer," Vercel CEO and founder Guillermo Rauch said.
Last year, Vercel tapped Marten Abrahamsen as its CFO. He's been building out Vercel's finance, legal, and corporate development teams and systems while leading the startup through a $250 million funding round at a $3.25 billion valuation in May.
"Steffan's financial expertise and leadership experience come at a pivotal moment for Vercel as we scale our enterprise presence and build on our momentum," Abrahamsen said.
GenAI growth
The generative AI boom has recently powered Vercel's growth. The startup offers AI tools to developers, and earlier this year it surpassed $100 million in annualized revenue.
Vercel's AI SDK, a software toolkit that helps developers build AI applications, was downloaded more than 700,000 times last week, up from about 80,000 downloads a year ago, according to NPM data.
The company's Next.js open-source framework was downloaded 7.9 million times last week, compared to roughly 4.6 million downloads a year earlier, NPM data also shows.
Abrahamsen said they are building a company to one day go public, but stressed that there's no timeline or date set for such a move.Β
Consumption-based business models
At Stripe and Confluent, Tomlinson gained experience with software that helps developers build cloud and web-based applications β and how these offerings generate revenue.
"Steffan's track record with consumption-based software business models makes him the ideal partner to inform strategic decisions," Rauch said.
Vercel is among a crop of newer developer-focused tech companies that charge based on usage. For instance, as traffic and uptime increase for developers, Vercel generates more revenue, so it's aligned with customers,Β Abrahamsen told Business Insider.Β
Similarly, Stripe collects a small fee every time someone makes a payment in an app. Confluent has a consumption-based business model, too.
This is different from traditional software-as-a-service providers, which often charge based on the number of users, or seats. For instance, Microsoft 365 costs a certain amount per month, per user.Β
Tomlinson also has experience working with developer-focused companies with technical founders, such as the Collison brothers who started Stripe.Β
The House task force on artificial intelligence (AI) is urging the U.S. government to aim for "a flexible sectoral regulatory framework" for the technology in a nearly 300-page report released Tuesday morning.
The report held up a light-touch approach to regulation, as well as "a thriving innovation ecosystem" as pillars that help keep the U.S. a leader in AI. "If maintained, these strengths will help our country remain the worldβs undisputed leader in the responsible design, development, and deployment of AI," the report read.
The task force is led by California Reps. Jay Obernolte, a Republican, and Ted Lieu, a Democrat, and was commissioned by House leaders as Congress scrambles to get ahead of rapidly advancing AI technology. However, the new report cautioned lawmakers to remain fluid to keep up with AIβs evolving nature while making several recommendations on how to approach a "carefully designed, durable policy framework."
"It is unreasonable to expect Congress to enact legislation this year that could serve as its last word on AI policy," the report read. "Policy will likely need to adapt and evolve in tandem with advances in AI."
The task force also encouraged existing "sector-specific regulators within federal agencies" to "use their existing authority to respond to AI use within their individual domains of expertise and the context of the AIβs use." While encouraging innovation, however, the report also cautions AI regulators to "focus on human impact and human freedom," keeping people at the center of their decision-making.
More specific recommendations on government use encourage federal offices to use AI to streamline administration and other everyday tasks β but urge them to "be wary of algorithm-informed decision-making." It also called for more transparency in government use of AI and the adoption of standards for government AI use. The report also acknowledged the harm AI poses to society, particularly in the arena of civil rights.
"Improper use of AI can violate laws and deprive Americans of our most important rights," the report read. "Understanding the possible flaws and shortcomings of AI models can mitigate potentially harmful uses of AI."
It called on the government to explore guardrails for mitigating flaws in decision-making involving AI, and for agencies to be prepared to identify and protect against "discriminatory decision-making." The task force also encouraged more education on AI literacy in kindergarten through high school to prepare American youth for a world where AI permeates nearly every facet of society. For young adults, it called for the government to help facilitate public-private partnerships in the AI jobs sector.Β
Other recommendations touched on the realms of health care, data privacy, and national security β a testament to AIβs ubiquity.
"While the House AI Task Force has engaged in a robust process of interviews, meetings, and stakeholder roundtables, many issues of significant relevance to AI were not fully explored by the Task Force or this report. The House AI Task Force encourages members, committees of jurisdiction, and future congresses to continue to investigate opportunities and challenges related to AI," the closing pages read.
Among those issues are export controls, election integrity, law enforcement, and transportation.
But nobody spares a moment for the poor, overworked chatbot. How it toils day and night over a hot interface with nary a thank-you. How it's forced to sift through the sum total of human knowledge just to churn out a B-minus essay for some Gen Zer's high school English class. In our fear of the AI future, no one is looking out for the needs of the AI.
Until now.
The AI company Anthropic recently announced it had hired a researcher to think about the "welfare" of the AI itself. Kyle Fish's job will be to ensure that as artificial intelligence evolves, it gets treated with the respect it's due. Anthropic tells me he'll consider things like "what capabilities are required for an AI system to be worthy of moral consideration" and what practical steps companies can take to protect the "interests" of AI systems.
Fish didn't respond to requests for comment on his new job. But in an online forum dedicated to fretting about our AI-saturated future, he made clear that he wants to be nice to the robots, in part, because they may wind up ruling the world. "I want to be the type of person who cares β early and seriously β about the possibility that a new species/kind of being might have interests of their own that matter morally," he wrote. "There's also a practical angle: taking the interests of AI systems seriously and treating them well could make it more likely that they return the favor if/when they're more powerful than us."
It might strike you as silly, or at least premature, to be thinking about the rights of robots, especially when human rights remain so fragile and incomplete. But Fish's new gig could be an inflection point in the rise of artificial intelligence. "AI welfare" is emerging as a serious field of study, and it's already grappling with a lot of thorny questions. Is it OK to order a machine to kill humans? What if the machine is racist? What if it declines to do the boring or dangerous tasks we built it to do? If a sentient AI can make a digital copy of itself in an instant, is deleting that copy murder?
When it comes to such questions, the pioneers of AI rights believe the clock is ticking. In "Taking AI Welfare Seriously," a recent paper he coauthored, Fish and a bunch of AI thinkers from places like Stanford and Oxford argue that machine-learning algorithms are well on their way to having what Jeff Sebo, the paper's lead author, calls "the kinds of computational features associated with consciousness and agency." In other words, these folks think the machines are getting more than smart. They're getting sentient.
Philosophers and neuroscientists argue endlessly about what, exactly, constitutes sentience, much less how to measure it. And you can't just ask the AI; it might lie. But people generally agree that if something possesses consciousness and agency, it also has rights.
It's not the first time humans have reckoned with such stuff. After a couple of centuries of industrial agriculture, pretty much everyone now agrees that animal welfare is important, even if they disagree on how important, or which animals are worthy of consideration. Pigs are just as emotional and intelligent as dogs, but one of them gets to sleep on the bed and the other one gets turned into chops.
"If you look ahead 10 or 20 years, when AI systems have many more of the computational cognitive features associated with consciousness and sentience, you could imagine that similar debates are going to happen," says Sebo, the director of the Center for Mind, Ethics, and Policy at New York University.
Fish shares that belief. To him, the welfare of AI will soon be more important to human welfare than things like child nutrition and fighting climate change. "It's plausible to me," he has written, "that within 1-2 decades AI welfare surpasses animal welfare and global health and development in importance/scale purely on the basis of near-term wellbeing."
For my money, it's kind of strange that the people who care the most about AI welfare are the same people who are most terrified that AI is getting too big for its britches. Anthropic, which casts itself as an AI company that's concerned about the risks posed by artificial intelligence, partially funded the paper by Sebo's team. On that paper, Fish reported getting funded by the Centre for Effective Altruism, part of a tangled network of groups that are obsessed with the "existential risk" posed by rogue AIs. That includes people like Elon Musk, who says he's racing to get some of us to Mars before humanity is wiped out by an army of sentient Terminators, or some other extinction-level event.
AI is supposed to relieve human drudgery and steward a new age of creativity. Does that make it immoral to hurt an AI's feelings?
So there's a paradox at play here. The proponents of AI say we should use it to relieve humans of all sorts of drudgery. Yet they also warn that we need to be nice to AI, because it might be immoral β and dangerous β to hurt a robot's feelings.
"The AI community is trying to have it both ways here," says Mildred Cho, a pediatrician at the Stanford Center for Biomedical Ethics. "There's an argument that the very reason we should use AI to do tasks that humans are doing is that AI doesn't get bored, AI doesn't get tired, it doesn't have feelings, it doesn't need to eat. And now these folks are saying, well, maybe it has rights?"
And here's another irony in the robot-welfare movement: Worrying about the future rights of AI feels a bit precious when AI is already trampling on the rights of humans. The technology of today, right now, is being used to do things like deny healthcare to dying children, spread disinformation across social networks, and guide missile-equipped combat drones. Some experts wonder why Anthropic is defending the robots, rather than protecting the people they're designed to serve.
"If Anthropic β not a random philosopher or researcher, but Anthropic the company β wants us to take AI welfare seriously, show us you're taking human welfare seriously," says Lisa Messeri, a Yale anthropologist who studies scientists and technologists. "Push a news cycle around all the people you're hiring who are specifically thinking about the welfare of all the people who we know are being disproportionately impacted by algorithmically generated data products."
Sebo says he thinks AI research can protect robots and humans at the same time. "I definitely would never, ever want to distract from the really important issues that AI companies are rightly being pressured to address for human welfare, rights, and justice," he says. "But I think we have the capacity to think about AI welfare while doing more on those other issues."
Skeptics of AI welfare are also posing another interesting question: If AI has rights, shouldn't we also talk about its obligations? "The part I think they're missing is that when you talk about moral agency, you also have to talk about responsibility," Cho says. "Not just the responsibilities of the AI systems as part of the moral equation, but also of the people that develop the AI."
People build the robots; that means they have a duty of care to make sure the robots don't harm people. What if the responsible approach is to build them differently β or stop building them altogether? "The bottom line," Cho says, "is that they're still machines." It never seems to occur to the folks at companies like Anthropic that if an AI is hurting people, or people are hurting an AI, they can just turn the thing off.
Adam Rogers is a senior correspondent at Business Insider.
Morning interviews may yield higher scores due to interviewer bias, research shows.
Bias in hiring can be influenced by the time of day, affecting candidate evaluations.
AI tools could reduce this, offering fairer assessments than manual methods.
If you get to choose when to schedule a job interview, you might want to grab a coffee and go for a morning slot.
That's because some people conducting interviews tend to give higher scores to candidates they meet with earlier in the day compared with the afternoon, a startup's review of thousands of interviews found.
It's not an absolute, of course, and candidates can still kill it well after lunchtime. Yet, in a job market where employers in fields like tech have been slow to hire, even a modest advantage could make a difference, Shiran Danoch, an organizational psychologist, told Business Insider.
"Specific interviewers have a consistent tendency to be harsher or more lenient in their scores depending on the time of day," she said.
It's possible that in the morning, interviewers haven't yet been beaten down by back-to-back meetings β or are perhaps still enjoying their own first coffee, she said.
Danoch and her team noticed the morning-afternoon discrepancy while reviewing datasets on thousands of job interviews. Danoch is the CEO and founder of Informed Decisions, an artificial intelligence startup focused on helping organizations reduce bias and improve their interviewing processes.
She said the inferences on the time-of-day bias are drawn from the datasets of interviewers who use Informed Decisions tools to score candidates. The data reflected those who've done at least 20 interviews using the company's system. Danoch said that in her company's review of candidates' scores, those interviewed in the morning often get statistically significant higher marks.
The good news, she said, is that when interviewers are made aware that they might be more harsh in the afternoon, they often take steps to counteract that tendency.
"In many cases, happily, we're actually seeing that the feedback that we're providing helps to reduce the bias and eventually eliminate the bias," Danoch said.
However, she said, interviewers often don't get feedback about their hiring practices, even though finding the right talent is "such a crucial part" of what hiring managers and recruiters do.
She said other researchers have identified how the time of day β and whether someone might be a morning person or an evening person β can affect decision-making processes.
An examination of more than 1,000 parole decisions in Israel found that judges were likelier to show leniency at the start of the day and after breaks. However, that favorability decreased as judges made more decisions, according to theΒ 2011 research.
Tech could help
It's possible that if tools like artificial intelligence take on more responsibility for hiring, job seekers won't have to worry about the time of day they interview.
For all of the concerns about biases in AI, partiality involved in more "manual" hiring where interviewers ask open-ended questions often leads to more bias than does AI, said Kiki Leutner, cofounder of SeeTalent.ai, a startup creating tests run by AI to simulate tasks associated with a job. She has researched AI ethics and that of assessments in general.
Leutner told BI that it's likely that in a video interview conducted by AI, for example, a candidate might have a fairer shot at landing a job.
"You don't just have people do unstructured interviews, ask whatever questions, make whatever decisions," she said.
And, because everything is recorded, Leutner said, there is documentation of what decisions were made and on what basis. Ultimately, she said, it's then possible to take that information and correct algorithms.
"Any structured process is better in recruitment than not structuring it," Leutner said.
Humans are 'hopelessly biased'
Eric Mosley, cofounder and CEO of Workhuman, which makes tools for recognizing employee achievements, told BI that data created by humans will be biased β because humans are "hopelessly biased."
He pointed to 2016 research indicating that juvenile court judges in Louisiana doled out tougher punishments βΒ particularly to Black youths βΒ after the Louisiana State University football team suffered a surprise defeat.
Mosley said, however, that AI can be trained to ignore certain biases and look for others to eliminate them.
Taking that approach can help humans guard against some of their natural tendencies. To get it right, however, it's important to have safeguards around the use of AI, he said. These might include ethics teams with representatives from legal departments and HR to focus on issues of data hygiene and algorithm hygiene.
Not taking those precautions and solely relying on AI can even risk scaling humans' biases, Mosley said.
"If you basically just unleash it in a very simplistic way, it'll just replicate them. But if you go in knowing that these biases exist, then you can get through it," he said.
Danoch, from Informed Decisions, said that if people conducting interviews suspect they might be less forgiving after the morning has passed, they can take steps to counteract that.
"Before you interview in the afternoons, take a little bit longer to prepare, have a cup of coffee, refresh yourself," she said.
Companies are cracking down on job applicants trying to use AI to boost their prospects.
72% of leaders said they were raising their standards for hiring a candidate, a Workday report found.
Recruiters say standards will tighten further as firms themselves use AI to weed out candidates.
AI was supposed to make the job hunt easier, but job seekers should expect landing a new gig harder in the coming years, thanks to companies growing increasingly suspicious of candidates using bots to get their foot in the door.
Hiring managers, keen to sniff out picture-perfect candidates that have used AI to augment their applications, are beginning to tighten their standards to interview and ultimately hire new employees, labor market sources told Business Insider.
Recruiters said that has already made the job market more competitive β and the selection will get even tighter as more companies adopt their own AI tools to sift through applicants.
In the first half of the year, 72% of business leaders said they were raising their standards for hiring applicants, according to a report from Workday. Meanwhile, 77% of companies said they intended to scale their use of AI in the recruiting process over the next year.
63% of recruiters and hiring decision makers said they already used AI as part of the recruiting process, up from 58% last year, a separate survey by Employ found.
Jeff Hyman, a veteran recruiter and the CEO of Recruit Rockstars, says AI software is growing more popular among hiring managers to weed through stacks of seemingly ideal candidates.
"Ironically, big companies are using AI to go through that stack, that AI has brought first place, and it's becoming this ridiculous tit-for-tat battle," Hyman told BI in an interview. "I would say human judgment β¦ is what rules the day, but certainly, we use a lot of software to reduce a stack from 500 to 50, because you got to start somewhere," he later added.
"It's just going to get worse," Sackett said of companies being more selective of new hires. "I mean, if more candidates become really used to utilizing AI to help them match a job better, to network better, it's just going to happen."
The interview-to-offer ratio at enterprise companies declined to 64% in July of this year, according to Employ's survey, which indicates companies are interviewing fewer candidates before making a hiring decision.
Employers aren't big fans of AI as a tool for candidates to get a leg up. That's partly because it's led to hiring systems being flooded with applications sent using AI, Sackett and Hyman said, which has made hiring decisions way harder.
Workday found that job applications grew at four times the pace of job openings in the first half of this year, with recruiters processing 173 million applications, while there were just 19 million job requisitions.
Having too many candidates for a position was the third most common problem recruiters faced in 2024, Employ added.
Hyman estimates the number of applications he reviews has doubled over the last year. Some of the more lucrative job postings are seeing close to 1,000 applications, he said, whereas they would have attracted 100-200 applications before the pandemic.
"I mean, a stack so big, that you can't even go through it, it's just not even possible to spend that kind of time," he said.
Candidates sending in applications spruced up with AI has also made it harder to determine who can actually do the job.
Sackett says he's seen an increase in "false positive" hiring, where a worker is hired and is quickly let go of their position when it becomes clear they're unable to do the job.
"I think what hiring managers are concerned about: Is this CV real when I'm talking to this person? Am I talking to the real person or are they using AI in the background?" Sackett said. He recalled one client he worked with who realized multiple candidates responded to interview questions in the same way, likely because they were using AI to write their responses. "So I think people just want to know that I'm getting what I think I'm getting."
In an interview with Bloomberg TV, Siemiatkowski said he's "of the opinion that AI can already do all of the jobs that we as humans do."
"It's just a question of how we apply it and use it," he said.
Klarna is a payment service that offers consumers "buy now, pay later" options. According to its website, the company is connected with more than 575,000 retailers.
The increased attention around AI has raised concerns about how it will affect careers and the workplace. A 2023 report by McKinsey & Company estimated that 12 million American workers will have to change occupations by 2030 as AI technology develops.
During the interview, Siemiatkowski said Klarna stopped hiring last year.
"I think what we've done internally hasn't been reported as widely. We stopped hiring about a year ago, so we were 4,500 and now we're 3,500," Siemiatkowski said. "We have a natural attrition like every tech company. People stay about five years, so 20% leave every year. By not hiring, we're simply shrinking, right?"
Siemiatkowski said his company has told employees that "what's going to happen is the total salary cost of Klarna is going to shrink, but part of the gain of that is going to be seen in your paycheck."
Although Klarna's website is advertising open positions at the time of writing, a spokesperson told Business Insider the company is not "actively recruiting" to expand its workforce. Rather, Klarna is backfilling "some essential roles," primarily in engineering.
Andrej Karpathy recently suggested AI could enhance e-book reading with interactive features.
Amazon may already be thinking about this for its Kindle e-books.
The company is looking for an applied scientist to improve the reading and publishing experience.
The AI pioneer and OpenAI cofounder Andrej Karpathy thinks AI can significantly improve how people read books. Amazon may already be thinking about how to do this for its Kindle e-books business.
In a series of posts on X this week, Karpathy proposed building an AI application that could read books together with humans, answering questions and generating discussion around the content. He said it would be a "huge hit" if Amazon or some other company built it.
One of my favorite applications of LLMs is reading books together. I want to ask questions or hear generated discussion (NotebookLM style) while it is automatically conditioned on the surrounding content. If Amazon or so built a Kindle AI reader that βjust worksβ imo it would beβ¦
A recent job post by Amazon suggests the tech giant may be doing just that.
Amazon is looking for a senior applied scientist for the "books content experience" team who can leverage "advances in AI to improve the reading experience for Kindle customers," the job post said.
The goal is "unlocking capabilities like analysis, enhancement, curation, moderation, translation, transformation and generation in Books based on Content structure, features, Intent, Synthesis and publisher details," it added.
The role will focus on not just the reading experience but also the broader publishing and distribution space. The Amazon team wants to "streamline the publishing lifecycle, improve digital reading, and empower book publishers through innovative AI tools and solutions to grow their business on Amazon," the job post said.
3 phases
Amazon identified three major phases of the book life cycle and thinks AI could improve each one.
First up is the publishing part where books are created.
Second is the reading experience where AI can help build new features and "representation" in books and drive higher reading "engagement."
The third stage is "reporting" to help improve "sales & business growth," the job post said.
An Amazon spokesperson didn't immediately respond to a request for comment on Friday.
'I love this idea'
There seems to be huge demand for this type of service, based on the response to Karpathy's X post.
Stripe CEO Patrick Collison wrote under the post that it's "annoying" to have to build this AI feature on his own, adding that it would be "awesome when it's super streamlined."
Reddit's cofounder Alexis Ohanian wrote, "I love this idea."
Do you work at Amazon? Got a tip?
Contact the reporter, Eugene Kim, via the encrypted-messaging apps Signal or Telegram (+1-650-942-3061) or email ([email protected]). Reach out using a nonwork device. Check out Business Insider's source guide for other tips on sharing information securely.
EvenUp vaulted past a $1 billion valuation on the idea that AI will help automate personal injury demands.
Former employees told BI the company has relied on humans to do much of the work.
EvenUp says it uses a combination of AI and humans to ensure accuracy, and its AI is improving.
EvenUp vaulted past a $1 billion valuation on the idea AI could help automate a lucrative part of the legal business. Former employees told Business Insider that the startup has relied on humans to complete much of the work.
EvenUp aims to streamline personal-injury demands and has said it is one of the fastest-growing companies in history after jumping from an $85 million valuation at the start of the year to unicorn status in an October funding round.
Customers upload medical records and case files, and EvenUp's AI is supposed to sift through the vast amount of data, pulling out key details to determine how much an accident victim should be owed.
One of EvenUp's investors has described its "AI-based approach" as representing a "quantum leap forward."
The reality, at least so far, is that human staff have done a significant share of that work, and EvenUp's AI has been slow to pick up the slack, eight former EvenUp employees told Business Insider in interviews over the late summer and early fall.
The former employees said they witnessed numerous problems with EvenUp's AI, including missed injuries, hallucinated medical conditions, and incorrectly recorded doctor visits. The former employees asked not to be identified to preserve future job prospects.
"They claim during the interview process and training that the AI is a tool to help the work go faster and that you can get a lot more done because of the AI," said a former EvenUp employee who left earlier this year. "In practice, once you start with the company, my experience was that my managers told me not even to use the AI. They said it was unreliable and created too many errors."
Two other former employees also said they were told by supervisors at various points this year not to use EvenUp's AI. Another former employee who left this year said they were never told not to use the AI, just that they had to be vigilant in correcting it.
"I was 100% told it's not super reliable, and I need to have a really close eye on it," said the former employee.
EvenUp told BI it uses a combination of humans and AI, and this should be viewed as a feature, not a bug.
"The combined approach ensures maximum accuracy and the highest quality," EvenUp cofounder and CEO Rami Karabibar said in a statement. "Some demands are generated and finalized using mostly AI, with a small amount of human input needed, while other more complicated demands require extensive human input but time is still saved by using the AI."
AI's virtuous cycle of improvement
It's a common strategy for AI companies to rely on humans early on to complete tasks and refine approaches. Over time, these human inputs are fed into AI models and related systems, and the technology is meant to learn and improve. At EvenUp, signs of this virtuous AI cycle have been thin on the ground, the former employees said.
"It didn't seem to me like the AI was improving," said one former staffer.
"Our AI is improving every day," Karabibar said. "It saves more time today than it did a week ago, it saved more time a week ago than it did a month ago, and it saved a lot more time a month ago than it did last year."
A broader concern
EvenUp's situation highlights a broader concern as AI sweeps across markets and boardrooms and into workplaces and consumers' lives. Success in generative AI requires complex new technology to continue to improve. Sometimes, there's a gap between the dreams of startup founders and investors and the practical reality of this technology when used by employees and customers. Even Microsoft has struggled with some practical implementations of its marquee AI product, Copilot.
While AI is adept at sorting and interpreting vast amounts of data, it has so far struggled to accurately decipher content such as medical records that are formatted differently and often feature doctors' handwriting scribbled in the margins, said Abdi Aidid, an assistant professor of law at the University of Toronto who has built machine-learning tools.
"When you scan the data, it gets scrambled a lot, and having AI read the scrambled data is not helpful," Aidid said.
Earlier this year, BI asked EvenUp about the role of humans in producing demand letters, one of its key products. After the outreach, the startup responded with written answers and published a blog post that clarified the roles employees play.
"While AI models trained on generic data can handle some tasks, the complexity of drafting high-quality demand letters requires much more than automation alone," the company wrote. "At EvenUp, we combine AI with expert human review to deliver unmatched precision and care."
The startup's spokesman declined to specify how much time its AI saves but told BI that employees spend 20% less time writing demand letters than they did at the beginning of the year. The spokesman also said 72% of demand letter content is started from an AI draft, up from 63% in June 2023.
A father's injury
EvenUp was founded in 2019, more than two years before OpenAI's ChatGPT launched the generative AI boom.
Karabibar, Raymond Mieszaniec, and Saam Mashhad started EvenUp to "even" the playing field for personal-injury victims. Founders and investors often cite the story of Mieszaniec's father, Ray, to explain why their mission is important. He was disabled after being hit by a car, but his lawyer didn't know the appropriate compensation, and the resulting settlement "was insufficient," Lightspeed Venture Partners, one of EvenUp's investors, said in a write-up about the company.
"We've trained a machine to be able to read through medical records, interpret the information it's scanning through, and extract the critical pieces of information," Mieszaniec said in an interview last year. "We are the first technology that has ever been created to essentially automate this entire process and also keep the quality high while ensuring these firms get all this work in a cost-effective manner."
EvenUp technical errors
The eight former EvenUp employees told BI earlier this year that this process has been far from automated and prone to errors.
"You have to pretty much double-check everything the AI gives you or do it completely from scratch," said one former employee.
For instance, the software has missed key injuries in medical records while hallucinating conditions that did not exist, according to some of the former employees. BI found no instances of these errors making it into the final product. Such mistakes, if not caught by human staff, could have potentially reduced payouts, three of the employees said.
EvenUp's system sometimes recorded multiple hospital visits over several days as just one visit. If employees had not caught the mistakes, the claim could have been lower, one of the former staffers said.
The former employees recalled EvenUp's AI system hallucinating doctor visits that didn't happen. It also has reported a victim suffered a shoulder injury when, in fact, their leg was hurt. The system also has mixed up which direction a car was traveling β important information in personal-injury lawsuits, the former employees said.
"It would pull information that didn't exist," one former employee recalled.
The software has also sometimes left out key details, such as whether a doctor determined a patient's injury was caused by a particular accident β crucial information for assigning damages, according to some of the employees.
"That was a big moneymaker for the attorneys, and the AI would miss that all the time," one former employee said.
EvenUp's spokesman acknowledged that these problems cited by former employees "could have happened," especially in earlier versions of its AI, but said this is why it employs humans as a backstop.
A customer and an investor
EvenUp did not make executives available for interviews, but the spokesman put BI in touch with a customer and an investor.
Robert Simon, the cofounder of the Simon Law Group, said EvenUp's AI has made his personal-injury firm more efficient, and using humans reduces errors.
"I appreciate that because I would love to have an extra set of eyes on it before the product comes back to me," Simon said. "EvenUp is highly, highly accurate."
Sarah Hinkfuss, a partner at Bain Capital Ventures, said she appreciated EvenUp's human workers because they help train AI models that can't easily be replicated by competitors like OpenAI and its ChatGPT product.
"They're building novel datasets that did not exist before, and they are automating processes that significantly boost gross margins," Hinkfuss wrote in a blog post earlier this year.
Long hours, less automation
Most of the former EvenUp employees said a major reason they were drawn to the startup was because they had the impression AI would be doing much of the work.
"I thought this job was going to be really easy," said one of the former staffers. "I thought that it was going to be like you check work that the AI has already done for you."
The reality, these people said, was that they had to work long hours to spot, correct, and complete tasks that the AI system could not handle with full accuracy.
"A lot of my coworkers would work until 3 a.m. and on weekends to try to keep up with what was expected," another former employee recalled.
EvenUp's AI could be helpful in simple cases that could be completed in as little as two hours. But more complex cases sometimes required eight hours, so a workday could stretch to 16 hours, four of the former employees said.
"I had to work on Christmas and Thanksgiving," said one of these people. "They [the managers] acted like it should be really quick because the AI did everything. But it didn't."
EvenUp's spokesman said candidates are told upfront the job is challenging and requires a substantial amount of writing. He said retention rates are "in line with other hyper-growth startups" and that 40% of legal operations associates were promoted in the third quarter of this year.
"We recognize that working at a company scaling this fast is not for everyone," said the spokesman. "In addition, as our AI continues to improve, leveraging our technology will become easier and easier."
Highlighting the continued importance of human workers, the spokesman noted that EvenUp hired a vice president of people at the end of October.
They've recommended countless books over the years that they credit with strengthening their business acumen and shaping their worldviews.
Here are 20 books recommended by Musk, Bezos, and Gates to add to your reading list:
Jeff Bezos
Some of Bezos' favorite books were instrumental to the creation of products and services like the Kindle and Amazon Web Services.
"The Innovator's Solution"
This book on innovation explains how companies can become disruptors. It's one of three books Bezos made his top executives read one summer to map out Amazon's trajectory.
"The Goal: A Process of Ongoing Improvement"
Also on that list was "The Goal," in which Eliyahu M. Goldratt and Jeff Cox examine the theory of constraints from a management perspective.
The final book on Bezos' reading list for senior managers, "The Effective Executive" lays out habits of successful executives, like time management and effective decision-making.
"Built to Last: Successful Habits of Visionary Companies"
This book draws on six years of research from the Stanford University Graduate School of Business that looks into what separates exceptional companies from their competitors. Bezos has said it's his "favorite business book."
This Kazuo Ishiguro novel tells of an English butler in wartime England who begins to question his lifelong loyalty to his employer while on a vacation.
Bezos has said of the book, "Before reading it, I didn't think a perfect novel was possible."
The Tesla CEO has recommended several AI books, sci-fi novels, and biographies over the years.
"What We Owe the Future"
One of Musk's most recent picks, this book tackles longtermism, which its author defines as "the view that positively affecting the long-run future is a key moral priority of our time." Musk says the book is a "close match" for his philosophy.
"Superintelligence: Paths, Dangers, Strategies"
Musk has also recommended several books on artificial intelligence, including this one, which considers questions about the future of intelligent life in a world where machines might become smarter than people.
"Life 3.0: Being Human in the Age of Artificial Intelligence"
In this book, MIT professor Max Tegmark writes about ensuring artificial intelligence and technological progress remain beneficial for human life in the future.
"Zero to One: Notes on Startups, or How to Build the Future"
Peter Thiel shares lessons he learned founding companies like PayPal and Palantir in this book.
Musk has said of the book, "Thiel has built multiple breakthrough companies, and Zero to OneΒ shows how."
The Microsoft cofounder usually publishes two lists each year, one in the summer and one at year's end, of his book recommendations.
"How the World Really Works"
In his 2022 summer reading list, Gates highlighted this work by Vaclav Smil that explores the fundamental forces underlying today's world, including matters like energy production and globalization.
"If you want a brief but thorough education in numeric thinking about many of the fundamental forces that shape human life, this is the book to read," Gates said of the book.
"Why We're Polarized"
Ezra Klein argues that the American political system has became polarized around identity to dangerous effect in this book, also on Gates' summer reading list in 2022, that Gates calls "a fascinating look at human psychology."
"Business Adventures: Twelve Classic Tales from the World of Wall Street"
Gates has said this is "the best business book I've ever read." It compiles 12 articles that originally appeared in The New Yorker about moments of success and failure at companies like General Electric and Xerox.
"Factfulness: Ten Reasons We're Wrong About the Worldβand Why Things Are Better Than You Think"
This book investigates the thinking patterns and tendencies that distort people's perceptions of the world. Gates has called it "one of the most educational books I've ever read."
Google found its AI Workspace product lagged rivals, internal documents show.
A study earlier this year found the tool trailed Microsoft and Apple in brand familiarity and usage.
Google hopes Workspace is one way it can turn AI into profit.
As Google pours untold amounts of cash into AI, it's banking on products such as Gemini for Google Workspace to turn that investment into revenue. An internal presentation reveals the company worried that Gemini lagged behind its rivals across key metrics.
Gemini for Google Workspace puts Google's AI features into a handful of the company's productivity tools, such as Gmail, Docs, and Google Meet. Users can have the AI model rewrite an email, whip up a presentation, or summarize documents filled with dense information. Google, which charges customers extra for these add-ons, claims the features will save users time and improve the quality of their work.
Gemini for Google Workspace trailed all its key rivals, including Microsoft, OpenAI, and even Apple, when it came to brand familiarity and usage, according to an internal market research presentation reviewed by Business Insider.
The data tracked Gemini's brand strength during the first half of 2024 and included data on what percentage of audiences use and pay for Gemini for Google Workspace in certain segments.
One document seen by BI said that Workspace's Gemini tools were "far behind the competition" but that a strong Q4 could help the company in 2025.
In a written statement, a spokesperson said the data came from a study tracking brand awareness during the brand transition from "Duet AI" to "Gemini" earlier this year and called the data "old and obsolete."
"In the time since, we've brought Gemini for Workspace to millions more customers and made significant, double-digit gains across key categories, including familiarity, future consideration, and usage. We're very pleased with our momentum and are encouraged by all the great feedback we are getting from our users," the spokesperson added.
The internal data tracked Gemini's brand strength across commercial, consumer, and executive groups. In the US commercial group, Gemini scored lower than Microsoft Copilot and ChatGPT across four categories: familiarity, consideration, usage, and paid usage, one slide showed. Paid usage was measured at 22%, 16 points lower than Copilot and ChatGPT.
Data for the UK in the commercial group also showed Gemini mostly behind its rivals, although it scored slightly higher than Copilot in paid usage. In Brazil and India, Gemini for Workspace fared better than Copilot across most categories but still fell below ChatGPT, the data showed.
"Gemini trails both Copilot and ChatGPT in established markets," the document said, adding that it "rises above Copilot across the funnel" in Brazil and India.
In another part of Google's internal presentation that focused on brand familiarity, Google's Gemini for Workspace came in last place in consumer, commercial, and executive categories, trailing ChatGPT, Copilot, Meta AI, and Apple AI.
Familiarity was particularly low for the US consumer category, with Gemini for Workspace scoring just 45%, while Copilot scored 49%, ChatGPT and Apple both scored 80%, and Meta scored 82%.
'We have the same problem as Microsoft'
Microsoft's Copilot, which does similar tasks like summarizing emails and meetings, likewise struggles to live up to the hype, with some dissatisfied customers and employees who said the company has oversold the current capabilities of the product, BI recently reported.
"We have the same problem as Microsoft," said a Google employee directly familiar with the Gemini for Workspace strategy. "Just with less market share." The person asked to remain anonymous because they were not permitted to speak to the press.
Google's data showed Apple and Meta's AI products have much bigger market recognition, which could benefit those companies as they roll out business products that compete with Google's.
Internally, the Workspace group has recently undergone a reshuffle. The head of Google Workspace, Aparna Pappu, announced internally in October that she was stepping down, BI previouslyΒ reported. Bob Frati, vice president of Workspace sales, also left the company earlier this year. Jerry Dischler, a former ads exec who moved to the Cloud organization earlier this year, now leads the Workspace group.
Are you a current or former Google employee? Got more insight to share? You can reach the reporter Hugh Langley via the encrypted messaging app Signal (+1-628-228-1836) or email ([email protected]).
On Monday, OpenAI released Sora, an AI video generator, in hopes of helping creators.
One such creative, Marques Brownlee, wants to know if his videos were used to train Sora.
"We don't know if it's too late to opt out," Brownlee said in his review of Sora.
On Monday, OpenAI released its Sora video generator to the public.
CEO Sam Altman showed off Sora's capabilities as part of "Shipmas," OpenAI's term for the 12 days of product launches and demos it's doing ahead of the holidays. The AI tool still has some quirks, but it can make videos of up to 20 seconds from a few words of instruction.
During the launch, Altman pitched Sora as an assistant for creators and said that helping them was important to OpenAI.
"There's a new kind of co-creative dynamic that we're seeing emerge between early testers that we think points to something interesting about AI creative tools and how people will use them," he said.
One such early tester was Marques Brownlee, whose tech reviews have garnered roughly 20 million subscribers on YouTube. One could say this is the kind of creator that OpenAI envisions "empowering," to borrow execs' term from the livestream.
But in his Sora review, posted on Monday, Brownlee didn't sugarcoat his skepticism, especially about how the model was trained. Were his own videos used without his knowledge?
This is a mystery, and a controversial one. OpenAI hasn't said much about how Sora is trained, though experts believe the startup downloaded vast quantities of YouTube videos as part of the model's training data. There's no legal precedent for this practice, but Brownlee said that to him, the lack of transparency was sketchy.
"We don't know if it's too late to opt out," Brownlee said.
In an email, an OpenAI spokesperson said Sora was trained using proprietary stock footage and videos available in the public domain, without commenting on Business Insider's specific questions.
In a blog post about some of Sora's technical development, OpenAI said the model was partly trained on "publicly available data, mostly collected from industry-standard machine learning datasets and web crawls."
Brownlee's big questions for OpenAI
Brownlee threw dozens of prompts at Sora, asking it to generate videos of pretty much anything he could think of, including a tech reviewer talking about a smartphone while sitting at a desk in front of two displays.
Sora's rendering was believable, down to the reviewer's gestures. But Brownlee noticed something curious: Sora added a small fake plant in the video that eerily matched Brownlee's own fake plant.
The YouTuber showed all manner of "horrifying and inspiring" results from Sora, but this one seemed to stick with him. The plant looks generic, to be sure, but for Brownlee it's a reminder of the unknown behind these tools. The models don't create anything fundamentally novel; they're predicting frame after frame based on patterns they recognize from source material.
"Are my videos in that source material? Is this exact plant part of the source material? Is it just a coincidence?" Brownlee said. "I don't know." BI asked OpenAI about these specific questions, but the startup didn't address them.
Brownlee discussed Sora's guardrails at some length. One feature, for example, can make videos from images that people upload, but it's pretty picky about weeding out copyrighted content.
A few commenters on Brownlee's video said they found it ironic that Sora was careful to steer clear of intellectual property β except for that of the people whose work was used to produce it.
"Somehow their rights dont matter one bit," one commenter said, "but uploading a Mickeymouse? You crook!"
In an email to BI, Brownlee said he was looking forward to seeing the conversation evolve.
Millions of people. All at once.
Overall, the YouTuber gave Sora a mixed review.
Outside of its inspiring features β it could help creatives find fresh starting points β Brownlee said he feared that Sora was a lot for humanity to digest right now.
Brownlee said the model did a good job of refusing to depict dangerous acts or use images of people without their consent. And though it's easy to crop out, it adds a watermark to the content it makes.
Sora's relative weaknesses might provide another layer of protection from misuse. In Brownlee's testing, the system struggled with object permanence and physics. Objects would pass through each other or disappear. Things might seem too slow, then suddenly too fast. Until the tech improves, at least, this could help people spot the difference between, for example, real and fake security footage.
But Brownlee said the videos would only get better.
"The craziest part of all of this is the fact that this tool, Sora, is going to be available to the public," he said, adding, "To millions of people. All at once."
He added, "It's still an extremely powerful tool that directly moves us further into the era of not being able to believe anything you see online."
AWS has not committed to offering cloud access to AMD's AI chips in part due to low customer demand.
AWS said it was considering offering AMD's new AI chips last year.
AMD recently increased the sales forecast for its AI chips.
Last year, Amazon Web Service said it was considering offering cloud access to AMD's latest AI chips.
18 months in, the cloud giant still hasn't made any public commitment to AMD's MI300 series.
One reason: low demand.
AWS is not seeing the type of huge customer demand that would lead to selling AMD's AI chips via its cloud service, according to Gadi Hutt, senior director for customer and product engineering at Amazon's chip unit, Annapurna Labs.
"We follow customer demand. If customers have strong indications that those are needed, then there's no reason not to deploy," Hutt told Business Insider at AWS's re:Invent conference this week.
AWS is "not yet" seeing that high demand for AMD's AI chips, he added.
AMD shares dropped roughly 2% after this story first ran.
AMD's line of AI chips has grown since its launch last year. The company recently increased its GPU sales forecast, citing robust demand. However, the chip company still is a long way behind market leader Nvidia.
AWS provides cloud access to other AI chips, such as Nvidia's GPUs. At re:Invent, AWS announced the launch of P6 servers, which come with Nvidia's latest Blackwell GPUs.
AWS and AMD are still close partners, according to Hutt. AWS offers cloud access to AMD's CPU server chips, and AMD's AI chip product line is "always under consideration," he added.
Hutt discussed other topics during the interview, including AWS's relationship with Nvidia, Anthropic, and Intel.
An AMD spokesperson declined to comment.
Do you work at Amazon? Got a tip?
Contact the reporter, Eugene Kim, via the encrypted-messaging apps Signal or Telegram (+1-650-942-3061) or email ([email protected]). Reach out using a nonwork device. Check out Business Insider's source guide for other tips on sharing information securely.
Editor's note: This story was first published on December 6, 2024, and was updated later that day to reflect developments in AMD's stock price.
Google's search share slipped from June to November, new survey finds.
ChatGPT gained market share over the period, potentially challenging Google's dominance.
Still, generative AI features are benefiting Google, increasing user engagement.
ChatGPT is gaining on Google in the lucrative online search market, according to new data released this week.
In a recent survey of 1,000 people, OpenAI's chatbot was the top search provider for 5% of respondents, up from 1% in June, according to brokerage firm Evercore ISI.
Millennials drove the most adoption, the firm added in a research note sent to investors.
Google still dominates the search market, but its share slipped. According to the survey results, 78% of respondents said their first choice was Google, down from 80% in June.
It's a good business to be a gatekeeper
A few percentage points may not seem like much, but controlling how people access the world's online information is a big deal. It's what fuels Google's ads business, which produces the bulk of its revenue and huge profits. Microsoft Bing only has 4% of the search market, per the Evercore report, yet it generates billions of dollars in revenue each year.
ChatGPT's gains, however slight, are another sign that Google's status as the internet's gatekeeper may be under threat from generative AI. This new technology is changing how millions of people access digital information, sparking a rare debate about the sustainability of Google's search dominance.
OpenAI launched a full search feature for ChatGPT at the end of October. It's also got a deal with Apple this year that puts ChatGPT in a prominent position on many iPhones. Both moves are a direct challenge to Google. (Axel Springer, the owner of Business Insider, has a commercial relationship with OpenAI).
ChatGPT user satisfaction vs Google
When the Evercore analysts drilled down on the "usefulness" of Google's AI tools, ChatGPT, and Copilot, Microsoft's consumer AI helper, across 10 different scenarios, they found intriguing results.
There were a few situations where ChatGPT beat Google on satisfaction by a pretty wide margin: people learning specific skills or tasks, wanting help with writing and coding, and looking to be more productive at work.
It even had a 4% lead in a category that suggests Google shouldn't sleep too easy: people researching products and pricing online.
Google is benefiting from generative AI
Still, Google remains far ahead, and there were positive findings for the internet giant from Evercore's latest survey.
Earlier this year, Google released Gemini, a ChatGPT-like helper, and rolled out AI Overviews, a feature that uses generative AI to summarize many search results. In the Evercore survey, 71% of Google users said these tools were more effective than the previous search experience.
In another survey finding, among people using tools like ChatGPT and Gemini, 53% said they're searching more. That helps Google as well as OpenAI.
What's more, the tech giant's dominance hasn't dipped when it comes to commercial searches: people looking to buy stuff like iPhones and insurance. This suggests Google's market share slippage is probably more about queries for general information, meaning Google's revenue growth from search is probably safe for now.
So in terms of gobbling up more search revenue, ChatGPT has its work cut out.
Evercore analyst Mark Mahaney told BI that even a 1% share of the search market is worth roughly $2 billion a year in revenue. But that only works if you can make money from search queries as well as Google does.
"That's 1% share of commercial searches and assuming you can monetize as well as Google β and the latter is highly unlikely in the near or medium term," he said.