❌

Normal view

There are new articles available, click to refresh the page.
Before yesterdayMain stream

Here's how Nvidia CEO Jensen Huang won over his wife

24 November 2024 at 19:42
Nvidia CEO Jensen Huang.
Jensen Huang is known to write short emails

Mohd Rasfan/AFP/Getty Images

  • Nvidia CEO Jensen Huang met his wife Lori Huang at Oregon State University.
  • In a recent interview, he said that he tried to use homework as an excuse to spend time with her.
  • Huang said he promised her he'd be CEO by 30 to ensure she'd marry him.

When Jensen Huang met his wife in college, the odds weren't in his favor.

He was 17 years old, and she was 19. "I was the youngest kid in school, in class. There were 250 students and 3 girls," he said in an interview at Hong Kong University of Science and Technology last week after receiving an honorary degree. He was also the only student who "looked like a child," he said.

Huang used his youthful appearance to approach his future wife, hoping she'd assume he was smart. "I walked up to her and I said, 'do you want to see my homework?'" Then he made a deal with her. "If you do homework with me every Sunday, I promise you, you will get straight As."

From that point on, he said he had a date every Sunday. And just to ensure that she would eventually marry him, he told her that by 30, he'd be a CEO.

Huang married Lori Mills five years after they first met at Oregon State University, according to his biography on OSU College of Engineering's website. The couple has two children, Madison, a director of marketing at Nvidia, and Spencer, a senior product manager at the company.

After graduating from OSU in 1984, Huang worked at chip companies LSI Logic and Advanced Micro Devices, according to his bio on Nvidia's website. He then pursued a master's degree in electrical engineering at Stanford University in 1992, a year before he founded Nvidia, which has grown into a $3.48 trillion company thanks to the artificial intelligence boom.

Huang was 30 years old when he founded Nvidia.

The CEO often shares the lore about Nvidia's origin: He conceived the idea for a graphics company while dining at Denny's, a US diner chain, with his friends. Huang said in aΒ 2010 New York Times interviewΒ that he also waited tables atΒ Denny's while he was a student.

Huang's net worth is now estimated to be $124 billion.

The CEO also credits his wife and daughter with establishing his signature style: the black leather jacket.

In an interview last year on HP's online show, "The Moment," host Ryan Patel asked Huang how he feels to become a style icon.

"Now, at Denny's I'm sure you weren't thinking you were gonna be the style star of the future, but now you are," Patel said. "What do you think? How do you feel?"

"Don't give me that," Huang replied. "I'm happy that my wife and my daughter dresses me."

A spokesperson for Nvidia did not respond to a request for comment.

Read the original article on Business Insider

This is the biggest question in AI right now

24 November 2024 at 09:08
AI

Qi Yang/Getty Images

  • AI leaders are rethinking data-heavy training for large language models.
  • Traditional models scale linearly with data, but this approach may hit a dead end.
  • Smaller, more efficient models and new training methods are gaining industry support.

For years, tech companies like OpenAI, Meta, and Google have focused on amassing tons of data, assuming that more training material would lead to smarter, more powerful models.

Now, AI leaders are rethinking the conventional wisdom about how to train large language models.

The focus on training data arises from research showing that transformers, the neural networks behind large language models, have a one-to-one relationship with the amount of data they're given. Transformer models "scale quite linearly with the amount of data and compute they're given," Alex Voica, a consultant at the Mohamed bin Zayed University of Artificial Intelligence, previously told Business Insider.

However, executives are starting to worry that this approach can only go so far, and they're exploring alternatives for advancing the technology.

The money going into AI has largely hung on the idea that this scaling law "would hold," Scale AI CEO Alexandr Wang said at the Cerebral Valley conference this week, tech newsletter Command Line reported. It's now "the biggest question in the industry."

Some executives say the problem with the approach is that it's a little mindless. "It's definitely true that if you throw more compute at the model, if you make the model bigger, it'll get better," Aidan Gomez, the CEO of Cohere, said on the 20VC podcast. "It's kind of like it's the most trustworthy way to improve models. It's also the dumbest."

Gomez advocates smaller, more efficient models, which are gaining industry support for being cost-effective.

Others worry this approach won't reach artificial general intelligence β€” a theoretical form of AI that matches or surpasses human intelligence β€” even though many of the world's largest AI companies are banking on it.

Large language models are trained simply to "predict the next token, given the previous set of tokens," Richard Socher, a former Salesforce executive and CEO of AI-powered search engine You.com, told Business Insider. The more effective way to train them is to "force" these models to translate questions into computer code and generate an answer based on the output of that code, he said. This will reduce hallucinations in quantitative questions and enhance their abilities.

Not all industry leaders are sold that AI has hit a scaling wall, however.

"Despite what other people think, we're not at diminishing marginal returns on scale-up," Microsoft chief technology officer Kevin Scott said in July in an interview with Sequoia Capital's Training Data podcast.

Companies like OpenAI are also seeking to improve on existing LLMs.

OpenAI's o1, released in September, still relies on the token prediction mechanism Socher refers to. Still, the model is specialized to better handle quantitative questions, including areas like coding and mathematics β€” compared to ChatGPT, which is considered a more general-purpose model.

Part of the difference between o1 and ChatGPT is that o1 spends more time on inference or "thinking" before it answers a question.

"To summarize, if we were to anthropomorphize, gpt-4 is like your super know-it-all friend who when you ask them a question starts talking stream-of-consciousness, forcing you to sift through what they're saying for the gems," Waleed Kadous, a former engineer lead at Uber and former Google principal software engineer, wrote in a blog post. "o1 is more like the friend who listens carefully to what you have to say, scratches their chin for a few moments, and then shares a couple of sentences that hit the nail on the head."

One of o1's trade-offs, however, is that it requires much more computational power, making it slower and costlier, according toΒ Artificial Analysis, an independent AI benchmarking website.

Read the original article on Business Insider

❌
❌