cover of episode Humanoid Robot Startups Are Hot. This AI Expert Cuts Through the Hype.

Humanoid Robot Startups Are Hot. This AI Expert Cuts Through the Hype.

2025/2/28
logo of podcast WSJ’s The Future of Everything

WSJ’s The Future of Everything

AI Deep Dive AI Chapters Transcript
People
A
Ayanna Howard
Topics
@Ayanna Howard : 我认为人形机器人是科技界和风投的下一个热门投资领域。这对于非机器人领域的人来说似乎是人工智能和生成式AI的下一步:赋予生成式AI以物理形态。因为生成式AI是关于人类知识的,所以将人类知识置于人体中必然是人形机器人。这对于非机器人专家来说是一种自然的逻辑发展。 当前人形机器人的能力有限,主要局限于特定环境和任务。它们在有限的环境、有限的任务中,在没有人类的情况下可以非常有效。但这并不能给公司带来数十亿美元的估值。我认为它们处于大约10年前自动驾驶汽车所处的阶段,当时初创公司层出不穷,宣称自动驾驶汽车将在未来24个月内普及。 我认为最大的问题是杀手级应用是什么。我认为仓库和医疗保健是人形机器人最具潜力的应用领域。在仓库中,它们可以进行打包、分拣、装箱和运输等工作,提高效率。在医疗保健领域,它们可以帮助老年人在家养老,解决护理人员短缺的问题。 新的AI软件架构将为人形机器人提供更多能力,但仍存在局限性。例如,在机器人领域,信息延迟可能导致机器人摔倒或伤害他人,这一问题尚未解决。 关于AI偏见,我认为它涉及多种类型,可能对美国社会产生负面影响。如果设计得当,AI可以比人类更少偏见,尤其是在医疗领域。AI无法区分不同人的现实,这可能导致错误的输出被误认为是事实。尽管AI存在错误,但人们的行为表明他们仍然愿意接受AI技术。 美国在AI领域的竞争力受到计算资源不足的限制。缺乏足够的资源让学生和工程师们进行实验和创新。虽然在资源匮乏的情况下,人们会更有创造力,但这并不意味着美国应该忽视对AI研发的投资。 我认为,当我们在多个场景中看到机器人并习以为常时,机器人时代才真正到来。

Deep Dive

Chapters
The recent surge in humanoid robot development is fueled by venture capital investment and the perception that it's the next logical step in AI, combining generative AI with a physical body. However, the technology is still in its early stages, comparable to the state of self-driving cars decades ago. Despite significant investment, the capabilities of these robots are currently limited.
  • Humanoid robots are seen as the logical next step in AI development.
  • Current technology is comparable to the early days of self-driving cars.
  • Significant investment has been made, but capabilities are still limited.

Shownotes Transcript

Translations:
中文

AI requires a lot of compute power, and the cost for your AI workloads can spiral. That is, unless you're running on OCI, Oracle Cloud Infrastructure. This was the cloud built for AI, a blazing fast enterprise-grade platform for your infrastructure, database, apps, and all of your AI workloads. Right now, Oracle can cut your current cloud bill in half if you move to OCI. Minimum financial commitment and other terms apply. Offer ends March 31st.

See if you qualify at oracle.com slash wallstreet. oracle.com slash wallstreet.

Ayanna Howard is the Dean at the Ohio State University College of Engineering, a former robotics researcher at NASA's Jet Propulsion Lab, a startup founder, and an advisor to our nation's highest officials on artificial intelligence and competitiveness. She has been working on humanoid robots and the software brains that power them for longer than almost anyone else in the U.S. These people-shaped bots could assist shorthanded warehouse crews by packing goods in delivery boxes,

They could help take care of the elderly or even do your dishes. But think more C-3PO than R2D2. Don't get technical with me. That makes her an expert at the intersection of two of the most hyped areas in tech right now.

an insider with a unique view of what's happening in these fields. Humanoid robots is the next shiny penny that's out there that tech folks and VCs, venture capitals can invest in. But I think it's also this logical, for those who aren't in robotics, it seems like the next logical step around artificial intelligence and generative AI is, oh, let's think about generative AI with a physical embodiment.

and the founder of at least one robotics company, Zyrobotics, which developed personalized education tech for children. But Howard is also an outsider to the current wave of startups.

and an academic who isn't obligated by her cap table to hype up whatever the latest thing is in these white-hot fields. If you're connected around an AI in a virtual environment and you don't have great bandwidth, you just wait a few seconds and your answer comes. In a robotics world, if you are losing a second of information, you have a robot that falls or hurts someone. We haven't really addressed that yet.

As Dean of Engineering, Howard oversees the education of 12,000 students. She's also a member of the National AI Advisory Committee, which gives the president policy recommendations for keeping American AI competitive. And she sits on the board of design software company Autodesk to name a few of the titles on her resume. All of which shows that Howard has the ears of some of the most powerful people in the U.S. making decisions about the future of AI.

I like to use healthcare because I think that's the holy grail. I'm more trusting that the AI can filter and provide a customized, personalized solution or decision than one person who is, you know, in say Atlanta. They're going to have a very different perspective than someone who's seeing patients in California or New York or Texas. I'm going to trust the AI a little bit more.

From the Wall Street Journal, I'm Tim Higgins. And I'm Christopher Mims. This is Bold Names, where you'll hear from the leaders of the bold-name companies featured in the pages of the Wall Street Journal. There are more than a dozen humanoid robot startups worldwide, and tech giants like Meta and NVIDIA are pouring money into the field. Today, we ask, what's the deal with all these companies, and how likely are they to succeed?

Ayanna Howard, thank you for joining us today. There's a lot that we could talk about, but I want to start with why you now and why you and I started talking in the first place is

which, like seemingly everything else on this show, comes back to Elon Musk. It must be humanoids. That's right. So suddenly humanoids are the hottest thing in tech other than AI. There are, by my count, more than a dozen startups worldwide that have racked up significant investment to work on these. Some of these startups, it's in the hundreds of millions of dollars.

So as in my estimation, the world expert on humanoid robots, who's not already employed by a humanoid robot company, I have to ask, why is this happening now? What is going on?

There's a couple of things. One is that humanoid robots is the next shiny penny that's out there that tech folks and VCs, venture capitals can invest in. But I think it's also this logical, for those who aren't in robotics, it seems like the next logical step

around artificial intelligence and generative AI is, "Oh, let's think about generative AI with the physical embodiment." If I'm thinking about the technology lineage, it's going to be humanoids because generative AI is about human knowledge. Human knowledge in a human body must be a humanoid. It's just a natural progression for non-roboticists to think that way.

Shiny Penny, there's clearly a lot of hype in this space right now, and probably no one's bigger at the hype than Tesla's own techno king, Elon Musk. And he's recently saying that he thinks his car company can build 10,000 humanoid robots by the year's end, a goal that...

Even he seems to concede they won't meet, but if they get to thousands, he'll be happy. And then there's a company called Figure, which is out recently talking about how it's going to deploy 100,000 robots in the next four years. So generally speaking...

Where is the technology currently? What are these kinds of humanoid robots capable of doing in reality in your estimation? If you think about the abilities of humanoids and what you would like them to do in theory, you like them to go from point A to point B. So navigation, they can do that pretty well, given that you can model the environment.

Once they get to another location, you want them to be able to manipulate. So move objects, pick up things, build things, and then go back from B to A or B to C, the last mile, as an example. Now what they're really- The last feet. The last feet, at least in our human environments. The reality though, is that just like with self-driving cars, is that they can be very, very, very effective

in limited situations, limited environments, limited tasks with humans not rampant around them, they can be positive. But that's not going to give you a billion dollar valuation of a company. It feels like these robots are where autonomous cars were, I want to say 10 years ago, but maybe it's more like 20 or 30 years ago. Is that a fair analogy? Yeah.

I would say that they are at the point when all of the investment incurred, which was, I would say that was about 10 years ago, where startups were just popping up. They were talking about self-driving cars are going to be rampant within the next 24 months, including well-established car companies. And, you know, as roboticists, we always like, do you remember that

that car that drove across America about 40 years ago, like,

That's where we are with humanoid robotics. We've seen humanoid robotics 40, 50 years ago, clunky, not well designed necessarily. And just like with self-driving cars, I always like to think about Hunger Games. There will be a winner at some point. We just don't know, one, who the winner is, and two, what is that application that is going to really set it apart from everyone else?

Well, you're talking about kind of the important part of the startup ecosystem. The hype brings the money, right? And then the money maybe will allow somebody to make it through the end of battle. I'm curious, you kind of alluded, these robots, right?

And humanoid robots are a lot like self-driving cars. Maybe driverless cars are robots in disguise, you know, more than meets the eyes. But I guess I wonder, 10 years ago when I was in a lot of these prototypes that were being used to raise money,

These cars, it turns out maybe, you know, doing a prototype is easy, easier going into production and actually putting and deploying them in a way more like fleet a lot harder. And I guess I wonder, as you look at some of these humanoid robots out there now.

Are you seeing kind of some similarities in the prototyping, the hype part of it versus what is actually deployable? There is a big gap. And when I think about deployability, I also think about the cost. If I had $100 billion deployed,

to build and deploy, and I was giving it for free, I didn't need a customer or client to purchase, then we have some opportunities. But that's not our reality. And so if you think about building a prototype that works in limited situation, pretty doable. Putting it now in anyone's home or in any warehouse, irrespective of where it is, that's almost impossible at this point. Just like with autonomous cars on safe streets,

where there's no other drivers, where there's no rain or snow.

Easy. As soon as you add in any dynamic, it becomes, oh, we need more data. We need more information. We need to think about how do we share information between cars, data, data, data, data, data collection. What's the business model, do you think? Are we going to have a Waymo of humanoid robots? Is this going to be a kind of a service that I pay for? Or am I going to have a robot that lives in my house and then walks my dog?

So I would say this is the biggest issue is what is the holy grail of an application? And I'm just going to give you my bias. If I think about where there are flows of need, one is in the warehouse. It's the packing, it's the sorting, it's the closing up the boxes, transporting the box to some other location, right?

That is really around the warehouse. That is one area. If you can figure that out, unfortunately, that also means loss of jobs. But you now have more efficiency. Ideally, that efficiency and lower costs will transfer to the consumer. But we, of course, know that that's not going to happen. But that's one area. The other, I truly believe, is with respect to health care.

One of the things that we know is that in most countries, including in the United States, there is not sufficient labor to address the older adults that want to live with their home.

For example, the baby boomers, once they retire, have some resources. There are not enough individuals that are going to be working that want to go into the home to do the caretaking that's necessary for them to stay in their home. So that's an area that, one, there's going to be a huge need.

Two, we don't have the personnel in terms of people to fill it. And three, there are resources in terms of actually paying for that if it's at a reasonable cost. I really think that that's one of the holy grail applications that these robotic platforms can social impact affordable and address a great need that is coming across the world. Yeah. And just a footnote, what you said recently, I was speaking with the

chief executive of Apptronic, which is an Austin, Texas based humanoid robot company, which spun out of a lab at the university of Texas at Austin. Like they've announced a deal with a GXO, which is a big logistics company, which does things like, um,

uses humanoid robots in a Spanx distribution warehouse in Atlanta, Georgia. But he said his long-term goal is he wants to help people age in place. He was inspired by, you know, seeing family members take care of his own grandfather. And he thinks that's where we can ultimately go. Of course, he also said that the unique enabler now is he's like, look, we're trying to build the best robot body. We've been waiting for the right

quote unquote, brain to put in it. And the enabler is artificial intelligence. It's these new transformer architectures. Is there a there there? I mean, is the new software actually going to enable more capabilities for these robots? It will enable more capability. And that's this element of the natural progression from

generative AI, large language models is this element of how do we use data in a more efficient way? How do we share data to associate human input, human desires to actions, motor actions? And what does this look like? I think that there is some possibility

positive trajectory in that space. And just so you know, when you look at the user transformer models in these applications, the videos are amazing, but it's still a really, really limited application. I will say one of the things that a lot of people aren't thinking about or considering is that, so it's one thing if you're connected around an AI in a virtual environment and you don't have great bandwidth, you just wait a few seconds and your answer comes.

In a robotics world, if you are losing a second of information, you have a robot that falls or hurts someone. And so I think that disconnect of do we have the infrastructure for doing this in orders of less than milliseconds, nanoseconds, and not losing any connectivity, we haven't really addressed that yet.

Ayanna Howard just described the capability gap between the AI models we have today and what robot bodies need to operate safely. But she's not just speaking as an academic. For the last few years, she's been a member of the National AI Advisory Committee and helped advise former President Joe Biden on the subject, especially on bias. But now, how might AI policy change under a second Trump administration?

When I think about bias, I think about it as all types of bias. It could be around ideology, it could be around gender, race. Bias is bias, and bias could be detrimental to Americans. And so I think when we think about extracting that one source of bias, the question is, is it an overarching term for all types of bias? I'm not quite sure.

Stay with us.

Explore the possibilities of AI with your data at explore.elastic.co. Elastic, the search AI company. Well, speaking of AI, you've been on the National AI Advisory Committee, I think, since 2022. Your term is supposed to run through April of this year, but...

The White House just took down the official page on that committee. Is it still operating? Have you heard anything? What's going on? We are still, until we are notified otherwise, still acting, still meeting. With the new administration, the old AI executive order was rescinded. There's a new one. I'll just give you a line from it. To maintain this leadership in AI...

we must develop AI systems that are free from ideological bias or engineered social agendas. And, you know, you've spoken with our colleagues before at the Journal about Bias and AI.

What do you make of that wording and what you interpret as the intent behind it? When I think about bias, I think about it as all types of bias. It could be around ideology. It could be around gender, race. Bias is bias, and bias could be detrimental to Americans. And so I think when we think about extracting that one source of bias—

The question is, is it an overarching term for all types of bias? Not quite sure. I'd have to actually read the entire piece to find out. But I think what I've seen, though, in AI in general is that it doesn't always reflect the different populations, our different mindsets. And it doesn't give us a choice.

to query it based on our own viewpoints and say, hey, I think that that framing is incorrect based on my lived experience. And question is, is this going to be a limited definition of bias or not? That I don't know. Yeah, you've said in the past that one of the unique things about AI

is that it can be built so that it's actually less biased than humans, which of course flies in the face of most of the coverage I read about AI, which is like bias means it's denying people healthcare or unjustly extending their prison sentences. But you seem like a glass half full person when it comes to AI bias. Could you expand on that? Yes. It's because when it's done right,

it can bring in all different viewpoints around any type of whatever it is. And I like to use healthcare because I think that's the holy grail. So imagine an AI system that has learned about correct treatments based on gender. The mass of information, but you think about all with this link of gender, socioeconomics, age, are you childbearing or not? Do you have kids, dot, dot, dot, dot, dot.

I'm more trusting that the AI can filter and provide a customized, personalized solution or decision than one person who is in, say, Atlanta. They're going to have a very different perspective than someone who's seeing patients in California or New York or Texas. I'm going to trust the AI a little bit more. There'll be less bias.

I think the idea or the discussion around bias in AI really has become more of a popular topic because we see these chat models who have had some interesting hallucinations. And, you know, I think the makers of these chat models are really aiming to be accurate, right? I mean, they don't want to be embarrassed, right?

But I think in the current political kind of climate, the real debate over kind of what is fact and the perception of fact is kind of hanging over some of this. And how does AI deal with that? I mean, when is a fact a fact and when is kind of it more than that? I would say this is a problem with our field of AI is because it's still a very human defined perspective of what is fact versus reality.

not fat, because it really is based on your lived reality. So I can say it's raining outside.

The question is, is it raining in Columbus? Is it raining in California? Like, no, it's raining outside. So that's my fact. But it may not be your fact if you're somewhere where it's nice and sunny and 90 degrees. AI is not able to distinguish between, you know, is this your reality or is this someone else's reality? And that's a problem. The other thing is, is that typical people tend to trust what AI is doing. The fact that people call it hallucinations.

What that basically means is at some point they realize that the AI has gone crazy, but they didn't realize it like the first incorrect output or the second incorrect output, which if you go back, you're like, oh yeah, that was wrong and that was wrong and that was wrong. It's only when it becomes entirely incorrect

Like, oh my gosh, this makes absolutely no sense. Do people then say, oh, that's a hallucination? But no, it was actually a mistake. And the mistake happened somewhere down the lineage of question and answers. And what happens is people tend to follow the thread of thought of the AI without questioning it until it violates their fundamental beliefs around whatever it is. That's a problem.

One of the things in driverless cars that's been out there, the companies behind it have been very concerned that if there's a crash and somebody was hurt or died, that this would turn people off to the technology more broadly. Whereas if

If I run into you, maybe you'd understand it because we're all human and we understand that sometimes car crashes occur, right? But I wonder, as we move in kind of this other field of AI, are these hallucinations or these kind of misstatements, is it going to turn people off? So unfortunately, it doesn't turn off too many people. And why I say that is, is that

What I say is not always what my actions dictate. I could say, oh, I'm angry at social media. You know, it's horrible, horrible. But are you going on social media and still posting and tweeting? Yes. Right. It's if I'm in a self-driving car and you hear all the news and so-and-so crashed, it's

I didn't see a drop in people going out and putting it in self-driving though. They may have talked about like, oh, this is really bad. And like, oh, this is really nice. This is very convenient. And so our actions dictate that we are okay with AI, even if our words are much more reactive in that space. Yeah. For the record, I don't fall into the groups that you just named because I don't have a self-driving car and I just quit social media for the

time in my life. Um,

While we're talking about the things that make people... Well, no, actually, this is interesting. So in our studies, we do these studies around trust and things like that. When we ask persons at initial, like, would you use it or would you not? It's actually about, depending on which demographic, it's about an 80 to 20 to a 60-40. So basically, you're in the 20% that you're never going to be convinced. I can be convinced. I can be convinced. But I'm a skeptic and a hater by nature.

But whether or not people trust AI models, or if the skeptics and haters can be flipped, the technology is developing at a rapid pace, especially after a Chinese company's model, DeepSeek, turned the AI industry on its head. Where does that leave American AI developers? If there had been resources or there had been, you know, data server farms for people just to play around with, you'd have had student AI

in their garage, finding out something, saying, oh, I tried this. And then a company's like, oh, that's great. Come work for us. But we didn't do that in this space. That's next. This podcast is brought to you by U.S. Cloud. Tired of slow response times and increasing costs from Microsoft Unified Support? Switch to the enterprise choice for up to 50% less than Unified. U.S. Cloud, faster Microsoft support for less. Learn more at uscloud.com.

Obviously, not too long ago, DeepSeek happened, took a trillion dollars of value off of

mostly U.S. tech stocks in a day. And there was something I thought was really prescient. You said in summer of 2024, you told the Senate Joint Economic Committee that you were concerned that if the U.S. didn't stay competitive, we might one day wake up to a Sputnik moment in AI. And here we are a half a year later. You have everyone from Senator Chuck Schumer to President

Mark Andreessen saying that DeepSeq was, quote unquote, a Sputnik moment in AI. Well, first off, are you Nostradamus? Do you feel like we're actually in a Sputnik moment? I feel like we are around it. Not quite at the level of Sputnik, but pretty close. But one of these things is...

It wasn't rocket science. This has always happened around technology. And the reason for that was that the academics, the ones that tinker around and try new ways, were being cut off from access because of the cost of computes. That was it. Whereas if there had been resources or there had been data server farms for people just to play around with, you'd have had the student

in their garage, finding out something, saying, oh, I tried this. And then a company's like, oh, that's great. Come work for us. But we didn't do that in this space. One of the storylines out of DeepSeek is this kind of idea that they didn't have the same kind of resources as an OpenAI or these other US tech companies. So

On one hand, I think I hear you saying that not enough in the U.S. is being spent to allow these up-and-coming engineers and coders to learn how to play with this stuff. On the other hand, there's an argument that is certainly becoming popular that DeepSeek was successful because it didn't have resources. It had to think more cleverly to kind of compete. How do you kind of square that?

So when you're hungry and have to think of alternative ways of doing things, you become more creative and more innovative. It's all relative, right? So even like they're saying it was less than training an open AI model, but there's not a lot of universities that could just devote five to $10 million. It just doesn't happen, especially for public institutions. So one, it's all relative.

But one of the reasons why in the U.S. startups do so well is because they start off, they don't have a ton of resources, just enough to, say, get a prototype that then allows them to get a CD or some other fund. It allows them to be really, really hungry. That's just the way innovation sometimes works. I'd love to end on a note of creativity.

kind of cautious optimism. Cause I think that's your forte. Um, what is going to be that Waymo moment that I have when I visit San Francisco and I'm like, there goes a car with no driver. Like when can we expect to have that moment? Um,

And might be different if you work in an Amazon warehouse, for example. When is C-3PO real? Yeah. Or R2D2. Yeah. So I will say when you go into an airport, maybe it's the hospital, maybe it's City Hall.

And you go there and not only is the receptionist a robot, but it's also going to be everywhere that you go. You go to the cafeteria and there is another robot. You go to the restroom and there's a robot that's cleaning the floors. Maybe not humanoid, but cleaning the floors. When you see more than one robot and we become accustomed to it, that's when we know robots have arrived.

So it sounds like 2030. 2030. Five years? I can see it in some spaces, yes. Awesome. Well, Dr. Howard, it's such a pleasure to speak with you. I just really value your insights as both an insider and an outsider. And, you know, it's been a pleasure. Thank you. And that's bold names for this week. Michael LaValle and Jessica Fenton are our sound designers. Jessica also wrote our theme music.

Our producer is Danny Lewis. We got help this week from Catherine Millsop, Scott Salloway, and Falana Patterson. For even more, check out our columns on WSJ.com. I'm Christopher Mims. And I'm Tim Higgins. Thanks for listening.

Robert Half Research indicates 9 out of 10 hiring managers are having difficulty hiring. Robert Half is here to help. Our recruiting professionals utilize our proprietary AI to connect businesses with highly skilled talent. At Robert Half, we know talent. Visit roberthalf.com today.