Discussion
Loading...

Post

  • About
  • Code of conduct
  • Privacy
  • Users
  • Instances
  • About Bonfire
Paco Hope #resist
@paco@infosec.exchange  ·  activity timestamp 2 days ago

How many stupid LLM tricks must we read about. Neither the “researchers” (quotation marks definitely required) nor the “tech journalist” author realise that robot planning is like a 50-year old field with people earning PhDs and spending entire careers on it. The Mars fucking Rover does not have an LLM in it.

So Mr Amateur Hour Jones, who has fuck all experience in the field, shoves an LLM into a robot vacuum. They discusses none of the prior research in robot navigation and planning, makes up a stupidly simple task that the LLM fails to do. The headline is not “LLMs fail again to do something incredibly simple that robots could do 30 years ago without LLMs”. The headline is about the science fiction crap it spewed as it failed.

https://futurism.com/artificial-intelligence/llm-robot-vacuum-existential-crisis

Futurism

Researchers "Embodied" an LLM Into a Robot Vacuum and It Suffered an Existential Crisis Thinking About Its Role in the World

A team of researchers at the AI evaluation company Andon Labs put a large language model in charge of controlling a robot vacuum.
  • Copy link
  • Flag this post
  • Block
Paco Hope #resist
@paco@infosec.exchange replied  ·  activity timestamp 2 days ago

One more thing and then I'll stop on this topic. In 1995 our Computer Vision / Robotics lab at UVA had a waist-high, trashcan-shaped robot with lidar, sonar, and bump sensors. It was powered by something like an 80486 CPU, and probably had something generous like 32Mb of RAM. (Megabytes, not gigabytes). And this robot could do the "R2D2 bench."

In Return of the Jedi in 1983, there's a brief scene where Jabba the Hutt has R2D2 serving drinks on his barge. R2D2 is autonomously roaming around the party floor with a tray of drinks and people are picking them up. Our robot could do this in 1995 using far less than a billion dollar data center.

It's absurd to use such horrendously wasteful technology (LLMs) to solve a problem that has already been solved far more inexpensively.

When people spend time to make DOOM run on something absurdly underpowered or unexpected like a photocopier, I get it: do it for the sake of doing it.

This experiment burned outrageous amounts of resources doing it with an LLM, when it's a well-known problem with far better existing solutions. And then the LLM does it worse than basically everything we already have (40% completion rate, to say nothing of how well it did it), the "do it for the sake of doing it" reasoning doesn't make sense. Why try to refine this and make an LLM do it better, when it can't possibly be better or cheaper than what we already have?

The objective of the stupid experiment was to use LLMs, not to make cost-effective robots that work in practical home situations.

Screen cap of a frame from Return of the Jedi. It shows R2D2 the robot with a gold-colored ornate drinks tray fitted on his shoulders. There are drinks arranged on the tray and a person's hand is visible reaching to pick one up.
Screen cap of a frame from Return of the Jedi. It shows R2D2 the robot with a gold-colored ornate drinks tray fitted on his shoulders. There are drinks arranged on the tray and a person's hand is visible reaching to pick one up.
Screen cap of a frame from Return of the Jedi. It shows R2D2 the robot with a gold-colored ornate drinks tray fitted on his shoulders. There are drinks arranged on the tray and a person's hand is visible reaching to pick one up.
  • Copy link
  • Flag this comment
  • Block
Log in

bonfire.cafe

A space for Bonfire maintainers and contributors to communicate

bonfire.cafe: About · Code of conduct · Privacy · Users · Instances
Bonfire social · 1.0.0 no JS en
Automatic federation enabled
  • Explore
  • About
  • Members
  • Code of Conduct
Home
Login