Figure AI, backed by OpenAI, Launched Figure 02

Plus: New AI model listens and speaks simultaneously.

 

Welcome HumansšŸ¤–,

Here is what we have Today:

  • Superhumanoid Robot ā€” FIGURE 02 Launched

  • New AI model listens and Speaks simultaneously

  • 5 New Tech Job Opoortunities

FIGURE 02

Figure AI, backed by OpenAI, has unveiled its redesigned humanoid robot, the Figure 02, an AI-powered humanoid robot capable of autonomous work and engage in natural speech conversations using OpenAI's language models.

Source: Figure AI

How it looks like?

  • Stands 5'6" tall and weighs 132 lbs

  • Lifts up to 44 lbs

  • Operates for 20 hours on a single charge

  • Navigates using 360-degree vision from six RGB cameras

What makes Figure 02 Special?

  • OpenAI Language Models Integration

  • Vision language model with 6 RGB cameras to make sense of surrounding and make quick decisions to act.

  • Can self-correct errors based on visual input.

  • 50% more battery life

What can it do?

  • Engage in fluid, context-aware conversations.

  • Understand and respond to visual cues in its environment.

  • Potentially assist in various tasks requiring human-like interaction.

Source: Figure

Is this the robot apocalypse?

Not quite. But Figure's CEO claims it's the "most advanced humanoid on the planet," challenging Tesla's Optimus.

What's the secret sauce?

Figure's partnership with OpenAI gives it a potential edge in the humanoid robot race.

What could this mean for the future?

These robots are literally working ā€” BMW recently conducted the successful trail with the Robots in South Carolina Manufacturing Unit.

The humanoid race is heating up ;-

  • Amazon joined hands with Agility Robotics to test Digit for warehouse.

  • Mercedes-Benz tested Apptronik Apollo

  • Hyundai put Boston Dynamics` new robot at manufacturing unit.

  • Tesla Optimus are already getting ready for field test with public launch in 2026.

TODAY`s PICK
  • 2 Big Technologies needed for the completion process of the quantum computer are coming closer to being achieved.

  • Groq Inc., secures $640 million in funding led by BlackRock, Cisco and Samsung to disrupt Nvidia's stronghold in AI hardware. The Nvidia competitor now has a $2.8B valuation.

  • OpenAI has announced plans for its 2024 Dev Day that are coming to London and Singapore in addition to San Francisco this year but itā€™ll be on a smaller scale than last yearā€™s. As far as expectations go, donā€™t expect GPT-5.

  • ByteDance, the parent company of TikTok, just launched Jimeng AI for Chinese users, a text-to-video AI app that directly competes with OpenAIā€™s (unreleased) Sora AI video model.

  • Reddit is planning to test AI-powered search result pages to ā€˜summarize and recommend contentā€ later this year using a combination of first-party and third-party technology to power the feature.

  • Elon Muskā€™s X is being taken to court in Ireland for using Europeansā€™ data to train AI models. The development relates to the Xā€™s decision to process user data to train its ā€œGrokā€ AI model without notifying or asking people if theyā€™re okay with that.

Recommended Reading
If WE had to recommend other newsletters

AI Tool Report is an Newsletter, we enjoy reading daily. AI Tool Report delivers top techniques on how AI can transform your business filled with practical tips, real-world examples.

A Smart Bear is for people who love thinking about strategy, startups, product, marketing, decision-making, and founder psychology.

Job Opportunities
  • Shield AI - RF Engineer - Apply

  • Palantir Technologies - Software Engineer - Defence Tech, Internship - Apply

  • Palantir Technologies - Deployment Strategist - Apply

  • Databricks - Manager, IT Support - Apply

  • Meta - AI Research Scientist - VLLM, Generative AI - Apply

AI RESEARCH

Researchers have developed a Listening-While-Speaking Language Model (LSLM) that can listen and speak simultaneously, advancing real-time, interactive AI conversations.

How does it work?

  • Uses a token-based decoder-only TTS for speech generation

  • Employs a streaming self-supervised learning encoder for real-time audio input

  • Enables full-duplex modeling in interactive speech-language models

What can it do that others can't?

  • Detects turn-taking in real-time

  • Responds to interruptions, mimicking natural conversation

  • Demonstrates robustness to noise

  • Shows sensitivity to diverse instructions

Is it better than existing voice AI?

While OpenAI's voice mode for ChatGPT is advancing AI conversations, LSLM takes it further by processing incoming speech while talking.

Why should we care?

  • Could revolutionize human-AI interactions

  • Makes conversations with machines feel more natural and responsive

  • Potential applications in customer service, virtual assistants, and more

What's next?

  • Further refinement and testing of the LSLM model

  • Potential integration into existing AI platforms

  • Exploration of new applications for this technology

This development could mark a significant leap forward in creating more human-like AI conversational interfaces, potentially transforming how we interact with machines in the future.

AI ART

Choose your favorite bike ride?

Ideas? Comments? Complaints?

We read your emails, comments and poll replies daily.

Get the most important AI, tech, and science news in a free daily email.
New Here? Subscribe!
Sponsorship Slots Open for August and Reach over 800+ active readers. (Now 40% off) šŸ¤Æ

What`d you think of today`s edition?

Login or Subscribe to participate in polls.

Until next time, Stay Informed!

Reply

or to participate.