Connect with us

Business

Facebook: Teaching AI to View the World Through Your Eyes

Today, we’re announcing Ego4D, a long-term project by Facebook AI that aims to solve research challenges around egocentric perception: the ability for AI to understand and interact with the world like we do, from a first-person perspective. AI typically learns from photos and videos captured in third-person, but next-generation AI will need to learn from…

Published

on

Today, we’re announcing Ego4D, a long-term project by Facebook AI that aims to solve research challenges around egocentric perception: the ability for AI to understand and interact with the world like we do, from a first-person perspective. AI typically learns from photos and videos captured in third-person, but next-generation AI will need to learn from videos that show the world from the center of action. AI that understands the world from this point of view could unlock a new era of immersive experiences, as devices like augmented reality (AR) glasses and virtual reality (VR) headsets become as useful in everyday life as smartphones.

For this project, we brought together a consortium of 13 universities and labs across nine countries, who collected more than 2,200 hours of first-person video in the wild, featuring over 700 participants going about their daily lives. This greatly increases the amount of egocentric data publicly available to the research community, as this data set is 20X greater than any other in terms of hours of footage.

We also developed five benchmark challenges for developing smarter, more useful AI assistants including:

  • Episodic memory: What happened when? (e.g., “Where did I leave my keys?”)
  • Forecasting: What am I likely to do next? (e.g., “Wait, you’ve already added salt to this recipe”)
  • Hand and object manipulation: What am I doing? (e.g., “Teach me how to play the drums”)
  • Audio-visual diarization: Who said what when? (e.g., “What was the main topic during class?”)
  • Social interaction: Who is interacting with whom? (e.g., “Help me better hear the person talking to me at this noisy restaurant”)

Learn more about our AI research and developments.

Source

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Business

Introducing CoreAI – Platform and Tools

Satya Nadella, Chairman and CEO, shared the below communication with Microsoft employees this morning. As we begin the new year, it’s clear that we’re entering the next innings of this AI platform shift. 2025 will be about model-forward applications that reshape all application categories. More so than any previous platform shift, every layer of the……

Published

on

By

Satya Nadella, Chairman and CEO, shared the below communication with Microsoft employees this morning. As we begin the new year, it’s clear that we’re entering the next innings of this AI platform shift. 2025 will be about model-forward applications that reshape all application categories. More so than any previous platform shift, every layer of the…

Source

Continue Reading

Business

Facebook: How Organizations Are Using Llama to Solve Industry Challenges

Organizations are using Llama to build tools that provide value to individuals, society and the economy, and saving time and money in the process. Source

Published

on

By

Organizations are using Llama to build tools that provide value to individuals, society and the economy, and saving time and money in the process.

Source

Continue Reading

Business

Facebook: How Open Source AI is Evolving Healthcare

Open source AI models like Llama are available for free for organizations to use, modify and build on, making this critical technology more accessible than commercial models.  Source

Published

on

By

Open source AI models like Llama are available for free for organizations to use, modify and build on, making this critical technology more accessible than commercial models. 

Source

Continue Reading

Trending

Copyright © 2021 Today's Digital.