The Times Australia
Google AI
The Times World News

.

Facebook wants AI to find your keys and understand your conversations

  • Written by Jumana Abu-Khalaf, Research Fellow in Computing and Security, Edith Cowan University
Facebook wants AI to find your keys and understand your conversations

Facebook has announced[1] a research project that aims to push the “frontier of first-person perception”, and in the process help you remember where your left your keys.

The Ego4D project[2] provides a huge collection of first-person video and related data, plus a set of challenges for researchers to teach computers to understand the data and gather useful information from it.

In September, the social media giant launched a line of “smart glasses” called Ray-Ban Stories[3], which carry a digital camera and other features. Much like the Google Glass[4] project, which met mixed reviews[5] in 2013, this one has prompted complaints of privacy invasion[6].

The Ego4D project aims to develop software that will make smart glasses far more useful, but may in the process enable far greater breaches of privacy.

Read more: Ray-Ban Stories let you wear Facebook on your face. But why would you want to?[7]

What is Ego4D?

Facebook describes the heart of the project as

a massive-scale, egocentric dataset and benchmark suite collected across 74 worldwide locations and nine countries, with over 3,025 hours of daily-life activity video.

Ego4D: Teaching AI to perceive the world through your eyes.

The “Ego” in Ego4D means egocentric (or “first-person” video), while “4D” stands for the three dimensions of space plus one more: time. In essence, Ego4D seeks to combine photos, video, geographical information and other data to build a model of the user’s world.

There are two components: a large dataset of first-person photos and videos, and a “benchmark suite” consisting of five challenging tasks that can be used to compare different AI models or algorithms with each other. These benchmarks involve analysing first-person video to remember past events, create diary entries, understand interactions with objects and people, and forecast future events.

The dataset includes more than 3,000 hours of first-person video from 855 participants going about everyday tasks, captured with a variety of devices including GoPro cameras and augmented reality (AR) glasses[8]. The videos cover activities at home, in the workplace, and hundreds of social settings.

What is in the data set?

Although this is not the first such video dataset to be introduced to the research community, it is 20 times larger than publicly available datasets. It includes video, audio, 3D mesh scans of the environment, eye gaze, stereo, and synchronized multi-camera views of the same event.

Most of the recorded footage is unscripted or “in the wild”. The data is also quite diverse as it was collected from 74 locations across nine countries, and those capturing the data have various backgrounds, ages and genders.

What can we do with it?

Commonly, computer vision models are trained and tested on annotated images and videos for a specific task. Facebook argues that current AI datasets and models represent a third-person or a “spectator” view, resulting in limited visual perception. Understanding first-person video will help design robots that better engage with their surroundings.

Future robotic agents will benefit from a better understanding of their environment
Wikimedia: Future robotic agents will benefit from a better understanding of their environment Furthermore, Facebook argues egocentric vision can potentially transform how we use virtual and augmented reality devices such as glasses and headsets. If we can develop AI models that understand the world from a first-person viewpoint, just like humans do, VR and AR devices may become as valuable as our smartphones. Can AI make our lives better? Facebook has also developed five benchmark challenges as part of the Ego4D project. The challenges aim to build better understanding of video materials to develop useful AI assistants. The benchmarks focus on understanding first person perception. The benchmarks are described as follows: Episodic memory (what happened when?): for example, figuring out from first-person video where you left your keys Hand-object manipulation (what am I doing and how?): this aims to better understand and teach human actions, such as giving instructions on how to play the drums Audio-visual conversation (who said what and when?): this includes keeping track of and summarising conversations, meetings or classes Social interactions (who is interacting with whom?): this is about identifying people and their actions, with a goal of doing things like helping you hear a person better if they’re talking to you Forecasting activities (what am I likely to do next?): this aims to anticipate your intentions and offer advice, like pointing out you’ve already added salt to a recipe if you look like you’re about to add some more. What about privacy? Obviously there are significant concerns regarding privacy. If this technology is paired with smart glasses constantly recording and analysing the environment, the result could be constant tracking and logging (via facial recognition) of people moving around in public. Read more: Face masks and facial recognition will both be common in the future. How will they co-exist?[9] While the above may sound dramatic, similar technology has already been trialled[10] in China, and the potential dangers have been explored by journalists[11]. Facebook says it will maintain high ethical and privacy standards for the data gathered for the project, including consent of participants, independent reviews, and de-identifying data[12] where possible. As such, Facebook says the data was captured in a “controlled environment with informed consent”, and in public spaces “faces and other PII [personally identifing information] are blurred”. But despite these reassurances (and noting this is only a trial), there are concerns over the future of smart-glasses technology coupled with the power of a social media giant whose intentions have not always been aligned to their users[13]. Read more: Artificial intelligence in Australia needs to get ethical, so we have a plan[14] The future? The ImageNet[15] dataset, a huge collection of tagged images, has helped computers learn to analyse and describe images over the past decade or more. Will Ego4D do the same for first-person video? We may get an idea next year. Facebook has invited the research community to participate in the Ego4D competition in June 2022, and pit their algorithms against the benchmark challenges to see if we can find those keys at last.

Read more https://theconversation.com/facebook-wants-ai-to-find-your-keys-and-understand-your-conversations-170092

Times Magazine

Epson launches ELPCS01 mobile projector cart

Designed for the EB-810E[1] projector and provides easy setup for portable displays in flexible ...

Governance Models for Headless CMS in Large Organizations

Where headless CMS is adopted by large enterprises, governance is the single most crucial factor d...

Narwal Freo Z Ultra Robotic Vacuum and Mop Cleaner

Rating: ★★★★☆ (4.4/5)Category: Premium Robot Vacuum & Mop ComboBest for: Busy households, ha...

Shark launches SteamSpot - the shortcut for everyday floor mess

Shark introduces the Shark SteamSpot Steam Mop, a lightweight steam mop designed to make everyda...

Game Together, Stay Together: Logitech G Reveals Gaming Couples Enjoy Higher Relationship Satisfaction

With Valentine’s Day right around the corner, many lovebirds across Australia are planning for the m...

AI threatens to eat business software – and it could change the way we work

In recent weeks, a range of large “software-as-a-service” companies, including Salesforce[1], Se...

The Times Features

Shark SteamSpot S2001 Review: A Chemical-Free Way to Tackle Messes and Stubborn Stains

If you're looking for a reliable steam mop that can handle both everyday spills and stubborn stains ...

How Businesses Are Generating Profits in a High-Inflation Economic Environment

Inflation in Australia and globally has surged to multi-decade highs since 2021, driven by pande...

The Effects of the War in the Middle East on Australian Small Businesses

The war in the Middle East is not a distant geopolitical event for Australia. In an interconnect...

Back at uni? How to help your wellbeing while you study

University can be a time of great opportunities, but it can also be very stressful[1]. Many stud...

Taste Port Douglas celebrates 10 years of world-class flavour in the tropics

30+ events, new sunrise and wellness experiences, 20+ chefs and a headline Michelin-star line-up...

Oztent RV tent range. Buy with caution

A review of the Oztent RV "30 second tent" range. Three years ago we bought an RV-4 from BCF Mack...

Essential Upgrades for a Smarter, Safer Australian Home

As we settle into 2026, the concept of the "dream home" has fundamentally shifted. The focus has m...

How To Modernise Your Home Without Overcapitalising

For many Australian homeowners, the dream of a "Grand Designs" transformation is often checked by ...

The Art of the Big Trip: Planning a Seamless Multi-Generational Getaway in Tropical North Queensland

There is a unique magic to the multi-generational holiday. It is a rare opportunity where gr...