Saturday, November 26, 2022

World No 1 Tech News Website

Microsoft to retire controversial facial recognition tool that claims to identify emotion

Must read

fatima khan
fatima khan
A brand new writer in the fields, Fatima has been taken under my electric spark's RGB- rich and ensures she doesn't engage in excessive snark on the website. It's unclear what command and Conquer are; however, she can talk for hours about the odd rhythm games, hardware, product reviews, and MMOs that were popular in the 2000s. Fatima has been creating various announcements, previews, and other content while here, but particularly enjoys writing regarding Products' latest news in the market she's currently addicted to. She is likely talking to an additional blogger with her current obsession right now.

Microsoft is phasing out public access to a number of AI-powered facial analysis tools — including one that claims to identify a subject’s emotion from videos and pictures.

Such “emotion recognition” tools have been criticized by experts. They say not only do facial expressions that are thought to be universal differ across different populations but that it is unscientific to equate external displays of emotion with internal feelings.

“Companies can say whatever they want, but the data are clear,” Lisa Feldman Barrett, a professor of psychology at Northeastern University who conducted a review into the subject of AI-powered emotion recognition, told The Verge in 2019. “They can detect a scowl, but that’s not the same thing as detecting anger.”

The decision is part of a larger overhaul of Microsoft’s AI ethics policies. The company’s updated Responsible AI Standards (first outlined in 2019) emphasize accountability to find out who uses its services and greater human oversight into where these tools are applied.

In practical terms, this means Microsoft will limit access to some features of its facial recognition services (known as Azure Face) and remove others entirely. Users will have to apply to use Azure Face for facial identification, for example, telling Microsoft exactly how and where they’ll be deploying its systems. Some use cases with less harmful potential (like automatically blurring faces in images and videos) will remain open-access.

In addition to removing public access to its emotion recognition tool, Microsoft is also retiring Azure Face’s ability to identify “attributes such as gender, age, smile, facial hair, hair, and makeup.”

“Experts inside and outside the company have highlighted the lack of scientific consensus on the definition of ‘emotions,’ the challenges in how inferences generalize across use cases, regions, and demographics, and the heightened privacy concerns around this type of capability,” wrote Microsoft’s chief responsible AI officer, Natasha Crampton, in a blog post announcing the news.

Microsoft says that it will stop offering these features to new customers from today, June 21st, while existing customers will have their access revoked on June 30th, 2023.

However, while Microsoft is retiring public access to these features, it will continue using them in at least one of its own products: an app named Seeing AI that uses machine vision to describe the world for people with visual impairments.

In a blog post, Microsoft’s principal group product manager for Azure AI, Sarah Bird, said that tools such as emotion recognition “can be valuable when used for a set of controlled accessibility scenarios.” It’s not clear if these tools will be used in any other Microsoft products.

Microsoft is also introducing similar restrictions to its Custom Neural Voice Feature, which lets customers create AI voices based on recordings of real people (sometimes known as an audio deepfake).

The tool “has exciting potential in education, accessibility, and entertainment,” writes Bird, but she notes that it “is also easy to imagine how it could be used to inappropriately impersonate speakers and deceive listeners.” Microsoft says that in the future, it will limit access to the Feature to “managed customers and partners” and “ensure the active participation of the speaker when creating a synthetic voice.”

More articles


Please enter your comment!
Please enter your name here

Latest article