Microsoft doesn't want AI recognizing your emotions anymore - mostly
Date:
Thu, 23 Jun 2022 15:46:30 +0000
Description:
Part of an overhaul of ethics policies, Microsoft makes a move as it worries over the potential of misuse.
FULL STORY ======================================================================
Microsoft is updating its Responsible AI Standard and revealed that it's retiring Azure Faces emotional and facial recognition abilities (for the most part).
The Responsible AI Standard is Microsofts internal ruleset when it comes to building AI systems. The company wants AI to be a positive force in the world and to never be misused by bad actors. Its a standard that's never been
shared with the public before. However, with this new change, Microsoft decided now would be the time.
Emotional and facial recognition software has been controversial, to say the least. There are multiple organizations calling for this technology to be banned. Fight for the Future , for example, wrote an open letter back in May asking Zoom to stop its own development of emotional tracking software and called it invasive and a violation of privacy and human rights. Policy
change
As its laid out, Microsoft will rework its Azure Face service to meet the requirements of its new Responsible AI Standard. First, the company is removing public access to the AI's emotion scanning capability. Second, Azure Face will no longer be able to identify a persons facial characteristics, including gender, age, [a] smile, facial hair, hair, and makeup.
The reason for the retirement is because the global science community still doesnt have a clear consensus on the definition of emotions. Natasha Cramption, Chief Responsible AI Officer at Microsoft, said that experts from inside and outside the company have voiced their concerns. The problem is the challenges in how inferences generalize across use cases, regions, and demographics, and the heightened privacy concerns
In addition to Azure Face, Microsofts Custom Neural Voice will be seeing similar restrictions. Custom Neural Voice is a text-to-speech app that is shockingly lifelike. Now the service will be limited to a select few managed customers and partners, which are people who work directly with Microsoft's account teams. The company states that while the technology has great potential, it may be used to impersonate. In order to keep having access to Neural Voice, all existing customers must submit an intake form and get approved by Microsoft. They have to be approved by June 30, 2023, and if they aren't selected, these customers will no longer have access to Neural Voice. Still in the works
Despite everything thats been said, Microsoft isnt totally abandoning its facial recognition tech. The announcement only pertains to public access. Sarah Bird, who is the Principal Group Project Manager at Azure AI, wrote about responsible facial recognition . And in that post, she states Microsoft recognizes these capabilities can be valuable when used for a set of controlled accessibility scenarios. One of these scenarios, according to a representative, is Seeing AI which is an iOS app that helps the visually impaired with identifying people and objects around them.
It's good to see another tech giant recognizing the problems with facial recognition and the potential for abuse. IBM did something similar back in 2020, although its approach was more absolute.
Back in 2020, IBM announced it was abandoning work on facial recognition because the company was afraid it could be misused for mass surveillance. Seeing these two titans of the industry get rid of this tech is a win for anti-facial recognition critics. If youre interested in learning more about AI, TechRadar recently published a piece on what it can do for cybersecurity .
======================================================================
Link to news story:
https://www.techradar.com/news/microsoft-doesnt-want-ai-recognizing-your-emoti ons-anymore-mostly/
--- Mystic BBS v1.12 A47 (Linux/64)
* Origin: tqwNet Technology News (1337:1/100)