• AI reduces 'communication gap' for nonve

    From ScienceDaily@1337:3/111 to All on Mon Jun 15 21:30:34 2020
    AI reduces 'communication gap' for nonverbal people by as much as half


    Date:
    June 15, 2020
    Source:
    University of Cambridge
    Summary:
    Researchers have used artificial intelligence to reduce the
    'communication gap' for nonverbal people with motor disabilities
    who rely on computers to converse with others.



    FULL STORY ========================================================================== Researchers have used artificial intelligence to reduce the 'communication
    gap' for nonverbal people with motor disabilities who rely on computers
    to converse with others.


    ==========================================================================
    The team, from the University of Cambridge and the University of Dundee, developed a new context-aware method that reduces this communication gap
    by eliminating between 50% and 96% of the keystrokes the person has to
    type to communicate.

    The system is specifically tailed for nonverbal people and uses a range
    of context 'clues' -- such as the user's location, the time of day or
    the identity of the user's speaking partner -- to assist in suggesting sentences that are the most relevant for the user.

    Nonverbal people with motor disabilities often use a computer with speech output to communicate with others. However, even without a physical
    disability that affects the typing process, these communication aids
    are too slow and error prone for meaningful conversation: typical typing
    rates are between five and 20 words per minute, while a typical speaking
    rate is in the range of 100 to 140 words per minute.

    "This difference in communication rates is referred to as the
    communication gap," said Professor Per Ola Kristensson from Cambridge's Department of Engineering, the study's lead author. "The gap is typically between 80 and 135 words per minute and affects the quality of everyday interactions for people who rely on computers to communicate." The method developed by Kristensson and his colleagues uses artificial intelligence
    to allow a user to quickly retrieve sentences they have typed in the
    past. Prior research has shown that people who rely on speech synthesis,
    just like everyone else, tend to reuse many of the same phrases and
    sentences in everyday conversation. However, retrieving these phrases
    and sentences is a time-consuming process for users of existing speech synthesis technologies, further slowing down the flow of conversation.



    ==========================================================================
    In the new system, as the person is typing, the system uses information retrieval algorithms to automatically retrieve the most relevant previous sentences based on the text typed and the context the conversation the
    person is involved in. Context includes information about the conversation
    such as the location, time of day, and automatic identification of
    the speaking partner's face. The other speaker is identified using
    a computer vision algorithm trained to recognise human faces from a front-mounted camera.

    The system was developed using design engineering methods typically used
    for jet engines or medical devices. The researchers first identified
    the critical functions of the system, such as the word auto-complete
    function and the sentence retrieval function. After these functions had
    been identified, the researchers simulated a nonverbal person typing a
    large set of sentences from a sentence set representative of the type
    of text a nonverbal person would like to communicate.

    This analysis allowed the researchers to understand the best method
    for retrieving sentences and the impact of a range of parameters on performance, such as the accuracy of word-auto complete and the impact
    of using many context tags. For example, this analysis revealed that only
    two reasonably accurate context tags are required to provide the majority
    of the gain. Word-auto complete provides a positive contribution but
    is not essential for realising the majority of the gain. The sentences
    are retrieved using information retrieval algorithms, similar to web
    search. Context tags are added to the words the user types to form
    a query.

    The study is the first to integrate context-aware information retrieval
    with speech-generating devices for people with motor disabilities, demonstrating how context-sensitive artificial intelligence can improve
    the lives of people with motor disabilities.

    "This method gives us hope for more innovative AI-infused systems to
    help people with motor disabilities to communicate in the future,"
    said Kristensson.

    "We've shown it's possible to reduce the opportunity cost of not doing innovative research with AI-infused user interfaces that challenge
    traditional user interface design mantra and processes." The research
    paper was published at CHI 2020, the world leading conference for Computer-Human Interaction research.

    The research was funded by the Engineering and Physical Sciences Research Council.


    ========================================================================== Story Source: Materials provided by University_of_Cambridge. The original
    story is licensed under a Creative_Commons_License. Note: Content may
    be edited for style and length.


    ========================================================================== Journal Reference:
    1. Per Ola Kristensson, James Lilley, Rolf Black, Annalu Waller. A
    Design
    Engineering Approach for Quantitatively Exploring Context-Aware
    Sentence Retrieval for Nonspeaking Individuals with Motor
    Disabilities.

    Proceedings of the 38th ACM Conference on Human Factors in Computing
    Systems, 2020 DOI: 10.1145/3313831.3376525 ==========================================================================

    Link to news story: https://www.sciencedaily.com/releases/2020/06/200615115729.htm

    --- up 20 weeks, 6 days, 2 hours, 34 minutes
    * Origin: -=> Castle Rock BBS <=- Now Husky HPT Powered! (1337:3/111)