Abstract:
A method, performed in an electronic device, for connecting to a target device is disclosed. The method includes capturing an image including a face of a target person associated with the target device and recognizing an indication of the target person. The indication of the target person may be a pointing object, a speech command, and/or any suitable input command. The face of the target person in the image is detected based on the indication and at least one facial feature of the face in the image is extracted. Based on the at least one facial feature, the electronic device is connected to the target device.
Abstract:
A method for activating a voice assistant function in a mobile device is disclosed. The method includes receiving an input sound stream by a sound sensor and determining a context of the mobile device. The method may determine the context based on the input sound stream. For determining the context, the method may also obtain data indicative of the context of the mobile device from at least one of an acceleration sensor, a location sensor, an illumination sensor, a proximity sensor, a clock unit, and a calendar unit in the mobile device. In this method, a threshold for activating the voice assistant function is adjusted based on the context. The method detects a target keyword from the input sound stream based on the adjusted threshold. If the target keyword is detected, the method activates the voice assistant function.
Abstract:
Embodiment systems and methods for presenting a facial expression in a virtual meeting may include detecting a user facial expression of a user based on information received from a sensor of the computing device, determining whether the detected user facial expression is approved for presentation on an avatar in a virtual meeting, generating an avatar exhibiting a facial expression consistent with the detected user facial expression in response to determining that the detected user facial expression is approved for presentation on an avatar in the virtual meeting, generating an avatar exhibiting a facial expression that is approved for presentation in response to determining that the detected user facial expression is not approved for presentation on an avatar in the virtual meeting, and presenting the generated avatar in the virtual meeting.
Abstract:
A method for controlling voice activation by a target keyword in a mobile device is disclosed. The method includes receiving an input sound stream. When the input sound stream indicates speech, the voice activation unit is activated to detect the target keyword and at least one sound feature is extracted from the input sound stream. Further, the method includes deactivating the voice activation unit when the at least one sound feature indicates a non-target keyword.
Abstract:
A method, performed in an electronic device, for connecting to a target device is disclosed. The method includes capturing an image including a face of a target person associated with the target device and recognizing an indication of the target person. The indication of the target person may be a pointing object, a speech command, and/or any suitable input command. The face of the target person in the image is detected based on the indication and at least one facial feature of the face in the image is extracted. Based on the at least one facial feature, the electronic device is connected to the target device.
Abstract:
A method for activating a voice assistant function in a mobile device is disclosed. The method includes receiving an input sound stream by a sound sensor and determining a context of the mobile device. The method may determine the context based on the input sound stream. For determining the context, the method may also obtain data indicative of the context of the mobile device from at least one of an acceleration sensor, a location sensor, an illumination sensor, a proximity sensor, a clock unit, and a calendar unit in the mobile device. In this method, a threshold for activating the voice assistant function is adjusted based on the context. The method detects a target keyword from the input sound stream based on the adjusted threshold. If the target keyword is detected, the method activates the voice assistant function.
Abstract:
According to an aspect of the present disclosure, a method for controlling access to a plurality of applications in an electronic device is disclosed. The method includes receiving a voice command from a speaker for accessing a target application among the plurality of applications, and verifying whether the voice command is indicative of a user authorized to access the applications based on a speaker model of the authorized user. In this method, each application is associated with a security level having a threshold value. The method further includes updating the speaker model with the voice command if the voice command is verified to be indicative of the user, and adjusting at least one of the threshold values based on the updated speaker model.
Abstract:
A method of detecting a target keyword from an input sound for activating a function in a mobile device is disclosed. In this method, a first plurality of sound features is received in a buffer, and a second plurality of sound features is received in the buffer. While receiving each of the second plurality of sound features in the buffer, a first number of the sound features are processed from the buffer. The first number of the sound features includes two or more sound features. Further, the method may include determining a keyword score for each of the processed sound features and detecting the input sound as the target keyword if at least one of the keyword scores is greater than a threshold score.
Abstract:
A device to process an audio signal representing input sound includes a hand detector configured to generate a first indication responsive to detection of at least a portion of a hand over at least a portion of the device. The device also includes an automatic speech recognition system configured to be activated, responsive to the first indication, to process the audio signal.
Abstract:
A device to process an audio signal representing input sound includes a user voice verifier configured to generate a first indication based on whether the audio signal represents a user's voice. The device includes a speaking target detector configured to generate a second indication based on whether the audio signal represents at least one of a command or a question. The device includes an activation signal unit configured to selectively generate an activation signal based on the first indication and the second indication. The device also includes an automatic speech recognition engine configured to be activated, responsive to the activation signal, to process the audio signal.