-
公开(公告)号:US11172122B2
公开(公告)日:2021-11-09
申请号:US16241438
申请日:2019-01-07
Applicant: Amazon Technologies, Inc.
Inventor: William Evan Welbourne , Ross David Roessler , Cheng-Hao Kuo , Jim Oommen Thomas , Paul Aksenti Savastinuk , Yinfei Yang
Abstract: Devices, systems and methods are disclosed for improving facial recognition and/or speaker recognition models by using results obtained from one model to assist in generating results from the other model. For example, a device may perform facial recognition for image data to identify users and may use the results of the facial recognition to assist in speaker recognition for corresponding audio data. Alternatively or additionally, the device may perform speaker recognition for audio data to identify users and may use the results of the speaker recognition to assist in facial recognition for corresponding image data. As a result, the device may identify users in video data that are not included in the facial recognition model and may identify users in audio data that are not included in the speaker recognition model. The facial recognition and/or speaker recognition models may be updated during run-time and/or offline using post-processed data.
-
公开(公告)号:US20150364109A1
公开(公告)日:2015-12-17
申请号:US14307284
申请日:2014-06-17
Applicant: Amazon Technologies, Inc.
Inventor: Kritarth Jain , Michal Marek Kozlowski , Michael Lee Sandige , Andrew Bartlett Leonard , Paul Savastinuk , Ross David Roessler , Geoffrey Scott Heller
CPC classification number: G09G5/006 , G06F3/011 , G06F3/012 , G06F3/017 , G06F3/0346 , G06F3/038 , G06F2203/0383
Abstract: A tracking architecture is provided that enables data for gestures and head positions to be provided to both native and non-native clients on a computing device. A pipeline component can obtain the raw image data and sensor data and synchronize that data to be processed to determine, for example, location and/or motion data that may correspond to device input. The data can be processed by separate components, such as an event publisher and an event provider, each capable of filtering the location, motion, and/or raw sensor data to generate a set of event data. The event data then can be published to registered listeners or provided in response to polling requests. Head coordinates, gesture data, and other such information can be passed through one or more interface layers enabling the data to be processed by a non-native client on the device.
Abstract translation: 提供了一种跟踪架构,可以将手势和头部位置的数据提供给计算设备上的本机和非本机客户端。 流水线部件可以获得原始图像数据和传感器数据,并使要处理的数据同步以确定例如可能对应于设备输入的位置和/或运动数据。 数据可以由单独的组件(例如事件发布者和事件提供者)处理,每个组件能够过滤位置,运动和/或原始传感器数据以生成一组事件数据。 事件数据然后可以发布到注册的听众或响应轮询请求提供。 头部坐标,手势数据和其他这样的信息可以被传递通过一个或多个界面层,使数据能够由设备上的非本机客户端处理。
-
公开(公告)号:US10477104B1
公开(公告)日:2019-11-12
申请号:US14929928
申请日:2015-11-02
Applicant: Amazon Technologies, Inc.
Inventor: Ross David Roessler
IPC: H04N5/232 , H04N7/18 , H04N13/189 , H04N13/324 , H04N13/398
Abstract: Various examples are directed to systems and methods for selecting image sensors in a multiple image sensor device. A control circuit may receive a first frame from the first image sensor and a second frame from the second image sensor. The control circuit may receive object data describing an object depicted in the first frame and may turn off the second image sensor.
-
公开(公告)号:US20190313014A1
公开(公告)日:2019-10-10
申请号:US16241438
申请日:2019-01-07
Applicant: Amazon Technologies, Inc.
Inventor: William Evan Welbourne , Ross David Roessler , Cheng-Hao Kuo , Jim Oommen Thomas , Paul Aksenti Savastinuk , Yinfei Yang
Abstract: Devices, systems and methods are disclosed for improving facial recognition and/or speaker recognition models by using results obtained from one model to assist in generating results from the other model. For example, a device may perform facial recognition for image data to identify users and may use the results of the facial recognition to assist in speaker recognition for corresponding audio data. Alternatively or additionally, the device may perform speaker recognition for audio data to identify users and may use the results of the speaker recognition to assist in facial recognition for corresponding image data. As a result, the device may identify users in video data that are not included in the facial recognition model and may identify users in audio data that are not included in the speaker recognition model. The facial recognition and/or speaker recognition models may be updated during run-time and/or offline using post-processed data.
-
公开(公告)号:US09754552B2
公开(公告)日:2017-09-05
申请号:US14307284
申请日:2014-06-17
Applicant: Amazon Technologies, Inc.
Inventor: Kritarth Jain , Michal Marek Kozlowski , Michael Lee Sandige , Andrew Bartlett Leonard , Paul Savastinuk , Ross David Roessler , Geoffrey Scott Heller
IPC: G06F3/01 , G09G5/00 , G06F3/038 , G06F3/0346
CPC classification number: G09G5/006 , G06F3/011 , G06F3/012 , G06F3/017 , G06F3/0346 , G06F3/038 , G06F2203/0383
Abstract: A tracking architecture is provided that enables data for gestures and head positions to be provided to both native and non-native clients on a computing device. A pipeline component can obtain the raw image data and sensor data and synchronize that data to be processed to determine, for example, location and/or motion data that may correspond to device input. The data can be processed by separate components, such as an event publisher and an event provider, each capable of filtering the location, motion, and/or raw sensor data to generate a set of event data. The event data then can be published to registered listeners or provided in response to polling requests. Head coordinates, gesture data, and other such information can be passed through one or more interface layers enabling the data to be processed by a non-native client on the device.
-
公开(公告)号:US10582125B1
公开(公告)日:2020-03-03
申请号:US14727782
申请日:2015-06-01
Applicant: Amazon Technologies, Inc.
Inventor: Ross David Roessler , Matthew Alan Townsend , Yinfei Yang , Jim Oommen Thomas , Deon Poncini , William Evan Welbourne , Geoff Hunter Donaldson , Paul Aksenti Savastinuk , Cheng-Hao Kuo
Abstract: A video capture device may include multiple cameras that simultaneously capture video data. The video capture device and/or one or more remote computing resources may stitch the video data captured by the multiple cameras to generate stitched video data that corresponds to 360° video. The remote computing resources may apply one or more algorithms to the stitched video data to identify one or more frames that depict content that is likely to be of interest to a user. The video capture device and/or the remote computing resources may generate one or more images from the one or more frames, and may send the one or more images to the user.
-
公开(公告)号:US10277813B1
公开(公告)日:2019-04-30
申请号:US14751024
申请日:2015-06-25
Applicant: Amazon Technologies, Inc.
Inventor: Jim Oommen Thomas , Paul Aksenti Savastinuk , Cheng-Hao Kuo , Tsz Ho Yu , Ross David Roessler , William Evan Welbourne , Yinfei Yang
Abstract: A viewing device, such as a virtual reality headset, allows a user to view a panoramic scene captured by one or more video capture devices that may include multiple cameras that simultaneously capture 360° video data. The viewing device may display the panoramic scene in real time and change the display in response to moving the viewing device and/or changing perspectives by switching to video data being captured by a different video capture device within the environment. Moreover, multiple video capture devices located within an environment can be used to create a three-dimensional representation of the environment that allows a user to explore the three-dimensional space while viewing the environment in real time.
-
公开(公告)号:US10178301B1
公开(公告)日:2019-01-08
申请号:US14750895
申请日:2015-06-25
Applicant: Amazon Technologies, Inc.
Inventor: William Evan Welbourne , Ross David Roessler , Cheng-Hao Kuo , Jim Oommen Thomas , Paul Aksenti Savastinuk , Yinfei Yang
Abstract: Devices, systems and methods are disclosed for improving facial recognition and/or speaker recognition models by using results obtained from one model to assist in generating results from the other model. For example, a device may perform facial recognition for image data to identify users and may use the results of the facial recognition to assist in speaker recognition for corresponding audio data. Alternatively or additionally, the device may perform speaker recognition for audio data to identify users and may use the results of the speaker recognition to assist in facial recognition for corresponding image data. As a result, the device may identify users in video data that are not included in the facial recognition model and may identify users in audio data that are not included in the speaker recognition model. The facial recognition and/or speaker recognition models may be updated during run-time and/or offline using post-processed data.
-
公开(公告)号:US10104286B1
公开(公告)日:2018-10-16
申请号:US14837793
申请日:2015-08-27
Applicant: Amazon Technologies, Inc.
Inventor: Tsz Ho Yu , Paul Aksenti Savastinuk , Yinfei Yang , Cheng-Hao Kuo , Ross David Roessler , William Evan Welbourne
Abstract: Systems and methods may be directed to de-blurring panoramic images and/or video. An image processor may receive a frame, where the frame comprises a plurality of pixel values arranged in a grid. The image processor may divide the frame into a first section and a second section. The image processor may determine a first motion kernel for the first section and apply the first motion kernel to the first section. The image processor may also determine a second motion kernel for the second section and apply the second motion kernel to the second section.
-
公开(公告)号:US09973711B2
公开(公告)日:2018-05-15
申请号:US14753826
申请日:2015-06-29
Applicant: Amazon Technologies, Inc.
Inventor: Yinfei Yang , William Evan Welbourne , Ross David Roessler , Paul Aksenti Savastinuk , Cheng-Hao Kuo , Jim Oommen Thomas , Tsz Ho Yu
CPC classification number: H04N5/2628 , G06K9/00711 , G06K9/00751 , G06K9/3233 , G06T3/40 , G11B27/031 , G11B27/06 , H04N5/23238
Abstract: Devices, systems and methods are disclosed for identifying content in video data and creating content-based zooming and panning effects to emphasize the content. Contents may be detected and analyzed in the video data using computer vision, machine learning algorithms or specified through a user interface. Panning and zooming controls may be associated with the contents, panning or zooming based on a location and size of content within the video data. The device may determine a number of pixels associated with content and may frame the content to be a certain percentage of the edited video data, such as a close-up shot where a subject is displayed as 50% of the viewing frame. The device may identify an event of interest, may determine multiple frames associated with the event of interest and may pan and zoom between the multiple frames based on a size/location of the content within the multiple frames.
-
-
-
-
-
-
-
-
-