-
公开(公告)号:US10210905B2
公开(公告)日:2019-02-19
申请号:US15394391
申请日:2016-12-29
Applicant: SONY INTERACTIVE ENTERTAINMENT INC.
Inventor: Dennis Dale Castleman , Ruxin Chen , Frank Zhao , Glenn Black
IPC: G01C23/00 , G11B27/10 , B64C39/02 , G05D1/00 , G06F3/0484
Abstract: A flight path management system manages flight paths for an unmanned aerial vehicle (UAV). The flight path management system receives a sequence of controller inputs for the UAV, and stores the sequence of controller inputs in a memory. The flight path management system accesses the memory and selects a selected section of the sequence of controller inputs corresponding to a time period. The flight management system outputs the selected section to a playback device in real time over a length of the time period.
-
公开(公告)号:US10191541B2
公开(公告)日:2019-01-29
申请号:US15385794
申请日:2016-12-20
Applicant: Sony Interactive Entertainment Inc.
Inventor: Ruxin Chen
IPC: G06F3/01 , G02B27/01 , G06T11/60 , G09G5/36 , G06F3/03 , G06F3/0481 , G06F3/0346
Abstract: Methods, devices, and computer programs for augmenting a virtual reality scene with real world content are provided. One example method includes an operation for obtaining sensor data from an HMD of a user to determine that a criteria is met to overlay one or more real world objects into the virtual reality scene to provide an augmented virtual reality scene. In certain examples, the criteria corresponds to predetermined indicators suggestive of disorientation of a user when wearing the HMD and being presented a virtual reality scene. In certain other examples, the one or more real world objects are selected based on their effectiveness at reorienting a disoriented user.
-
公开(公告)号:US20240198240A1
公开(公告)日:2024-06-20
申请号:US18589404
申请日:2024-02-27
Applicant: Sony Interactive Entertainment Inc.
Inventor: Javier Fernandez Rico , Erik Beran , Michael Taylor , Ruxin Chen
IPC: A63F13/90 , A63F13/213 , B25J5/00 , B25J9/04 , B25J11/00 , B25J13/08 , B25J15/04 , B25J15/10 , G06F3/01 , G06F3/042 , G06F3/04817 , G06F3/16 , H04N9/31
CPC classification number: A63F13/90 , A63F13/213 , B25J5/007 , B25J9/046 , B25J11/0005 , B25J11/003 , B25J11/008 , B25J13/08 , B25J15/04 , B25J15/10 , G06F3/017 , G06F3/0425 , G06F3/04817 , G06F3/167 , H04N9/3173 , H04N9/3182 , H04N9/3185 , H04N9/3194
Abstract: Methods and systems are provided for providing real world assistance by a robot utility and interface device (RUID) are provided. A method provides for identifying a position of a user in a physical environment and a surface within the physical environment for projecting an interactive interface. The method also provides for moving to a location within the physical environment based on the position of the user and the surface for projecting the interactive interface. Moreover, the method provides for capturing a plurality of images of the interactive interface while the interactive interface is being interacted with by the use and for determining a selection of an input option made by the user.
-
14.
公开(公告)号:US11829878B2
公开(公告)日:2023-11-28
申请号:US17852602
申请日:2022-06-29
Applicant: Sony Interactive Entertainment Inc.
Inventor: Ruxin Chen , Naveen Kumar , Haoqi Li
CPC classification number: G06N3/08 , G06F18/217 , G06N3/006 , G06N20/00 , G06V20/41 , G06V40/161 , G06V40/174
Abstract: In sequence level prediction of a sequence of frames of high dimensional data one or more affective labels are provided at the end of the sequence. Each label pertains to the entire sequence of frames. An action is taken with an agent controlled by a machine learning algorithm for a current frame of the sequence at a current time step. An output of the action represents affective label prediction for the frame at the current time step. A pool of actions taken up until the current time step including the action taken with the agent is transformed into a predicted affective history for a subsequent time step. A reward is generated on predicted actions up to the current time step by comparing the predicted actions against corresponding annotated affective labels.
-
15.
公开(公告)号:US11640519B2
公开(公告)日:2023-05-02
申请号:US16176949
申请日:2018-10-31
Applicant: Sony Interactive Entertainment Inc.
Inventor: Ruxin Chen , Min-Hung Chen , Jaekwon Yoo , Xiaoyu Liu
Abstract: A domain adaptation module is used to optimize a first domain derived from a second domain using respective outputs from respective parallel hidden layers of the domains.
-
公开(公告)号:US11494612B2
公开(公告)日:2022-11-08
申请号:US16176812
申请日:2018-10-31
Applicant: Sony Interactive Entertainment Inc.
Inventor: Ruxin Chen , Min-Hung Chen , Jaekwon Yoo , Xiaoyu Liu
Abstract: A domain adaptation module is used to optimize a first domain derived from a second domain using respective outputs from respective parallel hidden layers of the domains.
-
公开(公告)号:US11079999B2
公开(公告)日:2021-08-03
申请号:US16392589
申请日:2019-04-23
Applicant: Sony Interactive Entertainment Inc.
Inventor: Ruxin Chen
IPC: G09G5/00 , G06F3/14 , G06F3/01 , G02B27/01 , A63F13/26 , A63F13/211 , A63F13/213 , A63F13/5255 , A63F13/212 , G06K9/00
Abstract: Method for providing image of HMD user to a non-HMD user includes, receiving a first image of a user including the user's facial features captured by an external camera when the user is not wearing a head mounted display (HMD). A second image capturing a portion of the facial features of the user when the user is wearing the HMD is received. An image overlay data is generated by mapping contours of facial features captured in the second image with contours of corresponding facial features captured in the first image. The image overlay data is forwarded to the HMD for rendering on a second display screen that is mounted on a front face of the HMD.
-
公开(公告)号:US10885341B2
公开(公告)日:2021-01-05
申请号:US16171018
申请日:2018-10-25
Applicant: Sony Interactive Entertainment Inc.
Inventor: Ruxin Chen , Naveen Kumar , Haoqi Li
Abstract: Methods and systems for performing sequence level prediction of a video scene are described. Video information in a video scene is represented as a sequence of features depicted each frame. An environment state for each time step t corresponding to each frame is represented by the video information for time step t and predicted affective information from a previous time step t−1. An action A(t) as taken with an agent controlled by a machine learning algorithm for the frame at step t, wherein an output of the action A(t) represents affective label prediction for the frame at the time step t. A pool of predicted actions is transformed to a predicted affective history at a next time step t+1. The predictive affective history is included as part of the environment state for the next time step t+1. A reward R is generated on predicted actions up to the current time step t, by comparing them against corresponding annotated movie scene affective labels.
-
公开(公告)号:US10726062B2
公开(公告)日:2020-07-28
申请号:US16206439
申请日:2018-11-30
Applicant: Sony Interactive Entertainment Inc.
Inventor: Jian Zheng , Ruxin Chen
IPC: G06F16/383 , G06F16/583 , G06K9/32 , G06N5/04
Abstract: For image captioning such as for computer game images or other images, bottom-up attention is combined with top-down attention to provide a multi-level residual attention-based image captioning model. A residual attention mechanism is first applied in the Faster R-CNN network to learn better feature representations for each region by taking spatial information into consideration. In the image captioning network, taking the extracted regional features as input, a second residual attention network is implemented to fuse the regional features attentionally for subsequent caption generation.
-
公开(公告)号:US10714076B2
公开(公告)日:2020-07-14
申请号:US15645985
申请日:2017-07-10
Applicant: Sony Interactive Entertainment Inc.
Inventor: Xavier Menendez-Pidal , Ruxin Chen
Abstract: A method for improved initialization of speech recognition system comprises mapping a trained hidden markov model based recognition node network (HMM) to a Connectionist Temporal Classification (CTC) based node label scheme. The central state of each frame in the HMM are mapped to CTC-labeled output nodes and the non-central states of each frame are mapped to CTC-blank nodes to generate a CTC-labeled HMM and each central state represents a phoneme from human speech detected and extracted by a computing device. Next the CTC-labeled HMM is trained using a cost function, wherein the cost function is not part of a CTC cost function. Finally the CTC-labeled HMM is trained using a CTC cost function to produce a CTC node network. The CTC node network may be iteratively trained by repeating the initialization steps.
-
-
-
-
-
-
-
-
-