-
公开(公告)号:US20220129696A1
公开(公告)日:2022-04-28
申请号:US17647390
申请日:2022-01-07
Applicant: NVIDIA Corporation
Inventor: Sakthivel Sivaraman , Shagan Sah , Niranjan Avadhanam
Abstract: In various examples, the present disclosure relates to using temporal filters for automated real-time classification. The technology described herein improves the performance of a multiclass classifier that may be used to classify a temporal sequence of input signals—such as input signals representative of video frames. A performance improvement may be achieved, at least in part, by applying a temporal filter to an output of the multiclass classifier. For example, the temporal filter may leverage classifications associated with preceding input signals to improve the final classification given to a subsequent signal. In some embodiments, the temporal filter may also use data from a confusion matrix to correct for the probable occurrence of certain types of classification errors. The temporal filter may be a linear filter, a nonlinear filter, an adaptive filter, and/or a statistical filter.
-
12.
公开(公告)号:US20240265254A1
公开(公告)日:2024-08-08
申请号:US18605628
申请日:2024-03-14
Applicant: NVIDIA Corporation
Inventor: Nuri Murat Arar , Niranjan Avadhanam , Nishant Puri , Shagan Sah , Rajath Shetty , Sujay Yadawadkar , Pavlo Molchanov
IPC: G06N3/08 , G06F18/21 , G06F18/214 , G06N20/00 , G06V10/764 , G06V10/774 , G06V10/82 , G06V10/94 , G06V20/59 , G06V20/64 , G06V40/16 , G06V40/18
CPC classification number: G06N3/08 , G06F18/214 , G06F18/2193 , G06N20/00 , G06V10/764 , G06V10/774 , G06V10/82 , G06V10/95 , G06V20/597 , G06V20/647 , G06V40/171 , G06V40/193
Abstract: Systems and methods for more accurate and robust determination of subject characteristics from an image of the subject. One or more machine learning models receive as input an image of a subject, and output both facial landmarks and associated confidence values. Confidence values represent the degrees to which portions of the subject's face corresponding to those landmarks are occluded, i.e., the amount of uncertainty in the position of each landmark location. These landmark points and their associated confidence values, and/or associated information, may then be input to another set of one or more machine learning models which may output any facial analysis quantity or quantities, such as the subject's gaze direction, head pose, drowsiness state, cognitive load, or distraction state.
-
公开(公告)号:US20240022601A1
公开(公告)日:2024-01-18
申请号:US17863140
申请日:2022-07-12
Applicant: NVIDIA Corporation
Inventor: Manoj Kumar Yennapureddy , Shagan Sah , Rajath Shetty
CPC classification number: H04L63/1466 , H04L63/1416 , G06T7/50 , G06T2207/10028 , G06T2207/20084 , G06T2207/20076
Abstract: In various examples, techniques are described for detecting whether spoofing attacks are occurring using multiple sensors. Systems and methods are disclosed that include at least a first sensor having a first pose to capture a first perspective view of a user and a second sensor having a second pose to capture a second perspective view of the user. The first sensor and/or the second sensor may include an image sensor, a depth sensor, and/or the like. The systems and methods include a neural network that is configured to analyze first sensor data generated by the first sensor and second sensor data generated by the second sensor to determine whether a spoofing attack is occurring. The systems and methods may also perform one or more processes, such as facial recognition, based on whether the spoofing attack is occurring.
-
公开(公告)号:US11721089B2
公开(公告)日:2023-08-08
申请号:US17647390
申请日:2022-01-07
Applicant: NVIDIA Corporation
Inventor: Sakthivel Sivaraman , Shagan Sah , Niranjan Avadhanam
IPC: G06V10/00 , G06V10/764 , G06V20/40 , G06V40/20 , G06V40/10 , G06F18/20 , G06F18/2321 , G06N3/045
CPC classification number: G06V10/764 , G06F18/2321 , G06F18/285 , G06N3/045 , G06V20/47 , G06V40/113 , G06V40/28
Abstract: In various examples, the present disclosure relates to using temporal filters for automated real-time classification. The technology described herein improves the performance of a multiclass classifier that may be used to classify a temporal sequence of input signals—such as input signals representative of video frames. A performance improvement may be achieved, at least in part, by applying a temporal filter to an output of the multiclass classifier. For example, the temporal filter may leverage classifications associated with preceding input signals to improve the final classification given to a subsequent signal. In some embodiments, the temporal filter may also use data from a confusion matrix to correct for the probable occurrence of certain types of classification errors. The temporal filter may be a linear filter, a nonlinear filter, an adaptive filter, and/or a statistical filter.
-
公开(公告)号:US11704814B2
公开(公告)日:2023-07-18
申请号:US17319891
申请日:2021-05-13
Applicant: NVIDIA Corporation
Inventor: Nuri Murat Arar , Niranjan Avadhanam , Hairong Jiang , Nishant Puri , Rajath Shetty , Shagan Sah
CPC classification number: G06T7/20 , G06F18/211 , G06F18/2178 , G06N3/08 , G06V20/59 , G06V40/18
Abstract: In various examples, an adaptive eye tracking machine learning model engine (“adaptive-model engine”) for an eye tracking system is described. The adaptive-model engine may include an eye tracking or gaze tracking development pipeline (“adaptive-model training pipeline”) that supports collecting data, training, optimizing, and deploying an adaptive eye tracking model that is a customized eye tracking model based on a set of features of an identified deployment environment. The adaptive-model engine supports ensembling the adaptive eye tracking model that may be trained on gaze vector estimation in surround environments and ensemble based on a plurality of eye tracking variant models and a plurality of facial landmark neural network metrics.
-
公开(公告)号:US20220366568A1
公开(公告)日:2022-11-17
申请号:US17319891
申请日:2021-05-13
Applicant: NVIDIA Corporation
Inventor: Nuri Murat Arar , Niranjan Avadhanam , Hairong Jiang , Nishant Puri , Rajath Shetty , Shagan Sah
Abstract: In various examples, an adaptive eye tracking machine learning model engine (“adaptive-model engine”) for an eye tracking system is described. The adaptive-model engine may include an eye tracking or gaze tracking development pipeline (“adaptive-model training pipeline”) that supports collecting data, training, optimizing, and deploying an adaptive eye tracking model that is a customized eye tracking model based on a set of features of an identified deployment environment. The adaptive-model engine supports ensembling the adaptive eye tracking model that may be trained on gaze vector estimation in surround environments and ensemble based on a plurality of eye tracking variant models and a plurality of facial landmark neural network metrics.
-
公开(公告)号:US11222232B1
公开(公告)日:2022-01-11
申请号:US16907125
申请日:2020-06-19
Applicant: NVIDIA Corporation
Inventor: Sakthivel Sivaraman , Shagan Sah , Niranjan Avadhanam
Abstract: In various examples, the present disclosure relates to using temporal filters for automated real-time classification. The technology described herein improves the performance of a multiclass classifier that may be used to classify a temporal sequence of input signals—such as input signals representative of video frames. A performance improvement may be achieved, at least in part, by applying a temporal filter to an output of the multiclass classifier. For example, the temporal filter may leverage classifications associated with preceding input signals to improve the final classification given to a subsequent signal. In some embodiments, the temporal filter may also use data from a confusion matrix to correct for the probable occurrence of certain types of classification errors. The temporal filter may be a linear filter, a nonlinear filter, an adaptive filter, and/or a statistical filter.
-
公开(公告)号:US20210397885A1
公开(公告)日:2021-12-23
申请号:US16907125
申请日:2020-06-19
Applicant: NVIDIA Corporation
Inventor: Sakthivel Sivaraman , Shagan Sah , Niranjan Avadhanam
Abstract: In various examples, the present disclosure relates to using temporal filters for automated real-time classification. The technology described herein improves the performance of a multiclass classifier that may be used to classify a temporal sequence of input signals—such as input signals representative of video frames. A performance improvement may be achieved, at least in part, by applying a temporal filter to an output of the multiclass classifier. For example, the temporal filter may leverage classifications associated with preceding input signals to improve the final classification given to a subsequent signal. In some embodiments, the temporal filter may also use data from a confusion matrix to correct for the probable occurrence of certain types of classification errors. The temporal filter may be a linear filter, a nonlinear filter, an adaptive filter, and/or a statistical filter.
-
19.
公开(公告)号:US20210182625A1
公开(公告)日:2021-06-17
申请号:US17004252
申请日:2020-08-27
Applicant: NVIDIA Corporation
Inventor: Nuri Murat Arar , Niranjan Avadhanam , Nishant Puri , Shagan Sah , Rajath Shetty , Sujay Yadawadkar , Pavlo Molchanov
Abstract: Systems and methods for more accurate and robust determination of subject characteristics from an image of the subject. One or more machine learning models receive as input an image of a subject, and output both facial landmarks and associated confidence values. Confidence values represent the degrees to which portions of the subject's face corresponding to those landmarks are occluded, i.e., the amount of uncertainty in the position of each landmark location. These landmark points and their associated confidence values, and/or associated information, may then be input to another set of one or more machine learning models which may output any facial analysis quantity or quantities, such as the subject's gaze direction, head pose, drowsiness state, cognitive load, or distraction state.
-
20.
公开(公告)号:US12162418B2
公开(公告)日:2024-12-10
申请号:US18481603
申请日:2023-10-05
Applicant: NVIDIA Corporation
Inventor: Atousa Torabi , Sakthivel Sivaraman , Niranjan Avadhanam , Shagan Sah
IPC: B60R21/017 , B60R21/013 , B60W50/14 , B60W60/00 , G06N3/02 , B60R21/01 , B60W50/00
Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.
-
-
-
-
-
-
-
-
-