Reinforcement learning for guaranteed delivery of supplemental content

    公开(公告)号:US11546665B2

    公开(公告)日:2023-01-03

    申请号:US17315107

    申请日:2021-05-07

    申请人: HULU, LLC

    摘要: In some embodiments, a method receives a request for supplemental content to be provided in association with main content. The method selects an instance of supplemental content based on a long-term reward metric and a short-term reward metric. The long-term reward metric is based on feedback from delivery of a plurality of instances of supplemental content and a delivery status for a delivery constraint of one instance of supplemental content. The short-term reward metric is based on feedback from delivery of the one instance of supplemental content. The long-term reward metric is based on feedback from delivery of a plurality of instances of supplemental content and the short-term reward metric is based on feedback from delivery of one instance of supplemental content. The instance of supplemental content is provided to a client device.

    Image detection using multiple detection processes

    公开(公告)号:US11113537B2

    公开(公告)日:2021-09-07

    申请号:US16530597

    申请日:2019-08-02

    申请人: HULU, LLC

    IPC分类号: G06K9/00 G06K9/62 G06K9/46

    摘要: In some embodiments, a first detector generates a first output based on a first probability that an image was inserted in a video. The first detector is trained with a set of known images to detect the set of known images. A second detector generates a second output based on a second probability that an image was inserted in the video. The second detector is used to detect the set of unknown images without training. The method analyzes the first output from the first detector based on the probability of the image existing in the video and the second output from the second detector based on the probability of the image existing in the video to generate a combined score from the first output and the second output. An indication of whether the image is detected in the video is output based on the combined score.

    Frame Level And Video Level Text Detection In Video

    公开(公告)号:US20200349381A1

    公开(公告)日:2020-11-05

    申请号:US16399823

    申请日:2019-04-30

    申请人: HULU, LLC

    IPC分类号: G06K9/34 G06K9/00

    摘要: In some embodiments, a method detects a first set of frames in a video that include lines of text, the detecting performed at a frame level on each individual frame. A first representation is generated from the first set of frames and a second representation is generated from the first set of frames. The method filters the first representation based on a number of lines of text within a space in the space dimension to select a second set of frames and filters the second representation based on a number of frames within time intervals in the time dimension to select a third set of frames. Frames in both the second set of frames and the third set of frames are analyzed to determine whether the lines of text in both the second set of frames and the third set of frames are burned-in subtitles.

    Scene level video search
    5.
    发明授权

    公开(公告)号:US10755104B2

    公开(公告)日:2020-08-25

    申请号:US16011181

    申请日:2018-06-18

    申请人: HULU, LLC

    摘要: In some embodiments, a method trains a first prediction network to predict similarity between images in videos. The training uses boundaries detected in the videos to train the prediction network to predict images in a same scene to have similar feature descriptors. The first prediction network generates feature descriptors that describe library images from videos in a video library offered to users of a video delivery service. A search image is received and the prediction network predicts one or more library images for one or more videos that are predicted to be similar to the received image. The one or more library images for the one or more videos are provided as a search result.

    MASKED MODEL TRAINING OF A PREDICTION NETWORK

    公开(公告)号:US20240137620A1

    公开(公告)日:2024-04-25

    申请号:US18402643

    申请日:2024-01-02

    申请人: HULU, LLC

    摘要: In some embodiments, a method receives a first sequence of inputs for processing via a sub-model of a plurality of sub-model. The plurality of sub-models are part of a main model. An input in the sequence of inputs is masked with a masked value to generate a second sequence of inputs. The method processes the second sequence of inputs using the sub-model to generate a sequence of features that correspond to the second sequence of inputs and processes the sequence of features to generate a first output. The first output is processed to generate a second output of the main model. The sub-model is trained based on a feature in the sequence of features that corresponds to the masked input and the second output.

    Representation Of Content Based On Content-Level Features

    公开(公告)号:US20210044870A1

    公开(公告)日:2021-02-11

    申请号:US16535008

    申请日:2019-08-07

    申请人: HULU, LLC

    摘要: In some embodiments, a method maps attributes of metadata for a plurality of content instances to metadata nodes. The metadata nodes are connected to type nodes that define a type of metadata for each metadata node and a respective content node for a respective content instance. The method generates a plurality of sample paths using the content nodes, the metadata nodes, and the type nodes from the mapping of attributes of the metadata. A similarity of content nodes is analyzed using the plurality of sample paths. Then, the method generates a representation of each of the plurality of content instances using the similarity of the content nodes. The representation represents the similarity between content instances in the plurality of content instances.

    Video motion effect generation based on content analysis

    公开(公告)号:US10721388B2

    公开(公告)日:2020-07-21

    申请号:US15934583

    申请日:2018-03-23

    申请人: HULU, LLC

    摘要: In one embodiment, a system detects objects in an image and generates attention regions that are positioned in the image based on first positions of the objects in the image. Focus points for the objects are generated for the attention regions at one or more second positions. Focus boxes are generated using the second positions of the focus points. Then, the system generates information for a motion effect using content of the image based on a number of the focus boxes and third positions of the focus boxes.