PERSONALIZED RETRIEVAL SYSTEM
    1.
    发明公开

    公开(公告)号:US20240346084A1

    公开(公告)日:2024-10-17

    申请号:US18398495

    申请日:2023-12-28

    Applicant: Roku, Inc.

    CPC classification number: G06F16/9035 G06F16/9038 G06F40/40

    Abstract: Disclosed are system, method and/or computer program product embodiments that retrieve items for a user based on a query using a two-tower deep machine learning model. An example embodiment provides input to a context tower, wherein the input includes the query and one or more of a query embedding corresponding to the query or a graph user embedding corresponding to the user. The context tower generates a context embedding in a vector space based on the input. The model determines a measure of similarity between the context embedding and each of a plurality of item embeddings in the vector space that are generated by an item tower and represent a plurality of candidate items. A relevancy score is calculated for each candidate item based on the measure of similarity between the context embedding and the corresponding item embedding. The relevancy scores are used for item retrieval and/or ranking.

    ENHANCING TRANSFER LEARNING FOR LARGE LANGUAGE MODELS

    公开(公告)号:US20250045575A1

    公开(公告)日:2025-02-06

    申请号:US18423802

    申请日:2024-01-26

    Applicant: Roku, Inc.

    Abstract: Pre-trained large language models may be trained on a large data set which may not necessarily align with specific tasks, business goals, and requirements. Pre-trained large language models can solve generic semantic relationship or question-answering type problems but may not be suited for content item retrieval or recommendation of content items that are semantically relevant to a query. It is possible to build a machine learning model while using transfer learning to learn from pre-trained large language models. Training data can significantly impact the performance of machine learning models, especially machine learning models developed using transfer learning. The training data can impact a model's performance, generalization, fairness, and adaptation to specific domains. To address some of these concerns, a popularity bucketing strategy can be implemented to debias training data. Optionally, an ensemble of models can be used to generate diverse training data.

    HETEROGENEOUS GRAPH NEURAL NETWORK USING OFFSET TEMPORAL LEARNING FOR SEARCH PERSONALIZATION

    公开(公告)号:US20240346309A1

    公开(公告)日:2024-10-17

    申请号:US18582249

    申请日:2024-02-20

    Applicant: Roku, Inc.

    CPC classification number: G06N3/08 G06N3/042

    Abstract: Disclosed herein are system, apparatus, article of manufacture, method and/or computer program product embodiments, and/or combinations and sub-combinations thereof, for training a heterogenous graph neural network (GNN) to generate user embeddings corresponding to users and item embeddings corresponding to items. An example embodiment generates a first user interaction graph for a first time window and a second user interaction graph for a second time window, wherein each graph represents users and items as nodes and user-item interactions within the respective time window as edges, samples user-item node pairs from the second user interaction graph, and trains the heterogeneous GNN based on user-item node pairs from the first user interaction graph that correspond to the sampled user-item node pairs from the second user interaction graph. User and item embeddings generated by the trained GNN may be used to determine a relevancy of a given item with respect to a given user.

    USING A LARGE LANGUAGE MODEL TO IMPROVE TRAINING DATA

    公开(公告)号:US20250045535A1

    公开(公告)日:2025-02-06

    申请号:US18423789

    申请日:2024-01-26

    Applicant: Roku, Inc.

    Abstract: Training data can significantly impact the performance of machine learning models. Its impact may be more significant in transfer learning. Different data sources can be used to generate training data used in transfer learning. The training data originating from user interaction logs may be subject to presentation bias. The training data originating from model generated labeled data may have false positives. Poor quality training data may cause the machine learning model to perform poorly. To address some of these concerns, a checker having one or more models can check for false positives and for labeled data entries that may have been subject to presentation bias. Such entries may be removed or modified. In some cases, the checker can generate a test that can be used to test the machine learning model and penalize the machine learning model if the model generates an incorrect prediction.

Patent Agency Ranking