AUTOMATIC ML PIPELINE PLANNING FOR LIVE ML ANALYTICS

    公开(公告)号:US20240412096A1

    公开(公告)日:2024-12-12

    申请号:US18208173

    申请日:2023-06-09

    Abstract: Optimizing ML pipeline deployment using an ML pipeline management system. A method includes receiving an indication of an input data source and input data type from the input data source. An indication of a plurality filters to be included in the pipeline, an ML model, and predetermined performance criteria is received. The method includes determining a physical topology of the ML pipeline and configuration of the filters or the ML model. The determined physical topology includes placement of the filters and the model, and the configuration. The determined physical topology satisfies the performance criteria. The filters and ML model are placed across an infrastructure, comprising a plurality of tiers, according to the determined physical topology.

    MERGING MODELS ON AN EDGE SERVER
    4.
    发明申请

    公开(公告)号:US20220383188A1

    公开(公告)日:2022-12-01

    申请号:US17471816

    申请日:2021-09-10

    Abstract: Systems and methods are provided for merging models for use in an edge server under the multi-access edge computing environment. In particular, a model merger selects a layer of a model based on a level of memory consumption in the edge server and determines sharable layers based on common properties of the selected layer. The model merger generates a merged model by generating a single instantiation of a layer that corresponds to the sharable layers. A model trainer trains the merged model based on training data for the respective models to attain a level of accuracy of data analytics above a predetermined threshold. The disclosed technology further refreshes the merged model upon observing a level of data drift that exceeds a predetermined threshold. The refreshing of the merged model includes detaching and/or splitting consolidated sharable layers of sub-models in the merged model. By merging models, the disclosed technology reduces memory footprints of models used in the edge server, rectifying memory scarcity issues in the edge server.

Patent Agency Ranking