Container image distribution acceleration

    公开(公告)号:US10291706B1

    公开(公告)日:2019-05-14

    申请号:US15079424

    申请日:2016-03-24

    IPC分类号: G06F15/16 H04L29/08

    摘要: A container image registry is maintained at each host device in a set of host devices in a container hosting environment. A container image is composed of one or more container image layers such that each container image registry of each host device is configured to be able to store one or more different layers of one or more different container images. The method tracks which container image registry at which host device contains which container image layers. At each host device, one or more container images are obtained for launching one or more application programs. For a given host device, the one or more container images are obtained from one or more of: (i) one or more container image layers stored in its container image registry; and (ii) one or more container image layers stored in the container image registry of another host device.

    Queue-based GPU virtualization and management system

    公开(公告)号:US10109030B1

    公开(公告)日:2018-10-23

    申请号:US15391223

    申请日:2016-12-27

    IPC分类号: G06T1/20 G06T1/60

    摘要: A method implemented by a server enables sharing of GPU resources by multiple clients. The server receives a request from a first client for GPU services. The request includes a first block of GPU code of an application executing on the first client. A first task corresponding to the first block of GPU code is enqueued in a task queue. The task queue includes a second task that corresponds to a second block of GPU code of an application executing on a second client. The server schedules a time for executing the first task using a GPU device that is assigned to the first client, and dispatches the first task to a GPU worker process to execute the first task at the scheduled time using the GPU device. The GPU device is shared, either temporally or spatially, by the first and second clients for executing the first and second tasks.

    Managing access to a resource pool of graphics processing units under fine grain control

    公开(公告)号:US10262390B1

    公开(公告)日:2019-04-16

    申请号:US15487887

    申请日:2017-04-14

    IPC分类号: G06F9/48 G06F9/50 G06T1/20

    摘要: A graphics processing unit (GPU) service platform includes a control server, and a cluster of GPU servers each having one or more GPU devices. The control server receives a service request from a client system for GPU processing services, allocates multiple GPU servers nodes within the cluster to handle GPU processing tasks specified by the service request by logically binding the allocated GPU server nodes, and designating one of the at least two GPU servers as a master server, and send connection information to the client system to enable the client system to connect to the master server. The master GPU server node receives a block of GPU program code transmitted from the client system, which is associated with the GPU processing tasks specified by the service request, processes the block of GPU program code using the GPU devices of the logically bound GPU servers, and returns processing results to the client system.

    Method and device for data center management

    公开(公告)号:US11226882B2

    公开(公告)日:2022-01-18

    申请号:US15638794

    申请日:2017-06-30

    摘要: Embodiments of the present disclosure provide a method and device for data center management. For example, there is provided a method, comprising: obtaining information of hardware used in a data center, the information of the hardware including identification information describing an identifiable attribute of the hardware; identifying the hardware by matching the identification information with a resource profile, the resource profile recording identifiable attributes of a plurality of types of hardware; and updating a record associated with the hardware in a database of the data center using the obtained information. Corresponding device and computer program product are also provided.

    MANAGING ACCESS TO A RESOURCE POOL OF GRAPHICS PROCESSING UNITS UNDER FINE GRAIN CONTROL

    公开(公告)号:US20190197655A1

    公开(公告)日:2019-06-27

    申请号:US16287719

    申请日:2019-02-27

    IPC分类号: G06T1/20 G06F9/48 G06F9/50

    摘要: A graphics processing unit (GPU) service platform includes a control server, and a cluster of GPU servers each having one or more GPU devices. The control server receives a service request from a client system for GPU processing services, allocates multiple GPU servers within the cluster to handle GPU processing tasks specified by the service request by logically binding the allocated GPU servers, and designating one of the at least two GPU servers as a master server, and send connection information to the client system to enable the client system to connect to the master server. The master GPU server receives a block of GPU program code transmitted from the client system, which is associated with the GPU processing tasks specified by the service request, processes the block of GPU program code using the GPU devices of the logically bound GPU servers, and returns processing results to the client system.

    Intelligent data coordination for accelerated computing in cloud environment

    公开(公告)号:US10891156B1

    公开(公告)日:2021-01-12

    申请号:US15498055

    申请日:2017-04-26

    IPC分类号: G06F9/48 G06F9/50

    摘要: Systems and methods are provided to implement intelligent data coordination for accelerated computing in a distributed computing environment. For example, a method includes executing a task on a computing node, monitoring requests issued by the executing task, intercepting requests issued by the executing task which correspond to data flow operations to be performed as part of the task execution, and asynchronously executing the intercepted requests at scheduled times to coordinate data flow between resources on the computing node.

    Managing access to a resource pool of graphics processing units under fine grain control

    公开(公告)号:US10467725B2

    公开(公告)日:2019-11-05

    申请号:US16287719

    申请日:2019-02-27

    IPC分类号: G06T1/20 G06F9/50 G06F9/48

    摘要: A graphics processing unit (GPU) service platform includes a control server, and a cluster of GPU servers each having one or more GPU devices. The control server receives a service request from a client system for GPU processing services, allocates multiple GPU servers within the cluster to handle GPU processing tasks specified by the service request by logically binding the allocated GPU servers, and designating one of the at least two GPU servers as a master server, and send connection information to the client system to enable the client system to connect to the master server. The master GPU server receives a block of GPU program code transmitted from the client system, which is associated with the GPU processing tasks specified by the service request, processes the block of GPU program code using the GPU devices of the logically bound GPU servers, and returns processing results to the client system.

    Checkpointing for GPU-as-a-service in cloud computing environment

    公开(公告)号:US10275851B1

    公开(公告)日:2019-04-30

    申请号:US15496568

    申请日:2017-04-25

    IPC分类号: G06T1/20 G06T1/60 G09G5/00

    摘要: Techniques are provided to configure a graphics processing unit (GPU) service platform to provide GPU processing and checkpointing services to a client system. For example, when the GPU service platform receives a service request from a first client system for GPU processing services, a GPU server node is allocated to execute a first GPU processing task associated with the service request using one or more GPU devices. When a GPU checkpointing operation is performed during execution of the first GPU processing task to generate a checkpoint image of a current state of the first GPU processing task, the GPU server node is configured to execute at least a portion of a second GPU processing task associated with a second service request received from a second client system, using the one or more GPU devices concurrently with performing the GPU checkpointing operation for the first GPU processing task.

    Container image layer compaction
    10.
    发明授权

    公开(公告)号:US10534671B1

    公开(公告)日:2020-01-14

    申请号:US15195451

    申请日:2016-06-28

    申请人: EMC Corporation

    IPC分类号: G06F17/00 G06F7/00 G06F11/14

    摘要: A method for managing container images in a container hosting environment, comprising a set of machine nodes for hosting container images, comprises the following steps. At a machine node of the set of machine nodes, at least one image layer of a container image comprising multiple image layers is identified for compaction. The identified image layer is compacted by merging data of the identified layer with data of another image layer of the container image. The data of the compacted image layer is transmitted to a backup store remote from the machine node for storage. The set of machine nodes and the backup store are implemented via processing devices operatively coupled via a communication network to form the container hosting environment. The compacted image layer may then be recovered from the backup store.