-
公开(公告)号:US20240345901A1
公开(公告)日:2024-10-17
申请号:US18757299
申请日:2024-06-27
发明人: David Dice , Alex Kogan
CPC分类号: G06F9/52 , G06F9/5022 , G06F9/524
摘要: Generic Concurrency Restriction (GCR) may divide a set of threads waiting to acquire a lock into two sets: an active set currently able to contend for the lock, and a passive set waiting for an opportunity to join the active set and contend for the lock. The number of threads in the active set may be limited to a predefined maximum or even a single thread. Generic Concurrency Restriction may be implemented as a wrapper around an existing lock implementation. Generic Concurrency Restriction may, in some embodiments, be unfair (e.g., to some threads) over the short term, but may improve the overall throughput of the underlying multithreaded application via passivation of a portion of the waiting threads.
-
公开(公告)号:US20240345881A1
公开(公告)日:2024-10-17
申请号:US18751415
申请日:2024-06-24
申请人: Next Silicon Ltd
发明人: Elad RAZ , Ilan TAYARI , Dan SHECHTER
CPC分类号: G06F9/5016 , G06F9/5022 , G06F12/0253
摘要: There is provided a memory, comprising: issuing an allocation operation for allocation of a region of a memory by a first process of a plurality of first processes executed in parallel on a first processor, sending a message to a second processor indicating the allocation of the region of the pool of the memory, issuing a free operation for release of the allocated region of the pool of the memory by a second process of a plurality of second processes executed in parallel on a second processor, and releasing, by the first processor, the allocated region of the pool of the memory as indicated in the free operation, wherein a same region of memory is allocated by the first process and released by the second process, wherein the first processes are concurrently attempting to issue the allocation operation and the second processes are concurrently attempting to issue the free operation.
-
公开(公告)号:US20240320044A1
公开(公告)日:2024-09-26
申请号:US18673214
申请日:2024-05-23
申请人: REBELLIONS INC.
发明人: Minhoo Kang
IPC分类号: G06F9/50
CPC分类号: G06F9/5016 , G06F9/5022 , G06F9/5044
摘要: A method for allocating on-chip memory of a neural processing unit is performed by one or more processors, and includes deallocating an allocated chunk in an on-chip memory area, which is finished with the use of the memory, and converting it into a cached chunk, receiving an on-chip memory allocation request for specific data, determining whether there is a cached chunk of one or more cached chunks that is allocable for the specific data, based on a comparison between a size of the specific data and the size of the one or more cached chunks, and based on a result of determining whether there is the cached chunk that is allocable for the specific data, allocating the specific data to a specific cached chunk of the one or more cached chunks, or allocating the specific data to at least a portion of the free chunk.
-
公开(公告)号:US20240311185A1
公开(公告)日:2024-09-19
申请号:US18624288
申请日:2024-04-02
申请人: SONICWALL INC.
发明人: Miao Mao , Wei Zhou , Zhong Chen
CPC分类号: G06F9/4881 , G06F9/455 , G06F9/46 , G06F9/461 , G06F9/466 , G06F9/48 , G06F9/4806 , G06F9/4843 , G06F9/485 , G06F9/4887 , G06F9/50 , G06F9/5005 , G06F9/5022 , G06F9/5027 , G06F9/505
摘要: Methods and apparatus consistent with the present disclosure may be used in environments where multiple different virtual sets of program instructions are executed by shared computing resources. These methods may allow actions associated with a first set of virtual software to be paused to allow a second set of virtual software to be executed by the shared computing resources. In certain instances, methods and apparatus consistent with the present disclosure may manage the operation of one or more sets of virtual software at a point in time. Apparatus consistent with the present disclosure may include a memory and one or more processors that execute instructions out of the memory. At certain points in time, a processors of a computing system may pause a virtual process while allowing instructions associated with another virtual process to be executed.
-
公开(公告)号:US20240272997A1
公开(公告)日:2024-08-15
申请号:US18612770
申请日:2024-03-21
申请人: Shopify Inc.
发明人: Chen-Yui YANG , Paritosh BAJPAY , David H. LU , Chaoxin QIU
CPC分类号: G06F11/203 , G06F9/45558 , G06F9/5022 , G06F9/5077 , G06F11/1658 , G06F11/202 , G06F11/2033 , G06F2009/4557 , G06F2009/45595 , G06F2201/815
摘要: Concepts and technologies are disclosed herein for providing a network virtualization policy management system. An event relating to a service can be detected. A first policy that defines allocation of hardware resources to host the virtual network functions can be obtained, as can a second policy that defines deployment of the virtual network functions to the hardware resources. The hardware resources can be allocated based upon the first policy and the virtual network functions can be deployed to the hardware resources based upon the second policy.
-
公开(公告)号:US12026552B1
公开(公告)日:2024-07-02
申请号:US18389676
申请日:2023-12-19
申请人: REBELLIONS INC.
发明人: Minhoo Kang
CPC分类号: G06F9/5016 , G06F9/5022 , G06F9/5044
摘要: A method for allocating on-chip memory of a neural processing unit is performed by one or more processors, and includes deallocating an allocated chunk in an on-chip memory area, which is finished with the use of the memory, and converting it into a cached chunk, receiving an on-chip memory allocation request for specific data, determining whether there is a cached chunk of one or more cached chunks that is allocable for the specific data, based on a comparison between a size of the specific data and the size of the one or more cached chunks, and based on a result of determining whether there is the cached chunk that is allocable for the specific data, allocating the specific data to a specific cached chunk of the one or more cached chunks, or allocating the specific data to at least a portion of the free chunk.
-
公开(公告)号:US12020077B1
公开(公告)日:2024-06-25
申请号:US17246459
申请日:2021-04-30
申请人: SPLUNK INC.
发明人: Sumeet Rohatgi , Gaurav Chawla , Mehul Goyal , Tyler Bakke , Aditya Kamble , Hao Zhan , Rahul Merulinath , Ryan Flynn , Matteo Merli , Boyang Peng , Sanjeev Kulkarni , Ivan Kelly , Karthikeyan Ramasamy
CPC分类号: G06F9/505 , G06F9/5022 , G06F9/5038 , G06F11/1492 , G06F11/203 , G06F11/3428
摘要: A process for facilitating autoscaling in a stateful system is described herein. In embodiments, a set of metrics associated with a set of components of a stateful service is obtained. The set of metrics may generally indicate a utilization or a load of each of the components of the set of components (e.g., message managers and/or data stores). Thereafter, it is determined to initiate a scaling event at the stateful service in association with the set of components of the stateful service based on at least a portion of the set of metrics attaining a metric threshold indicating a threshold value for determining whether to scale stateful service components. A scaling request can then be provided to the stateful service to initiate the scaling event at the stateful service in association with the set of components of the stateful service.
-
公开(公告)号:US11995476B1
公开(公告)日:2024-05-28
申请号:US17482276
申请日:2021-09-22
发明人: Ramyanshu Datta , Ishaaq Chandy , Arvind Sowmyan , Wei You , Kunal Mehrotra , Kohen Berith Chia , Andrea Olgiati , Lakshmi Naarayanan Ramakrishnan , Saurabh Gupta
CPC分类号: G06F9/5038 , G06F9/5022 , G06F9/5055
摘要: A post-task-completion retention period for which a computing resource is to be retained, without de-activating the resource, on behalf of a set of requesters of machine learning tasks is determined at a machine learning service. A first task, identified at the service prior to expiration of the retention period at a first computing resource at which a second task has completed, is initiated at the first computing resource. In response to obtaining an indication of a third task and determining that a threshold criterion associated with the retention period satisfies a criterion, the third task is initiated at an additional computing resource. The additional computing resource is de-activated after the third task completes, without waiting for the retention period to expire.
-
公开(公告)号:US20240160458A1
公开(公告)日:2024-05-16
申请号:US18508155
申请日:2023-11-13
申请人: Nutanix, Inc.
发明人: Mohit ARON , Dheeraj PANDEY , Ajeet SINGH
IPC分类号: G06F9/455 , G06F3/06 , G06F9/48 , G06F9/50 , G06F16/11 , G06F16/174 , G06F16/18 , H04L49/356
CPC分类号: G06F9/455 , G06F3/06 , G06F3/0638 , G06F3/064 , G06F3/0641 , G06F9/45533 , G06F9/45558 , G06F9/48 , G06F9/4806 , G06F9/4843 , G06F9/485 , G06F9/4856 , G06F9/4881 , G06F9/50 , G06F9/5005 , G06F9/5011 , G06F9/5016 , G06F9/5022 , G06F9/5027 , G06F9/5061 , G06F9/5072 , G06F9/5077 , G06F9/5083 , G06F9/5088 , G06F16/128 , G06F16/1748 , G06F16/1873 , H04L49/356 , G06F2009/45562 , G06F2009/4557 , G06F2009/45575 , G06F2009/45579 , G06F2009/45583 , G06F2009/45595
摘要: Disclosed is an improved approach to implement I/O and storage device management in a virtualization environment. According to some approaches, a Service VM is employed to control and manage any type of storage device, including directly attached storage in addition to networked and cloud storage. The Service VM implements the Storage Controller logic in the user space, and can be migrated as needed from one node to another. IP-based requests are used to send I/O request to the Service VMs. The Service VM can directly implement storage and I/O optimizations within the direct data access path, without the need for add-on products.
-
公开(公告)号:US11972301B2
公开(公告)日:2024-04-30
申请号:US17229551
申请日:2021-04-13
发明人: Yuwen Yang , Gurpreet Virdi , Bo Qiao , Hang Dong , Karthikeyan Subramanian , Marko Lalic , Shandan Zhou , Si Qin , Thomas Moscibroda , Yunus Mohammed
CPC分类号: G06F9/5077 , G06F9/45558 , G06F9/50 , G06F9/5038 , G06F9/505 , G06F2009/4557 , G06F2009/45583 , G06F2009/45595 , G06F9/5022 , G06F2209/5019
摘要: The present disclosure relates to systems, methods, and computer readable media for predicting surplus capacity on a set of server nodes and determining a quantity of deferrable virtual machines (VMs) that may be scheduled over an upcoming period of time. This determination of VM quantity may be determined while minimizing risks associated with allocation failures on the set of server nodes. This disclosure described systems that facilitate features and functionality related to improving utilization of surplus resource capacity on a plurality of server nodes by implementing VMs having some flexibility in timing of deployment while also avoiding significant risk caused as a result of over-allocated storage and computing resources. In one or more embodiments, the quantity of deferrable VMs is determined and scheduled in accordance with rules of a scheduling policy.
-
-
-
-
-
-
-
-
-