Abstract:
A controller can receive first and second metrics respectively indicating distributed computing system servers' CPU, memory, or disk utilization, throughput, or latency for a first time. The controller can receive third and fourth metrics for a second time. The controller can determine a first graph including vertices corresponding to the servers and edges indicating data flow between the servers, a second graph including edges indicating the first metrics satisfy a first threshold, a third graph including edges indicating the second metrics satisfy a second threshold, a fourth graph including edges indicating the third metrics fail to satisfy the first threshold, and a fifth graph including edges indicating the fourth metrics fail to satisfy the second threshold. The controller can display a sixth graph indicating at least one of first changes between the second graph and the fourth graph or second changes between the third graph and the fifth graph.
Abstract:
Systems, methods, and computer-readable media are disclosed for graph based monitoring and management of network components of a distributed streaming system. In one aspect, a method includes generating, by a processor, a first metrics and a second metrics based on data collected on a system; generating, by the processor, a topology graph representing data flow within the system; generating, by the processor, at least one first metrics graph corresponding to the first metrics based in part on the topology graph; generating, by the processor, at least one second metrics graph corresponding to the second metrics based in part on the topology graph; identifying, by the processor, a malfunction within the system based on a change in at least one of the first metrics graph and the second metrics graph; and sending, by the processor, a feedback on the malfunction to an operational management component of the system.
Abstract:
In one embodiment, a method implements virtualized network functions in a serverless computing system having networked hardware resources. An interface of the serverless computing system receives a specification for a network service including a virtualized network function (VNF) forwarding graph (FG). A mapper of the serverless computing system determines an implementation graph comprising edges and vertices based on the specification. A provisioner of the serverless computing system provisions a queue in the serverless computing system for each edge. The provisioner further provisions a function in the serverless computing system for each vertex, wherein, for at least one or more functions, each one of said at least one or more functions reads incoming messages from at least one queue. The serverless computing system processes data packets by the queues and functions in accordance with the VNF FG. The queues and functions processes data packets in accordance with the VNF FG.
Abstract:
In one embodiment, a device in a network receives a first plurality of measurements for network metrics captured during a first time period. The device determines a first set of correlations between the network metrics using the first plurality of measurements captured during the first time period. The device receives a second plurality of measurements for the network metrics captured during a second time period. The device determines a second set of correlations between the network metrics using the second plurality of measurements captured during the second time period. The device identifies a difference between the first and second sets of correlations between the network metrics as a network anomaly.
Abstract:
Approaches are disclosed for improving performance of logical disks. A logical disk can comprise several storage devices. In an object storage system (OSS), when a logical disk stores a file, fragments of the file are stored distributed across the storage devices. Each of the fragments of the file is asymmetrically stored in (write) and retrieved from (read) the storage devices. The performance of the logical disk is improved by reconfiguring one or more of the storage devices based on an influence that each of the storage devices has on performance of the logical disk and the asymmetric read and write operations of each of the storage devices. For example, latency of the logical disk can be reduced by reconfiguring one or more of the plurality of storage disks based on a proportion of the latency of the logical device that is attributable to each of the plurality of storage devices.
Abstract:
The present disclosure describes a distributed, advertisement-based, solution for scheduling virtual resources in cloud infrastructures such as the OpenStack. The scheduling algorithm distributes the scheduling requirements and host state feasibility checks to the individual hosts in the datacenter, which can periodically send a summarized advertisement to the scheduler controller listing the number of instances of different type(s) of virtual resources that a particular host can support. The scheduler controller, thus no longer has to compute and maintain individual host states, and the scheduling problem is reduced to selecting the feasible advertisements that satisfy a given request. The solution can be extended to a scenario of multiple scheduler controllers using the same distributed, advertisement-based, approach.
Abstract:
Embodiments include receiving an indication of a data storage module to be associated with a tenant of a distributed storage system, allocating a partition of a disk for data of the tenant, creating a first association between the data storage module and the disk partition, creating a second association between the data storage module and the tenant, and creating rules for the data storage module based on one or more policies configured for the tenant. Embodiments further include receiving an indication of a type of subscription model selected for the tenant, and selecting the disk partition to be allocated based, at least in part, on the subscription model selected for the tenant. More specific embodiments include generating a storage map indicating the first association between the data storage module and the disk partition and indicating the second association between the data storage module and the tenant.
Abstract:
Techniques are provided to generate and store a network graph database comprising information that indicates a service node topology, and virtual or physical network services available at each node in a network. A service request is received for services to be performed on packets traversing the network between at least first and second endpoints. A subset of the network graph database is determined that can provide the services requested in the service request. A service chain and service chain identifier is generated for the service based on the network graph database subset. A flow path is established through the service chain by flow programming network paths between the first and second endpoints using the service chain identifier.
Abstract:
In one embodiment, data indicative of the size of an intermediate data set generated by a first resource device is received at a computing device. The intermediate data set is associated with a virtual machine to process the intermediate data set. A virtual machine configuration is determined based on the size of the intermediate data set. A second resource device is selected to execute the virtual machine based on the virtual machine configuration and on an available bandwidth between the first and second resource devices. The virtual machine is then assigned to the second resource device to process the intermediate data set.
Abstract:
An example method for leveraging hardware accelerators for scalable distributed stream processing in a network environment is provided and includes allocating a plurality of hardware accelerators to a corresponding plurality of bolts of a distributed stream in a network, facilitating a handshake between the hardware accelerators and the corresponding bolts to allow the hardware accelerators to execute respective processing logic according to the corresponding bolts, and performing elastic allocation of hardware accelerators and load balancing of stream processing in the network. The distributed stream comprises a topology of at least one spout and the plurality of bolts. In specific embodiments, the allocating includes receiving capability information from the bolts and the hardware accelerators, and mapping the hardware accelerators to the bolts based on the capability information. In some embodiments, facilitating the handshake includes executing a shadow process to interface between the hardware accelerator and the distributed stream.