Abstract:
Disclosed herein is a network selecting apparatus including: a determining unit checking an operation state of a cache server in which replicated data corresponding to a data request signal are stored and determining whether the cache server is normally operated, when the data request signal is received from a client; and a selecting unit selecting a first network so that the replicated data are transmitted to input and output devices when the cache server is normally operated and selecting a second network so that original data corresponding to the data request signal are transmitted from a storage server in which the original data are stored to the input and output devices when the cache server is not normally operated.
Abstract:
There are provided a system and method for providing a virtual desktop service using a cache server. A system for providing a virtual desktop service according to the invention includes a host server configured to provide a virtual desktop service to a client terminal using a virtual machine, a distributed file system configured to store data for the virtual machine, and a cache server that is provided for each host server group having at least one host server, and performs a read process or a write process of data using physically separate caches when the read process or write process of the data is requested from the virtual machine in the host server.
Abstract:
Disclosed is a quantum computing system including a first quantum chip including first physical qubits, a second quantum chip including second physical qubits, and a management device. The management device includes a physical qubit layer that manages physical qubit mapping including information about physical channels between first and second physical qubits, an abstraction qubit layer that manages abstraction qubit mapping including information about abstraction qubits and abstraction channels between the abstraction qubits based on the physical qubit mapping, a logical qubit layer that divides the abstraction qubits into logical qubits and to manage logical qubit mapping including information about logical channels between the logical qubits, based on the abstraction qubit mapping, and an application qubit layer that allocates at least one logical qubit corresponding to a qubit request received from a quantum application program based on the logical qubit mapping.
Abstract:
A method and apparatus are disclosed which relate generally to worker-scheduling technology in a serverless cloud-computing environment, and more particularly, to technology that allocates workers for executing functions on a micro-function platform which provides a function-level micro-service. The method and apparatus process a worker allocation task in a distributed manner as two-step pre-allocation schemes before a worker allocation request occurs, and pre-allocates workers required for a service using a function request period and a function execution time, thus minimizing scheduling costs incurred by worker allocation requests.
Abstract:
Disclosed herein is a method and apparatus for supporting virtualization. In the method, conversion of source code of a loadable module is initiated. A virtualization-sensitive instruction is searched for during the conversion of the source code. If the virtualization-sensitive instruction has been found, a virtualization-sensitive instruction table is generated based on the found virtualization-sensitive instruction. The virtualization-sensitive instruction is substituted with an instruction recognizable in a privileged mode, based on the generated virtualization-sensitive instruction table. The loadable module is loaded and executed in a kernel. Accordingly, the present invention supports virtualization, thus minimizing overhead occurring in full virtualization, and guaranteeing the high performance provided by para-virtualization without modifying a source.
Abstract:
Disclosed herein are an apparatus and method for setting the allocation rate of a parallel-computing accelerator. The method includes monitoring the utilization rate of the parallel-computing accelerator by an application and setting a start point, at which measurement of utilization data to be used for setting the allocation rate of the parallel-computing accelerator for the application is started, using the result of monitoring the utilization rate; setting an end point, at which the measurement of the utilization data is finished, based on the monitoring result; and setting the allocation rate of the parallel-computing accelerator using the utilization data measured during a time period from the start point to the end point.
Abstract:
Disclosed herein is a method and apparatus for supporting virtualization. In the method, conversion of source code of a loadable module is initiated. A virtualization-sensitive instruction is searched for during the conversion of the source code. If the virtualization-sensitive instruction has been found, a virtualization-sensitive instruction table is generated based on the found virtualization-sensitive instruction. The virtualization-sensitive instruction is substituted with an instruction recognizable in a privileged mode, based on the generated virtualization-sensitive instruction table. The loadable module is loaded and executed in a kernel. Accordingly, the present invention supports virtualization, thus minimizing overhead occurring in full virtualization, and guaranteeing the high performance provided by para-virtualization without modifying a source.
Abstract:
An apparatus and method for executing a function. The apparatus includes one or more processors and executable memory for storing at least one program executed by the one or more processors, and the at least one program is configured to determine whether it is possible to reengineer a user function source using interface description language (IDL) code, to generate a reengineered function source by reengineering the user function source, and to execute the reengineered function source.
Abstract:
Disclosed herein is a distributed file system using a torus network. The distributed file system includes multiple servers. The location of a master server may be determined to shorten the latency of data input/output. The location of the master server may be determined such that the distance between the master server and a node farthest away from the master server, among nodes, is minimized. When the location of the master server is determined, the characteristics of the torus network and the features of a propagation transmission scheme may be taken into consideration.