Abstract:
An application programming interface may receive workload identifiers and checkpoint identifiers from which bottleneck detection may be performed. Workloads may be tracked through various checkpoints in an application and timestamps collected at each checkpoint. From these data, bottlenecks may be identified in real time or by analyzing the data in a subsequent analysis. The workloads may be processed by multiple devices which may comprise a large application. In some cases, the workloads may be processed by different devices in sequence or in a serial fashion, while in other cases workloads may be processed in parallel by different devices. The application programming interface may be part of a bottleneck detection service which may be sold on a pay-per-use model, a subscription model, or some other payment scheme.
Abstract:
The disclosure provides a micro-processing system operable in a hardware decoder mode and in a translation mode. In the hardware decoder mode, the hardware decoder receives and decodes non-native ISA instructions into native instructions for execution in a processing pipeline. In the translation mode, native translations of non-native ISA instructions are executed in the processing pipeline without using the hardware decoder. The system includes a code portion profile stored in hardware that changes dynamically in response to use of the hardware decoder to execute portions of non-native ISA code. The code portion profile is then used to dynamically form new native translations executable in the translation mode.
Abstract:
A high-performance-computer system includes a statistics accumulation apparatus configured to efficiently accumulate system performance data from a variety of system components, and periodically write such data to processor local memory for efficient subsequent software processing of the thus acquired data, thereby reducing the system hardware and software overhead needed for collection of such data as compared to prior art systems.
Abstract:
A method and system for controlling access to an application being executed by a server. A time interval between the server's receipt times of a first request and a second request. The first and second requests are included in multiple requests for an access to the application. An adjustment value is selected based on the time interval and multiple historical time intervals. Each historical time interval is between the server's receipt times of two requests of the multiple requests. The two requests' receipt times are prior to the first and second requests' receipt times. An accumulator is adjusted with the adjustment value, resulting in the accumulator's value exceeding a predefined threshold. Access to the application is controlled via a warning, a delay in responding to the second request, a temporary halt in processing, or a permanent access denial in response to the accumulator's value exceeding the threshold.
Abstract:
Embodiments of the invention provide a method, system and computer program product for embedding a global barrier and global interrupt network in a parallel computer system organized as a torus network. The computer system includes a multitude of nodes. In one embodiment, the method comprises taking inputs from a set of receivers of the nodes, dividing the inputs from the receivers into a plurality of classes, combining the inputs of each of the classes to obtain a result, and sending said result to a set of senders of the nodes. Embodiments of the invention provide a method, system and computer program product for embedding a collective network in a parallel computer system organized as a torus network. In one embodiment, the method comprises adding to a torus network a central collective logic to route messages among at least a group of nodes in a tree structure.
Abstract:
Aspects of an application program's execution which might be subject to non-determinism are performed in a deterministic manner while the application program's execution is being recorded in a virtual machine environment so that the application program's behavior, when played back in that virtual machine environment, will duplicate the behavior that the application program exhibited when originally executed and recorded. Techniques disclosed herein take advantage of the recognition that only minimal data needs to be recorded in relation to the execution of deterministic operations, which actually can be repeated “verbatim” during replay, and that more highly detailed data should be recorded only in relation to non-deterministic operations, so that those non-deterministic operations can be deterministically simulated (rather than attempting to re-execute those operations under circumstances where the outcome of the re-execution might differ) based on the detailed data during replay.
Abstract:
A server computer system identifies change operations for an object in a file system. The object can be a file or a directory. The change operations can include a change to a local copy of the object and one or more remote copies of the object. The server computer system determines that one of the change operations is unsuccessful and creates tracking data that identifies the object that is associated with at least one change operation that is unsuccessful.
Abstract:
Software performance may be improved by collecting and correlating performance counter events and program execution state information. On each successive callback of performance counter data, a statistical exponential moving average association matrix may be updated until particular performance counter events and program execution states may be correlated.
Abstract:
The present invention measures an actual utilization frequency of data and controls a location of this data in a storage apparatus in a case where a host computer makes joint use of a storage apparatus and a cache apparatus. A portion of data used by an application program 1A is stored in a storage apparatus 2 and a cache apparatus 3. A management apparatus 4 detects an I/O load of a page (4A), and detects an I/O load of cache data (4B). The management apparatus 4 determines a corresponding relationship between the page and the cache data (4C), and adds the I/O load of the cache data to the I/O load of the page.
Abstract:
A process-requesting apparatus for requesting a process-performing apparatus to perform a predefined process and querying a progress status of the predefined process includes a progress status obtaining unit for obtaining, as a response to the query about the progress status to the process-performing apparatus, the progress status from the process-performing apparatus; a completion determining unit for determining whether the predefined process has been completed based on the obtained progress status; a time interval determining unit for determining a time interval from the last time the process-requesting apparatus queried the progress status to the next time the process-requesting apparatus queries the progress status according to an elapsed time from the start of the predefined process; and a progress status querying unit for, in the case where the predefined process has not been completed, querying the process-performing apparatus about the progress status at the determined time interval.