Abstract:
A multiple channel data transfer system (10) includes a source (12) that generates data packets with sequence numbers for transfer over multiple request channels (14). Data packets are transferred over the multiple request channels (14) through a network (16) to a destination (18). The destination (18) re-orders the data packets received over the multiple request channels (14) into a proper sequence in response to the sequence numbers to facilitate data processing. The destination (18) provides appropriate reply packets to the source (12) over multiple response channels (20) to control the flow of data packets from the source (12).
Abstract:
A high performance computing (HPC) system includes computing blades having a first region that includes processors for performing a computation, and a second region that includes non-volatile memory for use in performing the computation and another computing processor for performing data movement and storage. Because data movement and storage are offloaded to the secondary processor, the processors for performing the computation are not interrupted to perform these tasks. A method for use in the HPC system receives instructions in the computing processors and first data in the memory. The method includes receiving second data into the memory while continuing to execute the instructions in the computing processors, without interruption. A computer program product implementing the method is also disclosed.
Abstract:
A method of distributing threads among accelerators in a high performance computing system receives a request to assign an accelerator in the computing system to a thread. The request includes a mode indicative of location and exclusivity of the accelerator for use by the thread. The method selects the accelerator according to a processor assigned to the thread. The method also assigns the accelerator to the thread with the exclusivity specified in the request.
Abstract:
Processor clock signals are generated for each processor in a HPC system, such that all the processor clock signals are of the same frequency. Furthermore, as part of a startup (boot) procedure, a process sets all time stamp counters (TSCs) of the processors, such they indicate identical times. Each blade of the HPC system recovers a recovered clock signal from a synchronous communication network, to which the blade is coupled. The blade generates a processor clock from the recovered clock signal and provides the processor clock to processor(s) on the blade. Each chassis is coupled to a second, system-wide, synchronous communication network, and each chassis synchronizes its chassis synchronous communication network with the system-wide synchronous communication system. Thus, all the processor clock signals are generated with the same frequency.
Abstract:
A power consumption threshold is implemented to manage power consumed by a plurality of devices. A power consumption threshold may be selected for a data storage system having multiple drives. Policies may control operation of storage devices such as hard disk drives to ensure the power consumption threshold is not exceeded. The policies may implement procedures for scheduling hard disk drive operations based on disk drive power characteristics, scheduling maintenance tasks, managing device power states, and strategically scheduling device operations based on their current state. The policies may be implemented by a data manager application in communication with multiple tiers of a data storage system.
Abstract:
Embodiments of the invention includes identifying the priority of data sets based on how frequently they are accessed by data center compute resources or by other measures assigning latency metrics to data storage resources accessible by the data center, moving data sets with the highest priority metrics to data storage resources with the fastest latency metrics, and moving data sets with lower priority metrics to slower data storage resources with slower latency metrics. The invention also may be compatible with or enable new forms of related applications and methods for managing the data center.
Abstract:
A system reduces the impact of constrained bandwidth to long-term data storage without adding new data storage resources to the data center, typically by temporarily storing data on data storage devices that are contained within a desktop computer, a notebook computer, or other computing device. The invention stores lower priority data sets temporarily on data storage devices that are already purchased or expensed until lower priority data sets can be migrated to long-term data storage. The invention relieves the performance impact of congestion caused by slow communication interfaces, recording channels, and mechanical systems that move tape cartridges around. The invention may also be configured with security functions that restrict where or how certain data sets are stored temporarily.
Abstract:
Quotas are tracked for user usage of hard disk drive space and offline backup storage space. The quota is enforced against the total space utilized by a user, not just high tier hard drive space usage. When data is migrated from hard disk drive space to backup storage space, data metadata is updated to reflect data kept offline for the user. As such, when users request to store new data, the data usage of hard disk space and backup storage space is determined from the metadata that reflects both data types, and the total storage spaced for the user is used to grant or reject the user's request to store more data in the system.
Abstract:
An associative look-up instruction for an instruction set architecture (ISA) of a processor and methods for use of an associative look-up instruction. The associative look-up instruction of the ISA specifies one or more fields within a data unit that are used as a pattern of bits for identifying data content in a memory structure to be loaded into hardware registers or other storage components of the ISA. Specified parameters of the associative operation may be explicit within the instruction or indirectly pointed to via hardware registers or other storage components of the ISA. The memory structure may be content addressable memory (CAM).
Abstract:
Embodiments of the invention includes a plurality of connectors configured to connect a plurality of data storage host bus adaptors to a plurality of data storage device subassemblies such that at least one lane of low voltage differential signal pairs from each of the plurality of host bus adaptors is connected to each of the data storage device subassemblies. The invention improves the electrical interconnections in a data storage array such as a JBOD enclosure or data storage server. The invention minimizes the number of connectors by reducing the number of printed circuit boards, and eliminates the need to add signal repeaters to maintain signal quality. The invention also increases the cooling efficiency of the enclosure by increasing air flow by reducing the number of printed circuit boards in the data storage array.