Abstract:
PROBLEM TO BE SOLVED: To provide a method and a system for providing a processor book of a multiprocessor used as a building block for a large-scale data processing system. SOLUTION: The processor book is generated by using two 4 way-multichip modules (MCM). First and second MCM are constructed by using regular wiring between processors. Outer buses of respective chips in first MCM are connected to buses of corresponding chips of second MCM, and additional wiring which connects them reversely in a similar way is provided. The respective processors of first MCM can substantially directly access distributed memory structure elements of next MCM, which do not have affinity, by additional wiring. The processor book is plugged in a processor rack constituted to receive a plurality of the processor books. A plurality of the processor books collectively constitute the large-scale data processing system. COPYRIGHT: (C)2005,JPO&NCIPI
Abstract:
PROBLEM TO BE SOLVED: To provide an improved processor, a data processing system and a data processing method. SOLUTION: An integrated circuit such as the processor includes a substrate and an integrated circuit element formed in the substrate. The integrated circuit element includes a processor core executing an instruction, an interconnect interface coupled to the processor core and supporting a communication between the processor core and a system interconnect external to the integrated circuit and, at least, a part of an external communication adapter coupled to the processor core and supporting input/output communication via an input/output communication link. COPYRIGHT: (C)2004,JPO&NCIPI
Abstract:
PROBLEM TO BE SOLVED: To provide an improved data processing system architecture reducing waiting time of communication between physically separating processors, reducing bus bandwidth consumption, and releasing the bus bandwidth for a general data transfer between the processor and a hierarchical memory system. SOLUTION: The identical processing communication information useful in pipelined multiprocessing or parallel multiprocessing is stored in each processor communication register (PCR). Each processor possesses an exclusive right to store to a sector within each PCR within a cluster network and has continuous access to read the PCR contents of itself. Each processor updates its exclusive sector within all of the PCRs via a private protocol or dedicated wireless network, makes all other processors within the cluster network to be able to quickly see the change within the PCR data and bypasses a cache subsystem. COPYRIGHT: (C)2004,JPO&NCIPI
Abstract:
PROBLEM TO BE SOLVED: To provide an improved data processing system architecture reducing waiting time of communication between physically separating processors, reducing bus bandwidth consumption, and releasing the bus bandwidth for a general data transfer between the processor and a hierarchical memory system. SOLUTION: The identical processing communication information useful in pipelined multiprocessing or parallel multiprocessing is stored in each processor communication register (PCR). Each processor updates its exclusive sector within all of the PCRs, makes all other processors to be able to quickly see the change within the PCR data and bypasses a cache subsystem. BY temporarily restricting access to the information or by forcing all the processors to continuously compete and providing processor communication quickly transferred to all the processors, efficiency of the multiprocessor system can be improved. COPYRIGHT: (C)2004,JPO&NCIPI
Abstract:
PROBLEM TO BE SOLVED: To provide a method for avoiding livelock within a multiprocessor data processing system when there are multiple concurrent clone operations to same memory data locations within a data processing system. SOLUTION: A set of tokens are defined within the memory cloner use prior to conducting a clone operation on the fabric. The tokens include a source token and a destination token. The tokens are issued on the fabric by a memory cloner prior to initiating the data clone operation. The tokens discover the availability of a source address and a destination address required to complete the clone operation. Once the response to the destination address token indicates that the destination address is not currently being utilized by another memory cloner to conduct another data clone operation, the memory cloner issues a command to initiate the memory clone operation. COPYRIGHT: (C)2004,JPO&NCIPI
Abstract:
PROBLEM TO BE SOLVED: To provide an NUMA architecture having improved queuing, storage communication efficiency. SOLUTION: A non-uniform memory access(NUMA) computer system includes at least one remote node and a home node coupled by node mutual connection. The home node includes a home system memory and a memory controller. In response to the reception of a data request from the remote node, the memory controller determines whether to impart the exclusive ownership or non- exclusive ownership of request data designated in the data request by referring to history information showing previous data access occurring in the remote node. The memory controller transmits the request data and the instruction of the exclusive ownership or non-exclusive ownership to the remote node, next.
Abstract:
PROBLEM TO BE SOLVED: To obtain a cache coherence protocol which uses a tagged coherence state to increase the memory band width without immediately writing back a change value to a system memory. SOLUTION: When a tagged state is assigned to a cache line which is loaded with the change value latest, the history state related to the tagged state which is moved between caches (in the horizontal direction) can be used furthermore. This system is also applied to a multi-processor computer system having clustered processing units, and a tagged state is applied t one of cache lines in each group of caches which support different processing unit clusters. Priority levels are assigned to different cache states, and they include tagged states for response to requests which access corresponding memory blocks. Because of use of a crossbar, a tagged intermediary response is transferred to only selected caches which are affected by this intermediary response.
Abstract:
PROBLEM TO BE SOLVED: To obtain an improved method for maintaining data coherency by determining whether or not a 1st cache should be updated according to the operation mode of the 1st cache in response to the detection of data transfer to remotely sent and including a 2nd data item. SOLUTION: A L2 cache 14 includes a cache controller 36. The cache controller 36 manages the storage and retrieval of data in a data array 34 and manages the update of a cache directory 32 in response to a signal received from a relative L1 cache and transaction snooped through an interconnection line. Then, a read request is put in an entry in a read queue 50. The cache controller 36 services the read request by supplying requested data to the relative L1 cache and then, removes the read request from the read queue 50.
Abstract:
PROBLEM TO BE SOLVED: To improve the inclusivity of a vertical cache hierarchy by mounting a modified MESI cache coherency protocol in a cache which can be accessed. SOLUTION: A coherency state field 208 of each entry in a cache director 204 is initially set to an ineffective state when a system is powered on and indicates that the both of a tag field 206 and data stored in relative cache lines in a cache memory 202 are ineffective. Thereafter, the coherency state field 208 can be updated to the coherency state in a deformed MESI coherency protocol. A cache controller 214 responds variously to snooped system bus operation and an L3 cache releases the allocation of a designated cache line of the cache memory.
Abstract:
PROBLEM TO BE SOLVED: To maintain the coherency between a data cache and an instruction cache which are separated by cleaning a designated cache entry in the data cache and instructing to invalidate the designated cache entry of the instruction cache. SOLUTION: Combined instructions are executed repeatedly for each of cache blocks included in the whole page 224 of a memory or in the plural pages of the memory to update a graphic display and a display buffer. When a mode bit 214 is set, icbi from a local processor is handled as no operation. In different kind of a system, snooped icbi is handled as the icbi even when the mode bit 214 is set. Instead of the above, the contents at a cache position (x) are copied to another position (y) and the corresponding cache position in a horizontal cache is invalidated.