An integrated memory system with a spiral cache responds to requests for values at a first external interface coupled to a particular storage location in the cache in a time period determined by the proximity of the requested values to the particular storage location. The cache supports multiple outstanding in-flight requests directed to the same address using an issue table that tracks multiple outstanding requests and control logic that applies the multiple requests to the same address in the order received by the cache memory. The cache also includes a backing store request table that tracks push-back write operations issued from the cache memory when the cache memory is full and a new value is provided from the external interface, and the control logic to prevent multiple copies of the same value from being loaded into the cache or a copy being loaded before a pending push-back has been completed.

Patent
   9009415
Priority
Nov 13 2008
Filed
Feb 19 2013
Issued
Apr 14 2015
Expiry
Jan 02 2029
Extension
50 days
Assg.orig
Entity
Large
1
49
EXPIRED
4. A method of caching a plurality of values within a storage device, comprising:
storing the plurality of values in multiple storage elements;
in response to a request for any one of the plurality of values, moving the requested value to a front-most one of the storage elements along at least one information pathway according to a move-to-front network prior to providing the requested value in response to the request;
responsive to moving the requested value to the front-most storage element, providing the requested value in response to the request from an interface coupled to the front-most storage element;
swapping remaining ones of the plurality of values backwards between adjacent elements in order according to an ordered push-back network along the at least on information pathway to a corresponding next-backward neighbor so that the values are stored in the multiple storage elements along the information pathway according to most-recent access, wherein a least-recently-accessed one of the plurality of values is either stored in a front-most non-empty memory location or is pushed out a last one of the multiple storage elements to a backing store interface, and wherein the moving moves the requested values according to the move-to-front network in a different order than a reverse order of the push-back network, whereby individual ones of the requested values are not moved through at least some of their adjacent multiple storage elements in the reverse order of the push-back network during their movement to the front-most one of the multiple storage elements.
1. A memory circuit, comprising:
multiple storage elements for storing values; and
access circuitry coupled to the multiple storage elements forming at least one information pathway for moving values among the multiple storage elements and supporting a move-to-front network and an ordered push-back network, wherein requested values provided in response to requests containing addresses corresponding to the requested values, wherein the requested values may be stored in and retrieved from any of the multiple storage elements and are moved to a front-most one of the multiple storage elements along the move-to-front network in response to the requests prior to accessing the requested values from an interface coupled to the front-most storage element, wherein the values stored in remaining ones of the multiple storage elements are swapped backward between adjacent ones of the multiple storage elements in order along the at least one information pathway according to the push-back network at each access to locations other than the front-most location so that the values are stored in the multiple storage elements along the at least one information pathway according to most-recent access, wherein a least-recently-accessed value is either stored in a front-most empty one of the storage elements or is pushed out of the memory circuit to a backing store interface, and wherein the requested values are moved according to the move-to-front network in a different order than a reverse order of the push-back network, whereby individual ones of the requested values are not moved through at least some of their adjacent multiple storage elements in the reverse order of the push-back network during their movement to the front-most one of the multiple storage elements.
2. The spiral memory circuit of claim 1, wherein the front-most storage location is located at an edge of an array formed by the multiple storage elements, and wherein a back-most one of the multiple storage elements is also located at an edge of the array.
3. The spiral memory circuit of claim 2, wherein the front-most storage location is located at a corner of the array, and wherein a back-most one of the multiple storage elements is located at another corner of the array.
5. The method of claim 4, wherein the front-most storage location is located at an edge of an array formed by the multiple storage elements, and wherein a back-most one of the multiple storage elements is also located at an edge of the array.
6. The method of claim 5, wherein the front-most one of the storage elements is located at a corner of an array formed by the multiple storage elements, and wherein the last one of the multiple storage elements is located at another corner of the array.
7. The memory circuit of claim 1, wherein the access circuitry moves multiple requested values simultaneously toward the front-most tile along the at least one information pathway according to the move-to-front network.
8. The memory circuit of claim 7, wherein an order of movement of the multiple requested values into the front-most tile is dependent on a location of the requested value within the multiple storage elements.
9. The method of claim 4, wherein the moving the requested value to the front-most one of the storage element moves multiple requested values simultaneously toward the front-most tile along the at least one information pathway according to the move-to-front network.
10. The method of claim 9, wherein an order of moving of the multiple requested values into the front-most tile is dependent on a location of the requested value within the multiple storage elements.

The present Application is a Divisional of U.S. patent application Ser. No. 12/640,360, filed on Dec. 17, 2009, which is a Continuation-in-Part of U.S. patent application Ser. No. 12/270,095 entitled “A SPIRAL CACHE MEMORY AND METHOD OF OPERATING A SPIRAL CACHE,” and Ser. No. 12/270,249 entitled “SPIRAL CACHE POWER MANAGEMENT, ADAPTIVE SIZING AND INTERFACE OPERATIONS”, both of which were filed on Nov. 13, 2008, have at least one common inventor, and are assigned to the same Assignee. The disclosures of the above-referenced U.S. Patent Applications are incorporated herein by reference.

1. Field of the Invention

The present invention is related to hierarchical memory systems, and more particularly to a memory interface that couples a spiral cache memory to other members of a memory hierarchy.

2. Description of Related Art

A spiral cache memory as described in the above-referenced Parent U.S. Patent application supports multiple in-flight requests referencing the same or different values by their address. In order to integrate a spiral cache memory in a hierarchical memory system, while permitting the next lower-order level of the memory hierarchy or a processor to access the same value repeatedly before a request for the same value is completed, a way to ensure that writes to the value are satisfied before subsequent reads is needed. It is desirable to do so without constraining the activity of the processor or lower-order level of the memory hierarchy that is coupled to the front-most storage tile, as to do so would introduce performance penalties, or require the processor architecture and/or program code to constrain the order of accesses. Also, in particular because the backing store will generally have a much higher latency that the spiral cache itself, queues as described in the above-incorporated parent U.S. Patent application are needed between the memory hierarchy levels, and in order to not constrain the activity of the spiral cache with respect to the backing store, at least at the internal level of the storage tiles, it is desirable to provide a mechanism to coordinate requests to the backing store so that push-back write values can be coordinated with read requests issued to the backing store. Further, read requests issued to the backing store return values from the backing store into the spiral cache. Without checking the address of each value and tracking all of the values present in the spiral cache, multiple copies of the same value could be read into the spiral cache. Therefore, a mechanism to prevent multiple copies of the same value being returned to the spiral cache is needed.

Therefore, it would be desirable to provide a spiral cache interface to a memory hierarchy and an integrated memory hierarchy including a spiral cache, in which multiple outstanding requests for the same value can be issued into the spiral cache without constraining the processor, program code, or lower-order level of the memory hierarchy. It would further be desirable to provide an interface from the spiral cache to a backing store without constraining the behavior of the network of tiles in the spiral cache or having multiple copies of the same value returned to the spiral cache.

The invention is embodied in a spiral cache memory, a hierarchical memory system including the spiral cache memory and methods of operation of the system. The spiral cache memory has multiple tiles with storage locations for storing values, each of which may be a smaller cache memory such as a direct-mapped cache or an associative cache.

Multiple requests accessing the same value can be issued into the spiral cache and to prevent erroneous reads due to the requests directed to the same value not being satisfied in the order in which they are issued into the spiral cache, but the order in which the requests are returned, an issue table is used to track the requests and control logic within the spiral cache memory interface controls the order of application of the returned requests to the interface that couples the spiral cache to the lower-order level of the memory hierarchy or processor.

Prevention of multiple copies of the same value from being returned to the spiral cache from the backing store is performed by maintaining a backing store request table that prevents multiple read requests to the same value (address) being issued to the backing store. The backing store request table also tracks push-back write operations issued from the spiral cache by giving priority to write operations coming from the push-back spiral over read requests issued from the spiral cache due to a miss.

The memory interface also provides a number of queues to buffer operations and values/requests for preventing overflow of the backing store, for ordering of operations on values and serializing requests, as are described in further detail below.

The foregoing and other objectives, features, and advantages of the invention will be apparent from the following, more particular, description of the preferred embodiment of the invention, as illustrated in the accompanying drawings.

The novel features believed characteristic of the invention are set forth in the appended claims. The invention itself, however, as well as a preferred mode of use, further objectives, and advantages thereof, will best be understood by reference to the following detailed description of the invention when read in conjunction with the accompanying Figures, wherein like reference numerals indicate like components, and:

FIG. 1 is a block diagram of a system incorporating a spiral cache according to an embodiment of the present invention.

FIG. 2 is illustrates the relationship between the spiral cache and the backing store in FIG. 1.

FIG. 3 is a block diagram illustrating further details of a spiral cache integrated within a memory hierarchy in accordance with an embodiment of the present invention.

FIGS. 4-5 are processing flow diagrams illustrating sequences of operations within the system of FIG. 3.

FIG. 6 is a flowchart illustrating a method of operation of control logic within memory interface 106 of FIG. 3.

The present invention encompasses techniques for effectively integrating a spiral cache memory into a memory hierarchy. A memory interface having a number of tables and queues provides unconstrained operation by the adjacent levels of the system hierarchy, by controlling the order of application of values returned from the spiral cache according to the order of the issued requests and not the order of the returned values, which may not match. The memory interface also ensures that the backing store input does not overflow and that multiple copies of the same value are not loaded into the spiral cache due to multiple requests issued at the front of the spiral. The memory interface also ensures that backing store read requests do not bypass push-back values that are propagating backwards through the spiral, returning invalid values that are not identified as such. An arrangement of a spiral cache that locates the lower-order and higher-order hierarchy member interfaces at edges of the spiral is also further illustrated, and while the cache type is still referred to as “spiral”, since the front-most tile is not located near the center of the array, the push-back network follows a meandering path that zig-zags in segments of increasing length.

Black-Box Behavior of the Spiral Cache Referring now to FIG. 1, a hierarchical system including a spiral cache is illustrated. The illustrated spiral cache is composed of tiles 0-63 and resembles a single quadrant of the spiral caches disclosed in the above-incorporated parent U.S. Patent Application “SPIRAL CACHE POWER MANAGEMENT, ADAPTIVE SIZING AND INTERFACE OPERATIONS”, and is a larger version of the cache depicted in FIG. 8A of the above-incorporated parent U.S. Patent Application “A SPIRAL CACHE MEMORY AND METHOD OF OPERATING A SPIRAL CACHE.” The illustrated spiral cache includes two systolic networks, a push-back (spiral) network 114 and a move-to-front network 116. Push-back network 114 imposes a linear structure on the tiles and is coupled at front-most tile 0 to a processor 100 and at a back-most tile 63 to a backing store 112. Each of tiles 0-63 contains a fast cache memory, such as a direct-mapped cache, and the unit of data transfer across the networks in such a configuration is a cache line. The move-to-front heuristic places cache lines into tiles. A processor issues load or store operations to front-most tile 0. Independent of the particular operation being performed on a value, the spiral cache fetches the corresponding cache line and places it in front-most tile 0. For load operations, the desired data are also sent to processor 100. For store operations, the cache line is patched with the store data before writing the cache line into the cache of tile 0. If the corresponding cache line storage of tile 0 is occupied, the line currently stored in the corresponding location is pushed back to empty the cache ling storage to accept the new line. The push-back operation effectively swaps values (e.g., cache lines) back at each tile along push-back network 114. Data being pushed back travel along push-back network 114 from front-most tile 0 towards the tail end of the spiral, until an empty line is located, or until the data is pushed out of the spiral cache into backing store 114.

The cache line being accessed by a load or store operation may be located within the spiral cache, or the cache line may be absent. If the cache line is present, the spiral cache can report a hit, which completes the associated operation successfully. Otherwise, if the accessed cache line is not present in the spiral cache, we incur a miss. A miss requires fetching the cache line from backing store 112 and moving the cache line to front-most tile 0. The move-to-front (M2F) operation involves not only the move-to-front network 114 inside the spiral cache, but requires an additional connection to backing store 112. Referring now to FIG. 2, a simplified block diagram illustrates the relationship between spiral cache 104 and backing store 112. If a request hits in spiral cache 104, no communication occurs outside the spiral cache, with the exception of sending the requested load value to processor 100. When a hit occurs, the search for the requested item succeeds inside the spiral cache. The associated cache line is moved to front-most tile 0, leaving an empty (invalid) cache line behind. In the worst case, a push-back operation initiated in front-most tile 0 to make space for the moved cache line generates swap operations that propagate all the way to the tile where the cache line was found. If there is an invalid cache line in a tile closer to the front of the spiral, then the push-back operation terminates at that tile. If a request misses in the spiral cache, the requested cache line is moved from the backing store into front-most tile 0. When a miss occurs, the associated push-back operation may cause a cache line to be pushed back into backing store 112, since a new cache line is being brought into spiral cache 104 and all of the corresponding cache line storage in tiles 0-63 may already contain valid entries.

When spiral cache 104 reports a miss, a single-copy invariant condition imposed on spiral cache 104 guarantees that the requested cache line does not exist anywhere in spiral cache 104. Therefore, the cache line is fetched from backing store 112, and written into front-most tile 0. The associated push-back operation causes a cache line to be written into backing store 112 if all tile caches contain non-empty (valid) cache lines. The black-box communication behavior of spiral cache 104 is described below. Data are communicated between spiral cache 104 and backing store 112 only in case of a miss. A miss requires a cache line to be moved from backing store 112 into front-most tile 0. The associated push-back operation may cause a cache line to be written into backing store 112. It is noted that cache lines are initially loaded into spiral cache 104 only at front-most tile 0, and leave spiral cache 104 only from the tail end of spiral cache 104. A pushed-back cache line exits a spiral cache of N tiles at the tail end after a delay of at least N−1 duty cycles has elapsed since the writing of the cache line fetched from backing store 112 into front-most tile 0. In order for the above-described black-box behavior of the spiral cache to operate, the ordering of requests and responses must be considered. Spiral cache 104 does not inherently preserve any ordering. Multiple requests to different cache lines may return in arbitrary order depending on the location of the values being requested. Requests to the same cache line may also return in a different order, depending on the location of the cache line, the operation of the geometric retry mechanism, and the collision resolution mechanism of new requests on the diagonal in the move-to-front network, as described in the above-incorporated parent U.S. Patent Application“A SPIRAL CACHE MEMORY AND METHOD OF OPERATING A SPIRAL CACHE.” Therefore, any ordering guarantees of the responses with respect their requests must be implemented outside of the spiral cache tile array. The present invention provides mechanisms to guarantee the completion order of load and store operations to the same cache line as issued by processor 100, without imposing any ordering restrictions on operations to different cache lines. The ordering behavior described above is consistent with that of contemporary processor architectures, which are capable of accepting multiple outstanding memory operations.

System Integration of a Spiral Cache Referring now to FIG. 3, a block diagram of a hierarchical system including a spiral cache 104 is shown in accordance with an embodiment of the present invention. The system includes a number of queues that interconnect processor 100 and backing store 112 with spiral cache 104. The system of FIG. 3 also includes a memory interface 106 that manages the above-described operations occurring at front-most tile 0 and back-most tile 63 in spiral cache 104, in order to meet the ordering requirements for processor 100 and preserve the single-copy invariant condition within spiral cache 104. With the exception of the push-back output of the spiral cache, memory interface 106 controls all dataflow between processor 100, backing store 112, spiral cache 104 and the various queues. The processor is de-coupled from the memory system by means of two queues, a load-store queue ldstq and a memory-to-processor queue m2pq. The same de-coupling interface can be employed if spiral cache 104 implements a higher-order level of the memory hierarchy, in which case the next lower-ordered cache would be connected in place of processor 100. However, when integrating spiral cache 104 with contemporary processors, the processors already include some of the functionality of memory interface 106 to track ordering of memory operations. Therefore, portions of memory interface 106 and load-store queue ldstq, memory-to-processor queue m2pq, and bypass queue bypq could be integrated into processor 100. In the following description, it is assumes that other than in load-store queue ldstq and memory-to-processor queue m2pq, the unit of data transfer within spiral cache 104, as well as across all queues is a cache line.

Queues and Tables The various queues and tables included in the system of FIG. 3 are summarized below and are implemented in hardware. The processor issues load and store requests to memory by enqueuing them into load-store queue ldstq. The memory system returns load values and potentially store acknowledgements to the processor via the memory-to-processor queue m2pq. When the spiral cache misses, memory interface 106 sends a read request to backing store 112 by enqueuing it into a read queue rdq. Two types of requests may be sent from spiral cache 104 to backing store 112, read requests as mentioned above and write requests originating from the tail end of the push-back network. A push-back/read queue pbrdq serializes these two types of requests and presents them to backing store 112. As will be explained in further detail below, read requests enqueued in read queue rdq are immediately passed through to push-back/read queue pbrdq, unless a backing store write request (due to a push-back leaving the tail end of spiral cache 104) is being enqued at the same time, causing a conflict. If a conflict exists, multiplexer M1 “favors” enqueing a backing-store write request into push-back/read queue pbrdq and read requests are pended in read queue rdq. Backing store 112 responds with cache lines retrieved in response to read requests by enqueuing them into a backing store queue bsq. A bypass queue bypq enables memory interface 106 to accept multiple outstanding read requests to the same cache line in backing store 112, without incurring the presumably high backing store access latency for each request.

In addition to the queues described above, memory interface 106 contains two tables: An issue table itab keeps track of all outstanding memory operations, and ensures that memory interface 106 performs load and store operations to the same cache line in the order issued by the processor into load-store queue ldstq. A backing store request table mtab keeps track of all outstanding backing-store read operations, and guarantees that multiple read requests directed to the same cache line result in a single read operation from backing store 112, which preserves the single-copy invariant condition. A primary function of the queueing system architecture depicted in FIG. 3 is to coordinate the dataflow between components having different access latencies and throughput, e.g., backing store 112 vs. spiral cache 104. Spiral cache 104 has variable access latency and throughput. Backing store 112 generally has much larger latency, typically about three orders of magnitude larger than the best-case access latency of spiral cache 104. Also, backing store 112 may or may not permit pipelined operation. Processor 100 may produce memory operations at a rate as high as one operation per clock cycle. The following description limits processor operations to load and store operations, and assumes that processor 100 can stall if load-store queue ldstq is full.

The dataflow of a memory operation through the memory system depicted in FIG. 3 is described below. Processor 100 enqueues a load or store request in load-store queue ldstq. Memory interface 106 dequeues the request, inserts the request into issue table itab, and issues the address into the move-to-front network of spiral cache 104. Depending on the placement of the requested cache line, it may require multiple retries before the request succeeds with a hit or the request results in a miss. If the spiral cache responds with a hit, the next action depends on the type of the operation. If the request is associated with a load operation, the value is returned to the processor by enqueuing it in the memory-to-processor queue m2pq. Otherwise, if the request is associated with a store operation, the store value is retrieved from issue table itab and patched into the cache line, before storing the value in front-most tile 0. If processor 100 requires a store acknowledgment, the store acknowledgement is enqueued into memory-to-processor queue m2pq and the value is stored in front-most tile 0. For either type of operation, the corresponding entry in issue table itab is then removed, and the request is completed. If the spiral cache 104 responds with a miss, memory interface 106 issues a read request to backing store 112. Memory interface 104 enqueues the read request into read queue rdq and inserts the address into backing store request table mtab. Backing store 112 responds with the requested cache line by enqueuing it into backing store queue bsq. Memory interface 106 dequeues the cache line from backing store queue bsq, and stores it in front-most tile 0, potentially patched with a store value that was stored in issue table itab if the operation requesting the cache line was a store operation. The entry in backing store request table mtab entry is deleted, and bypass queue bypq entries are updated, as will be described in further detail below. As when the operation hits in spiral cache 104, a reply to processor 100 is enqueued in memory-to-processor queue m2pq and the corresponding entry in issue table itab is removed, which completes the request. The above description applies to a single memory operation. System operation becomes more complex when multiple operations are in flight. Due to concurrency, three issues must be tackled to guarantee correct operation, ordering of responses, and proper flow-control: (1) ordering of responses to the same cache line that hit in spiral cache 104; (2) serialization of push-back write requests from spiral cache 104 and read requests from the memory interface 106 before the requests are sent to backing store 112; and (3) handling multiple requests to the same cache line for cache lines that miss in spiral cache 104.

Ordering of the Spiral responses The ordering problem of a sequence of load and store operations to the same cache line that hit in spiral cache 104 is as follows. Assume, for example, that processor 100 issues a store operation and subsequently a load operation to the same address. For correctness, it is expected that the load operation responds with the previously stored value. Problems can arise within the system, because requests may return out of order from spiral cache 104. For example, assume in a hypothetical system that a request issued into the spiral cache comprises all the information needed to service the request, including an op-code to distinguish loads from stores, the address, and the store value if it applies. It should be noted that this request differs from the requests used in the exemplary system of FIG. 3, as within spiral cache 104, requests due to load operations are not distinguished from requests due to store operations. If memory interface 106 first issues an unrelated request into spiral cache 104, then issues the store request, and finally issues the load request, the timing may be such that the unrelated request is travelling toward front-most tile 0 on the move-to-front network, and conflicts with the store request at a diagonal tile, forcing the store request outwards to enter the move-to-front network beyond its intended retry radius. In the subsequent duty cycle, the load request enters the move-to-front network at its intended retry radius, and thereby overtakes the store request. If the load request finds the cache line, it moves it to the front, and memory interface 106 delivers the current value to processor 100. The trailing store request finds the cache line in tile 0, where the store is performed only after the load, which is incorrect memory behavior. To guarantee correct ordering, issue table itab is introduced, which enables memory interface 106 to issue typeless requests into spiral cache 104. Requests issued into the spiral cache are oblivious to the type of memory operation, and serve to move the requested cache line to front-most tile 0. An entry in issue table itab includes the address of a memory operation, its type (load or store), and the store value in case of a store operation. Issue table itab is organized as a first-in-first-out (FIFO) queue with an enqueue operation that appends a new entry. To lookup or remove an entry, issue table itab can be implemented with a priority decoder that identifies the closest entry to the head that is associated with a given address.

A request issued by memory interface 106 into spiral cache 104 includes the address and a retry radius. When the corresponding response (reply) arrives on the M2F network at front-most tile 0, the address portion of the response is used to retrieve the corresponding entry from issue table itab. It is the entry in issue table itab that provides the operational context, and for store operations, the entry provides the store value. Support for multiple outstanding requests per cache line is provided by organizing issue table itab as a FIFO queue. The implicit ordering of the issue table itab FIFO maintains the order of memory operations. Therefore, when a store operation is issued before a load operation to the same address, the store operation entry precedes the load operation entry in issue table itab, and will be completed before the load operation is completed.

Referring now to FIG. 4, an exemplary re-ordering operation in the system of FIG. 3 is shown. Operations are notated below in the form: Id <addr> or st <addr>, <val>. Processor 100 issues three memory operations in the order: st 100,7; ld 200 and ld 100, which are tracked in issue table itab as illustrated in table T1A. Spiral cache 104 returns the responses out of order, and the order in which the cache lines arrive at frontmost tile 0 is according to addresses 200, 100, 100, with the requests for address 100 returning out of order. However, since spiral cache 104 does not maintain the association with the requesting operations, the order in which the requests return does not determine the execution order. The execution order is determined by retrieving the first entry with a matching address from issue table itab which produces: ld 200 as shown in the transition from table T1A to table T1B, which shows issue table itab having two entries remaining. Next, in response to the first response with address 100, issue table itab produces st 100,7 as illustrated by the transition from table T1B to table T1C. Finally, in response to spiral cache returning the next response with address 100, issue table produces ld 100 as shown in the transition from table T1C to table T1D, showing issue table itab as being empty. Thus, issue table itab ensures that the store to address 100 completes before the load. FIG. 4 illustrates the functionality of issue table itab. The requests traveling through storage tile network 104 are not explicitly associated with the original processor operation or corresponding entry in issue table itab. Instead, a request carries an address, such as address 100, only. Although memory interface 106 dequeues an entry from load-store queue ldstq, inserts the entry in issue table itab, and issues a request into spiral cache, the request is not associated with the particular entry in issue table itab except through a matching address. There can be as many outstanding requests for one address in spiral cache 104 as there are entries in issue table itab. When spiral cache 104 returns a response, issue table itab is searched for the first entry from the head having a matching address. In FIG. 4, the first response is due to the second request, issued when handling the load operation. Nevertheless, the first entry in issue table itab that is associated with address 100 is the store operation. Thus, issue table itab enables the spiral cache to complete memory operations affecting the same cache line in the order in which they were issued by the processor.

Serialization of Backing Store Operations Backing store 112 serves read operations issued by memory interface 106 into read queue rdq if a request misses in spiral cache 104. Backing store 112 also serves write operations emitted by the push-back network of spiral cache 104. Since there are two distinct sources for these operations: memory interface 106 for reads; and the push-back network for writes, the operations must be serialized. Serialization of read and write requests to backing store 112 must respect the following ordering constraint: if a read operation issued by memory interface 104 to backing store 112 contains the same address as a write operation issued by the push-back network, then the write operation must precede the read operation. The reason for the ordering constraint is described below. A write operation to backing store 112 contains a modified (dirty) cache line because push-back requests containing clean cache lines are discarded at the tail end tile of spiral cache 104. (There is no reason to return a clean cache line to backing store 112, as by definition, the clean cache line is already identically present in backingstore 112.) The backing store write operation originates at the tail end of the push-back network of spiral cache 104, when tail-end tile 63 (tile N−1) pushes a dirty value out. The dirty value was produced earlier by a store operation that stored the modified value in front-most tile 0. Subsequent memory accesses cause the dirty value to be pushed back through the push-back network. An example of such memory accesses is: N accesses that missed in spiral cache 104 and have the same direct mapping as the dirty line, causing the corresponding values to be read from backing store 112 and loaded into tile 0 in the same cache line that the dirty line occupied. As each value is pushed-back to make room for the next, because their mapping is the same, they will push the previous occupants of that storage, including the dirty line, backward at each access. As another example, spiral cache 104 could have received N−1 requests, again mapped to the same cache lines, that hit in spiral cache 104, causing the corresponding values to be moved into front-most tile 0, causing the dirty line to be pushed back by N−1 tiles into tail-end tile 63. One subsequent request that maps to the same cache line, but misses in spiral cache 104, causes the corresponding value to be loaded from backing store 112 and stored in front-most tile 0, causing the dirty line to be pushed out of tail-end tile 63. If processor 100 issues a load operation for the dirty cache line while the dirty cache line is pushed back on the push-back network toward backing store 112, a race condition occurs if spiral cache 104 reports a miss, and memory interface 106 initiates a read operation to backing store 112 before the dirty line has been written back into backing store 112.

The move-to-front request of the load operation traverses spiral cache 104 while the requested cache line, modified by the preceding store operation, is pushed back on the push-back network within spiral cache 104 or has been pushed out of spiral cache at tail-end tile 63. If the cache line is in spiral cache 104, the single-copy invariant condition guarantees that the move-to-front request will move the cache line to front-most tile 0. Otherwise, the cache line must have been pushed out of spiral cache 104 via the push-back network. In the extreme case of timing a spiral cache hit, the move-to-front request meets the requested cache line during the same duty cycle that the push-back value arrives at tail-end tile 63. For a miss to occur, the requested cache line must have been pushed out at least one duty cycle before the move-to-front request reaches tail-end tile 63. Since the M2F request must travel to memory interface 106 before a miss can be reported and a read request issued to backing store 112, the travel time of the M2F request from tail-end tile 63 to frontmost tile 0 enables ordering of backing store operations such that the write operation will reach the backing store before the read operation. To prevent a race condition between backing store write and read requests, push-back read queue pbrdq forms the master queue of read queue rdq. As such, direct insertions into push-back read queue pbrdq have priority over entries in read queue rdq. Thus, write operations emitted by the push-back network have priority over read operations originating from the M2F network, and are enqueued immediately into push-back read queue pbrdq. Read operations are enqueued into push-back read queue pbrdq when possible, that is during clock cycles when no push-back request is being enqueued. Collisions are resolved by enqueuing read operations in read queue rdq. The organization of the push-back read queue pbrdq and read queue rdq guarantees that a read request to backing store 112 trails a potential write request. Thus, backing store 112 serves the above described exemplary read operation correctly with the cache line written during the preceding push-back write operation.

Multiple Spiral Misses When spiral cache 104 accepts multiple outstanding requests, one or more of them may miss. Backing store request table mtab and bypass queue bypq are included to prevent duplication of lines in the spiral cache when multiple misses to the same cache line require retrieving the cache line from the backing store. The potential for duplication of cache lines due to multiple outstanding backing-store read requests exists due to multiple operations to the same address. For example, assume that processor 100 issues a store followed by a load operation to the same cache line, as discussed above, and that both spiral responses result in a miss, but are returned in order. Without including logic for handling such conditions, memory interface 106 would enqueue two read requests to backing store 112, the first associated with the store operation, and the second with the load operation. Assuming that backing store 112 preserves the order of the requests, it first returns the requested cache line associated with the store operation. Memory interface 106 would then patch the cache line with the store value and write the cache line into front-most tile 0. When backing store 112 returns the same cache line again, now associated with the load operation, memory interface 106 would return the requested load value to processor 100, and write the cache line into front-most tile 0, overwriting the previously written store value. Not only is the load value returned to processor 100 different from the expected value, but all subsequent load operations will return the wrong value as well. If the first copy of the cache line returned by backing store 112 is pushed back and out of tile 0 before memory interface 106 writes the second copy into tile 0, the problem is further exacerbated. Spiral cache 104 now contains two copies of the same cache line, violating the single-copy invariant condition. Therefore, memory interface 106 prevents duplication of cache lines due to multiple outstanding read requests to the backing store. In the illustrated embodiment of FIG. 3, memory interface 106 permits at most one outstanding read request per cache line, and delays additional requests to the cache line until backing store 112 has returned the cache line. Backing store request table mtab and bypass queue bypq are used to implement this strategy.

Backing store request table mtab is an associative memory that maintains one entry per cache-line address for each outstanding backing store read request. An address entry is inserted into backing store request table mtab when the spiral cache 104 responds with a miss. Memory interface 106 also enqueues a read request with the associated address into read queue rdq. The entry is deleted from the backing store request table mtab when memory interface 106 dequeues the backing store response from backing store queue bsq, and stores the cache line in front-most tile 0 of the spiral cache. Bypass queue bypq is a FIFO queue with additional functionality resembling that of an associative memory. Each queue entry contains an address plus a ready bit. Insertion of an entry into bypass queue bypq corresponds to a conventional enqueue operation. When inserting an address, its associated ready bit is initialized to not-ready. However, dequeuing an entry from bypass queue bypq is not performed according to a conventional dequeue operation. Instead, to dequeue an entry associated with an address, a priority decoder is included, which identifies the first ready entry having the requested address from the head of the queue, as has been described above for the operation of issue table itab. Bypass queue bypq also includes circuitry that implements a “ready” operation that sets the ready bits of all entries associated with an address from not-ready to ready.

Referring now to FIG. 5, exemplary operation of backing store request table mtab and bypass queue bypq in the system of FIG. 3 are illustrated. FIG. 5 shows a sequence of operations as coordinated by memory interface 106 with support of the backing store request table mtab and bypass queue bypq. In the example, the processor issues three memory operations (load or store) to the cache line at address 100, and two operations to the cache line at address 200. Assuming that memory interface 106 dequeues all five requests from the ldstq, inserts them into the itab, and issues the associated requests into spiral cache 104. All requests miss eventually, and are returned, potentially out of order, by spiral cache 104. For example, if the first response from spiral cache 104 is a miss report for address 100, then memory interface 106 sends a read request with address 100 to backing store 112 and inserts address 100 into backing store request table mtab. Next, the second miss report for address 100 may arrive at memory interface 106. Memory interface 106 examines backing store request table mtab and determines that an entry for address 100 already exists. Therefore, memory interface 106 does not send another read request to backing store 112 but enqueues an entry in bypass queue bypq containing address 100 with the ready bit clear (not-ready). Subsequently, two miss reports for address 200 are handled in a similar manner, and finally, another miss report for address 100 causes an entry to be enqueued in bypass queue bypq. The state of backing store request table mtab and bypass queue bypq after the above-described operations is illustrated in table T2A. Assuming that backing store 112 maintains the order of the read requests it receives, the first response contains the cache line at address 100. After memory interface 106 dequeues the cache line from backing store queue bsq, the information stored in issue table itab is used to perform the associated load or store operation, and the cache line is written into front-most tile 0, as described above. The entries in issue table itab and backing store request table mtab are then deleted. Also, the ready operation of bypass queue bypq is applied to address 100, and the remaining outstanding operations associated with this address are now marked ready, as illustrated in table T2B.

Memory interface 106 is responsible for dequeuing ready entries from the bypass queue bypq in FIFO order. There are two ready entries associated with address 100 illustrated in table T2B. The first entry corresponds to the second memory operation associated with address 100. After the first entry is dequeued, the state of bypass queue bypq is as shown in table T2C. Memory interface 106 issues a request for the address of the entry dequeued from bypass queue bypq into spiral cache 104. When spiral cache 104 responds, issue table itab provides the information needed to handle the response as for any other spiral cache responses. Backing store request table mtab and bypass queue bypq not only serve to enforce correctness by preventing duplication of cache lines, but also to improve performance. If multiple requests to a particular memory address occur in close succession, backing store request table mtab and bypass queue bypq reduce the overall latency from multiple, presumably high-latency accesses to the backing store to just one. This capability also improves the throughput of the overall memory system.

Memory interface 106 also handles dequeuing and processing entries from its associated input queues. The selection of the queues determines the order in which actions are scheduled. An exemplary scheduling loop that may be implemented by memory interface 106 is illustrated in FIG. 6. The algorithm illustrated in FIG. 6 performs two activities: (1) monitoring the M2F network of spiral cache 104 for incoming requests; and (2) monitoring input queues backing store queue bsq, bypass queue bypq, and load-store queue ldstq. The algorithm of FIG. 6 is executed once per duty cycle of the systolic spiral cache. If a M2F reply arrives at front-most tile 0 (decision 140), then if the reply indicates a hit (decision 141) then the load or store operation indicated in issue table itab is completed in the front-most tile 0, the load value is enqued in memory-to-processor queue m2pq and the entry in issue table itab is deleted (step 142). If the M2F reply indicates a miss (decision 143) then the address of the reply is inserted in backing store request table mtab and a backing store read request is enqueued in read queue rdq (step 144). If the M2F reply is neither a miss or a hit, indicating that a request has not been sent to the maximum retry bound then the request is re-issued into spiral cache 104 with a doubled retry bound (step 145). Next, the input queues to memory interface 106 are monitored according to a priority scheme. Backing store queue bsq has priority over bypass queue bypq, because backing store read operations constitute the critical path due to high access latencies. Therefore, if backing store bsq has an entry (decision 146) then the entry is deleted, other entries in bypass queue bypq having the same address are set to a ready state, the corresponding entry in backing store request table mtab is deleted and the operation is completed in front-most tile 0 (step 147). Only if there was no entry in backing store bsq, bypass queue bypq is examined for ready entries (decision 148) and if bypass queue bypq contains a ready entry, the first ready entry is dequeued and a request is issued into spiral cache 104. Only if there was no entry in backing store bsq (decision 146), and no ready entry in bypass queue bypq (decision 148), is the load-store queue examined for entries, and if an entry is present (decision 150), the first entry is dequeued, a corresponding entry is inserted into issue table itab and a request is issued into spiral cache 104 (step 151). Bypass queue bypq has priority over load-store queue ldstq, because entries in bypass queue bypq are necessarily older and should be serviced before dequeuing newer requests from load-store queue ldstq.

Another important priority consideration not exposed in the scheduling loop of FIG. 6 is that if any action due to handling a queue entry coincides with the arrival of a move-to-front signal at front-most tile 0, the move-to-front signal is given priority, or the systolic dataflow of the spiral cache 104 would be violated. Therefore, the implementation of queue actions includes the following fall-back mechanism. Rather than fully dequeuing entries in steps 147, 149 and 151, the queues are “peeked” i.e., the entries are read without removal. If no collision with a move-to-front signal occurs at tile 0, the entry is dequeued. Otherwise, the action is aborted. Memory interface 106 will then retry the queue operation during the next duty cycle, because the entry is still at the head of the queue.

As in many queueing systems, the system depicted in FIG. 3 requires flow control to prevent queue overflows and the associated loss of entries. To support a storage tile network that does not have its own flow control, queue overflows in the surrounding infrastructure are prevented by the action of memory interface 106. In fact, overflows can be prevented by proper queue sizing and controlled issuing of new requests from memory interface 106 into spiral cache 104. Since backing store 112 is generally significantly slower than spiral cache 104, the critical queue of the system is push-back read queue pbrdq. To avoid requiring a push-back read queue pbrdq of unreasonably large queue size, additional overflow protection is needed. As explained above, in the worst case, the spiral cache generates one push-back for each miss. The spiral cache generates the push-back after the cache line returned by the backing store has been written into front-most tile 0. Thus, the rate at which the spiral cache can generate push-backs is equal to the rate at which backing store 112 services requests. Therefore the loop from spiral cache 104 to push-back read queue pbrdq to backing store 112 to backing store queue bsq to memory interface 106 and returning to spiral cache 104 does not pose overflow problems. A small number of entries in the push-back read queue pbrdq suffice to handle push-backs that are generated with variable delay with respect to the corresponding duty cycle when the cache line was moved to front-most tile 0.

The rate at which spiral cache 104 can generate misses and cause memory interface 106 to enqueue the associated read requests via read queue rdq to the backing store is much greater than the push-back rate, because spiral cache 104 operates at a much higher clock frequency than backing store 112. Therefore, to prevent overflow of push-back read queue pbrdq, which is the master queue of read queue rdq, the number of outstanding requests issued into spiral cache 104 must be controlled. A “one-quadrant” cache such as that illustrated in FIG. 1 having N tiles and a side length of √{square root over (N)} tiles contains at most 3√{square root over (N)}−2 outstanding requests. In the worst case, each of these requests misses, causing one read request per miss to be enqueued into read queue rdq, and from there into push-back read queue pbrdq. Hence, overflow of push-back ready queue pbrdq can be prevented, since memory interface 106 only issues a request into spiral cache 104 if there are at least 3√{square root over (N)}−2 empty entries in push-back read queue pbrdq. By adding a safety for a few additional empty entries, for example 8, for push-back operations, memory interface 106 will issue a new request into the storage tile network 104 only if there are at least 3√{square root over (N)}+6 free entries in push-back read queue pbrdq. Note that the above condition depends on the number of tiles in the spiral cache 104, but is independent of the access latency of backing store 112. Thus, for a spiral cache such as that illustrated in FIG. 1, a reasonable choice for the size of push-back read queue pbrdq is 6(√{square root over (N)}+1), for example. The remaining queues can be sized in a straightforward fashion, and do not require associated flow control. Given the flow control constraint for push-back read queue pbrdq, the scheduling algorithm in FIG. 6 can be augmented such that a new request is issued from load-store queue ldstq into the spiral cache only if the conjunction of the following conditions holds: issue table itab is not full, backing store request table mtab is not full, push-back read queue pbrdq has at least 3√{square root over (N)}+6 free entries, backing store queue bsq is empty, and bypass queue bypq has no ready entries. Assuming that all queues are implemented as part of memory interface 106, each of the above conditions can be checked easily within a single clock cycle.

While the invention has been particularly shown and described with reference to the preferred embodiments thereof, it will be understood by those skilled in the art that the foregoing and other changes in form, and details may be made therein without departing from the spirit and scope of the invention.

Schaub, Jeremy D., Gebara, Fadi H., Strumpen, Volker

Patent Priority Assignee Title
9542315, Nov 13 2008 International Business Machines Corporation Tiled storage array with systolic move-to-front organization
Patent Priority Assignee Title
5339268, May 08 1992 Mitsubishi Denki Kabushiki Kaisha Content addressable memory cell and content addressable memory circuit for implementing a least recently used algorithm
5355345, Oct 04 1993 CHASE MANHATTAN BANK, AS ADMINISTRATIVE AGENT, THE Fully scalable memory apparatus
5539893, Nov 16 1993 Unisys Corporation Multi-level memory and methods for allocating data most likely to be used to the fastest memory level
5564041, Apr 20 1990 Hitachi, Ltd.; Hitachi Microcomputer System Ltd. Microprocessor for inserting a bus cycle in an instruction set to output an internal information for an emulation
5640339, May 11 1993 International Business Machines Corporation Cache memory including master and local word lines coupled to memory cells
6370620, Dec 10 1998 International Business Machines Corporation Web object caching and apparatus for performing the same
6418525, Jan 29 1999 International Business Machines Corporation Method and apparatus for reducing latency in set-associative caches using set prediction
6430654, Jan 21 1998 Oracle America, Inc Apparatus and method for distributed non-blocking multi-level cache
6665767, Jul 15 1999 Texas Instruments Incorporated Programmer initiated cache block operations
6763426, Dec 27 2001 AVAGO TECHNOLOGIES INTERNATIONAL SALES PTE LIMITED Cascadable content addressable memory (CAM) device and architecture
6839809, May 31 2000 Cisco Technology, Inc. Methods and apparatus for improving content quality in web caching systems
6961821, Oct 16 2002 International Business Machines Corporation Reconfigurable cache controller for nonuniform memory access computer systems
6996117, Sep 19 2001 BAY MICROSYSTEMS, INC Vertical instruction and data processing in a network processor architecture
7050351, Dec 30 2003 Intel Corporation Method and apparatus for multiple row caches per bank
7107399, May 11 2001 International Business Machines Corporation Scalable memory
7370252, Nov 10 2004 INTELLECTUAL DISCOVERY CO , LTD Interleaving apparatus and method for orthogonal frequency division multiplexing transmitter
7461210, Apr 14 2006 EZCHIP SEMICONDUCTOR LTD ; Mellanox Technologies, LTD Managing set associative cache memory according to entry type
7498836, Sep 19 2003 XILINX, Inc. Programmable low power modes for embedded memory blocks
7805575, Sep 29 2006 EZCHIP SEMICONDUCTOR LTD ; Mellanox Technologies, LTD Caching in multicore and multiprocessor architectures
8015357, Nov 13 2008 International Business Machines Corporation Storage array tile supporting systolic movement operations
8060699, Nov 13 2008 International Business Machines Corporation Spiral cache memory and method of operating a spiral cache
8271728, Nov 13 2008 International Business Machines Corporation Spiral cache power management, adaptive sizing and interface operations
8364895, Dec 17 2009 International Business Machines Corporation Global instructions for spiral cache management
8370579, Dec 17 2009 International Business Machines Corporation Global instructions for spiral cache management
20020083266,
20020116579,
20020188781,
20030033500,
20030074505,
20030128702,
20030145239,
20030236961,
20040148482,
20050114618,
20050125702,
20050132140,
20050160132,
20060143384,
20060212654,
20070022309,
20090178052,
20100057948,
20100064108,
20100115204,
20100122012,
20100122033,
20100122057,
20100122100,
RE43301, May 10 1996 Apple Inc. Method and apparatus for an improved stack arrangement and operations thereon
////
Executed onAssignorAssigneeConveyanceFrameReelDoc
Feb 14 2013GEBARA, FADI H International Business Machines CorporationASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0327250971 pdf
Feb 14 2013SCHAUB, JEREMY D International Business Machines CorporationASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0327250971 pdf
Feb 19 2013International Business Machines Corporation(assignment on the face of the patent)
Feb 19 2013STRUMPEN, VOLKERInternational Business Machines CorporationASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0327250971 pdf
Date Maintenance Fee Events
Dec 03 2018REM: Maintenance Fee Reminder Mailed.
May 20 2019EXP: Patent Expired for Failure to Pay Maintenance Fees.


Date Maintenance Schedule
Apr 14 20184 years fee payment window open
Oct 14 20186 months grace period start (w surcharge)
Apr 14 2019patent expiry (for year 4)
Apr 14 20212 years to revive unintentionally abandoned end. (for year 4)
Apr 14 20228 years fee payment window open
Oct 14 20226 months grace period start (w surcharge)
Apr 14 2023patent expiry (for year 8)
Apr 14 20252 years to revive unintentionally abandoned end. (for year 8)
Apr 14 202612 years fee payment window open
Oct 14 20266 months grace period start (w surcharge)
Apr 14 2027patent expiry (for year 12)
Apr 14 20292 years to revive unintentionally abandoned end. (for year 12)