A data storage system comprising a matrix of intelligent storage nodes interconnected to communicate with each other via a network of busses. The network of busses includes a plurality of first busses for conducting data from and to a corresponding plurality of host system processors and a plurality of second busses, each one of the second busses intersecting with each one of the first busses. The nodes are located at each intersection. The storage nodes each include a data storage device, such as a magnetic disk drive unit, a processor and buffer memory, whereby the node processor controls the storage and retrieval of data at the node as well as being capable of coordinating the storage and retrieval of data at other nodes within the network.

Patent
   5550986
Priority
Sep 07 1993
Filed
Oct 05 1995
Issued
Aug 27 1996
Expiry
Sep 07 2013
Assg.orig
Entity
Large
39
12
all paid
8. A data storage system comprising:
a network of nodes including a first node, a second node, and a third node, said nodes being interconnected to communicate with each other via a plurality of busses;
each one of said nodes including:
a data storage device connected to receive data from and provide data to at least one of said plurality of busses; and
a node processor connected to said at least one of said busses for controlling the storage and retrieval of data at said data storage device and at data storage devices associated with other nodes of said plurality of nodes through communications between said node processor and said other nodes, for receiving first data from a host system for storage in said data storage system, for dividing said first data into a plurality of second data for storage on said other nodes, wherein each of said plurality of second data is different from said first data, and for providing said second data to said other nodes for storage thereby distributing said first data across said other nodes.
14. A method for transferring data between a host processor and a matrix of data storage nodes, said matrix of data storage nodes including a first bus connecting said host processor to a first and a second data storage node, a second bus connecting said first data storage node to a first plurality of data storage nodes, and a third bus connecting said second data storage node to a second plurality of data storage nodes; each data storage node in said matrix including a data storage device and control logic for controlling the storage and retrieval of data for said matrix of data storage nodes, said method comprising the steps of:
selecting said first data storage node connected to said first bus for receiving first data from said host processor for storage in said first plurality of data storage nodes connected to said second bus
transferring said first data from said host processor to said first data storage node;
dividing said first data into a plurality of second data, wherein each of said plurality of second data is different from said first data; and
transferring said plurality of second data from said first data storage node to said first plurality of data storage nodes so that said first data storage node distributes said first data across said first plurality of data storage nodes.
1. A data storage system comprising:
a plurality of nodes including a first node, a second node, and a third node;
a plurality of first busses, at least one of said first busses for conducting data from and to a corresponding host system processor; and
a plurality of second busses;
wherein each one of said plurality of nodes connects with a different pairing of one of said first busses with one of said second busses;
each one of said plurality of nodes including:
a data storage device connected to receive data from and provide data to the first and second busses associated with said one of said plurality of nodes; and
a node processor connected to said first and second busses associated with said one of said plurality of nodes for controlling the storage and retrieval of data at said data storage device and for receiving a first command from a host system relating to one of storage and retrieval of data in said data storage system and developing a second command from said first command that is different from said first command for controlling one of storage and retrieval of data at said data storage device associated with at least one other node of said plurality of nodes through communications between said node processor and said at least one other node, thereby distributing the control of the storage and retrieval of data in said data storage system from a host system to said plurality of nodes.
12. A method for communicating data between a host processor and a data storage system comprising a network of a plurality of nodes connected together via a plurality of first busses and a plurality of second busses; each one of said nodes being connected to one of said plurality of first busses and one of said plurality of second busses; each one of said plurality of nodes including a data storage device connected to receive data from and provide data to said one of said first and the second- busses associated with said one of said plurality of nodes; and a node processor connected to said one of said first and second busses associated with said one of said plurality of nodes for controlling the storage and retrieval of data at said data storage device and other nodes of said plurality of nodes through communications between said node processor and said other nodes; said method comprising the steps of:
selecting a first selected node for receiving a first command relating to one of storage and retrieval of data within said data storage system from a host processor connected to one of said plurality of first busses;
controlling the operation of nodes connected to said second bus associated with said first selected node by said node processor of said first selected node developing a second command from said first command that is different from said first command and providing said second command to nodes connected to said second bus associated with said first selected node, thereby distributing the control of the storage and retrieval of data within said data storage system from the host system to said plurality of nodes.
7. A data storage system comprising:
a network of nodes connected together via a plurality of first busses and a plurality of second busses, each one of said nodes being connected to one of said plurality of first busses and one of said plurality of second busses,
each one of said nodes including:
a data storage device connected to receive data from and provide data to said one of said plurality of first and second busses associated with said one of said nodes; and
a node processor connected to said one of said plurality of first and second busses associated with said one of said nodes for controlling the storage and retrieval of data at said data storage device and at data storage devices associated with other of said nodes;
said network of nodes including:
a first node for receiving a first command from a host system relating to one of storage and retrieval of data in said data storage system and for developing a second command from said first command that is different from said first command for controlling one of storage and retrieval of data at said data storage device associated with a second node; and
a third node for receiving a third command from a host system relating to one of storage and retrieval of data within said data storage system and for developing a fourth command from said third command that is different from said third command, said second command, and said first command for controlling one of storage and retrieval of data at said data storage device associated with a fourth node, wherein said first and third nodes concurrently control said second and fourth nodes, respectively, thereby providing increased performance for storage and retrieval operations between a host system and said data storage system.
2. The data storage system according to claim 1, wherein:
said additional nodes comprise nodes connected to the second data bus associated with said one of said plurality of nodes.
3. The data storage system according to claim 1, wherein:
at least one of said second busses conducts data from one of said first, second, and third nodes to each of the others of said first, second, and third nodes without conducting data through any node before said data is received by said each of the others of said first, second, and third nodes.
4. The data storage system according to claim 1, wherein:
said first command relates to storing data from a host system to said data storage system; and
said second command relates to retrieving old parity information from said at least one other node.
5. The data storage system according to claim 1, wherein:
a number of commands, including said second command, are developed by said node processor based on said first command and in which said number of commands depends upon the manner in which said data is stored in said data storage system.
6. The data storage system according to claim 5, wherein:
said number of commands is related to the number of said plurality of nodes in which data is stored.
9. The data storage system according to claim 8, wherein said plurality of busses comprises:
a plurality of first busses for conducting data from and to a corresponding plurality of host system processors; and
a plurality of second busses, each one of said plurality of second busses intersecting with each one of said plurality of first busses;
said nodes being located at the intersections of said first busses with said second busses.
10. The data storage system according to claim 9, wherein within each one of said nodes:
said data storage device is connected to receive data from and provide data to the first and the second busses associated with said one of said nodes; and
said node processor is connected to said first and second busses associated with said one of said nodes for controlling the storage and retrieval of data at said data storage device.
11. The data storage system according to claim 8, wherein:
at least one of said plurality of busses conducts data from one of said first, second, and third nodes to each of the others of said first, second, and third nodes without conducting data through any node before said data is received by said each of the others of said first, second, and third nodes.
13. The method according to claim 12, further comprising the steps of:
selecting a second selected node for receiving a third command relating to one of storage and retrieval of data within said data storage system from a host processor connected to one of said plurality of first busses;
controlling the operation of nodes connected to said second bus associated with said second selected node by said node processor of said second selected node developing a fourth command from said third command that is different from said third command and providing said fourth command to nodes connected to said second bus associated with said second selected node, wherein said step of controlling the operation of nodes associated with said first selected node and said step of controlling the operation of nodes associated with said second selected node are performed concurrently, thereby providing increased performance when communicating data between a host processor and said data storage system.
15. The method according to claim 14, further comprising the steps of:
selecting said second data storage node for receiving third data from said host processor for storage in said second plurality of data storage nodes connected to said third bus;
dividing said third data into a plurality of fourth data, wherein each of said plurality of fourth data is different from said third data, said second data, and said first data; and
transferring said plurality of fourth data from said second data storage node to said second plurality of data storage nodes so that said second data storage node distributes said third data across said second plurality of data storage nodes;
wherein said step of transferring each of said plurality of second data and said step of transferring each of said plurality of fourth data are performed concurrently thereby providing increased performance when transferring data between a host processor and said matrix of data storage nodes.

This is a continuation of application Ser. No. 08/124,653 filed Sep. 7, 1993, now abandoned.

The present invention relates to data storage systems and, more particularly, to a data storage system architecture which enables the distribution of protected storage tasks among a matrix of data storage devices.

Disk array storage systems include a plurality of hard disk drives, such as the 51/4 or 31/2 inch disk drives currently used in personal computers and workstations, which operate in parallel and appear to the host system as a single large disk drive. Numerous disk array design alternatives are possible, incorporating a few to many disk drives. Several array alternatives, each possessing different attributes, benefits and shortcomings, are presented in an article titled "A Case for Redundant Arrays of Inexpensive Disks (RAID)" by David A. Patterson, Garth Gibson and Randy H. Katz; University of California Report No. UCB/CSD 87/391, December 1987. The article, incorporated herein by reference, discusses disk arrays and the improvements in performance, reliability, power consumption and scalability that disk arrays provide in comparison to single large magnetic disks.

Complex storage management techniques are required in order to coordinate the operation of the multitude of data storage devices within an array to perform read and write functions, parity generation and checking, data restoration and reconstruction, and other necessary or optional operations. Array operation can be managed by a dedicated hardware controller constructed to control array operations, i.e., a hard array architecture, or through software routines executed by the host computer system, i.e., a software array architecture.

A hardware array improves data storage reliability and availability, and increases system performance without modifying user applications. A hardware controller removes many of the data storage tasks from the host system processor, permitting greater accommodation of other applications and tasks by the host system.

A software architecture allows existing system hardware and storage devices to be utilized as an array storage system. The increasing speed and power of advanced computer systems provides performance that is competitive with many hardware array products. A software array architecture can deliver this functionality at a lower cost of implementation than a hardware array. Greater configuration flexibility may also be provided. However, the host processor remains burdened with data storage tasks.

The hardware and software array alternatives discussed above provide improvements in performance, reliability, power consumption, scalability and capacity in comparison to single large magnetic disks. However, coincident with these improvements there exists a need to store and manage ever increasing amounts of data, and to utilize the host processor, array controller and storage devices in the most efficient and effective manner. Additional improvements to data storage systems are desired to better utilize the speed and power of current and next generation computer systems, particularly multiple processor computer systems.

It is therefore an object of the present invention to provide a new and useful data storage system which overcomes many of the disadvantages associated with existing storage options.

It is another object of the present invention to provide a new and useful data storage system which includes a network of interconnected nodes, each node having a storage device and a processor possessing the capability for coordinating data storage and retreival operations at several nodes within the network.

It is yet another object of the present invention to provide such a data storage system wherein multiple storage operations may be performed concurrently, each operation being coordinated by a different node within the network.

There is provided, in accordance with the present invention, a data storage system comprising a network of nodes interconnected to communicate with each other via a plurality of busses. Each node includes a data storage device, such as a magnetic disk drive unit, connected to receive data from and provide data to at least one of the plurality of busses, and a node processor connected to the bus for controlling the storage and retrieval of data at the node. The node processor is further capable of controlling the storage and retrieval of data at data storage devices associated with additional nodes within the network through communications between node processor and the additional nodes via the plurality of busses.

In the described embodiment, the plurality of busses includes a plurality of first busses for conducting data from and to a corresponding plurality of host system processors and a plurality of second busses, each one of the second busses intersecting with each one of the first busses. The network nodes are provided at each intersection. Each one of the nodes includes a disk storage device connected to receive data from and provide data to the first and the second busses associated with the node, and a processor connected to the first and second busses associated with the node for controlling the storage and retrieval of data at the node as well as being capable of controlling the storage and retrieval of data at other nodes within the network.

The above and other objects, features, and advantages of the present invention will become apparent from the following description and the attached drawings.

FIG. 1 is a diagrammatic illustration of a data storage system including a plurality of disk drives and inexpensive processors located within a matrix network, constructed in accordance with the present invention.

FIG. 2 is a block diagram showing the processor, disk drive, and associated elements located within each node of the matrix network illustrated in FIG. 1.

Referring now to FIGS. 1 and 2, there is seen a data storage system in accordance with the present invention. The architecture shown in FIG. 1 includes a host processor connection block 12 providing connection to one or more host system processors, not shown. The host processors are identified by reference numerals H0, H1, H2, . . . Hm. Connection block 12 couples host processors H0, H1, H2, . . . Hm to a network 14 of data storage nodes. Network 14 includes several busses, R0 through Rm, arranged in rows, each bus connecting one of host processors H0 through Hm with a group of storage nodes. Network 14 further includes several busses, C0 through Cn, arranged in columns. A node is formed at every intersection between a row and column bus. The nodes are identified by pairs of coordinates, the first coordinate referring to the number of the row bus to which it connects, and the second coordinate identifying the column bus to which the node connects. The network includes nodes from (0, 0), at the intersection of busses R0 and C0, through (m, n), at the intersection of busses Rm and Cn.

H0 in the configuration shown is connected directly to storage nodes (0, 0) through (0, n) through bus R0. In addition, H0 is provided access to all the storage nodes on bus C0, i.e., nodes (1, 0) through (m, 0) by passing through node (0, 0). Nodes (0, 1 ) through (0, n) similarly provide access for processor H0 to the nodes on busses C1 through Cn, respectfully. Each one of host processors H1 through Hm has direct access to all the storage nodes on busses R1 through Rm, respectively, and access through interconnecting nodes to all the storage nodes on network 14.

Host processor connection block 12 may include logic for executing group array algorithms, such as the RAID algorithms that are necessary for issuing I/O operations, handling error exception conditions, and performing data reconstruction, when a storage device in network 14 fails. Other functions of the logic included in connection block 12 may include diagnostic and group algorithm initialization executed in response to input provided by a system administration. In a high performance configuration, a host processor connection block will exist for every row bus (R0 through Rm) that is shown in node network 14. The high performance configuration allows multiple I/O commands and data to flow over the attached row busses simultaneously. In a lower performance, lower cost configuration command and data flow over one row bus.

Each of storage nodes (0, 0) through (m, n) includes a storage device, node processor, buffers and interface logic as shown in FIG. 2. A block diagram showing the processor, disk drive, and associated elements located within node (m, n) of network 14 is shown.

Node (m, n) is seen to include an interface I/F 1 to column bus Cn, a second interface I/F 2 to row bus Rm, an inexpensive processor P, data buffers B1, B2, I and B3, and a storage element D, such as a Head Disk Assembly (HDA) for storing and retrieving data. Node processor P and data buffers B1, B2, I and B3 are connected to interface I/F 1 and thereby to network bus Cn by a node bus identified as BUS 1. A second bus, identified as BUS 2, provides connection between node processor P and data buffers B1, B2, I and B3 and interface I/F 2, which thereby provides connection to network bus Rm. Read/write buffer B3 also provides the node connection to storage element D. Nodes (0, 0) through (m, n) are similarly constructed.

Node processor P, in a conventional sense, controls the network protocols, buffer management, error recovery and storage media control such as head positioning, data encoding/decoding and defect handling. A typical example of the network node could be a Small Computer System Interface (SCSI) disk drive.

In operation, array storage requests are received from one or more host processors and directed to designated nodes within network 14 for execution. An exemplary array operation could be for H0 to issue a RAID level 5 write operation. The command is formed in a packetized mode for serial connections, or in handshake mode for parallel connections, and issued to appropriate nodes over bus R0. H0 could issue a write to any desired node (0,0) to (0,n) residing on bus R0. The node that receives the command will be referred to in the discussion which follows as the primary node. Remaining network nodes will be referred to as secondary nodes. The command contains information about secondary nodes that will be involved in subsequent read/write operations spawned by the primary node. Spawned operations are necessary to complete the RAID level 5 write command. The primary node upon receiving a command takes responsibility for the operation if no error conditions occur. The primary node will report status conditions to the appropriate host processors for irregular conditions.

The data storage system described above permits the distribution of the compute power necessary to execute the array algorithms and functions to the nodes of a generalized network. The network can consist of intelligent disk drives such that the array algorithms and most common functions are executed at the array nodes.

The host system is relieved of many of the array storage operations. Additionally, several array requests may be executed concurrently, each request being processed by a different primary node. The system thereby realizes increased performance beyond the capabilities of a storage system employing a single hardware controller.

The two main attributes of the described system are:

1. Increase in performance because each node contains sufficient processor power to relieve either the Host processor or the H/W array processor; and

2. Relieves the bandwidth bottleneck of the I/O connection since multiple I/O paths can be used to connect the array nodes.

The invention, therefore is very adaptable to various network architectures and provides improvements in network storage performance. This is due to the compute power which is available independent of host system application load. The invention is also intended to improve the incremental capacity and the reliability of computer networks.

It is important to note that network 14 can be a generalized switching arrangement that would provide a multitude of paths to the individual storage devices coupled to the network.

Listed below are example scenarios to show the execution of the exemplary operations by the storage system according to the present invention.

______________________________________
Operation Primary Secondary
Number Host Node Node Operation
______________________________________
1 H0 (0,1) (1,1) Write
2 H1 (1,0) -- Read
3 H2 (2,2) (1,2) Write
______________________________________

Operation 1:

H0 issues a RAID level 5 write to node (0,1). H0 passes commands and data to node (0, 1 ) processor P and buffer B1, respectively, over network bus R0 and node bus BUS 1. Node (0, 1) processor P decodes the command and determines a read-modify-write cycle is necessary involving secondary node (1,1). Node (0, 1) processor P issues a read command with node (0, 1) identified as the source to node (1,1). The command is communicated to node (1, 1) via bus C1.

Simultaneously processor P in node (0,1) issues a read to HDA device D in node (0,1) to read old data from HDA device D into Buffer I.

Node processor P in node (1,1) receives the read command via bus C1, interface block I/F 1, and node bus BUS 1. Node (1,1) processor P decodes the received read command and retrieves read data from HDA device D into buffer I. Node (0,1) and (1,1) complete their respective reads asynchronously. When the reads are complete, node (0,1) contains new data in buffer B1 and old data in buffer I. Node (1,1) contains old parity in its buffer I. Node (1,1) informs node (0,1) that old parity data is in buffer. Node (0,1) reads old parity data over column bus C1 into node (0,1) buffer B2. Node (0,1) now has new data, old data and old parity in its buffer.

To complete the RAID 5 write operation, node processor (0,1) orders an exclusive-OR of the data stored within buffers B1, B2 and I to generate the new parity data. The new parity is placed in buffer I and readied for transmission to node (1,1) for parity update. Simultaneously, node (0,1) writes the new data from buffer B1 to buffer B3 for writing to storage device D. Node (0,1) issues a normal write command of new parity from buffer I.

Node (1,1) informs node (0,1) that parity write is complete and, in turn, when node (0,1) completes write of new data, informs host processor H0 that the RAID level 5 write is complete.

Operation 2:

Host processor H1 issues a normal read to node (1,0) over row bus R1. Upon completion of the read, node (1,0) reports over bus R1 to processor H1 that the operation has completed.

Operation 3:

Operation 3 occurs identical to operation 1 except command and data is passed over row bus R2 and column bus C2 and report operation complete messages provided to host H2 over bus R2.

Operations 1, 2 and 3 may be performed concurrently.

As shown by the scenarios described above, the architecture enables multiple concurrent operations that distributes the RAID algorithms over the array of nodes. The nodes act as peers and operate in a dynamic client/server mode. This invention facilitates expansion of nodes in both row and column directions. Such expansion permits improvement in performance and capacity without impacting the host processor performance.

The node operation is generalized and could be implemented so that each node can manage as a primary or secondary mode and communicate over a multiplicity of channels.

It can thus be seen that there has been provided by the present invention a data storage system which provides increased performance beyond the capabilities of a host system managed storage system or a storage system employing a single hardware controller. The system described above permits the execution of multiple storage operations concurrently, each operation being coordinated by a different node within the storage network.

This architecture is scalable by design and may be expanded by the addition of nodes in both the row and column direction. In addition, the architecture is not limited to use with magnetic disk drive devices. It can be used to provide RAID technology on sequential access devices (e.g. QIC tapes, DAT tapes, etc.) as well as other direct access devices (e.g., optical disks and media changers) and robotic media changer storage devices. The system can be connected to a single host processor or may be interconnected with several host processors within a multiple processor computer system.

Although the presently preferred embodiment of the invention has been described, it will be understood that various changes may be made within the scope of the appended claims.

DuLac, Keith B.

Patent Priority Assignee Title
10019176, Oct 30 2012 Hewlett Packard Enterprise Development LP Smart memory buffers
10585602, Oct 30 2012 Hewlett Packard Enterprise Development LP Smart memory buffers
5745709, Jul 05 1994 Matsushita Electric Industrial C., Ltd. Data transfer apparatus and system providing high speed switching to allow for high speed data transfer between one device and multiple devices
5815649, Oct 20 1995 Alcatel Lucent Distributed fault tolerant digital data storage subsystem for fault tolerant computer system
6134630, Nov 14 1997 Summit Data Systems LLC High-performance bus architecture for disk array system
6353446, Jan 25 1999 Network General Technology Method and system for integrated network management applications
6356803, May 19 2000 International Business Machines Corporation Automated data storage library distributed control system
6421760, Nov 14 1997 Summit Data Systems LLC Disk array controller, and components thereof, for use with ATA disk drives
6549981, Nov 14 1997 Summit Data Systems LLC Disk array system with controllers that automate host side of ATA interface
6633996, Apr 13 2000 STRATUS TECHNOLOGIES IRELAND LTD Fault-tolerant maintenance bus architecture
6687851, Apr 13 2000 STRATUS TECHNOLOGIES IRELAND LTD Method and system for upgrading fault-tolerant systems
6691257, Apr 13 2000 STRATUS TECHNOLOGIES IRELAND LTD Fault-tolerant maintenance bus protocol and method for using the same
6708283, Apr 13 2000 STRATUS TECHNOLOGIES IRELAND LTD System and method for operating a system with redundant peripheral bus controllers
6735715, Apr 13 2000 STRATUS TECHNOLOGIES IRELAND LTD System and method for operating a SCSI bus with redundant SCSI adaptors
6766413, Mar 01 2001 STRATUS TECHNOLOGIES IRELAND LTD Systems and methods for caching with file-level granularity
6766479, Feb 28 2001 STRATUS TECHNOLOGIES IRELAND LTD Apparatus and methods for identifying bus protocol violations
6802022, Apr 14 2000 STRATUS TECHNOLOGIES IRELAND LTD Maintenance of consistent, redundant mass storage images
6820213, Apr 13 2000 STRATUS TECHNOLOGIES IRELAND LTD Fault-tolerant computer system with voter delay buffer
6862689, Apr 12 2001 STRATUS TECHNOLOGIES IRELAND LTD Method and apparatus for managing session information
6874102, Mar 05 2001 STRATUS TECHNOLOGIES IRELAND LTD Coordinated recalibration of high bandwidth memories in a multiprocessor computer
6886171, Feb 20 2001 STRATUS TECHNOLOGIES IRELAND LTD Caching for I/O virtual address translation and validation using device drivers
6901481, Apr 14 2000 STRATUS TECHNOLOGIES IRELAND LTD Method and apparatus for storing transactional information in persistent memory
6948010, Dec 20 2000 STRATUS TECHNOLOGIES IRELAND LTD Method and apparatus for efficiently moving portions of a memory block
6996750, May 31 2001 STRATUS TECHNOLOGIES IRELAND LTD Methods and apparatus for computer bus error termination
7065672, Mar 28 2001 STRATUS TECHNOLOGIES IRELAND LTD Apparatus and methods for fault-tolerant computing using a switching fabric
7203731, Mar 03 2000 Intel Corporation Dynamic replication of files in a network storage system
7266555, Mar 03 2000 Intel Corporation Methods and apparatus for accessing remote storage through use of a local device
7266556, Dec 29 2000 SCALE EIGHT, INC Failover architecture for a distributed storage system
7281168, Mar 03 2000 Intel Corporation Failover architecture for local devices that access remote storage
7428540, Mar 03 2000 Intel Corporation Network storage system
7496654, Jun 29 2001 Microsoft Technology Licensing, LLC Multi-threaded system for activating a process using a script engine and publishing data descriptive of the status of the process
7506034, Mar 03 2000 Intel Corporation Methods and apparatus for off loading content servers through direct file transfer from a storage center to an end-user
7509645, Oct 17 2002 Intel Corporation Methods and apparatus for load balancing storage nodes in a distributed network attached storage system
7590747, Mar 03 2000 Intel Corporation Distributed storage cluster architecture
7774325, Oct 17 2002 Intel Corporation Distributed network attached storage system
7774466, Oct 17 2002 Intel Corporation Methods and apparatus for load balancing storage nodes in a distributed storage area network system
8145736, Jun 30 2003 Symantec Corporation Fast distributed object lookup for a computer network
9442877, Jan 10 2013 Kioxia Corporation Storage device
9891841, Dec 27 2013 Kioxia Corporation Storage system including a plurality of memory nodes connected through first and second groups of interfaces
Patent Priority Assignee Title
4195344, Apr 08 1977 The President of the Agency of Industrial Science and Technology Computer system with a configuration monitor
4633431, Apr 22 1982 Siemens Aktiengesellschaft Arrangement for coupling digital processing units
4797882, Oct 02 1985 American Telephone and Telegraph Company, AT&T Bell Laboratories Mesh-based switching network
4821170, Apr 17 1987 HEWLETT-PACKARD DEVELOPMENT COMPANY, L P Input/output system for multiprocessors
4933933, Dec 19 1986 The California Institute of Technology Torus routing chip
5041971, Nov 30 1988 Level 3 Communications, LLC Memory accessing switch network
5124987, Apr 16 1990 Storage Technology Corporation; STORAGE TECHNOLOGY CORPORAITON, A CORP OF DELAWARE Logical track write scheduling system for a parallel disk drive array data storage subsystem
5168572, Mar 10 1989 The Boeing Company; BOEING COMPANY, THE, System for dynamic selection of globally-determined optimal data path
5181017, Jul 27 1989 IBM Corporation Adaptive routing in a parallel computing system
5223968, Dec 20 1990 The United States of America as represented by the Secretary of the Air First come only served communications network
5317735, Jun 14 1990 U S PHILIPS CORPORATION A CORP OF DE System for parallel computation with three phase processing in processor tiers in which new instructions trigger execution and forwarding
5319638, Sep 12 1991 TTI Inventions A LLC Link-by-link congestion control for packet transmission systems
///////////////
Executed onAssignorAssigneeConveyanceFrameReelDoc
Oct 05 1995AT&T Global Information Solutions Company(assignment on the face of the patent)
Oct 05 1995Hyundai Electronics America(assignment on the face of the patent)
Oct 05 1995Symbios Logic Inc.(assignment on the face of the patent)
Dec 10 1997SYMBIOS LOGIC INC SYMBIOS, INC CHANGE OF NAME SEE DOCUMENT FOR DETAILS 0090890936 pdf
Feb 26 1998HYUNDAI ELECTRONICS AMERICA, A CORP OF CALIFORNIALEHMAN COMMERCIAL PAPER INC , AS ADMINISTRATIVE AGENTSECURITY AGREEMENT0093960441 pdf
Feb 26 1998SYMBIOS, INC , A CORP OF DELAWARELEHMAN COMMERCIAL PAPER INC , AS ADMINISTRATIVE AGENTSECURITY AGREEMENT0093960441 pdf
Aug 06 1998SYMBIOS, INC Hyundai Electronics AmericaTERMINATION AND LICENSE AGREEMENT0095960539 pdf
Apr 12 2001Hyundai Electronics AmericaHYNIX SEMICONDUCTOR AMERICA INC CHANGE OF NAME SEE DOCUMENT FOR DETAILS 0152460599 pdf
Sep 20 2004HYNIX SEMICONDUCTOR AMERICA, INC Hynix Semiconductor IncASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0152790556 pdf
Oct 04 2004Hynix Semiconductor, IncMagnaChip Semiconductor, LtdASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0162160649 pdf
Dec 23 2004MagnaChip Semiconductor, LtdU S BANK NATIONAL ASSOCIATION, AS COLLATERAL TRUSTEESECURITY INTEREST SEE DOCUMENT FOR DETAILS 0164700530 pdf
Jan 07 2005LEHMAN COMMERICAL PAPER INC SYMBIOS, INC RELEASE OF SECURITY INTEREST0166020895 pdf
Jan 07 2005LEHMAN COMMERICAL PAPER INC Hyundai Electronics AmericaRELEASE OF SECURITY INTEREST0166020895 pdf
Nov 14 2007MagnaChip Semiconductor, LtdTAIWAN SEMICONDUCTOR MANUFACTURING CO , LTD ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0213980702 pdf
Nov 15 2007NCR CorporationTAIWAN SEMICONDUCTOR MANUFACTURING CO , LTD ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0213980702 pdf
Date Maintenance Fee Events
Sep 14 1999ASPN: Payor Number Assigned.
Feb 25 2000M183: Payment of Maintenance Fee, 4th Year, Large Entity.
Jan 21 2004M1552: Payment of Maintenance Fee, 8th Year, Large Entity.
Mar 03 2008REM: Maintenance Fee Reminder Mailed.
Mar 13 2008M1553: Payment of Maintenance Fee, 12th Year, Large Entity.
Mar 13 2008M1556: 11.5 yr surcharge- late pmt w/in 6 mo, Large Entity.


Date Maintenance Schedule
Aug 27 19994 years fee payment window open
Feb 27 20006 months grace period start (w surcharge)
Aug 27 2000patent expiry (for year 4)
Aug 27 20022 years to revive unintentionally abandoned end. (for year 4)
Aug 27 20038 years fee payment window open
Feb 27 20046 months grace period start (w surcharge)
Aug 27 2004patent expiry (for year 8)
Aug 27 20062 years to revive unintentionally abandoned end. (for year 8)
Aug 27 200712 years fee payment window open
Feb 27 20086 months grace period start (w surcharge)
Aug 27 2008patent expiry (for year 12)
Aug 27 20102 years to revive unintentionally abandoned end. (for year 12)