Disclosed herein are apparatuses and methods related to an artificial intelligence accelerator in memory. An apparatus can include a number of registers configured to enable the apparatus to operate in an artificial intelligence mode to perform artificial intelligence operations and an artificial intelligence (ai) accelerator configured to perform the artificial intelligence operations using the data stored in the number of memory arrays. The ai accelerator can include hardware, software, and or firmware that is configured to perform operations associated with ai operations. The hardware can include circuitry configured as an adder and/or multiplier to perform operations, such as logic operations, associated with ai operations.
|
15. An apparatus, comprising:
a number of memory arrays;
a controller; and
a number of artificial intelligence (ai) registers, wherein the number of artificial intelligence registers define an input, an output, a first temporary block, a second temporary block, a neural network, a number of activation functions, a number of bias values for artificial intelligence operations performed by the apparatus.
24. A method, comprising:
programming a first bit of a first artificial intelligence (ai) register to a particular state, wherein the first bit at the particular state enables an apparatus to operate in an ai mode and perform artificial intelligence operations; and
programing a second bit of the first ai register to a particular state, wherein the second bit at the particular state starts an ai operation.
10. A system, comprising:
a host; and
a memory device coupled to the host, wherein the memory device includes a number of memory arrays and a controller configured to:
receive a command from the host to enable the memory device to perform artificial intelligence (ai) operations using data stored in the number of memory arrays,
wherein a first bit of an ai register is programmed to a first state to start the artificial intelligence operations;
enable the memory device to perform the artificial intelligence operations by writing an indicator in a register.
1. A memory device, comprising:
a number of memory arrays;
a controller configured to receive a command to enable the memory device to perform artificial intelligence operations using data stored in the number of memory arrays,
wherein a particular bit of an artificial intelligence register is programmed to a particular state to indicate the artificial intelligence operations have started;
a number of registers configured to enable the apparatus to operate in an artificial intelligence mode to perform the artificial intelligence operations; and
an artificial intelligence (ai) accelerator configured to perform the artificial intelligence operations using data stored in the number of memory arrays.
2. The apparatus of
3. The apparatus of
4. The apparatus of
5. The apparatus of
6. The apparatus of
7. The apparatus of
8. The apparatus of
9. The apparatus of
11. The system of
12. The system of
13. The system of
14. The system of
16. The apparatus of
17. The apparatus of
18. The apparatus of
19. The apparatus of
20. The apparatus of
21. The apparatus of
22. The apparatus of
23. The apparatus of
25. The method
26. The method
|
The present disclosure relates generally to memory devices, and more particularly, to apparatuses and methods for memory with an artificial intelligence (AI) mode.
Memory devices are typically provided as internal, semiconductor, integrated circuits in computers or other electronic devices. There are many different types of memory including volatile and non-volatile memory. Volatile memory can require power to maintain its data and includes random-access memory (RAM), dynamic random access memory (DRAM), and synchronous dynamic random access memory (SDRAM), among others. Non-volatile memory can provide persistent data by retaining stored data when not powered and can include NAND flash memory, NOR flash memory, read only memory (ROM), Electrically Erasable Programmable ROM (EEPROM), Erasable Programmable ROM (EPROM), and resistance variable memory such as phase change random access memory (PCRAM), resistive random access memory (RRAM), and magnetoresistive random access memory (MRAM), among others.
Memory is also utilized as volatile and non-volatile data storage for a wide range of electronic applications. Non-volatile memory may be used in, for example, personal computers, portable memory sticks, digital cameras, cellular telephones, portable music players such as MP3 players, movie players, and other electronic devices. Memory cells can be arranged into arrays, with the arrays being used in memory devices.
The present disclosure includes apparatuses and methods related to memory with an artificial intelligence (AI) mode. An example apparatus can include a number of registers configured to enable the apparatus to operate in an artificial intelligence mode to perform artificial intelligence operations and an artificial intelligence (AI) accelerator configured to perform the artificial intelligence operations using the data stored in the number of memory arrays. The AI accelerator can include hardware, software, and or firmware that is configured to perform operations associated with AI operations. The hardware can include circuitry configured as an adder and/or multiplier to perform operations, such as logic operations, associated with AI operations.
The apparatus can be configured to operate in normal mode, where the apparatus performs read and write operations, like a memory device would traditionally operate. The apparatus can be configured to operate in an AI accelerator mode, where the apparatus can be configured to perform AI operations. The apparatus can include a number of registers that can place the apparatus normal mode and/or AI mode. The number of registers can also define the location and status of the inputs, outputs, bias information of a neural network, matrixes of weights representing a neural network, and/or activation functions used by the AI accelerator to perform AI operations.
The AI accelerator can reduce latency and power consumption associated with AI operations when compared to AI operations that are performed on a host. AI operations performed on a host use data that is exchanged between a memory device and the host, which adds latency and power consumption to the AI operations. While AI operations performed according to embodiments of the present disclosure can be performed on a memory device using the AI accelerator and the memory arrays, where data is not transferred from the memory device while performing the AI operations.
In the following detailed description of the present disclosure, reference is made to the accompanying drawings that form a part hereof, and in which is shown by way of illustration how a number of embodiments of the disclosure may be practiced. These embodiments are described in sufficient detail to enable those of ordinary skill in the art to practice the embodiments of this disclosure, and it is to be understood that other embodiments may be utilized and that process, electrical, and/or structural changes may be made without departing from the scope of the present disclosure. As used herein, the designator “N” indicates that a number of the particular feature so designated can be included with a number of embodiments of the present disclosure.
As used herein, “a number of” something can refer to one or more of such things. For example, a number of memory devices can refer to one or more of memory devices. Additionally, designators such as “N”, as used herein, particularly with respect to reference numerals in the drawings, indicates that a number of the particular feature so designated can be included with a number of embodiments of the present disclosure.
The figures herein follow a numbering convention in which the first digit or digits correspond to the drawing figure number and the remaining digits identify an element or component in the drawing. Similar elements or components between different figures may be identified by the use of similar digits. As will be appreciated, elements shown in the various embodiments herein can be added, exchanged, and/or eliminated so as to provide a number of additional embodiments of the present disclosure. In addition, the proportion and the relative scale of the elements provided in the figures are intended to illustrate various embodiments of the present disclosure and are not to be used in a limiting sense.
As illustrated in
Host 102 includes a host controller 108 to communicate with memory device 120. The host controller 108 can send commands to the memory device 120. The host controller 108 can communicate with the memory device 120, memory controller 122 on memory device 120, and/or the AI accelerator 124 on memory device 120 to perform AI operations, read data, write data, and/or erase data, among other operations. AI accelerator 124 can also include components described in associated with
A physical host interface can provide an interface for passing control, address, data, and other signals between memory device 120 and host 102 having compatible receptors for the physical host interface. The signals can be communicated between host 102 and memory device 120 on a number of buses, such as a data bus and/or an address bus, for example.
Memory device 120 can include controller 120, AI accelerator 124, and memory arrays 125-1, . . . , 125-N. Memory device 120 can be a low-power double data rate dynamic random access memory, such as a LPDDR5 device, and/or a graphics double data rate dynamic random access memory, such as a GDDR6 device, among other types of devices. Memory arrays 125-1, . . . , 125-N can include a number of memory cells, such as volatile memory cells (e.g., DRAM memory cells, among other types of volatile memory cells) and/or non-volatile memory cells (e.g., RRAM memory cells, among other types of non-volatile memory cells). Memory device 120 can read and/or write data to memory arrays 125-1, . . . , 125-N. Memory arrays 125-1, . . . , 125-N can store data that is used during AI operations performed on memory device 120. Memory arrays 125-1, . . . , 125-N can store inputs, outputs, weight matrix and bias information of a neural network, and/or activation functions information used by the AI accelerator to perform AI operations on memory device 120.
The host controller 108, memory controller 122, and/or AI accelerator 124 on memory device 120 can include control circuitry, e.g., hardware, firmware, and/or software. In one or more embodiments, the host controller 108, memory controller 122, and/or AI accelerator 124 can be an application specific integrated circuit (ASIC) coupled to a printed circuit board including a physical interface. Also, memory controller 122 on memory device 120 can include registers 130. Registers 130 can be programmed to provide information for the AI accelerator to perform AI operations. Registers 130 can include any number of registers. Registers 130 can be written to and/or read by host 102, memory controller 122, and/or AI accelerator 124. Registers 130 can provide input, output, neural network, and/or activation functions information for AI accelerator 124. Registers 130 can include mode register 131 to select a mode of operation for memory device 120. The AI mode of operation can be selected by writing the word 0xAA and/or 0x2AA, for example, to register 131, which inhibits access to the registers associated with normal operation of memory device 120 and allows access to the registers associated with AI operations. Registers 130 can also be located in memory arrays 125-1, . . . , 125-N and be accessible by controller 122.
AI accelerator 124 can include hardware 126 and/or software/firmware 128 to perform AI operations. Also, AI accelerator 124 can also include components described in associated with
The memory arrays 125-1, . . . , 125-N can provide main memory for the memory system or could be used as additional memory or storage throughout the memory system. Each memory array 125-1, . . . , 125-N can include a number of blocks of memory cells. The blocks of memory cells can be used to store data that is used during AI operations performed by memory device 120. Memory arrays 125-1, . . . , 125-N can include DRAM memory cells, for example. Embodiments are not limited to a particular type of memory device. For instance, the memory device can include RAM, ROM, DRAM, SDRAM, PCRAM, RRAM, 3D Xpoint and flash memory, among others.
By way of example, memory device 120 may perform an AI operation that is or includes one or more inference steps. Memory arrays 125 may be layers of a neural network or may each be individual nodes and memory device 120 may be layer; or memory device 120 may be a node within a larger network. Additionally or alternatively, memory arrays 125 may store data or weights, or both, to be used (e.g., summed) within a node. Each node (e.g., memory array 125) may combine an input from data read from cells of the same or a different memory array 125 with weights read from cells of memory array 125. Combinations of weights and data may, for instance, be summed within the periphery of a memory array 125 or within hardware 126 using adder/multiplier 127. In such cases, the summed result may be passed to an activation function represented or instantiated in the periphery of a memory array 125 or within hardware 126. The result may be passed to another memory device 120 or may be used within AI accelerator 124 (e.g., by software/firmware 128) to make a decision or to train a network that includes memory device 120.
A network that employs memory device 120 may be capable of or used for supervised or unsupervised learning. This may be combined with other learning or training regimes. In some cases, a trained network or model is imported or used with memory device 120, and memory device's 120 operations are primarily or exclusively related to inference.
The embodiment of
Register 232-0 can define parameters associated with AI mode of the memory device. Bits in register 232-0 can start AI operations, restart AI operations, indicate content in registers is valid, clear content from registers, and/or exit from AI mode.
Registers 232-1, 232-2, 232-3, 232-4, and 232-5 can define the size of inputs used in AI operations, the number of inputs used in AI operations, and the start address and end address of the inputs used in AI operations. Registers 232-7, 232-8, 232-9, 232-10, and 232-11 can define the size of outputs of AI operations, the number of outputs in AI operations, and the start address and end address of the outputs of AI operations.
Register 232-12 can be used to enable the usage of the input banks, the neuron banks, the output banks, the bias banks, the activation functions, and the temporary banks used during AI operations.
Registers 232-13, 232-14, 232-15, 232-16, 232-17, 232-18, 232-19, 232-20, 232-21, 232-22, 232-23, 232-24, and 232-25 can be used to define the neural network used during AI operations. Registers 232-13, 232-14, 232-15, 232-16, 232-17, 232-18, 232-19, 232-20, 232-21, 232-22, 232-23, 232-24, and 232-25 can define the size, number, and location of neurons and/or layers of the neural network used during AI operations.
Register 232-26 can enable a debug/hold mode of the AI accelerator and output to be observed at a layer of AI operations. Register 232-26 can indicate that an activation should be applied during AI operations and that the AI operation can step forward (e.g., perform a next step in an AI operation) in AI operations. Register 232-26 can indicate that the temporary blocks, where the output of the layer is located, is valid. The data in the temporary blocks can be changed by a host and/or a controller on the memory device, such that the changed data can be used in the AI operation as the AI operation steps forward. Registers 232-27, 232-28, and 232-29 can define the layer where the debug/hold mode will stop the AI operation, change the content of the neural network, and/or observe the output of the layer.
Registers 232-30, 232-31, 232-32, and 232-33 can define the size of temporary banks used in AI operations and the start address and end address of the temporary banks used in AI operations. Register 232-30 can define the start address and end address of a first temporary bank used in AI operations and register 232-33 can define the start address and end address of a first temporary bank used in AI operations. Registers 232-31, and 232-32 can define the size of the temporary banks used in AI operations.
Registers 232-34, 232-35, 232-36, 232-37, 232-38, and 232-39 can be associated with the activation functions used in AI operations. Register 232-34 can enable usage of the activation function block, enable usage of the activation function for each neuron, the activation function for each layer, and enables usage of an external activation function. Registers 232-35 can define the start address and the end address of the location of the activation functions. Registers 232-36, 232-37, 232-38, and 232-39 can define the resolution of the inputs (e.g., x-axis) and outputs (e.g., y-axis) of the activation functions and/or a custom defined activation function.
Registers 232-40, 232-41, 232-42, 232-43, and 232-44 can define the size of bias values used in AI operations, the number of bias values used in AI operations, and the start address and end address of the bias values used in AI operations.
Register 232-45 can provide status information for the AI calculations and provide information for the debug/hold mode. Register 232-45 can enable debug/hold mode, indicate that the AI accelerator is performing AI operations, indicate that the full capability of the AI accelerator should be used, indicate only matrix calculations of the AI operations should be made, and/or indicate that the AI operation can proceed to the next neuron and/or layer.
Register 232-46 can provide error information regarding AI operations. Register 232-46 can indicate that there was an error in a sequence of an AI operation, that there was an error in an algorithm of an AI operations, that there was an error in a page of data that ECC was not able to correct, and/or that there was an error in a page of data that ECC was able to correct.
Register 232-47 can indicate an activation function to use in AI operations. Register 232-47 can indicated one of a number of pre-define activation function can be used in AI operations and/or a custom activation function located in a block can be used in AI operations.
Registers 232-48, 232-49, and 232-50 can indicate the neuron and/or layer where the AI operation is executing. In the case where errors occur during the AI operations, registers 232-48, 232-49, and 232-50 the neuron and/or layer where an error occurred.
Register 332-0 can define parameters associated with AI mode of the memory device. Bit 334-5 of register 332-0 can be a read/write bit and can indicate that an elaboration of an AI operation can restart 360 at the beginning when programmed to 1b. Bit 334-5 of register 332-0 can be reset to 0b once the AI operation has restarted. Bit 334-4 of register 332-0 can be a read/write bit and can indicate that an elaboration of an AI operation can start 361 when programmed to 1b. Bit 334-4 of register 332-0 can be reset to 0b once the AI operation has started.
Bit 334-3 of register 332-0 can be a read/write bit and can indicate that the content of the AI registers is valid 362 when programmed to 1b and invalid when programmed to 0b. Bit 334-2 of register 332-0 can be a read/write bit and can indicate that the content of the AI registers is to be cleared 363 when programmed to 1b. Bit 334-1 of register 332-0 can be a read only bit and can indicate that the AI accelerator is in use 363 and performing AI operations when programmed to 1b. Bit 334-0 of register 332-0 can be a write only bit and can indicate that the memory device is to exit 365 AI mode when programmed to 1b.
Registers 332-1, 332-2, 332-3, 332-4, and 332-5 can define the size of inputs used in AI operations, the number of inputs used in AI operations, and the start address and end address of the inputs used in AI operations. Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, 334-6, and 334-7 of registers 332-1 and 332-2 can define the size of the inputs 366 used in AI operations. The size of the inputs can indicate the width of the inputs in terms of number of bits and/or the type of input, such as floating point, integer, and/or double, among other types. Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, 334-6, and 334-7 of registers 332-3 and 332-4 can indicate the number of inputs 367 used in AI operations. Bits 334-4, 334-5, 334-6, and 334-7 of register 332-5 can indicate a start address 368 of the blocks in memory arrays of the inputs used in AI operations. Bits 334-0, 334-1, 334-2, and 334-3 of register 332-5 can indicate an end address 369 of the blocks in memory arrays of the inputs used in AI operations. If the start address 368 and the end address 369 is the same address, only one block of input is indicated for the AI operations.
Registers 332-7, 332-8, 332-9, 332-10, and 332-11 can define the size of outputs of AI operations, the number of outputs in AI operations, and the start address and end address of the outputs of AI operations. Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, 334-6, and 334-7 of registers 332-7 and 332-8 can define the size 370 of the outputs used in AI operations. The size of the outputs can indicate the width of the outputs in terms of number of bits and/or the type of output, such as floating point, integer, and/or double, among other types. Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, 334-6, and 334-7 of registers 332-9 and 332-10 can indicate the number of outputs 371 used in AI operations. Bits 334-4, 334-5, 334-6, and 334-7 of register 332-11 can indicate a start address 372 of the blocks in memory arrays of the outputs used in AI operations. Bits 334-0, 334-1, 334-2, and 334-3 of register 332-11 can indicate an end address 373 of the blocks in memory arrays of the outputs used in AI operations. If the start address 372 and the end address 373 is the same address, only one block of output is indicated for the AI operations.
Register 332-12 can be used to enable the usage of the input banks, the neuron banks, the output banks, the bias banks, the activation functions, and the temporary banks used during AI operations. Bit 334-0 of register 332-12 can enable the input banks 380, bit 334-1 of register 332-12 can enable the neural network banks 379, bit 334-2 of register 332-12 can enable the output banks 378, bit 334-3 of register 332-12 can enable the bias banks 377, bit 334-4 of register 332-12 can enable the activation function banks 376, and bit 334-5 and 334-6 of register 332-12 can enable a first temporary 375 banks and a second temporary bank 374.
Registers 332-13, 332-14, 332-15, 332-16, 332-17, 332-18, 332-19, 332-20, 332-21, 332-22, 332-23, 332-24, and 332-25 can be used to define the neural network used during AI operations. Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, 334-6, and 334-7 of registers 332-13 and 332-14 can define the number of rows 381 in a matrix used in AI operations. Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, 334-6, and 334-7 of registers 332-15 and 332-16 can define the number of columns 382 in a matrix used in AI operations.
Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, 334-6, and 334-7 of registers 332-17 and 332-18 can define the size of the neurons 383 used in AI operations. The size of the neurons can indicate the width of the neurons in terms of number of bits and/or the type of input, such as floating point, integer, and/or double, among other types. Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, 334-6, and 334-7 of registers 332-19, 332-20, and 322-21 can indicate the number of neurons 384 of the neural network used in AI operations. Bits 334-4, 334-5, 334-6, and 334-7 of register 332-22 can indicate a start address 385 of the blocks in memory arrays of the neurons used in AI operations. Bits 334-0, 334-1, 334-2, and 334-3 of register 332-5 can indicate an end address 386 of the blocks in memory arrays of the neurons used in AI operations. If the start address 385 and the end address 386 is the same address, only one block of neurons is indicated for the AI operations. Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, 334-6, and 334-7 of registers 332-23, 332-24, and 322-25 can indicate the number of layers 387 of the neural network used in AI operations.
Register 332-26 can enable a debug/hold mode of the AI accelerator and an output to be observed at a layer of AI operations. Bit 334-0 of register 332-26 can indicate that the AI accelerator is in a debug/hold mode and that an activation function should be applied 391 during AI operations. Bit 334-1 of register 332-26 can indicate that the AI operation can step forward 390 (e.g., perform a next step in an AI operation) in AI operations. Bit 334-2 and bit 334-3 of register 232-26 can indicate that the temporary blocks, where the output of the layer is located, is valid 388 and 389. The data in the temporary blocks can be changed by a host and/or a controller on the memory device, such that the changed data can be used in the AI operation as the AI operation steps forward.
Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, 334-6, and 334-7 of registers 332-27, 332-28, and 332-29 can define the layer where the debug/hold mode will stop 392 the AI operation and observe the output of the layer.
Registers 332-30, 332-31, 332-32, and 332-33 can define the size of temporary banks used in AI operations and the start address and end address of the temporary banks used in AI operations. Bits 334-4, 334-5, 334-6, and 334-7 of register 332-30 can define the start address 393 of a first temporary bank used in AI operations. Bits 334-0, 334-1, 334-2, and 334-3 of register 332-30 can define the end address 394 of a first temporary bank used in AI operations. Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, 334-6, and 334-7 of registers 332-31 and 332-32 can define the size 395 of the temporary banks used in AI operations. The size of the temporary banks can indicate the width of the temporary banks in terms of number of bits and/or the type of input, such as floating point, integer, and/or double, among other types. Bits 334-4, 334-5, 334-6, and 334-7 of register 332-33 can define the start address 396 of a second temporary bank used in AI operations. Bits 334-0, 334-1, 334-2, and 334-3 of register 332-34 can define the end address 397 of a second temporary bank used in AI operations.
Registers 332-34, 332-35, 332-36, 332-37, 332-38, and 332-39 can be associated with the activation functions used in AI operations. Bit 334-0 of register 332-34 can enable usage of the activation function block 3101. Bit 334-1 of register 332-34 can enable holding that AI at a neuron 3100 and usage of the activation function for each neuron. Bit 334-2 of register 332-34 can enable holding the AI at a layer 399 and the usage of the activation function for each layer. Bit 334-3 of register 332-34 can enable usage of an external activation function 398.
Bits 334-4, 334-5, 334-6, and 334-7 of register 332-35 can define the start address 3102 of activation function banks used in AI operations. Bits 334-0, 334-1, 334-2, and 334-3 of register 332-35 can define the end address 3103 of activation functions banks used in AI operations. Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, 334-6, and 334-7 of registers 332-36 and 332-37 can define the resolution of the inputs (e.g., x-axis) 3104 of the activation functions. Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, 334-6, and 334-7 of registers 332-38 and 332-39 can define the resolution and/or the outputs (e.g., y-axis) 3105 of the activation functions for a given x-axis value of a custom activation function.
Registers 332-40, 332-41, 332-42, 332-43, and 332-44 can define the size of bias values used in AI operations, the number of bias values used in AI operations, and the start address and end address of the bias values used in AI operations. Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, 334-6, and 334-7 of registers 332-40 and 332-41 can define the size of the bias values 3106 used in AI operations. The size of the bias values can indicate the width of the bias values in terms of number of bits and/or the type of bias values, such as floating point, integer, and/or double, among other types. Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, 334-6, and 334-7 of registers 332-42 and 332-43 can indicate the number of bias values 3107 used in AI operations. Bits 334-4, 334-5, 334-6, and 334-7 of register 332-44 can indicate a start address 3108 of the blocks in memory arrays of the bias values used in AI operations. Bits 334-0, 334-1, 334-2, and 334-3 of register 332-44 can indicate an end address 3109 of the blocks in memory arrays of the bias values used in AI operations. If the start address 3108 and the end address 3109 is the same address, only one block of bias values is indicated for the AI operations.
Register 332-45 can provide status information for the AI calculations and provide information for the debug/hold mode. Bit 334-0 of register 332-45 can activate the debug/hold mode 3114. Bit 334-1 of register can indicate that the AI accelerator is busy 3113 and performing AI operations. Bit 334-2 of register 332-45 can indicate that the AI accelerator is on 3112 and/or that the full capability of the AI accelerator should be used. Bit 334-3 of register 332-45 can indicate only matrix calculations 3111 of the AI operations should be made. Bit. 334-4 of register 332-45 can indicate that the AI operation can step forward 3110 and proceed to the next neuron and/or layer.
Register 332-46 can provide error information regarding AI operations. Bit 334-3 of register 332-46 can indicate that there was an error in a sequence 3115 of an AI operation. Bit 334-2 of register 332-46 can indicate that there was an error in an algorithm 3116 of an AI operation. Bit 334-1 of register 332-46 can indicate there was an error in a page of data that ECC was not able to correct 3117. Bit 334-0 of register 332-46 can indicate there was an error in a page of data that ECC was able to correct 3118.
Register 332-47 can indicate an activation function to use in AI operations. Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, and 334-6 of register 332-47 can indicate one of a number of pre-define activation functions 3120 can be used in AI operations. Bit 334-7 of register 332-47 can indicate a custom activation function 3119 located in a block can be used in AI operations.
Registers 332-48, 332-49, and 332-50 can indicate the neuron and/or layer where the AI operation is executing. Bits 334-0, 334-1, 334-2, 334-3, 334-4, 334-5, 334-6, and 334-7 of registers 332-48, 332-49, and 332-50 can indicate the address of the neuron and/or layer where the AI operation is executing. In the case where errors occur during the AI operations, registers 332-48, 332-49, and 332-50 can indicate the neuron and/or layer where an error occurred.
Output block 420 is a block in the memory arrays where output data from AI operations is stored. Data in output block 442 can be used store the output from AI operations and sent to the host. The address of output block 442 can be indicated in register 11 (e.g. register 232-11 in
Data in output block 442 can be sent to host upon completion and/or holding of an AI operation. Temporary blocks 444-1 and 444-2 can be blocks in memory arrays where data is stored temporarily while AI operations are being performed. Data can be stored in temporary blocks 444-1 and 444-2 while the AI operations are iterating through the neuron and layers of the neural network used for the AI operations. The address of temporary block 448 can be indicated in registers 30 and 33 (e.g. registers 232-30 and 232-33 in
Activation function block 446 is a block in the memory arrays where the activations functions for the AI operations are stored. Activation function block 446 can store pre-defined activation functions and/or custom activation functions that are created by the host and/or AI accelerator. The address of activation function block 448 can be indicated in register 35 (e.g. register 232-35 in
Bias values block 448 is a block in the memory array where the bias values for the AI operations are stored. The address of bias values block 448 can be indicated in register 44 (e.g. register 232-44 in
Neural network blocks 450-1, 450-2, 450-3, 450-4, 450-5, 450-6, 450-7, 450-8, 450-9, and 450-10 are a block in the memory array where the neural network for the AI operations are stored. Neural network blocks 450-1, 450-2, 450-3, 450-4, 450-5, 450-6, 450-7, 450-8, 450-9, and 450-10 can store the information for the neurons and layers that are used in the AI operations. The address of neural network blocks 450-1, 450-2, 450-3, 450-4, 450-5, 450-6, 450-7, 450-8, 450-9, and 450-10 can be indicated in register 22 (e.g. register 232-22 in
In
As used herein, the additional latches are intended to mean additional functionalities (e.g., amplifiers, select logic) that sense, couple, and/or move (e.g., read, store, cache) data values of memory cells in an array and that are distinct from the plurality of compute components 5148-1, . . . , 5148-M and/or logic stripes 5152-1, . . . , 5152-N in a compute unit in a data path of the plurality of shared I/O lines 5144 shown in
System 500 in
For clarity, the system 500 has been simplified to focus on features with particular relevance to the present disclosure. The memory array 525 can be a DRAM array, SRAM array, STT RAM array, PCRAM array, TRAM array, RRAM array, NAND flash array, and/or NOR flash array, among other types of arrays. The array 525 can include memory cells arranged in rows coupled by access lines (which may be referred to herein as word lines or select lines) and columns coupled by sense lines (which may be referred to herein as data lines or digit lines). Although a single array 525 is shown in
The memory device 520 can include address circuitry 5525 to latch address signals provided over a data bus 556 (e.g., an I/O bus connected to the host 502) by I/O circuitry 5134 (e.g., provided to external ALU circuitry and/or to DRAM DQs via local I/O lines and global I/O lines). As used herein, DRAM DQs can enable input of data to and/or output of data from a bank (e.g., from and/or to the controller 540 and/or host 502) via a bus (e.g., data bus 556). During a write operation, a voltage (high=1, low=0) can be applied to a DQ (e.g., a pin). This voltage can be translated into an appropriate signal and stored in a selected memory cell. During a read operation, a data value read from a selected memory cell can appear at the DQ once access is complete and the output is enabled (e.g., by the output enable signal being low). At other times, DQs can be in a high impedance state, such that the DQs do not source or sink current and do not present a signal to the system. This also may reduce DQ contention when two or more devices (e.g., banks) share a combined data bus, as described herein. Such DQs are separate and distinct from the plurality of shared I/O lines 5144 (in
Status and exception information can be provided from the controller 540 of the memory device 520 to a channel controller 504, for example, through an out-of-band (00B) bus 557, e.g., high-speed interface (HSI), which in turn can be provided from the channel controller 504 to the host 502. The channel controller 504 can include a logic component to allocate a plurality of locations (e.g., controllers for subarrays) in the arrays of each respective bank to store bank commands, application instructions (e.g., for sequences of operations), and arguments (PIM commands) for various banks associated with operations for each of a plurality of memory devices 520. The channel controller 504 can dispatch commands (e.g., PIM commands) to the plurality of memory devices 520 to store those program instructions within a given bank 5146 (
Address signals are received through address circuitry 5525 and decoded by a row decoder 5132 and a column decoder 5142 to access the memory array 525. Data can be sensed (read) from memory array 525 by sensing voltage and/or current changes on sense lines (digit lines) using a number of sense amplifiers, as described herein, of the sensing circuitry 5138. A sense amplifier can read and latch a page (e.g., a row) of data from the memory array 525. Additional compute circuitry, as described herein, can be coupled to the sensing circuitry 5138 and can be used in combination with the sense amplifiers to sense, store (e.g., cache and/or buffer), perform compute functions (e.g., operations), and/or move data. The I/O circuitry 5134 can be used for bi-directional data communication with host 502 over the data bus 556 (e.g., a 64 bit wide data bus). The write circuitry 5136 can be used to write data to the memory array 525.
Controller 540 (e.g., bank control logic, sequencer and timing circuitry shown in
Examples of the sensing circuitry 5138 are described further below (e.g., in
In some embodiments, the sensing circuitry 5138 can be used to perform operations using data stored in memory array 525 as inputs and participate in movement of the data for copy, transfer, transport, writing, logic, and/or storage operations to a different location in the memory array 525 without transferring the data via a sense line address access (e.g., without firing a column decode signal). As such, various compute functions can be performed using, and within, sensing circuitry 5138 rather than (or in association with) being performed by processing resources external to the sensing circuitry 5138 (e.g., by a processor associated with host 502 and/or other processing circuitry, such as ALU circuitry, located on device 520, such as on controller 540 or elsewhere). However, in addition, embodiments according to the present disclosure perform compute functions on data values moved to a plurality of compute components 5148-1, . . . , 5148-M and/or logic stripes 5152-1, . . . , 5152-N in a compute unit from the rows of the array. And as an example, according to embodiments, compute operations may be controlled in the compute unit at speeds of 2 nanoseconds (ns) without having to move the data values back into the rows, as compared to an example time required to fire the rows in the array of 60 nanoseconds (ns).
In various previous approaches, data associated with an operand, for instance, would be read from memory via sensing circuitry and provided to external ALU circuitry via I/O lines (e.g., via local I/O lines and/or global I/O lines). The external ALU circuitry could include a number of registers and would perform compute functions using the operands, and the result would be transferred back to the array via the I/O circuitry 5134.
In contrast, embodiments according to the present disclosure perform compute functions on data values, moved to a plurality of compute components 5148-1, . . . , 5148-M and/or logic stripes 5152-1, . . . , 5152-N via a plurality of shared I/O lines 5144 from the rows of the array, in a compute unit in a data path local to the array. Additionally, sensing circuitry 5138 may be configured to perform operations on data stored in memory array 525 and store the result back to the memory array 525 without enabling an I/O line (e.g., a local I/O line) coupled to the sensing circuitry 5138. However, once loaded, compute operations may be controlled in the compute unit much faster, e.g., at speeds of 2 nanoseconds (ns), without having to move the data values back into the rows, as compared to an example time required to fire the rows in the array, e.g., 60 nanoseconds (ns). The sensing circuitry 5138 can be formed on pitch with the memory cells of the array. The plurality of compute components 5148-1, . . . , 5148-M and/or logic stripes 5152-1, . . . , 5152-N, associated with the data path of the plurality of shared I/O lines 5144, have a pitch equal to that of the data path and that is a function of a pitch of digit lines to the array of memory cells. For example, the compute component has a pitch that is an integer multiple of the pitch of digit lines to the array of memory cells.
For example, the sensing circuitry 5138 described herein can be formed on a same pitch as a pair of complementary sense lines (e.g., digit lines). As an example, a pair of complementary memory cells may have a cell size with a 6F2 pitch (e.g., 3F×2F), where F is a feature size. If the pitch of a pair of complementary sense lines for the complementary memory cells is 3F, then the sensing circuitry being on pitch indicates the sensing circuitry (e.g., a sense amplifier and corresponding compute component per respective pair of complementary sense lines) is formed to fit within the 3F pitch of the complementary sense lines. Likewise, the compute components 5148-1, . . . , 5148-M and/or logic stripes 5152-1, . . . , 5152-N, associated with the data path of the plurality of shared I/O lines 5144, have a pitch that is a function of the 3F pitch of the complementary sense lines. For example, the compute components 5148-1, . . . , 5148-M and/or logic stripes 5152-1, . . . , 5152-N will have a pitch that is an integer multiple of the 3F pitch of digit lines to the array of memory cells.
By contrast, the circuitry of the processing resource(s) (e.g., a compute engine, such as an ALU) of various prior systems may not conform to pitch rules associated with a memory array. For example, the memory cells of a memory array may have a 4F2 or 6F2 cell size. As such, the devices (e.g., logic gates) associated with ALU circuitry of previous systems may not be capable of being formed on pitch with the memory cells (e.g., on a same pitch as the sense lines), which can affect chip size and/or memory density, for example. In the context of some computing systems and subsystems (e.g., a central processing unit (CPU)), data may be processed in a location that is not on pitch and/or on chip with memory (e.g., memory cells in the array), as described herein. For example, the data may be processed by a processing resource associated with a host, for instance, rather than on pitch with the memory.
As such, in a number of embodiments, circuitry external to array 525 and sensing circuitry 5138 is not needed to perform compute functions as the sensing circuitry 5138 can perform the appropriate operations to perform such compute functions or can perform such operations in a data path of a plurality of shared I/O lines local to the array without the use of an external processing resource. Therefore, the sensing circuitry 5138 and/or the plurality of compute components 5148-1, . . . , 5148-M and/or logic stripes 5152-1, . . . , 5152-N in a compute unit in a data path of the plurality of shared I/O lines 5144 may be used to complement or to replace, at least to some extent, such an external processing resource (or at least the bandwidth consumption of such an external processing resource). In some embodiments, the sensing circuitry 5138 and/or the plurality of compute components 5148-1, . . . , 5148-M and/or logic stripes 5152-1, . . . , 5152-N in a compute unit in a data path of the plurality of shared I/O lines 5144 may be used to perform operations (e.g., to execute instructions) in addition to operations performed by an external processing resource (e.g., host 502). For instance, host 502 and/or sensing circuitry 5138 may be limited to performing only certain operations and/or a certain number of operations.
Operations described herein can include operations associated with a processing in memory (PIM) capable device. PIM capable device operations can use bit vector based operations. As used herein, the term “bit vector” is intended to mean a physically contiguous number of bits on a bit vector memory device (e.g., a PIM device) stored physically contiguous in a row of an array of memory cells. Thus, as used herein a “bit vector operation” is intended to mean an operation that is performed on a bit vector that is a contiguous portion of virtual address space (e.g., used by a PIM device). For example, a row of virtual address space in the PIM device may have a bit length of 16K bits (e.g., corresponding to 16K complementary pairs of memory cells in a DRAM configuration). Sensing circuitry 5138, as described herein, for such a 16K bit row may include a corresponding 16K processing elements (e.g., compute components, as described herein) formed on pitch with the sense lines selectably coupled to corresponding memory cells in the 16 bit row. A compute component in the PIM device may operate as a one bit processing element (PE) on a single bit of the bit vector of the row of memory cells sensed by the sensing circuitry 5138 (e.g., sensed by and/or stored in a sense amplifier paired with the compute component, as described herein). Similarly, the plurality of compute components 5148-1, . . . , 5148-M and/or logic stripes 5152-1, . . . , 5152-N in a compute unit in a data path of the plurality of shared I/O lines 5144 may operate as a one bit processing element (PE) on a single bit of the bit vector of the row of memory cells sensed in an array.
Enabling an I/O line can include enabling (e.g., turning on, activating) a transistor having a gate coupled to a decode signal (e.g., a column decode signal) and a source/drain coupled to the I/O line. However, embodiments are not limited to not enabling an I/O line. For instance, in a number of embodiments, the sensing circuitry (e.g., 5138) can be used to perform operations without enabling column decode lines of the array.
However, the plurality of shared I/O lines 5144 may be enabled in order to load data values to the plurality of compute components 5148-1, . . . , 5148-M and/or logic stripes 5152-1, . . . , 5152-N in a compute unit in a data path of the plurality of shared I/O lines 5144 where compute operations may be controlled much faster. For example, in the plurality of compute components 5148-1, . . . , 5148-M and/or logic stripes 5152-1, . . . , 5152-N in the compute unit, the compute operation may be performed at speeds of 2 nanoseconds (ns). This enhancement of speed can be attributed to not having to move the data values back into the rows with the associated time used in firing the rows in the array, e.g., 60 nanoseconds (ns).
As shown in
As shown in the example embodiment of
The cells of the memory array 625 can be arranged in rows coupled by access (word) lines 6164-X (Row X), 6164-Y (Row Y), etc., and columns coupled by pairs of complementary sense lines (e.g., digit lines DIGIT(D) and DIGIT(D)_ shown in
Although rows and columns are illustrated as orthogonally oriented in a plane, embodiments are not so limited. For example, the rows and columns may be oriented relative to each other in any feasible three-dimensional configuration. For example, the rows and columns may be oriented at any angle relative to each other, may be oriented in a substantially horizontal plane or a substantially vertical plane, and/or may be oriented in a folded topology, among other possible three-dimensional configurations.
Memory cells can be coupled to different digit lines and word lines. For example, a first source/drain region of a transistor 6160-1 can be coupled to digit line 6168-1 (D), a second source/drain region of transistor 6160-1 can be coupled to capacitor 6162-1, and a gate of a transistor 6160-1 can be coupled to word line 6164-Y. A first source/drain region of a transistor 6160-2 can be coupled to digit line 6168-2 (D)_, a second source/drain region of transistor 6160-2 can be coupled to capacitor 6162-2, and a gate of a transistor 6160-2 can be coupled to word line 6164-X. A cell plate, as shown in
The memory array 625 is configured to couple to sensing circuitry 6138 in accordance with a number of embodiments of the present disclosure. In this embodiment, the sensing circuitry 6138 comprises a sense amplifier 6170 and a compute component 6148 corresponding to respective columns of memory cells (e.g., coupled to respective pairs of complementary digit lines). The sense amplifier 6170 can be coupled to the pair of complementary digit lines 6168-1 and 6168-2. The compute component 6148 can be coupled to the sense amplifier 6170 via pass gates 6172-1 and 6172-2. The gates of the pass gates 6172-1 and 6172-2 can be coupled to operation selection logic 6178.
The operation selection logic 6178 can be configured to include pass gate logic for controlling pass gates that couple the pair of complementary digit lines un-transposed between the sense amplifier 6170 and the compute component 6148 and swap gate logic for controlling swap gates that couple the pair of complementary digit lines transposed between the sense amplifier 6170 and the compute component 6148. The operation selection logic 6178 can also be coupled to the pair of complementary digit lines 6168-1 and 6168-2. The operation selection logic 6178 can be configured to control continuity of pass gates 6172-1 and 6172-2 based on a selected operation.
The sense amplifier 6170 can be operated to determine a data value (e.g., logic state) stored in a selected memory cell. The sense amplifier 6170 can comprise a cross coupled latch, which can be referred to herein as a primary latch. In the example illustrated in
In operation, when a memory cell is being sensed (e.g., read), the voltage on one of the digit lines 6168-1 (D) or 6168-2 (D)_ will be slightly greater than the voltage on the other one of digit lines 6168-1 (D) or 6168-2 (D)_. An ACT signal and an RNL* signal can be driven low to enable (e.g., fire) the sense amplifier 6170. The digit lines 6168-1 (D) or 6168-2 (D)_ having the lower voltage will turn on one of the PMOS transistor 6184-1 or 6184-2 to a greater extent than the other of PMOS transistor 6184-1 or 6184-2, thereby driving high the digit line 6168-1 (D) or 6168-2 (D)_ having the higher voltage to a greater extent than the other digit line 6168-1 (D) or 6168-2 (D)_ is driven high.
Similarly, the digit line 6168-1 (D) or 6168-2 (D)_ having the higher voltage will turn on one of the NMOS transistor 6182-1 or 6182-2 to a greater extent than the other of the NMOS transistor 6182-1 or 6182-2, thereby driving low the digit line 6168-1 (D) or 6168-2 (D)_ having the lower voltage to a greater extent than the other digit line 6168-1 (D) or 6168-2 (D)_ is driven low. As a result, after a short delay, the digit line 6168-1 (D) or 6168-2 (D)_ having the slightly greater voltage is driven to the voltage of the supply voltage VCC through a source transistor, and the other digit line 6168-1 (D) or 6168-2 (D)_ is driven to the voltage of the reference voltage (e.g., ground) through a sink transistor. Therefore, the cross coupled NMOS transistors 6182-1 and 6182-2 and PMOS transistors 6184-1 and 6184-2 serve as a sense amplifier pair, which amplify the differential voltage on the digit lines 6168-1 (D) and 6168-2 (D)_ and operate to latch a data value sensed from the selected memory cell on nodes 6177-1 and/or 6177-2.
Embodiments are not limited to the sense amplifier 6170 configuration illustrated in
The sense amplifier 6170 can, in conjunction with the compute component 6148, be operated to perform various operations using data from an array as input. In a number of embodiments, the result of an operation can be stored back to the array without transferring the data via a digit line address access (e.g., without firing a column decode signal such that data is transferred to circuitry external from the array and sensing circuitry via local I/O lines). As such, a number of embodiments of the present disclosure can enable performing operations and compute functions associated therewith using less power than various previous approaches. Additionally, since a number of embodiments eliminate the need to transfer data across local and global I/O lines in order to perform compute functions (e.g., between memory and discrete processor), a number of embodiments can enable an increased (e.g., faster) processing capability as compared to previous approaches.
The sense amplifier 6170 can further include equilibration circuitry 6174, which can be configured to equilibrate the digit lines 6168-1 (D) and 6168-2 (D)_. In this example, the equilibration circuitry 6174 comprises a transistor 6188 coupled between digit lines 6168-1 (D) and 6168-2 (D)_. The equilibration circuitry 6174 also comprises transistors 6180-1 and 6180-2 each having a first source/drain region coupled to an equilibration voltage (e.g., VDD/2), where VDD is a supply voltage associated with the array. A second source/drain region of transistor 6180-1 can be coupled digit line 6168-1 (D), and a second source/drain region of transistor 6180-2 can be coupled digit line 6168-2 (D)_. Gates of transistors 6188, 6180-1, and 6180-2 can be coupled together, and to an equilibration (EQ) control signal line 6186. As such, activating EQ enables the transistors 6188, 6180-1, and 6180-2, which effectively shorts digit lines 6168-1 (D) and 6168-2 (D)_ together and to the equilibration voltage (e.g., VCC/2).
Although
As described further below, in a number of embodiments, the sensing circuitry 6138 (e.g., sense amplifier 6170 and compute component 6148) can be operated to perform a selected operation and initially store the result in one of the sense amplifier 6170 or the compute component 6148 without transferring data from the sensing circuitry via a local or global I/O line (e.g., without performing a sense line address access via activation of a column decode signal, for instance).
However, further to embodiments described herein, sensing circuitry 6138 having sense amplifiers, and which in some embodiments may also include compute components as shown in
Performance of operations (e.g., Boolean logical operations involving data values) is fundamental and commonly used. Boolean logical operations are used in many higher level operations. Consequently, speed and/or power efficiencies that can be realized with improved operations, can translate into speed and/or power efficiencies of higher order functionalities.
As shown in
As described herein, a memory device (e.g., 520 in
Each bank section 5150 can be associated with a plurality of logic stripes (e.g., 5152-0, 5152-1, . . . , 5152-N−1 in
The memory device can include a logic stripe (e.g., 5152 in
In some embodiments, a number of a plurality of logic stripes (e.g., 5152-1, . . . , 5152-N in
As described herein, the array of memory cells can include an implementation of DRAM memory cells where the controller is configured, in response to a command, to move (e.g., copy, transfer, and/or transport) data from the source location to the destination location via a shared I/O line. In various embodiments, the source location can be in a first bank and the destination location can be in a compute unit 7198 (
As described in
Data values present on the pair of complementary digit lines 7168-1 and 7168-2 can be loaded into the compute component 7148-0 as described in connection with
The sense amplifiers 7170-0, 7170-1, . . . , 7170-7 in
The configurations of embodiments illustrated in
The circuitry illustrated in
Controller 740 can be coupled to column select circuitry 7194 to control select lines (e.g., select line 0) to access data values stored in the sense amplifiers, compute components, and/or present on the pair of complementary digit lines (e.g., 7168-1 and 7168-2 when selection transistors 7196-1 and 7196-2 are activated via signals from select line 0). Activating the selection transistors 7196-1 and 7196-2 (e.g., as directed by the controller 540) enables coupling of sense amplifier 7170-0, compute component 7148-0, and/or complementary digit lines 7168-1 and 7168-2 of column 0 (7192-0) to move data values on digit line 0 and digit line 0* to shared I/O line 7144. For example, the moved data values may be data values from a particular row 319 stored (cached) in sense amplifier 7170-0 and/or compute component 7148-0. Data values from each of columns 0 through 7 can similarly be selected by controller 540 activating the appropriate selection transistors.
Moreover, enabling (e.g., activating) the selection transistors (e.g., selection transistors 7196-1 and 7196-2) can enable a particular sense amplifier and/or compute component (e.g., 7170-0 and/or 7148-0, respectively) to be coupled with a shared I/O line 7144 such that data values stored by an amplifier and/or compute component can be moved to (e.g., placed on and/or transferred to) the shared I/O line 7144. In some embodiments, one column at a time is selected (e.g., column 7192-0) to be coupled to a particular shared I/O line 7144 to move (e.g., copy, transfer, and/or transport) the stored data values. In the example configuration of
In the example of
According to embodiments, the controller 540 (
According to some embodiments, a butterfly network 8202 can be used to connect the 1K bit data values to a respective one of the plurality of compute components 8148-1, . . . , 8148-Z in a respective one of the plurality of logic stripes 8152-1, . . . , 8152-N. By way of example, and not by way of limitation, 1K bits of data values can be moved in parallel to a logic strip associate with each of the 32 subarrays 8145-1, . . . , 8145-32 in each of 4 quadrants to a bank section 8150. In this example, 128 logic stripes 8152-1, . . . , 8152-N having 1K compute components 8148-1, . . . , 8148-Z each can be included in the compute unit 8198. Data values loaded to the plurality of compute components 8148-1, . . . , 8148-Z in the logic stripes 8152-1, . . . , 8152-N of the compute unit 8198 can be operated on according to microcode instructions from the controller 540 (
In the example embodiment of
In the example of
As shown in the example of
Although specific embodiments have been illustrated and described herein, those of ordinary skill in the art will appreciate that an arrangement calculated to achieve the same results can be substituted for the specific embodiments shown. This disclosure is intended to cover adaptations or variations of various embodiments of the present disclosure. It is to be understood that the above description has been made in an illustrative fashion, and not a restrictive one. Combination of the above embodiments, and other embodiments not specifically described herein will be apparent to those of skill in the art upon reviewing the above description. The scope of the various embodiments of the present disclosure includes other applications in which the above structures and methods are used. Therefore, the scope of various embodiments of the present disclosure should be determined with reference to the appended claims, along with the full range of equivalents to which such claims are entitled.
In the foregoing Detailed Description, various features are grouped together in a single embodiment for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that the disclosed embodiments of the present disclosure have to use more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive subject matter lies in less than all features of a single disclosed embodiment. Thus, the following claims are hereby incorporated into the Detailed Description, with each claim standing on its own as a separate embodiment.
Patent | Priority | Assignee | Title |
11605420, | Aug 28 2019 | Lodestar Licensing Group LLC | Memory with artificial intelligence mode |
11922995, | Aug 28 2019 | Lodestar Licensing Group LLC | Memory with artificial intelligence mode |
Patent | Priority | Assignee | Title |
5313586, | Mar 06 1992 | Intel Corporation | Co-processor de-coupling bus structure |
6504780, | Oct 01 1998 | Invensas Corporation | Method and apparatus for completely hiding refresh operations in a dram device using clock division |
6928590, | Dec 14 2001 | SanDisk Technologies LLC | Memory device and method for storing bits in non-adjacent storage locations in a memory array |
8395950, | Oct 15 2010 | Taiwan Semiconductor Manufacturing Co., Ltd. | Memory device having a clock skew generator |
20170024644, | |||
20180189638, | |||
20180218257, | |||
20180225056, | |||
20190042538, | |||
20190057302, | |||
20190094946, | |||
20190114078, | |||
20190205737, | |||
20190235867, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Aug 26 2019 | TROIA, ALBERTO | Micron Technology, Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 050197 | /0663 | |
Aug 28 2019 | Micron Technology, Inc. | (assignment on the face of the patent) | / | |||
Mar 23 2023 | Micron Technology, Inc | Lodestar Licensing Group LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 065071 | /0434 |
Date | Maintenance Fee Events |
Aug 28 2019 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Oct 22 2024 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Date | Maintenance Schedule |
May 11 2024 | 4 years fee payment window open |
Nov 11 2024 | 6 months grace period start (w surcharge) |
May 11 2025 | patent expiry (for year 4) |
May 11 2027 | 2 years to revive unintentionally abandoned end. (for year 4) |
May 11 2028 | 8 years fee payment window open |
Nov 11 2028 | 6 months grace period start (w surcharge) |
May 11 2029 | patent expiry (for year 8) |
May 11 2031 | 2 years to revive unintentionally abandoned end. (for year 8) |
May 11 2032 | 12 years fee payment window open |
Nov 11 2032 | 6 months grace period start (w surcharge) |
May 11 2033 | patent expiry (for year 12) |
May 11 2035 | 2 years to revive unintentionally abandoned end. (for year 12) |