It is possible to reduce the latency attributable to memory protection in shared memory systems by performing access protection at a central data ownership manager (DOM), rather than at distributed memory management units in the central processing unit (CPU) elements (ces) responsible for parallel thread processing. In particular, the DOM may monitor read requests communicated over a data plane between the ces and a memory controller, and perform access protection verification in parallel with the memory controller's generation of the data response. The DOM may be separate and distinct from both the ces and the memory controller, and therefore may generally be able to make the access determination without interfering with data plane processing/generation of the read requests and data responses exchanged between the memory controller and the ces.
|
16. A method for managing a cache, the method comprising:
sending, by a central processing unit (CPU) element (ce), a read request to a memory controller without independently verifying whether a thread being processed by the ce is permitted to access a shared memory resource, the read request requesting access to the shared memory resource, the ce being one of a plurality of ces, and the plurality of ces having a common centralized data ownership manager (DOM) providing memory access protection verification;
receiving, by the ce, a data response from the memory controller, the data response including a response code that indicates whether the thread being processed by the ce is permitted, by the centralized DOM, to access the shared memory resource; and
in response to the response code in the data response indicating whether the thread is permitted to access the shared memory resource, processing, by the ce, the data response based on the response code without independently verifying whether the thread is permitted to access the shared memory resource.
11. A method for providing memory protection for shared memory, the method comprising:
detecting, by a centralized data ownership manager (DOM), a read request communicated from a central processing unit (CPU) element (ce) to a memory controller, the read request requesting access to a shared memory resource, the ce being one of a plurality of ces, and the centralized DOM being common to each of the ces; and
determining, by the centralized DOM, whether a thread assigned to the ce is permitted to access the shared memory resource based on one or more identifiers (IDs) in the read request; and
sending, by the centralized DOM, a control signal to the memory controller, the control signal indicating whether the thread being executed by the ce is permitted to access the shared memory resource,
the control signal prompting the memory controller to set a response code in a data response sent from the memory controller to the ce, the response code indicating whether the thread being executed by the ce is permitted to access the shared memory resource, and
the response code in the data response prompting a local control block in the ce to process the data response without independently verifying that the thread is permitted to access the shared memory resource.
1. An apparatus comprising:
a data plane;
a plurality of central processing unit (CPU) elements (ces) coupled to the data plane, the plurality of ces including a ce configured to send a read request to a memory controller over the data plane, the read request requesting access to a shared memory resource;
a centralized data ownership manager (DOM) coupled to the data plane, wherein the centralized DOM is common to each of the ces, and wherein the centralized DOM is configured to:
detect the read request,
determine whether a thread assigned to the ce is permitted to access the shared memory resource based on one or more identifiers (IDs) in the read request, and
send a control signal to the memory controller indicating whether the thread being executed by the ce is permitted to access the shared memory resource; and
the memory controller, coupled to the data plane and configured to send a data response to the ce, the data response including a response code that indicates whether the thread being executed by the ce is permitted to access the shared memory resource;
wherein the ce includes a local control block that is configured to process the data response based on the response code in the data response without independently verifying that the thread is permitted to access the shared memory resource.
2. The apparatus of
comparing both a context ID and a job ID in the read request with a context ID and a job ID in a data ownership table in response to the read request being a read/write (RW) request; or
comparing the context ID in the read request with the context ID in the data ownership table without comparing the job ID in the read request with the job ID in the data ownership table in response to the read request being a read-only (RO) request.
3. The apparatus of
4. The apparatus of
5. The apparatus of
6. The apparatus of
7. The apparatus of
8. The apparatus of
9. The apparatus of
10. The apparatus of
12. The method of
comparing both a context ID and a job ID in the read request with a context ID and a job ID in a data ownership table in response to the read request being a read/write (RW) request; being a read-only (RO) request.
13. The method of
14. The method of
15. The method of
storing, by the DOM, faulty access information corresponding to the read request in debug registers of the DOM in response to the thread not being permitted to access the shared memory resource, the faulty access information including one or more of a ce ID, a physical address, a time stamp, and access attributes carried by the read request in the debug registers.
17. The method of
wherein processing the data response comprises storing data, carried by the data response, into a local cache of the ce.
18. The method of
wherein processing the data response comprises generating an exception.
19. The method of
wherein processing the data response comprises performing a self-correction.
|
The present invention relates generally to a system and method for shared memory ownership, and, in particular embodiments, to a system and method for shared memory ownership using context.
Next-generation wireless devices may use multi-core and/or parallel processing architectures to support the heightened signal processing requirements of Fifth Generation (5G) wireless network telecommunication protocols. In such architectures, memory resources may be shared by multiple central processing unit (CPU) elements (CEs) so that two or more CEs can simultaneously access the same memory resource. In this way, shared memory resources may provide an efficient means of passing data/messages between CEs that are executing different threads of an application. As used herein, the term “thread” refers to a sequence of programming instructions that is being executed by, or is otherwise scheduled for execution by, a single CE. In some instances, CEs executing different threads of an application may need to exchange data with one another in order to execute their respective threads. Shared memory provides an efficient means for passing data between CEs by allowing one CE to write data into a memory resource, and another CE to subsequently read the data from the memory resource. This is generally referred to as message passing.
Memory errors may occur when multiple CEs access the same shared memory resource if cache coherency is not maintained. More specifically, CEs will generally read data from a shared memory resource into a local cache of the CE, and then use the local copy of the data during execution of a corresponding thread. If another CE subsequently modifies the data stored in the shared memory resource via a write instruction, and then reads the modified data into its own cache, the local caches of the respective CEs will be inconsistent, meaning that cache coherency was not be maintained. Other scenarios, such as false sharing, may also affect cache coherency.
Technical advantages are generally achieved, by embodiments of this disclosure which describe a system and method for shared memory ownership using context.
In accordance with an embodiment, an apparatus for shared memory processing is provided. In this example, the apparatus includes a data plane, a memory controller coupled to the data plane, and a plurality of central processing unit (CPU) elements (CEs) coupled to the data plane. The plurality of CEs include a CE configured to send a read request to the memory controller over the data plane that requests access to a shared memory resource. The apparatus further includes a Data Ownership Manager (DOM) coupled to the data plane. The DOM is configured to detect the read request, to determine whether a thread assigned to the CE is permitted to access the shared memory resource based on one or more identifiers (IDs) in the read request, and to send a control signal to the memory controller indicating whether the thread being executed by the CE is permitted to access the shared memory resource.
In accordance with another embodiment, a method of for providing memory protection for shared memory is provided. In this example, the method includes detecting a read request requesting access to a shared memory resource by a DOM. The read request is communicated from a central processing unit (CPU) element (CE) to a memory controller. The method further includes determining whether a thread assigned to the CE is permitted to access the shared memory resource based on one or more identifier (IDs) in the read request, and sending a control signal from the DOM to the memory controller that indicates whether the thread being executed by the CE is permitted to access the shared memory resource.
In accordance with yet another embodiment, a method for managing a cache is provided. In this example, the method includes sending a read request from central processing unit (CPU) element (CE) to a memory controller without independently verifying whether a thread is permitted to access a shared memory resource. The read request requests access to the shared memory resource. The method further includes receiving a data response from the memory controller that includes a response code that indicates whether the thread being processed by the CE is permitted to access the shared memory resource, and processing the data response based on the response code.
In accordance with yet another embodiment, a method for memory control is provided. In this example, the method includes storing a plurality of data ownership table (DOT) entries of a DOT for a memory. Each of the plurality of DOT entries comprises a context identifier identifying a shared context, a job identifier identifying at least one task of a plurality of computing tasks that share the shared context, and an access permission for a shared memory block associated with the shared context.
For a more complete understanding of the present invention, and the advantages thereof, reference is now made to the following descriptions taken in conjunction with the accompanying drawings, in which:
The structure, manufacture and use of the presently preferred embodiments are discussed in detail below. It should be appreciated, however, that the present invention provides many applicable inventive concepts that can be embodied in a wide variety of specific contexts. The specific embodiments discussed are merely illustrative of specific ways to make and use the invention, and do not limit the scope of the invention. As used herein, the term “CE” may generally refer to any component that attempts to access a shared memory resource. For example, a CE may be a core of a multi-core processor or a processor in a parallel processing system. A CE may be a component that does not directly execute threads, such as a direct memory access controller (DMA).
Memory protection techniques may be employed to avoid and/or mitigate memory errors in shared memory architectures. Conventional memory protection schemes rely primarily on distributed memory management units (MMUs) in the CEs to verify access permissions and maintain cache coherency. In particular, an MMU may verify that a thread being executed by a host CE has permission to access a given memory resource before the data is made available to a CPU in the CE for execution of a thread. By way of example, a CE executing a particular thread may send a read request to a memory controller in order to access data stored in shared memory resource. As used herein, the term “read request” refers to a read-only (RO) request or a read-write (RW) request. The read request may generally include an address assigned to the shared memory location that is being accessed. Prior to sending the read request, an MMU in the CE may perform an access check to determine whether the thread being executed by the CE has the appropriate access permission. In general, the MMU performs the access check by searching for the address, carried by the read request, in a page table. The page table may list addresses, or a range of address, that a given thread is permitted to access. If the MMU determines that the address is not listed in the page table, or within a range of addresses listed in the page table, the MMU generates a page fault, and the read request is not forwarded to the MMU. A page fault is an interrupt that may cause the CE to suspend its processing of the thread, and perform a self-correcting operation.
If the MMU determines that the address is listed in the page table, or within a range of addresses listed in the page table, then access permission is confirmed, and the read request is sent to the memory controller. The memory controller then processes the read request, and returns a data response to the CE that includes the data stored in the memory resource. Upon reception, the MMU or a local control block in the CE may store the data into the local cache, at which time the data is made available to a CPU in the CE for execution of the thread.
It may take several clock cycles (e.g., roughly four clock cycles) for an MMU to verify an access permission prior to sending the read request to the memory controller. This generally delays the CPU's access to the data, thereby introducing latency into execution of the thread. Additionally, the CE/CPU may need to accommodate, or adjust for, the clock cycles used to verify the access permission by the MMU, which adds complexity to the design of the CE/CPU.
Embodiments provided herein reduce latency attributable to memory protection in shared memory systems by performing access protection at a central Data Ownership Manager (DOM), rather than at distributed MMUs in the CEs. In particular, the DOM may monitor read requests communicated over a data plane between the CEs and the memory controller, and perform access protection verification in parallel with the memory controller's generation of the data response. The DOM is separate and distinct from both the CEs and the memory controller, and therefore generally is able to make the access determination without interfering with data plane processing/generation of the read requests and data responses exchanged between the memory controller and the CEs. After making the access permission determination, the DOM transmits a control signal to the memory controller indicating whether a thread being executed by the requesting CE is permitted to access the corresponding shared memory resource. The control signal may also include cache attributes. Cache attributes may qualify how a particular unit, or piece, of memory is to be cached, such as read only cacheable, read-write cacheable, non-cacheable, write through, write back, etc. The memory controller then sets a response code in the data response to indicate whether the thread being executed by the requesting CE is permitted to access the shared memory resource, and sends the response code to the requesting CE.
The response code in the data response improves efficiency of the CE in so far as it allows the CE to send the read request without having to independently verify that the thread is permitted to access the shared memory resource. This saves several clock cycles when compared with conventional memory protection techniques, which typically require an MMU in the CE to verify access permissions prior to the data being stored in the local cache.
The DOM may determine whether a given thread is permitted to access a shared memory resource based on one or more identifiers (IDs) in the read request. In some embodiments, the DOM makes the access permission determination by comparing one or more IDs in the read request with one or more IDs associated with the corresponding shared memory location in a data ownership table. The data ownership table may be maintained by the DOM according to context instructions received from a hypervisor as well as read requests received from CEs. In one example, the data ownership table associates a context identifier (ID), a job ID, and an access privilege with each shared memory resource. The context ID in the data ownership table may identify threads that are permitted to read data from a given memory resource.
The Job ID may provide a mechanism for allowing multiple threads to access a shared memory resource via a read-write (RW) request at the same time. In particular, shared memory access protocols often include a general rule that prohibits multiple threads from accessing a memory resource with a read-write request at the same time, even when the threads share the same context ID. This general rule is intended to maintain coherency between the shared memory resource and the local caches of the CEs. However, in some instances, an underlying application may be designed to maintain coherence and ordering between threads when the threads are simultaneously modifying data in a shared memory resource via a RW request.
Embodiments of this disclosure assign a job ID to those threads in order to allow an application to override the general rule prohibiting multiple threads from accessing a memory resource with a RW request at the same time. More specifically, a DOM may store the context ID and job ID carried by an initial RW request in the data ownership table when an initial thread accesses a memory resource using a RW request. Thereafter, if a subsequent thread attempts to access the memory resource with a subsequent RW request, the DOM may compare both the context ID and the job ID of the subsequent RW request with the context ID and the job ID stored in the data ownership table. If both pairs of IDs match, then the DOM may grant RW access permission to the subsequent thread before the memory resource is released by the initial thread.
To improve system scalability, the DOM may only compare context IDs for RO requests. Hence, on one hand, the DOM may compare a context ID and a job ID in a RW request with the context ID and the job ID in the data ownership table. On the other hand, the DOM may only compare the context ID in a RO request to the context ID associated with the corresponding memory resource in the data ownership table. This may improve scalability by allowing the DOM to process RO requests more quickly/efficiently than RW requests. The DOM may also store faulty access information in debug registers of the DOM to improve the ability to debug impermissible access attempts.
An additional benefit of performing access verification at the DOM is that it allows the caches to be managed at a finer granularity. When access verification is performed at distributed MMUs, cache attributes are set/stored in the page tables maintained by the MMU at the CE. For practical reasons, memory is generally written into the page tables on a page-by-page basis, meaning that the minimum block-memory size is approximately 4 kilobytes (KBs), which is the size of a single page. When access verification is performed at the DOM, the cache attributes are set/stored in the DOT table maintained by the DOM at the cache line boundary. Memory may generally be written to a much lower granularity at the cacheline boundary, e.g., 64 bytes, 128 bytes, or 256 bytes. Thus, performing the access verification at the DOM allows for the attributes to be written into the corresponding table at a much lower granularity, thereby allowing the memory resources to be used more efficiently. These and other aspects are described in greater detail below.
The CEs 110, 120, 130 may be independent processors and/or independent cores of a multi-core processor. As shown, the CE 110 includes a local cache 112, an MMU 114, a CPU 116, and a control block 118. Likewise, the CE 120 includes a local cache 122, an MMU 124, a CPU 126, and a control block 128, and the CE 130 includes a local cache 132, an MMU 134, a CPU 136, and a control block 138. The local caches 112, 122, 132 are used to store data that is read from the pool of memory resources. The control blocks 118, 128, 138 are configured to manage the local caches, and the MMUs 114, 124, 134 are configured to perform various tasks, such as memory protection, address translation, and cache attribute verification. The CPUs 116, 126, 136 are configured to execute a thread based on data stored in the local cache 112. The CPUs 116, 126, 136 may be general-purpose processors and/or specialized processors, such as digital signal processors (DSPs), hardware accelerators (HACs), etc.
The CEs 110, 120, 130 access the pool of memory resources 170 by sending read requests to the memory controller 160 via the data plane iso. The data plane 150 is an interface (e.g., bus, etc.) between the CEs 110, 120, 130 and the memory controller 160. The memory controller 160 manages the flow of data going into, and out of, the pool of memory resources 170. The memory controller 160 may be a digital circuit that is implemented on a stand-alone chip or otherwise integrated into another chip, e.g., as part of a multi-core microprocessor.
As mentioned above, the MMUs 114, 124, 134 may perform access verification prior to sending a read request to the memory controller 260 to ensure that the threads being executed by the CEs 110, 120, 130 are permitted to access the corresponding shared memory resource. This may introduce latency into thread execution, as well as syphon off processing resources that could otherwise be used by the CPUs 116, 126, 136 when executing the thread.
Embodiments of this disclosure offload access protection and/or other tasks (e.g., cache coherency verification, etc.) from the distributed MMUs to a centralized DOM in order to reduce latencies and/or processing inefficiencies that result from MMU operation of the MMUs.
The DOM 290 includes a data ownership table (DOT) 292 and one or more debug registers 294. The DOM 290 may use the DOT 292 to perform access permissions. In the event that an access permission is deemed invalid, the DOM 290 may store a faulty access information corresponding to the read request in the debug registers 294. The faulty access information may include any information that may prove useful during debugging, such as a CE ID, a physical address, a time stamp, and access attributes carried by the read request.
A DOM 290 may perform different access checks for RO requests and RW requests. In particular, the DOM 290 may compare both a context ID and a job ID in a RW request with a context ID and a job ID in the DOT 292. If both the context IDs and the job IDs match, then the DOM 290 may verify that the thread being processed by the requesting CE has permission to access, and modify, data stored in the memory resource. In comparison, the DOM 290 may compare a context ID in a RO request with a context ID in the DOT 292 without comparing the job ID in the RO request with the job ID in the DOT 292. If the context IDs match, then the DOM 290 may verify that the thread being processed by the requesting CE has permission to read data stored in the memory resource. This may improve scalability by reducing the number of steps/tasks the DOM 290 must perform to verify access permission for RO requests.
A DOM may be a context aware module that is responsible for setting and changing contexts and permissions associated with shared memory resources based on qualified context requests. Upon receiving a context request, the DOM may check a context table to verify that the transaction context exists in the table and that the requesting party is qualified to set or change permissions. An example of a context table is provided in Table 1. As shown, the context ID table may include N-supervisor contexts. At reset, all the supervisor contexts may be programmed as hypervisor context. For example, a DOM may have 16 contexts programmed as hypervisor context at reset time. The DOM may look at shared memory as a set of variable size blocks with context attached to it. Each context may have its own permissions. Each context may have multiple buffers. The buffers may have different permissions. All memory blocks may be cache-line aligned.
TABLE 1
Context
(8-bits)
Description
Comments
0
Global permissions
Special context
1
Hypervisor
Special context
At boot time programmed as
supervisor for all IPs
xxxx
Supervisor
16-supervisor contexts and could
Hypervisor(1)
be different for different IP
A DOM may validate all shared memory access attempts without being in the critical path and without adding any extra cycles for access validation. A DOM may perform all access checks in parallel with the memory controller data path and flags data path if the access has passed protection checks or failed before accessing physical memory.
A DOM may maintain a common data ownership (context aware) table for shared memory. The data ownership table may be scalable to add any number of CEs. All protection checks may be done at the cacheline boundary. Cache attributes may be assigned at a cacheline granularity. The end CEs may configure different cache policies at cacheline granularity.
In some embodiments, multiple CEs sharing the same context may access a memory block with RO permissions. No additional checks will be performed. CEs accessing a memory block with RW permissions may have to go through additional job ID check to ensure that the application is aware of the sharing of RW buffers and will take care of coherence and ordering. In some instance, if a DOM receives a RW request and it is not aware of any other CEs accessing the corresponding memory resource, then the DOM only checks the context and job-id. This promotes scalability and flexibility, multiple CEs can be added removed as per need to access a buffer at any time
In various embodiments, a multicore system having real-time data-flow with message passing includes data that is isolated in each of multiple parallel computing tasks (e.g., threads, processes, and the like) and that is consumed and shared by sending and receiving messages using, e.g., buffer pointers. The multicore system may be, for example, a Wireless Base Band (WBB) device. To make shared memory behave like a private memory for each time slice and to maintain coherency in a manner that is appropriate to the data application, the WBB includes a Data Ownership Manager (DOM) that performs protection checking of memory access requests. The DOM is context-aware and aligned with each cache line at its source in memory. In some embodiments, the DOM works in parallel with the pipeline of the shared memory controller and does not add any extra latency in the data access path.
In various embodiments, the DOM resides in a portion of shared memory that tracks all accesses and looks for sharing violations. Even when multiple tasks have the same permissions, the DOM may still keep track of which task has permission to access and/or modify each cache line worth of data in the memory. The DOM may use a data ownership table to track the ownership of shared memory blocks that has been established through access requests. As an example, for each access request that established ownership of a particular memory block, the data ownership table may store an entry that includes a context identifier (ID) that identifies the context of a task that requested the memory access (which is also the context of the shared memory block), a job ID that identifies the requesting task among different tasks that share the same context, and a resulting access permission of that task for accessing the memory block.
In various embodiments using a DOM, multiple tasks can own memory with a Read Only (RO) access permission, but only one task can own memory with a Read-Write (RW) access permission. Ownership is established automatically when memory is accessed. If a task attempts to read from a memory location, then that thread is assigned RO sharing ownership for the memory location, but if another task owns the memory location as RW at that time, a permission exception occurs. When memory is written by a task, it is promoted to RW sharing for that task, but if other tasks own the memory as RO or RW at that time, then a permission exception occurs. When a task terminates or when it releases a memory location, the Operating System (OS) informs the DOM and ownership is revoked for that task.
Each of the cores 802 communicates with the memory controller 828 via master ports 824 of bus interfaces 826. In an embodiment, core 802 has multiple master ports 824, while the remaining cores 802 have one master port 824 each. In other embodiments, any of the cores 802 may have multiple master ports 824.
Referring again to
The master ports 824 interface with a bus interface 830 of the memory controller 828. This bus interface 830 may be any interface that allows signaling between the master ports 824 and the memory controller, such as an AXI/ACE INTF. The shared memory 804 includes M physical memory banks 832 to 832. Each of these physical memory banks 832 to 832 is written to and read by bank controllers 834 of the memory controller 828. The accuracy of the data that is read and written by these bank controllers 834 is checked by Error-Correcting Code (ECC) modules 836 of the memory controller 828 that receive memory access requests from other components of the memory controller 828 and provide responses indicating whether these requests were successful.
The DOM 810 includes configuration registers 838, debug registers 840, a DOT 842, and an access check module 844. The debug registers 840 may be, for example, trace registers. The DOM 810 also provides a protection pass/fail signal to components of the memory controller 828.
In the embodiment of
As an example, the DOM 810 may have a global permissions context having a context ID of 0, a special boot time context called a hypervisor context that has a context ID of 8 and is programmed as a supervisor for all the cores 802, and 16 supervisor contexts that are normally identified by different context IDs but are all programmed with the hypervisor context ID of 1 at a reset time. In an embodiment, the context ID is an eight-bit identifier.
In some embodiments, the processing system 900 is included in a network device that is accessing, or part otherwise of, a telecommunications network. In one example, the processing system 900 is in a network-side device in a wireless or wireline telecommunications network, such as a base station, a relay station, a scheduler, a controller, a gateway, a router, an applications server, or any other device in the telecommunications network. In other embodiments, the processing system 900 is in a user-side device accessing a wireless or wireline telecommunications network, such as a mobile station, a user equipment (UE), a personal computer (PC), a tablet, a wearable communications device (e.g., a smartwatch, etc.), or any other device adapted to access a telecommunications network.
In some embodiments, one or more of the interfaces 910, 912, 914 connects the processing system 900 to a transceiver adapted to transmit and receive signaling over the telecommunications network.
The transceiver 1000 may transmit and receive signaling over any type of communications medium. In some embodiments, the transceiver 1000 transmits and receives signaling over a wireless medium. For example, the transceiver 1000 may be a wireless transceiver adapted to communicate in accordance with a wireless telecommunications protocol, such as a cellular protocol (e.g., long-term evolution (LTE), etc.), a wireless local area network (WLAN) protocol (e.g., Wi-Fi, etc.), or any other type of wireless protocol (e.g., Bluetooth, near field communication (NFC), etc.). In such embodiments, the network-side interface 1002 comprises one or more antenna/radiating elements. For example, the network-side interface 1002 may include a single antenna, multiple separate antennas, or a multi-antenna array configured for multi-layer communication, e.g., single input multiple output (SIMO), multiple input single output (MISO), multiple input multiple output (MIMO), etc. In other embodiments, the transceiver 1000 transmits and receives signaling over a wireline medium, e.g., twisted-pair cable, coaxial cable, optical fiber, etc. Specific processing systems and/or transceivers may utilize all of the components shown, or only a subset of the components, and levels of integration may vary from device to device.
It should be appreciated that one or more steps of the embodiment methods provided herein may be performed by corresponding units or modules. For example, a signal may be transmitted by a transmitting unit or a transmitting module. A signal may be received by a receiving unit or a receiving module. A signal may be processed by a processing unit or a processing module. Other steps may be performed by a generating unit/module, a comparing unit/module, a calculating unit/module, and/or a detecting unit/module. The respective units/modules may be hardware, software, or a combination thereof. For instance, one or more of the units/modules may be an integrated circuit, such as field programmable gate arrays (FPGAs) or application-specific integrated circuits (ASICs).
While this invention has been described with reference to illustrative embodiments, this description is not intended to be construed in a limiting sense. Various modifications and combinations of the illustrative embodiments, as well as other embodiments of the invention, will be apparent to persons skilled in the art upon reference to the description. It is therefore intended that the appended claims encompass any such modifications or embodiments.
Gatherer, Alan, Luan, Hao, Wokhlu, Sushma, Mcfearin, Lee Dobson
Patent | Priority | Assignee | Title |
11048648, | Aug 10 2018 | Shenzhen Goodix Technology Co., Ltd. | SoC chip and method for controlling bus access |
Patent | Priority | Assignee | Title |
5953530, | Feb 07 1995 | Sun Microsystems, Inc. | Method and apparatus for run-time memory access checking and memory leak detection of a multi-threaded program |
9747218, | Mar 20 2015 | MILL COMPUTING, INC | CPU security mechanisms employing thread-specific protection domains |
20020002654, | |||
20120047516, | |||
20130160114, | |||
20140223047, | |||
20140281110, | |||
JP2009251967, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Jun 22 2016 | GATHERER, ALAN | FUTUREWEI TECHNOLOGIES, INC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 039031 | /0460 | |
Jun 23 2016 | MCFEARIN, LEE DOBSON | FUTUREWEI TECHNOLOGIES, INC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 039031 | /0460 | |
Jun 23 2016 | LUAN, HAO | FUTUREWEI TECHNOLOGIES, INC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 039031 | /0460 | |
Jun 24 2016 | Futurewei Technologies, Inc. | (assignment on the face of the patent) | / | |||
Jun 27 2016 | WOKHLU, SUSHMA | FUTUREWEI TECHNOLOGIES, INC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 039031 | /0460 |
Date | Maintenance Fee Events |
Apr 05 2023 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Date | Maintenance Schedule |
Oct 22 2022 | 4 years fee payment window open |
Apr 22 2023 | 6 months grace period start (w surcharge) |
Oct 22 2023 | patent expiry (for year 4) |
Oct 22 2025 | 2 years to revive unintentionally abandoned end. (for year 4) |
Oct 22 2026 | 8 years fee payment window open |
Apr 22 2027 | 6 months grace period start (w surcharge) |
Oct 22 2027 | patent expiry (for year 8) |
Oct 22 2029 | 2 years to revive unintentionally abandoned end. (for year 8) |
Oct 22 2030 | 12 years fee payment window open |
Apr 22 2031 | 6 months grace period start (w surcharge) |
Oct 22 2031 | patent expiry (for year 12) |
Oct 22 2033 | 2 years to revive unintentionally abandoned end. (for year 12) |