A dns proxy may tag dns requests from a virtual network with a separating identifier, such as a virtual network ID, before forwarding them to a shared dns server. This can allow each virtual network to have its own namespace and avoid naming conflicts. The shared dns server can examine the tagged dns requests to respond to the dns requests using the local namespace as identified by the separating identifier.
|
12. A domain name system (dns) proxy coupled between a shared multi-tenant dns server and a plurality of virtual networks, comprising:
one or more storage devices; and
one or more processors in communication with the one or more storage devices, wherein the one or more processors are configured to:
receive a dns request from a virtual network of the plurality of virtual networks, the shared multi-tenant dns server storing dns data for each of the plurality of virtual networks, the shared multi-tenant dns server hosting isolated domain name databases, each domain name database being associated with a different virtual network of the plurality of virtual networks, each of the plurality of virtual networks having a dedicated dns local namespace;
tag, by the dns proxy, a separating identifier to the dns request; and
forward, by the dns proxy, the dns request to the shared multi-tenant dns server.
1. A method for use by at least one processor and at least one memory, wherein computer-executable instructions are stored in the at least one memory which, when executed by the at least one processor, perform a method comprising:
receiving, by a domain name system (dns) proxy coupled between a shared multi-tenant dns server and a plurality of virtual networks, a dns request from a virtual network of the plurality of virtual networks, the shared multi-tenant dns server storing dns data for each of the plurality of virtual networks, the shared multi-tenant dns server hosting isolated domain name databases, each domain name database being associated with a different virtual network of the plurality of virtual networks, each of the plurality of virtual networks having a dedicated dns local namespace;
tagging, by the dns proxy, a separating identifier to the dns request; and
forwarding, by the dns proxy, the dns request to the shared multi-tenant dns server.
6. A method for use by at least one processor and at least one memory, wherein computer-executable instructions are stored in the at least one memory which, when executed by the at least one processor, perform a method comprising:
receiving, by a multi-tenant domain name system (dns) server, a dns request that was tagged by a proxy with a separating identifier, the multi-tenant dns server configured to serve a plurality of virtual networks, the multi-tenant dns server hosting isolated domain name databases, each domain name database being associated with a different virtual network of the plurality of virtual networks;
accessing, by the multi-tenant dns server, a dns local namespace dedicated to a virtual network of the plurality of virtual networks based on the separating identifier;
using, by the multi-tenant dns server, the separating identifier to process the request for the virtual network to resolve the dns request; and
if the dns request resolves at the dns local namespace, providing, by the multi-tenant dns server, a response to the dns request to the proxy.
4. The method of
5. The method of
7. The method of
9. The method of
accessing a shared dns cache for the dns request if the request does not resolve at the dns local namespace; and
if the dns request resolves at the dns cache, providing a response to the dns request.
10. The method of
11. The method of
14. The proxy of
16. The proxy of
17. The method of
receiving, by the dns proxy, a dns response from the shared multi-tenant dns server; and
forwarding, by the dns proxy, the dns response to the virtual network of the plurality of virtual networks.
19. The method of
20. The method of
creating, by the multi-tenant dns server, a partition for the virtual network of the plurality of virtual networks to store the dns local namespace.
|
This Application claims the benefit of Indian Patent Application Serial No. 6691/CHE/2015, filed Dec. 14, 2015, the disclosure of which is hereby incorporated by reference as set forth in its entirety.
The Domain Name System (DNS) is a hierarchical distributed naming system for resources connected to networks, such as the Internet. The Domain Name System associates numerical IP addresses with domain names of network entities. DNS servers translate domain names, which can be easily remembered and used by humans, to numerical IP addresses. DNS servers can also store and provide other record types (such as Canonical Name (CNAME), free form text (TXT), mail exchanger record (MX record), Pointer records (such as PTR records) that do different functions.
The Domain Name System maintains the domain name hierarchy and provides translation services between it and the Internet Protocol (IP) address space. Internet name servers and a communication protocol implement the Domain Name System. A DNS name server is a server that stores the DNS records for a domain name, receives DNS requests and responds with answers to queries against its database.
In addition to the Domain names that it is responsible for, DNS servers typically cache DNS requests and responses that have been received from other DNS name servers. Caching name servers (DNS caches) store DNS query results for a period of time determined in an associated time-to-live of a domain-name record. DNS caches improve the efficiency of the DNS by reducing DNS traffic across the Internet, and by reducing load on authoritative name-servers, particularly root name-servers. Because they can answer questions more quickly, they also increase the performance of end-user applications that use the DNS.
Embodiments concern a DNS server that can be shared by multiple virtual networks in a cloud computing environment.
A DNS proxy may tag DNS requests from a virtual network with an identifier, such as a virtual network ID, before forwarding them to a shared DNS server. This can allow each virtual network to have its own namespace and avoid naming conflicts.
The shared DNS server can examine the tagged DNS requests to respond to the DNS requests using the local namespace as identified by the virtual network ID.
This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter.
A more detailed understanding may be had from the following description, given by way of example in conjunction with accompanying drawings wherein:
Embodiments concern a system and method for providing a shared domain name resolution services for a multitenant network.
A cloud service provider (CSP) is a company that offers some component of cloud computing, such as Infrastructure as a Service (IaaS), Software as a Service (SaaS) or Platform as a Service (PaaS), to other businesses or individuals. Cloud service providers often aim to facilitate onboarding of the tenants' private networks into their environments for which they need to keep the addressing and naming properties of these networks intact while providing appropriate network isolation between tenants. The network virtualization technologies allow tenants to bring their own IP addresses or IP address space in a managed cloud. Embodiments can solve the issue of providing name resolution services for the virtual networks where customers can bring their own domain names. The systems can also ensure privacy by preventing private domains from being seen outside the virtual network.
Embodiments define the capabilities and interfaces for a shared DNS server composed of different virtualization instances that host isolated domain name databases, catering to different virtual networks. Embodiments also define a DNS proxy that listens to the virtual networks; identifies and adds tenant metadata into the DNS packets (in form of a EDNS0 option) which are received by the DNS server which processes the metadata to identify the virtualization instance.
Benefits can include:
1. Ability to allow tenants of a managed cloud to bring their own domain names.
2. Ability to use a shared authoritative DNS server that provides name resolution services for tenants with overlapping domain name spaces with appropriate tenant isolations.
3. Ability to provide the above services irrespective of the virtualization technologies used by the provider.
4. Isolation of tenant domain name databases from the query and management perspective
The cloud computing platform 110 is configured to execute cloud applications 120 requested by the client devices 130. The cloud computing platform 110 maintains computing devices that provide virtual machines, which execute the cloud application 120. The cloud computing platform also includes storage resources that store applications and system information. The cloud computing platform 110 connects to the client devices 130 via a communications network, such as a wireless network, local area network, wired network, or the Internet.
The cloud applications 120 are available to the client devices 130. The software executed on the cloud computing platform 110 implements the cloud applications 120. In one embodiment, virtual machines provided by the cloud computing platform 110 execute the cloud applications 120. The cloud applications 120 may include, but are not limited to, editing applications, network management applications, finance applications, or any application requested or developed by the client devices 130. In certain embodiments, some functionality of the cloud application 120 may be executed on the client devices 130.
The client devices 130 are utilized by a user to interact with cloud applications 120 provided by the cloud computing platform 110. The client devices 130, in some embodiments, must register with the cloud computing platform 110 to access the cloud applications 120. Any client device 130 with an account from the cloud computing platform 110 may access the cloud applications 120 and other resources provided in the cloud computing platform 110. The client devices 130 include, without limitation, personal digital assistants, smart phones, laptops, personal computers, gaming systems, set-top boxes, or any other suitable client computing device. The client devices 130 include user and system information storage to store user and system information on the client devices 130. The user information may include search histories, cookies, and passwords. The system information may include internet protocol addresses, cached Web pages, and system utilization. The client devices 130 communicate with the cloud computing platform 110 to receive results from the cloud applications 120.
Accordingly, the computing system 100 is configured with a cloud computing platform 110 that provides cloud applications 120 to the client devices 130. The cloud applications 120 remove the burden of updating and managing multiple local client applications on the client devices 130. Cloud applications 120 can also handle necessary scale, security and availability issues that the client devices 130 may not be setup for.
An exemplary cloud computing platform is Microsoft Azure. Microsoft Azure uses a specialized operating system, also called Microsoft Azure, to run its “fabric layer”: a cluster hosted at Microsoft's data centers that manage computing and storage resources of the computers and provisions the resources (or a subset of them) to applications running on top of Microsoft Azure.
The cloud computing system can use cloud computing devices and software at a client data center, such as with Microsoft Azure Stack. Moving cloud computing techniques to a company datacenter can provide agility and productivity for application owners, flexibility and control for Information technology (IT) units, and assurance that corporate assets are protected.
In
The computer 241 may also include other removable/non-removable, volatile/nonvolatile computer storage media. By way of example only,
The drives and their associated computer storage media discussed above provide storage of computer readable instructions, data structures, program modules and other data for the computer 241. In
The computer 241 may operate in a networked environment using logical connections to one or more remote computers, such as a remote computer 246. The remote computer 246 may be a personal computer, a server, a router, a network PC, a peer device or other common network node, and typically includes many or all of the elements described above relative to the computer 241, although only a memory storage device 247 has been illustrated in
When used in a LAN networking environment, the computer 241 is connected to the LAN 245 through a network interface or adapter 237. When used in a WAN networking environment, the computer 241 typically includes a modem 250 or other means for establishing communications over the WAN 249, such as the Internet. The modem 250, which may be internal or external, may be connected to the system bus 221 via the user input interface 236, or other appropriate mechanism. In a networked environment, program modules depicted relative to the computer 241, or portions thereof, may be stored in the remote memory storage device. By way of example, and not limitation,
A virtual network is a computer network that consists, at least in part, of virtual network links. A virtual network link is a link that does not consist of a physical (wired or wireless) connection between two computing devices but is implemented using methods of network virtualization.
Multi-tenant architectures may use multiple virtual networks. As disclosed below, each virtual network may have need of a DNS server.
One solution is for each tenant (virtual network) to deploy one virtual machine (VM) as a DNS server in its network, where it manages and hosts the required domains. The tenant virtual machines point to this DNS server for name resolution. This DNS server may double up as an authoritative server as well as a forwarder/recursive resolver for external queries. This solution is non-optimal as it requires one dedicated VM per tenant which is dedicated for DNS and puts the management overhead on the clients. This solution also doesn't provide fault tolerance, e.g. if the VM dies.
Another solution is for each tenant virtual network to use restricted name spaces. In this solution, the clients cannot bring their own domain name into the cloud. They are forced to keep some specific DNS suffixes (say in format <VM-Name>.<Tenant-ID>.myclouddomain.com). In this solution, the DNS resolution works fine, but an extra access control list (ACL) has to be created on the query path to ensure that one tenant cannot resolve names for another tenant.
These solutions may not be cost effective or may limit the capacity or functionality of the system. The following describes a solution that allows tenants to use a single DNS server to serve multiple clients with full isolation and allows customers to bring their own domain name to cloud. This facilitates seamless migration of tenant domains to/from private-public cloud.
In one embodiment, as shown in
DNS proxy 302 can comprise computer or server including one or more storage devices 302a and one or more processors 302b in communication with the one or more storage devices 302a. The one or more processors 302b may be configured to do operations as described below.
The DNS proxy 302 can be considered to be a gateway for forwarding DNS traffic. It may listen to the DNS packets emanating from tenant networks and forwards the query to the DNS server 304 on the physical network. Apart from forwarding the multitenant proxy can also add an EDNS0 option ID, as defined by Internet standards track protocol RFC 2671 (Extension Mechanisms for DNS (EDNS0)), into the outgoing DNS packet. This optional ID can contain a separating identifier such as the Virtual Network ID (VNET identifier). This identifier is used by DNS server 304 to ensure isolation. On the return path the proxy may clean up the EDNS0 option and routes the packet back to the tenant VM.
In step 402, a Domain Name System (DNS) request is received from a virtual network. The DNS proxy 302 may intercept the DNS request as it is being sent to the shared DNS server 304.
In step 404, a DNS request is tagged with a separating identifier. The separating identifier can be any identifier that separates any two parties, e.g. it could be per company, per person, per application, per anything else. The separating identifier can be a virtual network ID.
In one example, the separating identifier, such as the virtual network ID, is placed in an option field. For example, the separating identifier may be placed in an EDNS0 option field. Alternately, the separating identifier can be appended in some other way to the DNS request.
In step 406, the DNS request is forwarded to a shared DNS server 304 that stores DNS data for multiple virtual networks including the virtual network.
Looking again at
The DNS proxy 302 may also be configured to receive a DNS response to the DNS request from the shared DNS server 304. The DNS response can be also tagged with the separating identifier, such as the virtual network ID. This DNS response can then be forwarded to the correct virtual network as indicated by the separating identifier, such as the virtual network ID. In another embodiment, the routing of the reply doesn't need to be done using the separating identifier, such as the virtual network ID.
The DNS proxy 302 can examine all DNS requests and DNS responses at a location. In
A DNS server 304 can comprise a shared DNS server that serves multiple virtual networks. The shared DNS server can be provisioned and maintained with independent operation for each associated virtual network. The DNS server 304 can comprise one or more storage devices 304a (memories) and one or more processors 304b in communication with one another.
The DNS Server 302 may be a DNS server capable of providing:
1. Support for a “bring your own domain name” scenario, allowing DNS zones to be configured by multiple tenants. A DNS zone may be any contiguous portion of the domain name space in the Domain Name System (DNS) for which administrative responsibility has been delegated to a single manager.
2. Ability to perform name resolution for customer hosted domains, while providing tenant isolation during name resolution
3. Isolation of management interfaces and persistent store for DNS zones. Application Programming Interfaces (APIs) can be provided to create virtual instance on DNS server. The virtual instances can have DNS zones that contain DNS records. These APIs can be isolated so that other users cannot access them. The APIs can send info to an underlying system which can perform a Remote Procedure Call (RPC) to the DNS Server 302.
4. Isolation of server settings (forwarders, conditional forwarders etc.). Configuration settings for each of the virtual networks at the DNS Server 302 may be independent.
5. Ability to perform tenant specific Dynamic DNS (DDNS) updates.
6. Isolation of DNS zone transfers. DNS zone transfers allow two DNS servers to synchronize. A zone transfer request can include a tag of the virtual network ID to allow the correct local namespace information to be synchronized.
7. Support for traffic management within the tenant domain
8. Ability to perform public name resolution.
The DNS server can be composed of multiple virtualization instances with tenant specific data and settings, each mapped to a virtual network. Using the virtual network ID in the incoming query, the DNS server may perform the name resolution tasks. The DNS server may also be capable of performing public name resolution.
In step 502, a DNS request that was tagged by a proxy with a separating identifier, such as a virtual network ID, is received. A separating identifier, such as the virtual network ID, allows the DNS server 304 to determine which virtual network the DNS request is for.
In step 504, a DNS local namespace dedicated to the virtual network is checked using the separating identifier to process that request for the virtual network. Since local namespaces are used, each virtual network can independently map the same domain name to different IP address.
In step 506 if the DNS request resolves at the DNS local namespace, a response is provided to the DNS request. Otherwise, the request can be resolved at a DNS cache or an external DNS server.
The example of
The DNS request can be replied to with a response that is tagged with the separating identifier, such as the virtual network ID. Alternately, other addressing methods can be used.
The DNS server 304 can check a shared DNS cache for the DNS request if the request does not resolve at the DNS local namespace. The shared DNS cache can be used for multiple virtual networks since it stores IP addresses for the external network. For example, when the DNS request concerns a request for an IP address of a location on an external network, the IP address may or may not be stored in this shared cache. If the IP address is stored in this shared cache, the DNS server 304 can respond with the IP address. The DNS server 304 can check with an external DNS server if the request does not resolve at the shared cache. The shared cache can then be updated with the IP address from the external DNS server.
As shown in
To provision the system, once a virtual network is created, the DNS proxy and DNS server are instructed about the correspondence of the separating identifiers (such as the virtual network IDs) and the virtual network. The DNS server then creates a partition for the virtual network to store the local namespace. A management process can set up and update the local namespace information at the DNS server.
It should be appreciated that any software components described herein may, when loaded into a processor and executed, transform the processor from a general-purpose computing system into a special-purpose computing system customized to facilitate the functionality presented herein. The processor may be constructed from any number of transistors or other discrete circuit elements, which may individually or collectively assume any number of states. More specifically, the processor may operate as a finite-state machine, in response to executable instructions contained within the software modules disclosed herein. These processor-executable instructions may transform the processor by specifying how the processor transitions between states, thereby transforming the transistors or other discrete hardware elements constituting the processor.
Encoding the software modules presented herein also may transform the physical structure of the computer-readable storage media presented herein. The specific transformation of physical structure may depend on various factors, in different implementations of this description. Examples of such factors may include, but are not limited to, the technology used to implement the computer-readable storage media, whether the computer-readable storage media is characterized as primary or secondary storage, and the like. For example, if the computer-readable storage media is implemented as semiconductor-based memory, the software disclosed herein may be encoded on the computer-readable storage media by transforming the physical state of the semiconductor memory. For example, the software may transform the state of transistors, capacitors, or other discrete circuit elements constituting the semiconductor memory. The software also may transform the physical state of such components in order to store data thereupon.
As another example, the computer-readable storage media disclosed herein may be implemented using magnetic or optical technology. In such implementations, the software presented herein may transform the physical state of magnetic or optical media, when the software is encoded therein. These transformations may include altering the magnetic characteristics of particular locations within given magnetic media. These transformations also may include altering the physical features or characteristics of particular locations within given optical media, to change the optical characteristics of those locations. Other transformations of physical media are possible without departing from the scope and spirit of the present description, with the foregoing examples provided only to facilitate this discussion.
In light of the above, it should be appreciated that many types of physical transformations take place in the architecture in order to store and execute the software components presented herein. It also should be appreciated that the architecture may include other types of computing devices, including hand-held computers, embedded computer systems, smartphones, PDAs, and other types of computing devices known to those skilled in the art. It is also contemplated that the architecture may not include all of the components shown in the figures, may include other components that are not explicitly shown in the figures, or may utilize an architecture completely different from that shown the figures.
Alternatively, or in addition, the functionally described herein may be performed, at least in part, by one or more hardware logic components. For example, and without limitation, illustrative types of hardware logic components that may be used include Field-programmable Gate Arrays (FPGAs), Application-specific Integrated Circuits (ASICs), Application-specific Standard Products (ASSPs), System-on-a-chip systems (SOCs), Complex Programmable Logic Devices (CPLDs), etc.
Computer-readable storage media may provide storage for instructions for the one or more processors. Although the description of computer-readable storage media contained herein refers to a mass storage device, such as a hard disk or CD-ROM drive, it should be appreciated by those skilled in the art that computer-readable media may be any available storage media.
By way of example, and not limitation, computer-readable storage media may include volatile and non-volatile, removable and non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules or other data. For example, computer-readable media includes, but is not limited to, RAM, ROM, EPROM (erasable programmable read only memory), EEPROM (electrically erasable programmable read only memory), Flash memory or other solid state memory technology, CD-ROM, DVDs, HD-DVD (High Definition DVD), BLU-RAY, or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which may be used to store the desired information and instructions. For purposes of this specification and the claims, the phrase “computer-readable storage medium” and variations thereof, does not include waves, signals, and/or other transitory and/or intangible communication media.
Although the subject matter presented herein has been described in language specific to computer structural features, methodological and transformative acts, specific computing machinery, and computer-readable storage media, it is to be understood that the methods and devices defined in the appended claims is not necessarily limited to the specific features, acts, or media described herein. Rather, the specific features, acts, and mediums are disclosed as example forms of implementing the claims.
Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example forms of implementing the claims. It is intended that the scope of the technology be defined by the claims appended hereto.
Gupta, Tushar, Sharma, Vijay Kumar, Gaitonde, Vithalprasad, Sain, Sourav, Ashutosh, Kumar, Mahodaya, Ashray, Bradshaw, Gareth, Costache, Cristian
Patent | Priority | Assignee | Title |
11019022, | Jan 28 2020 | F5 Networks, Inc | Processing packets with returnable values |
11329959, | Dec 21 2018 | Fortinet, Inc. | Virtual routing and forwarding (VRF)-aware socket |
Patent | Priority | Assignee | Title |
6480508, | May 12 1999 | NETGEAR, Inc; NETGEAR HOLDINGS LIMITED, A LIMITED LIABILITY | Router-based domain name system proxy agent using address translation |
6587455, | May 27 1999 | Telefonaktiebolaget LM Ericsson (publ) | Automatic discovery of nodes associated with a virtual subnet |
7853714, | Nov 15 2002 | Juniper Networks, Inc. | Providing services for multiple virtual private networks |
7970765, | Mar 14 2006 | Pulse Secure, LLC | Network device for providing integrated DNS caching services |
8707429, | Mar 31 2011 | AKAMAI TECHNOLOGIES, INC | DNS resolution, policies, and views for large volume systems |
8762574, | Aug 09 2007 | KOUNT INC | Establishing unique sessions for DNS subscribers |
8825839, | Nov 24 2010 | Unisys Corporation | Snooping DNS messages in a server hosting system providing overlapping address and name spaces |
8874790, | Dec 30 2011 | Verisign, Inc | DNS package in a partitioned network |
8935427, | Sep 23 2010 | Microsoft Technology Licensing, LLC | Providing virtual networks using multi-tenant relays |
20020002622, | |||
20020112076, | |||
20060013209, | |||
20070233851, | |||
20080189415, | |||
20090043900, | |||
20130024503, | |||
20140089523, | |||
20140181306, | |||
20140269712, | |||
20150063351, | |||
20150074221, | |||
20150131674, | |||
20150244643, | |||
20150256508, | |||
20150264121, | |||
20150281111, | |||
20160072847, | |||
WO2012065641, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
May 03 2016 | BRADSHAW, GARETH | Microsoft Technology Licensing, LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 038489 | /0376 | |
May 05 2016 | GUPTA, TUSHAR | Microsoft Technology Licensing, LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 038489 | /0376 | |
May 05 2016 | MAHODAYA, ASHRAY | Microsoft Technology Licensing, LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 038489 | /0376 | |
May 06 2016 | Microsoft Technology Licensing, LLC | (assignment on the face of the patent) | / | |||
May 06 2016 | COSTACHE, CRISTIAN | Microsoft Technology Licensing, LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 038489 | /0376 | |
May 06 2016 | ASHUTOSH, KUMAR | Microsoft Technology Licensing, LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 038489 | /0376 | |
May 06 2016 | SAIN, SOURAV | Microsoft Technology Licensing, LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 038489 | /0376 | |
May 06 2016 | GIATONDE, VITHALPRASAD | Microsoft Technology Licensing, LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 038489 | /0376 | |
May 06 2016 | SHARMA, VIJAY KUMAR | Microsoft Technology Licensing, LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 038489 | /0376 | |
Oct 26 2016 | GAITONDE, VITHALPRASAD | Microsoft Technology Licensing, LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 040162 | /0242 |
Date | Maintenance Fee Events |
Jan 06 2023 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Date | Maintenance Schedule |
Jul 16 2022 | 4 years fee payment window open |
Jan 16 2023 | 6 months grace period start (w surcharge) |
Jul 16 2023 | patent expiry (for year 4) |
Jul 16 2025 | 2 years to revive unintentionally abandoned end. (for year 4) |
Jul 16 2026 | 8 years fee payment window open |
Jan 16 2027 | 6 months grace period start (w surcharge) |
Jul 16 2027 | patent expiry (for year 8) |
Jul 16 2029 | 2 years to revive unintentionally abandoned end. (for year 8) |
Jul 16 2030 | 12 years fee payment window open |
Jan 16 2031 | 6 months grace period start (w surcharge) |
Jul 16 2031 | patent expiry (for year 12) |
Jul 16 2033 | 2 years to revive unintentionally abandoned end. (for year 12) |