A method and apparatus are disclosed for classifying materials utilizing a computerized touch sensitive screen or other computerized pointing device for operator identification and electronic marking of spatial coordinates of materials to be extracted. An operator positioned at a computerized touch sensitive screen views electronic images of the mixture of materials to be sorted as they are conveyed past a sensor array which transmits sequences of images of the mixture either directly or through a computer to the touch sensitive display screen. The operator manually “touches” objects displayed on the screen to be extracted from the mixture thereby registering the spatial coordinates of the objects within the computer. The computer then tracks the registered objects as they are conveyed and directs automated devices including mechanical means such as air jets, robotic arms, or other mechanical diverters to extract the registered objects.
|
1. A method for identifying and sorting selected material objects from a mixture of material objects comprising steps of:
(a) conveying a mixture of material objects into and through an inspection zone;
(b) irradiating said mixture of material objects with incident electromagnetic radiation while in said inspection zone;
(c) measuring the electromagnetic radiation emanating from said irradiated mixture of material objects;
(d) processing said measured measurements of electromagnetic radiation to produce electronic images of said irradiated material objects and presenting said electronic images for visual display;
(e) interactively selecting designating, from said visual display of material objects, selected material objects to be sorted from said mixture of material objects by using a computerized pointing device operable by a human operator; and
(f) using an automated device to separate separating said selected material objects from said mixture of material objects.
33. An apparatus for identifying and sorting selected material objects from a mixture of material objects comprising:
(a) a conveyor that conveys a mixture of material objects into and through an inspection zone;
(b) an electromagnetic radiation source that irradiates said mixture of material objects within said inspection zone with incident electromagnetic radiation;
(c) a sensor that examines said material objects within said inspection zone and measures electromagnetic radiation emanating from said material objects;
(d) a microprocessor computer system that processes said measured measurements of electromagnetic radiation to produce electronic images of said material objects as they pass through said inspection zone and a display coupled to said microprocessor computer system that presents a visual display of said electronic images; and
(e) a computerized pointing device; and
(e) (f) a human operator interface to said visual display such that an operator may select designate selected material objects to be sorted from said mixture of objects using the computerized pointing device; and
(f) an automated device that separates said selected material objects from said mixture of material objects .
2. The method according to
3. The method according to
4. The method according to
5. The method according to
6. The method according to
7. The method according to
8. The method according to
9. The method according to
10. The method according to
11. The method according to
12. The method according to
13. The method according to
14. The method according to
15. The method according to
16. The method according to
17. The method according to
18. The method according to
19. The method according to
20. The method according to
21. The method according to
22. The method according to
23. The method according to
24. The method according to
25. The method according to
26. The method according to
27. The method according to
28. The method according to
29. The method according to
30. The method according to
31. The method according to
32. The method according to
34. The apparatus according to
36. The apparatus according to
37. The apparatus according to
39. The apparatus according to
42. The apparatus according to
43. The apparatus according to
44. The apparatus according to
45. The apparatus according to
46. The apparatus according to
47. The apparatus according to
48. The apparatus according to
49. The apparatus according to
50. The apparatus according to
51. The apparatus according to
0. 52. The apparatus according to
0. 53. The apparatus according to
0. 54. The apparatus according to
0. 55. The method according to
0. 56. The apparatus according to
0. 57. The method according to
0. 58. The apparatus according to
0. 59. The method according to
|
This application is a 371 a PCT/US97/19680, filed Nov. 3, 1997 which is a continuation of provisional application 60/030,183 filed Nov. 4, 1996.
The present invention relates generally to a robotic sorting system, and, more particularly to a robotic sorting system suitable for separating recyclable or waste material. This invention was made with Government support under Contract No. DE-FG02-95ER82037, having an effective date of Sep. 1, 1995, a ward ed by the United States Department Of Energy. The Government has certain rights in the invention.
In the recycling of waste secondary materials it is very useful to be able to separate mixtures of materials into usable fractions. Such separations are sometimes performed by fully automated systems which use automated sensors for materials identification with subsequent automated extraction of selected materials from mixtures such as disclosed in U.S. Pat. Nos. 5,260,576 and 5,555,984. In many instances automated materials handling and sorting systems are combined with manual handsorting in a semi-automated process such as that disclosed in U.S. Pat. No. 5,411,147. In many other systems, particularly at smaller recycling facilities, most of the sorting is done by manual handsorting.
In a fully automated sorting system identification of materials to be extracted from a conveyed stream is performed by automated sensors which are specific to identification of certain materials. Advantages to these systems are often high speed and the lack of need for manual labor. A disadvantage to these systems is that automated sensors are generally limited in their ability to identify a wide range of materials and therefore have limited applicability to only selected sorting tasks. On the other hand, in manual handsorting the human visual sensory system is used to make identifications and advantageously is capable of efficiently identifying a wide range of various materials to be sorted. However, also in manual handsorting, human hands are used to make the sorting extractions from the waste stream with disadvantages of relatively low capacity compared to the speed of identifications that a human can perform, and secondly, handsorting requires that humans come into contact with a waste stream which in general is unsanitary and often contains hazardous objects such as broken glass and other objects which can easily puncture or cut.
Present sorting technologies for waste materials generally require the presence of an operator on the sorting floor or, in the case of manual handsorting, at the sorting line manually extracting recyclable materials from the waste stream. This generally requires that the equipment operator or manual handsorters be able-bodied. Therefore this type of work has generally not been available to the physically handicapped since the physical demands of the work would be beyond their capabilities or would place them in a dangerous environment because of their physical condition. This situation has lessened job opportunities for the physically handicapped.
Robotic systems have been applied in industry for a number of years typically to reduce human labor, reduce human presence in hazardous or potentially hazardous situations, and to replace humans in tedious repetitive tasks. In U.S. Pat. No. 5,299,693 a robotic system is described for extracting recyclable materials from a waste stream where the recyclable materials physically have a tag coupled to them which provides a non-visual identifying signal which can be received by a sensor for identification and with a robotic arm subsequently guided into the waste stream to retrieve the tagged item. However, the requirement for physical attachment of a signal generating tag to the items to be sorted limits the usefulness of the process and would require a massive change in practices by the packaging industry to provide such signal markers on packaging materials typically found in the waste stream. U.S. Pat. No. 5,411,147 discloses using robotic arms to extract materials from a waste stream although details of the robotic systems and interactive functions with humans are not provided. U.S. Pat. No. 4,942,538 discloses a teleoperated robotic tracker which is guided by a human operator using a joystick type hand controller with video feedback of robotic arm motion to the operator from a camera mounted on the robotic arm. The use of such a system for recyclables sorting would be awkward and slow since the human operator would need to provide continual guidance to the robotic arm throughout the whole sorting process. There are numerous other robotic systems with interactive human operation disclosed in the prior art. However, nowhere in the prior art have the inventors found descriptions of robotic systems for sorting materials where the identification of selected materials to be extracted from a waste stream or other conveyed stream of materials is provided by a human operator utilizing a computerized pointing device such as a touch screen for electronically registering the spatial coordinates of the selected materials with subsequent fully computerized control of a mechanical or other robotic system to acquire and extract the selected materials from the conveyed stream.
One of the objectives of the present invention is to alleviate one or more of the problems identified above. Another objective of the present invention is to provide a sorting technology which incorporates the sensing flexibility and sensing speed of a human being combined with high speed high capacity mechanical material extraction systems to provide a highly flexible capacity sorting system. A second objective is to provide a sorting technology which incorporates the sensing flexibility and sensing speed of a human being while insulating the human being from contact with the material stream. A third objective is to provide a sorting technology which can be operated by the physically handicapped. A fourth objective is to provide an automated sorting technology capable of being trained by a human operator in order to become fully automated.
These and other objectives are achieved by providing a method for identifying and sorting material objects from a mixture of material objects in which the mixture of material objects is conveyed through an inspection zone. The mixture of material objects is irradiated with incident electromagnetic radiation in the inspection zone. The electromagnetic radiation emanating from the irradiated material objects is measured and processed to produce electronic images suitable for visual display. Interactive selection of material objects is performed and the selected material objects are then separated from the mixture of material objects by an automated device.
Therefore, the herein disclosed invention overcomes the limitations discussed above in automated sorting systems, manual handsorting, and in robotic sorting systems by providing for rapid human identification and selection of objects to be sorted with subsequent rapid extraction of selected objects by fully computerized mechanical means. The invention further provides for use of the human capability for rapid and efficient identification and selection of materials to be sorted without subjecting the human operator to direct contact with the material stream and without requiring that the operator be present on the sorting floor. The invention further provides for human training of the sorting system so that the sorting system retains the flexibility of the human for identifying selected materials for sorting while being fully automated.
The disclosed invention classifies materials by utilizing a computerized touch screen or other computerized pointing device for operator identification and electronic marking of spatial coordinates of materials to be extracted from a mixture of materials with subsequent computerized position tracking and extraction of the marked materials by computer controlled mechanical means. More specifically, an operator positioned at a computerized touch screen views electronic images of the mixture of materials to be sorted as they are conveyed past a sensor array which transmits a sequence of images of the mixture to a touch screen either directly or through a computer. The operator views the touch screen images and manually “touches” objects displayed on the screen to be extracted from the mixture thereby registering the spatial coordinates of the objects within the computer. The computer then tracks the registered objects as they are further conveyed and directs mechanical means such as air jets, robotic arms, or other mechanical diverters to extract the registered objects from the mixture at an appropriate position downstream from the camera position. High speed communications between the touch screen, computer, and mechanical sorting equipment allows that the touch screen monitor can be located remote from the sorting environment such as in an air conditioned office or even at a remote location such as across town or in another locally altogether. Therefore there is no requirement that the operator be in contact with or even near the material stream.
The computer is also capable of “learning” the properties of those objects being selected for extraction by the operator and capable of taking over the selecting process after a sufficient learning process. At that time the system becomes a fully automated sorting system for extraction of those types of selected objects. At any time the computer can be retrained to select different or additional objects by the operator selecting the different or additional objects through the touch screen or other computerized pointing device.
The accompanying drawings, which are incorporated in and constitute a part of the specification, illustrate presently preferred embodiments of the invention, and, together with the general description given above and the detailed description of the preferred embodiments given below, serve to explain the principles of the invention.
A first preferred embodiment of the invention is shown in
A sensor array 4 (for example, Sony Series 9000 CCD video camera when radiation from sources 5 is in the visible light range) is positioned to view the inspection zone so to provide data arrays corresponding to measurements of electromagnetic radiation emanating from inspection zone 3 and from any materials 1 being conveyed through the inspection zone. The electromagnetic radiation emanating from materials 1 may be reflected radiation, radiation transmitted through materials 1, radiation emitted from materials 1 through fluorescence, or any other forms of radiation resulting from interaction of materials 1 with the incident radiation from sources 5. Sensor array 4 is shown positioned above conveying surface 2 although in practice it may be located at any position required to give the desired view. Sensor array 4 is selected to be sensitive in the wavelength range of electromagnetic radiation emanating from material objects within the inspection zone 3 when irradiated by sources 5 which may be a single source or multiple sources. The geometry of sensor array 4 is determined by the application. For instance the sensor array 4 may be a linear array of sensors or an area array of sensors. It may physically span the full width and/or length of the inspection zone 3 or it may be more compact and use optics to scan the width and/or length of the inspection zone such as with a CCD camera. Sensor array 4 may be positioned on the same side of the inspection zone 3 as is irradiation sources 5 or it may be positioned on the opposite side of inspection zone 3 from sources 5 or positioned at any other location with respect to irradiation sources 5 and inspection zone 3. The effective wavelength ranges of sensor array 4 and sources 5 may each be in any one of the microwave wavelength range, the ultraviolet wavelength range, the visible light wavelength range, the infrared wavelength range, the x-ray wavelength range, or the gamma ray wavelength range or any combination thereof. Sensor array 4 may be fitted with special filters which allow only certain wavelengths of electromagnetic radiation to reach sensor array 4 for measurement.
Sensor array 4 data corresponding to electromagnetic radiation measurements emanating from the inspection zone 3 and from material objects 1 within the inspection zone are transmitted from sensor array 4 to computer 7 and/or touch sensitive screen 9 over transmitting cables 6 or by wireless means. Control of sensor array 4 operation may also be provided by computer 7 over transmitting cables 6 or by wireless transmission. Sensor array 4 data received by computer 7 are processed for analog to digital conversion if not already digital by nature of sensor array 4 and microcomputer processed into digitized electronic images which are transmitted over cables 8 to touch sensitive screen 9 which is capable of electronically registering the coordinates on the screen of a manual touch by a human operator 26. Touch screen 9 displays the digitized images corresponding to the sensor data from sensor array 4 of the inspection zone 3 and the materials to be sorted within the inspection zone 3. Alternatively, the sensor array 4 can transmit the images directly to the touch sensitive screen 9.
Human operator 26 views the electronic images on touch screen 9 and manually touches the image of any material object, in this case object 16, which the operator wishes to be removed from the stream of materials 1. Spatial coordinates describing the location on the touch sensitive screen 9 of the touch by operator 26 are registered by touch sensitive screen 9 and transmitted over cables 8 to computer 7. Computer 7 associates the touch screen coordinates of the registered touch with corresponding spatial location coordinates on conveying surface 2 within inspection zone 3 and further associates any object on conveying surface 2 at that location (in this case object 16) with the touch. Computer 7 then electronically tracks the motion of selected object 16 as it is further conveyed along conveying surface 2 utilizing signals transmitted over cables 28 indicating conveyor speed generated by conveyor speed encoder 27. Similarly, the motion could be tracked by programming into computer 7 a predetermined speed for conveying surface 2 in which case conveyor speed encoder 27 would not be needed. As object 16 leaves the inspection zone 3 and drops off the discharge end of conveying surface 2 it is diverted from its falling trajectory by air blast 15 as shown in FIG. 2. Air blast 15 is generated by computer 7 by sending at the right time control signals over cables 10 (
Further details of the present invention are described in the applicants' report to the U.S. Department of Energy, Phase II application under Grant application number 35853-95-I, titled “Use of Computer Robotics to Reduce Human Contact with the Waste Stream and Lower the Costs for Recyclable Materials,” (hereafter “DOE Report”) and in provisional application 60/030,183 filed Nov. 4, 1996, the disclosures of which are incorporated herein, in their entireties.
Computer 7 may contain a pre-compiled pattern database or identification and pattern recognition algorithms which can perform learning of selections by operator 26 as the operator makes the selections. Such identification and pattern recognition algorithms may be accomplished by computerized neural networks or other such pattern recognition computer code. Identification by pattern recognition of the objects can be performed by using, for example, the edge enhancement and image contour extraction techniques disclosed in the DOE Report at pages 22-29. Further details of pattern recognition and its interaction with robotic systems is described in the published text titled “Robot Vision,” Berthold Klaus Paul Horn, MIT Press, Cambridge, Mass. (1991), the disclosure of which is incorporated herein, in its entirety.
Learning the recognized object patterns can be performed using known neural network or other pattern recognition and learning systems. Neural network techniques for identifying and learning patterns are described in, for example, the published text “Neural Networks for Pattern Recognition,” Christopher M. Bishop, Oxford University Press, New York (1995), (hereafter “Bishop”), the disclosure of which is incorporated herein, in its entirety. Bishop chapters 3-6 describe neural network techniques including single and multiple layer perception as well as different error functions that can be used for training neural networks. Bishop, chapters 9 and 10 describe techniques for learning and generalization in a neural network.
In this case operator 26 will initially make selections of items to be extracted from the mixture of materials such as object 16. As operator 26 makes selections the associated electronic images will be processed through the computer algorithms with the imaging patterns distinctive to the selected items noted by the algorithms. As similar items are repetitively selected by operator 26 the computer algorithms associate the distinctive properties of the imaging patterns with objects to be selected for extraction and begin to electronically select similar patterns for extraction without input from the human operator 26. In this way the computerized system learns those objects to be extracted and after sufficient learning experience can begin sorting without input from operator 26.
The choice of using a touch screen 9 for making the selection of objects to be extracted from the mixture is a matter of preference. Similar pointing devices interfaced to a display screen could be used as a computer mouse, a track ball, a joystick, a touch pad, a light pen, or other such device. The inventors have chosen the touch screen as the preferred pointing device based upon their intensive studies of some of these various types of devices for sorting applications.
The inventors have found that a human operator can comfortably view computerized images of a mixture of materials conveyed past a camera and identify and select items to be sorted out of the mixture at a rates up to 2.5 selections per second using a computerized touch screen to make the identifications and selections (see FIG. 4). This is two to five times as fast as industry established typical manual handsorting rates of one item every one to two seconds.
A sensor array 4 (for example, a Sony Series 9000 CCD video camera when radiation from sources 5 is in the visible light range) is positioned to view the inspection zone 3 so to provide data arrays corresponding to measurements of electromagnetic radiation emanating from inspection zone 3 and from any materials 1 being conveyed through the inspection zone. The electromagnetic radiation emanating from materials 1 may be reflected radiation, radiation transmitted through materials 1, radiation emitted from materials 1 through fluorescence, or any other forms of radiation resulting from interaction of materials 1 with the incident radiation from sources 5. Sensor array 4 is shown positioned above conveyor 2 although in practice it may be located at any position required to give the desired view. Sensor array 4 is selected to be sensitive in the wavelength range of electromagnetic radiation emanating from material objects within the inspection zone 3 when irradiated by sources 5 which may be a single source or multiple sources. The geometry of sensor array 4 is determined by the application. For instance the sensor array 4 may be a linear array of sensors or an area array of sensors. It may physically span the full width and/or length of the inspection zone 3 or it may be more compact and use optics to scan the width and/or length of the inspection zone such as with a CCD camera. Sensor array 4 may be positioned on the same side of the inspection zone 3 as is irradiation sources 5 or it may be positioned on the opposite side of inspection zone 3 from sources 5 or positioned at any other location with respect to irradiation sources 5 and inspection zone 3. The effective wavelength ranges of sensor array 4 and sources 5 may each be in any one of the microwave wavelength range, the ultraviolet wavelength range, visible light wavelength range, the infrared wavelength range, the x-ray wavelength range, or the gamma ray wavelength range or any combination thereof. Sensor array 4 may be fitted with special filters which allow only certain wavelengths of electromagnetic radiation to reach sensor array 4 for measurement.
Sensor array 4 data corresponding to electromagnetic radiation measurements emanating from the inspection zone 3 and from material objects 1 within the inspection zone are transmitted from sensor array 4 to computer 7 over transmitting cables 6 or by wireless transmission. Control of sensor array 4 operation may also be provided by computer 7 over transmitting cables 6. Sensor array 4 data received by computer 7 are processed for analog to digital conversion if not already digital by nature of sensor array 4 and microcomputer processed into electronic images which are transmitted over cables 8 to touch sensitive screen 9 which is capable of electronically registering the coordinates on the screen of a manual touch by a human operator 26. Touch screen 9 displays the images corresponding to the sensor data from sensor array 4 of the inspection zone 3 and the materials to be sorted within the inspection zone 3.
Human operator 26 views the images on touch screen 9 and manually touches the images of any material object, in this case object 25, which the operator wishes to be removed from the stream of materials 1. Spatial coordinates describing the location on the touch screen 9 of the touch by operator 26 are registered by touch screen 9 and transmitted over cables 8 to computer 7. Computer 7 associates the touch screen coordinates of the registered touch with corresponding spatial location coordinates on the surface of conveyor 2 within inspection zone 3 and further associates any object on conveyor 2 at that location (in this case object 25) with the touch. Computer 7 then electronically tracks the motion of selected object 25 as it is further conveyed along conveyor 2 utilizing signals transmitted over cables 28 indicating conveyor speed generated by conveyor speed encoder 27. Similarly, the motion could be tracked by programming into computer 7 a predetermined speed for conveyor 2 in which case conveyor speed encoder 27 would not be needed.
As the selected object 25 is conveyed into the vicinity of robotic arms 18 the computer 7 performs scheduling algorithms to determine which of the robotic arms 18 will be used to most efficiently extract the object 25 from the waste stream. Such scheduling algorithms are specific to the equipment used and materials to be sorted and can be suitably devised by one skilled in the art using, for example, the techniques discussed in “Schedule Efficiency in a Robotic Cell” by Irina Ioachim and Francois Soumis, International Journal of Flexible Manufacturing Systems, Vol. 7, No. 1 (March 1995). The entire contents of this publication is incorporated herein by reference. There may be only one robotic arm 18 or as many robotic arms 18 as required to efficiently sort materials as selected by operator 26 on touch screen 9. For instance if the average acquisition and retrieval time of a robotic arm 18 for the materials to be sorted 1 is two seconds and the operator 26 can make an average of 2.5 selections per second then a minimum of five robotic arms 18 would be required to keep up with operator 26 selections.
As the selected object 25 approaches, the robotic arm 18 chosen by the computer to make the extraction is commanded by computer 7 over cables 10 to rotate (shown as motion 24) into proper position for most efficient timing to make an interception of object 25. Vertical actuator 19 is positioned by computer 7 through control signals over cables 10 to intercept object 25 as it passes under robotic arm 18. Vertical actuator 19 is placed by computer 7 in retracted (lifted) position so the object 25 can pass under it. As object 25 passes under actuator 19 computer 7 signals the actuator via cables 10 to extend downward until end effector 20 (or 21) contacts object 25. Upon contact with object 25 end effector 20 (or 21) acquires object 25 at which time actuator 19 retracts (lifts) and extracts object 25 from the mixture of objects 1. When object 25 is sufficiently lifted so that its lower extent is higher than the top of the other objects 1 the computer 7 signals actuator 19 to translate along robotic arm 18 in one of the directions toward the edge of the conveyor belt 2 as indicted by directional arrows 22. Positioned along each side of conveyor 2 are receiver chutes 29 and 30. Actuator 19 holding object 25 via its end effector 20 (or 21) continues moving until it is above either chute 29 or chute 30 at which time end effector 20 (or 21) releases object 25 into chute 20 or chute 30 which segregates object 25 from the other objects 1 on the conveyor 2 therefore having sorted object 25 from the other objects 1. Actuator 19 then translates back to a position over the conveyor 2 and awaits commands from computer 7 to make another object acquisition and retrieval in response to touches by operator 26 on the touch screen 9.
Computer 7 may contain identification and pattern recognition algorithms which can perform learning of selections by operator 26 as the operator makes the selections. Such identification and pattern recognition algorithms may be accomplished by computerized neural networks or other such pattern recognition computer code, as discussed earlier herein. In this case operator 26 will initially make selections of items to be extracted from the mixture of materials such as object 25. As operator 26 makes selections the associated electronic images will be processed through the computer algorithms with the imaging patterns distinctive to the selected items noted by the algorithms. As similar items are repetitively selected by operator 26 the computer algorithms associate the distinctive properties of the imaging patterns with objects to be selected for extraction and begin to electronically select similar patterns for extraction without input from the human operator 26. In this way the computerized system learns those objects to be extracted and after sufficient learning experience can begin sorting without input from operator 26.
Other embodiments of the invention will be apparent to those skilled in the art from consideration of the specification and practice of the invention disclosed herein. It is intended that the specification be considered as exemplary only, with the true scope and spirit of the invention being indicated by the following claims.
Sommer, Jr., Edward J., Parrish, Robert H., Roos, Charles E., Russell, James R.
Patent | Priority | Assignee | Title |
Patent | Priority | Assignee | Title |
4805778, | Sep 21 1984 | Nambu Electric Co., Ltd. | Method and apparatus for the manipulation of products |
4942538, | Nov 07 1986 | Spar Aerospace Limited | Telerobotic tracker |
5044002, | Jul 14 1986 | HOLOGIC, INC , A MASSACHUSETTS CORP | Baggage inspection and the like |
5141110, | Feb 09 1990 | UNILOY MILACRON INC | Method for sorting plastic articles |
5260576, | Oct 29 1990 | National Recovery Technologies, Inc.; NATIONAL RECOVERY TECHNOLOGIES, INC , NASHVILLE, TN A CORP OF TN | Method and apparatus for the separation of materials using penetrating electromagnetic radiation |
5299693, | Apr 12 1991 | UBALDI, RICHARD A ; SMITH, GARRETT A ; HREHOVCIK, MARK W ; RAUEN, DOUGLAS P ; MARION, MICHAEL E | Method and apparatus for extracting selected materials |
5339962, | Oct 29 1990 | NATIONAL RECOVERY TECHNOLOGIES, INC A CORP OF DELAWARE | Method and apparatus for sorting materials using electromagnetic sensing |
5411147, | Jan 28 1993 | Dynamic landfill recycling system | |
5423431, | Mar 23 1989 | Sellsberg Engineering AB | Method and an apparatus for waste handling |
5520290, | Dec 30 1993 | Huron Valley Steel Corporation | Scrap sorting system |
5555984, | Jul 23 1993 | NATIONAL RECOVERY TECHNOLOGIES, INC | Automated glass and plastic refuse sorter |
5600700, | Sep 25 1995 | L-3 COMMUNICATIONS SECURITY AND DETECTION SYSTEMS INCORPORATION DELAWARE | Detecting explosives or other contraband by employing transmitted and scattered X-rays |
5699400, | May 08 1996 | L-3 Communications Security and Detection Systems Corporation Delaware | Operator console for article inspection systems |
5752607, | Mar 18 1996 | Moen Incorporated | Process for distinguishing plumbing parts by the coatings applied thereto |
6124560, | Nov 04 1996 | National Recovery Technologies, LLC | Teleoperated robotic sorting system |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Nov 03 1997 | National Recovery Technologies, Inc. | (assignment on the face of the patent) | / | |||
Apr 26 2012 | NATIONAL RECOVERY TECHNOLOGIES, INC | CALTIUS PARTNERS III, LP, AS AGENT | SECURITY AGREEMENT | 028130 | /0530 | |
Jun 28 2012 | NATIONAL RECOVERY TECHNOLOGIES, INC | National Recovery Technologies, LLC | CHANGE OF NAME SEE DOCUMENT FOR DETAILS | 034995 | /0458 | |
Mar 20 2018 | CALTIUS PARTNERS III, LP | National Recovery Technologies, LLC | RELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS | 045293 | /0583 | |
Mar 26 2018 | National Recovery Technologies, LLC | TRUE WEST CAPITAL PARTNERS FUND II, LP | SECURITY INTEREST SEE DOCUMENT FOR DETAILS | 045353 | /0387 | |
May 01 2019 | Emerging Acquisitions, LLC | PNC Bank, National Association | SECURITY INTEREST SEE DOCUMENT FOR DETAILS | 049513 | /0198 | |
May 01 2019 | National Recovery Technologies, LLC | PNC Bank, National Association | SECURITY INTEREST SEE DOCUMENT FOR DETAILS | 049513 | /0198 | |
May 01 2019 | NIHOT RECYCLING TECHNOLOGY B V | PNC Bank, National Association | SECURITY INTEREST SEE DOCUMENT FOR DETAILS | 049513 | /0198 | |
May 01 2019 | Zero Waste Energy, LLC | PNC Bank, National Association | SECURITY INTEREST SEE DOCUMENT FOR DETAILS | 049513 | /0198 |
Date | Maintenance Fee Events |
Nov 29 2011 | ASPN: Payor Number Assigned. |
Mar 22 2012 | M2553: Payment of Maintenance Fee, 12th Yr, Small Entity. |
Date | Maintenance Schedule |
Jun 24 2011 | 4 years fee payment window open |
Dec 24 2011 | 6 months grace period start (w surcharge) |
Jun 24 2012 | patent expiry (for year 4) |
Jun 24 2014 | 2 years to revive unintentionally abandoned end. (for year 4) |
Jun 24 2015 | 8 years fee payment window open |
Dec 24 2015 | 6 months grace period start (w surcharge) |
Jun 24 2016 | patent expiry (for year 8) |
Jun 24 2018 | 2 years to revive unintentionally abandoned end. (for year 8) |
Jun 24 2019 | 12 years fee payment window open |
Dec 24 2019 | 6 months grace period start (w surcharge) |
Jun 24 2020 | patent expiry (for year 12) |
Jun 24 2022 | 2 years to revive unintentionally abandoned end. (for year 12) |