A wireless substrate-like sensor is provided to facilitate alignment and calibration of semiconductor processing systems. The wireless substrate-like sensor includes an optical image acquisition system that acquires one or more images of targets placed within the semiconductor processing system. Analysis of images of the targets obtained by the wireless substrate-like sensor provides position and/or orientation information in at least three degrees of freedom. An additional target is affixed to a known location within the semiconductor processing system such that imaging the reference position with the wireless substrate-like sensor allows the measurement and compensation for pick-up errors.

Patent
   7456977
Priority
Feb 06 2002
Filed
Mar 15 2006
Issued
Nov 25 2008
Expiry
Jan 31 2023
Assg.orig
Entity
Small
12
148
all paid
1. A method for determining position and orientation of a flat target in three dimensions, the method comprising:
providing at least four reference target indicia on a reference target and fixing the reference target's position;
removably coupling an image acquisition system to a robot such that the image acquisition system is movable by the robot;
digitizing an image of the reference target with the image acquisition system;
identifying the at least four reference target indicia in the image of the reference target;
providing at least four station target indicia on a station target;
digitizing an image of the station target with the image acquisition system;
identifying the at least four station target indicia in the image of the station target;
calculating the relative position and orientation between the station target and the robot based upon identification of the at least four reference target indicia and the at least four station target indicia.
17. A method for determining position and orientation of a station component in three dimensions, the method comprising:
providing a reference target and fixing the reference target's position with respect to a robot;
providing a station target having at least four indicia on a station component;
removably coupling an image acquisition system to the robot such that the image acquisition system is movable by the robot;
digitizing an image of the reference target with the image acquisition system;
calculating the relative position and orientation between the image acquisition system and the robot based on the digitized image of the reference target;
digitizing an image of the station target with the image acquisition system;
identifying the at least four indicia in the image of the station target; and
calculating the relative position and orientation between the station component and the robot as a function of the identification of the at least four indicia in the image of the station target and as a function of the calculated relative position between the image acquisition system and the robot.
15. A method for determining position and orientation of a flat target in three dimensions, the method comprising:
providing at least four reference target indicia on a reference target and fixing the reference target's position;
removably coupling acquisition system to a robot such that the image acquisition system is movable by the robot;
digitizing an image of the reference target with the image acquisition system;
identifying the at least four reference target indicia in the image of the reference target;
providing at least four station target indicia on a station target;
digitizing an image of the station target with the image acquisition a system;
identifying the at least four station tar et indicia in the image of the station target;
calculating the relative position and orientation between the station target and the robot based upon identification of the at least four station target indicia in the image of the station target; and
calculating the relative position and orientation between the image acquisition system and the robot based upon identification of the at least four reference target indicia in the image of the reference target.
2. The method of claim 1, wherein the station target indicia are shaped circularly.
3. The method of claim 1, wherein the robot is a material handling robot.
4. The method of claim 1, wherein the image acquisition system is embodied within a substrate-like sensor.
5. The method of claim 1, wherein the station target is fixedly attached relative to a processing tool.
6. The method of claim 1 wherein digitizing the image of the reference target with the image acquisition system comprises moving the image acquisition system with the robot to a position proximate the reference target, and wherein digitizing the image of the station target with the image acquisition system comprises moving the image acquisition system with the robot to a position proximate the station target.
7. The method of claim 1, wherein the station target is fixedly attached relative to a station component.
8. The method of claim 1, and further comprising decoupling the image acquisition system from the robot.
9. The method of claim 1, wherein the image acquisition system includes an area array device.
10. The method of claim 1, wherein the at least four station target indicia have high contrast with respect to a background of the station target.
11. The method of claim 9, wherein the area array device is a Charge Coupled Device (CCD).
12. The method of claim 9, wherein the area array device is a Complementary Metal Oxide Semiconductor (CMOS) image device.
13. The method of claim 10, wherein the at least four station target indicia are dark, and the background of the station target is light.
14. The method of claim 10, wherein the at least four station target indicia are raised above a surface of the station target.
16. The method of claim 15, wherein calculating the relative position and orientation between the station target and the robot comprises:
calculating the relative position and orientation between the station target and the robot as a function of the calculated relative position between the image acquisition system and the robot.
18. The method of claim 17, wherein the image acquisition system is embedded within a substrate-like sensor.
19. The method of claim 17, and further comprising decoupling the image acquisition system from the robot.
20. The method of claim 17, wherein calculating the relative position and orientation between the image acquisition system and the robot comprises:
identifying at least four reference indicia in the image of the reference target; and
calculating the relative position and orientation between the image acquisition system and the robot based upon identification of the at least four reference target indicia in the image of the reference target.

The present application is a divisional of and claims priority of U.S. patent application Ser. No. 10/356,684, filed Jan. 31, 2003, the content of which is hereby incorporated by reference in its entirety, which application claims priority to previously filed co-pending provisional application Ser. No. 60/354,551, filed Feb. 6, 2002, entitled WAFER-LIKE SENSOR, which application is incorporated herein by reference in its entirety.

Semiconductor processing systems are characterized by extremely clean environments and extremely precise semiconductor wafer movement. Industries place extensive reliance upon high-precision robotic systems to move substrates, such as semiconductor wafers, about the various processing stations within a semiconductor processing system with the requisite precision.

Reliable and efficient operation of such robotic systems depends on precise positioning, alignment, and/or parallelism of the components. Accurate wafer location minimizes the chance that a wafer may accidentally scrape against the walls of a wafer processing system. Accurate wafer location on a process pedestal in a process chamber may be required in order to optimize the yield of that process. Precise parallelism between surfaces within the semiconductor processing systems is important to ensure that minimal substrate sliding or movement during transfer from a robotic end effector to wafer carrier shelves, pre-aligner vacuum chucks, load lock elevator shelves, process chamber transfer pins and/or pedestals. When a wafer slides against a support, particles may be scraped off that cause yield loss. Misplaced or misaligned components, even on the scale of fractions of a millimeter, can impact the cooperation of the various components within the semiconductor processing system, causing reduced product yield and/or quality.

This precise positioning must be achieved in initial manufacture, and must be maintained during system use. Component positioning can be altered because of normal wear, or as a result of procedures for maintenance, repair, alteration, or replacement. Accordingly, it becomes very important to automatically measure and compensate for relatively minute positional variations in the various components of a semiconductor processing system.

In the past, attempts have been made to provide substrate-like sensors in the form of a substrate, such as a wafer, which can be moved through the semiconductor processing system to wirelessly convey information such as substrate inclination and acceleration within the semiconductor system. As used herein, “substrate-like” is intended to mean a sensor in the form of substrate such as a semiconductor wafer, a Liquid Crystal Display glass panel or reticle. Attempts have been made to provide wireless substrate-like sensors that include additional types of detectors to allow the substrate-like sensor to measure a host of internal conditions within the processing environment of the semiconductor processing system. Wireless substrate-like sensors enable measurements to be made at various points throughout the processing equipment with reduced disruption of the internal environment as well as reduced disturbance of the substrate handling mechanisms and fabrication processes (e.g.: baking, etching, physical vapor deposition, chemical vapor deposition, coating, rinsing, drying etc.). For example, the wireless substrate-like sensor does not require that a vacuum chamber be vented or pumped down; nor does it pose any higher contamination risk to an ultra-clean environment than is suffered during actual processing. The wireless substrate-like sensor form factor enables measurements of process conditions with minimal observational uncertainty.

A dire need currently exists for systems that offer the benefits of wireless substrate-like sensors while facilitating the acquisition of and compensation for information related to positional variations of components within a semiconductor processing system. Although wireless substrate-like sensors currently provide limited information such as inclination and acceleration, they do not provide the required positional information. Technicians must still make subjective judgments to adjust the relative positions of the various components within the semiconductor processing system in order to ensure that such components cooperate to provide extremely careful substrate processing. Currently available sensors do not enable automatic adjustment of positional offsets between components of a semiconductor processing system.

A wireless substrate-like sensor is provided to facilitate alignment and calibration of semiconductor processing systems. The wireless substrate-like sensor includes an optical image acquisition system that acquires one or more images of targets or objects within the semiconductor processing system. Analysis of images of the targets obtained by the wireless substrate-like sensor provides useful information such as position, presence/absence, value and/or orientation in at least three degrees of freedom. An additional target can be affixed to a known location within the semiconductor processing system such that analyzing the reference position image with the wireless substrate-like sensor allows the measurement and compensation for pickup induced positional errors.

FIG. 1 is a diagrammatic view of a semiconductor wafer process environment.

FIG. 2 is a top perspective view of a wireless substrate-like sensor in accordance with embodiments of the present invention.

FIG. 3 is a bottom view of a wireless substrate-like sensor in accordance with embodiments of the present invention.

FIG. 4 is a diagrammatic view of central portion 120 in accordance with embodiments of the present invention.

FIG. 5 is a top perspective view of a holster for maintaining a wireless substrate-like sensor in accordance with embodiments of the present invention.

FIG. 6 is a top plan view of a target for use with embodiments of the present invention.

FIG. 7 is a diagrammatic view of a vector transformation in accordance with embodiments of the present invention.

While aspects the prior art have provided wireless substrate-like semiconductor sensors, the information provided by such sensors has been limited. To significantly facilitate semiconductor processing system alignment and calibration requires substantially more functionality than has been heretofore provided by wireless substrate-like sensors. Specifically, no wireless substrate-like sensors have provided information allowing calculation of very precise positions and orientations of components within the semiconductor processing system. This feature as well as many others will be apparent upon reading the discussion below.

FIG. 1 is a diagrammatic view of a semiconductor wafer processing environment including a wafer container 100, robot 102 and system component station 104 illustrated diagrammatically as simply a box. Wafer container 100 is illustrated containing three wafers 106, 108, 110 and wireless substrate-like sensor 112 in accordance with embodiments of the present invention. As is apparent from FIG. 1, sensor 112 is preferably embodied in a form factor allowing it to be moveable within the semiconductor wafer processing environment in the same manner as wafers themselves. Accordingly, embodiments of the present invention provide a substrate-like wireless sensor having a height low enough to permit the substrate-like sensor to move through the system as if it were a substrate such as a wafer. For example, a height of less than about 9.0 mm is believed to be acceptable. Preferably, the sensor has a weight of less than two wafers, for example, a weight of less than about 250 grams is believed to be acceptable. A stand-off distance of about 25 mm is believed to meet the requirements of most applications; however some applications may require a different stand-off. As used herein “stand-off” is the nominal distance from the bottom of the sensor to the target. The diameter of the sensor preferably matches one of the standard semiconductor wafer diameters, such as, 300 mm, 200 mm or 150 mm.

Sensor 112 is preferably constructed from dimensionally stable materials. In order for the substrate-like sensor to accurately measure a three-dimensional offset, it is important for the sensor to deform in a manner similar to that of an actual substrate. Common wafer dimensions and characteristics may be found in the following specification: SEMI M1-0302, “Specification for Polished Monocrystaline Silicon Wafers”, Semiconductor Equipment and Materials International, www.semi.org. The center of a 300 mm silicon wafer supported at its edges will sag approximately 0.5 mm under its own weight. The difference in the deformation of the sensor and the deformation of an actual wafer should be much less than the accuracy of sensor measurement. In a preferred embodiment, the stiffness of the substrate-like sensor results in a deflection that is nearly identical to that of an actual silicon wafer. Therefore, no compensation is required to correct for any differential deflection. Alternatively, a compensation factor may be added to the measurement. Similarly, the weight of the substrate-like sensor will also deflect its support. Substrate supports include, but are not limited to: end effectors, pedestals, transfer pins, shelves, etc. The differential support deflection will be a function both of the difference in weights of the sensor and a substrate as well as the mechanical stiffness of the substrate support. The difference between deflection of the support by the sensor and that by a substrate should also be much less than the accuracy of sensor measurement, or the deflection difference should be compensated by a suitable calculation.

In the prior art, technicians have iteratively adjusted the alignment of a vacuum transfer robot end effector with a process chamber pedestal by viewing them after removing the lid of the process chamber or through a transparent window in the lid. Sometimes a snuggly fitting fixture or jig must first be placed on the process pedestal to provide a suitable reference mark. The substrate-like sensor enables an improved, technician assisted, alignment method. The substrate-like sensor provides an image of the objects being aligned without the step of removing the cover and with greater clarity than viewing through a window. The wireless substrate-like sensor saves significant time and improves the repeatability of alignment.

A wireless substrate-like sensor can transmit an analog camera image by radio.

A preferred embodiment uses a machine vision sub-system of a substrate-like wireless sensor to transmit all or a portion of the digital image stored in its memory to an external system for display or analysis. The external system can also be configured to store a number of such digital images. The display may be located near the receiver or the image data may be relayed through a data network for remote display. In a preferred embodiment, the camera image is transmitted encoded as a digital data stream to minimize degradation of image quality caused by communication channel noise. The digital image may be compressed using any of the well known data reduction methods in order to minimize the required data rate. The data rate may also be significantly reduced by transmitting only those portions of the image that have changed from the previous image. The substrate-like sensor or the display may overlay an electronic cross hair or other suitable mark to assist the technician with evaluating the alignment quality.

While vision-assisted teaching is more convenient than manual methods, technician judgment still affects the repeatability and reproducibility of alignment. The image acquired by a substrate-like wireless sensor camera may be analyzed using many well-known methods, including two-dimensional normalized correlation, to measure the offset of a pattern from its expected location. The pattern may be an arbitrary portion of an image that the vision system is trained to recognize. The pattern may be recorded by the system. The pattern may be mathematically described to the system. The mathematically described pattern may be fixed at time of manufacture or programmed at the point of use. Conventional two-dimensional normalized correlation is sensitive to changes in the pattern image size. When a simple lens system is used, magnification varies in proportion to object distance. Enhanced pattern offset measurement performance may be obtained by iteratively scaling either the image or the reference. The scale that results in the best correlation indicates the magnification, provided the size of the pattern is known, or the magnification, as used when the reference pattern was recorded, is known.

When the correspondence between pixels in the image plane to the size of pixels in the object plane is known, offsets may be reported in standard units of measure that are easier for technicians or machine controllers to interpret than arbitrary units such as pixels. For example, the offset may be provided in terms of millimeters such that the operator can simply adjust the systems by the reported amount. The computations required to obtain the offset in standard units may be performed manually, by an external computer, or preferentially within the sensor itself. When the sensor extracts the required information from an image, the minimum amount of information is transmitted and the minimum computational burden is placed on the technician or external controller. In this way objective criteria may be used to improve the repeatability and reproducibility of the alignment. Automated offset measurement improves the reproducibility of alignment by removing variation due to technician judgment.

During alignment and calibration of semiconductor processing equipment, it is not only important to correctly position an end effector relative to a second substrate supporting structure, it is also important to ensure that both substrate supporting structures are parallel to one another. In a preferred embodiment, a machine vision subsystem of a wireless substrate-like sensor is used to measure the three dimensional relationship between two substrate supports. For example: a robotic end effector may hold a wireless substrate-like sensor in close proximity to the transfer position and a measurement of the three dimensional offset with six degrees of freedom may be made from the sensor camera to a pattern located on an opposing substrate support. One set of six degrees of freedom includes yaw, pitch, and roll as well as displacement along the x, y, and z axes of the Cartesian coordinate system. However, those skilled in the art will appreciate that other coordinate systems may be used without departing from the spirit and scope of the invention. Simultaneous measurement of both parallelism and Cartesian offset allows a technician or a controller to objectively determine satisfactory alignment. When a controller is used, alignments that do not require technician intervention may be fully automated. Automated alignments may be incorporated into scheduled preventive maintenance routines that optimize system performance and availability.

In a very general sense, operation and automatic calibration of robotic system 102 is performed by instructing robot 102 to select and convey sensor 112 to reference target 114. Once instructed, robot 102 suitably actuates the various links to slide end effector 116 under sensor 112 to thereby remove sensor 112 from container 100. Once removed, robot 102 moves sensor 112 directly over reference target 114 to allow an optical image acquisition system (not shown in FIG. 1) within sensor 112 to obtain an image of reference target 114. Based upon a-priori knowledge of the target pattern, a three dimensional offset between the sensor and target 114 is measured. The measurement computation may occur within the sensor or an external computer. Based upon a-priori knowledge of the precise position and orientation of reference target 114, the three dimensional offset thereof can be analyzed to determine the pick-up error generated by robot 102 picking up sensor 112. Either internal or external computation allows the system to compensate for any error introduced by the pick-up process of sensor 112.

This information allows sensor 112 to be used to acquire images of additional targets, such as target 117 on system component 104 to calculate a precise position and orientation of system component 104. Repeating this process allows the controller of robot 102 to precisely map exact positions of all components within a semiconductor processing system. This mapping preferably generates location and orientation information in at least three and preferably six degrees of freedom (x, y, z, yaw, pitch and roll). The mapping information can be used by a technician to mechanically adjust the six degree of freedom location and orientation of any component with respect to that of any other component. Accurate measurements provided by the substrate-like wireless sensor are preferably used to minimize or reduce variability due to technician judgment. Preferably, this location information is reported to a robot or system controller which automates the calibration process. After all mechanical adjustments are complete; the substrate-like sensor may be used to measure the remaining alignment error. The six degrees of freedom offset measurement may be used to adjust the coordinates of points stored in the memories of the robot and/or system controllers. Such points include, but are not limited to: the position of an atmospheric substrate handling robot when an end effector is located at a FOUP slot #1 substrate transfer point; the position of an atmospheric substrate handling robot when an end effector is located at a FOUP slot #25 substrate transfer point; the position of an atmospheric substrate handling robot when an end effector is located at a substrate pre-aligner substrate transfer point; the position of an atmospheric substrate handling robot when an end effector is located at a load lock substrate transfer point; the position of an atmospheric substrate handling robot when an end effector is located at a reference target attached to the frame of an atmospheric substrate handling system; the position of a vacuum transfer robot when its end effector is located at a load lock substrate transfer point; the position of a vacuum transfer robot when an end effector is located at a process chamber substrate transfer point; and the position of a vacuum transfer robot when an end effector is located at a target attached to the frame of a vacuum transfer system.

An alternative embodiment of the present invention stores and reports the measurements. Real-time wireless communication may he impractical in some semiconductor processing systems. The structure of the system may interfere with wireless communication. Wireless communication energy may interfere with correct operation of a substrate processing system. In these cases, sensor 112 can preferably record values as it is conveyed to various targets, for later transmission to a host. When sensor 112, using its image acquisition system, or other suitable detectors, recognizes that it is no longer moving, sensor 112 preferably records the time and the value of the offset. At a later time, when sensor 112 is returned to its holster (shown in FIG. 5) sensor 112 can recall the stored times and values and transmit such information to the host. Such transmission may be accomplished by electrical conduction, optical signaling, inductive coupling or any other suitable means. Store and report operation of the wireless substrate-like sensor potentially increases the reliability, lowers the cost and shortens a regulatory approval cycle for the system. Moreover, it avoids any possibility that the RF energy could interact with sensitive equipment in the neighborhood of the sensor and its holster. Store and report operation can also be used to overcome temporary interruptions of a real-time wireless communication channel.

FIG. 2 is a top perspective view of a wireless substrate-like sensor 118 in accordance with embodiments of the present invention. Sensor 118 differs from sensor 112 illustrated in FIG. 1 solely in regard to the manner in which weight reduction is effected. Specifically, sensor 112 employs a number of struts 119 to suspend a central sensor portion 120 within an outer periphery 122 that can accommodate standard wafer sizes, such as 300 millimeter diameter wafers. In contrast, sensor 118 employs a number of though-holes 124 which also provide weight reduction to sensor 118. Other patterns of holes may be used to accomplish the necessary weight reduction. Additional weight reduction designs are also contemplated including, for example, portions of the sensor that are hollow, and/or portions that are filled with light-weight materials. Both sensor 112 and sensor 118 employ central region 120. A portion of the underside of central portion 120 is disposed directly over an access hole 126 as illustrated in FIG. 3. Access hole 126 allows illuminator 128 and image acquisition system 130 to acquire images of targets disposed below sensor 118 as sensor 118 is moved by robot 102.

FIG. 4 is a diagrammatic view of portion 120 in accordance with embodiments of the present invention. Portion 120 preferably includes a circuit board 140 upon which a number of components are mounted. Specifically, battery 142 is preferably mounted on circuit board 140 and coupled to digital signal processor (DSP) 144 via power management module 146. Power management module 146 ensures that proper voltage levels are provided to digital signal processor 144. Preferably, power management module 146 is a power management integrated circuit available from Texas Instrument under the trade designation TPS5602. Additionally, digital signal processor 144 is preferably a microprocessor available from Texas Instruments under the trade designation TMS320C6211. Digital signal processor 144 is coupled to memory module 148 which can take the form of any type of memory. Preferably, however, memory 148 includes a module of Synchronous Dynamic Random Access Memory (SDRAM) preferably having a size of 16M×16. Module 148 also preferably includes flash memory having a size of 256K×8. Flash memory is useful for storing such non-volatile data as programs, calibration data and/or additional other non-changing data as may be required. The random access memory is useful for storing volatile data such as acquired images or data relevant to program operation.

Illumination module 150, which preferably comprises a number of Light Emitting Diodes (LEDs), and image acquisition system 152 are coupled to digital signal processor 144 through camera controller 154. Camera controller 154 facilitates image acquisition and illumination thus providing relevant signaling to the LEDs and image acquisition system 152 as instructed by digital signal processor 144. Image acquisition system 152 preferably comprises an area array device such as a Charge Coupled Device (CCD) or Complementary Metal Oxide Semiconductor (CMOS) image device coupled preferably to an optical system 156, which focuses images upon the array. Preferably, the image acquisition device is available from Kodak under the trade designation KAC-0310. Digital signal processor 144 also preferably includes a number of I/O ports 158, 160. These ports are preferably serial ports that facilitate communication between digital signal processor 144 and additional devices. Specifically, serial port 158 is coupled to radio-frequency module 162 such that data sent through port 158 is coupled with external devices via radio frequency module 162. In one preferred embodiment, radio frequency module 162 operates in accordance with the well-known Bluetooth standard, Bluetooth Core Specification Version 1.1 (Feb. 22, 2001), available from the Bluetooth SIG (www.bluetooth.com). One example of module 162 is available from Mitsumi under the trade designation WML-C11.

Detectors 164 may take any suitable form and provide relevant information regarding any additional conditions within a semiconductor processing system. Such detectors can include one or more thermometers, accelerometers, inclinometers, compasses (Magnetic field direction detectors), light detectors, pressure detectors, electric field strength detectors, magnetic field strength detectors, acidity detectors, acoustic detectors, humidity detectors, chemical moiety activity detectors, or any other types of detector as may be appropriate.

FIG. 5 illustrates an optional holster 180 which can be used to store and maintain a wireless substrate-like sensor when such sensor is not in use. Holster 180 provides a convenient way to recharge the internal power storage device of the wireless sensor. Preferably, holster 180 includes suitable contacts to electrically couple to the wireless substrate-like sensor to thereby recharge the power source within the sensor. Such coupling can occur via any suitable methods including: inductive, photovoltaic, capacitive, and conductive methods.

FIG. 6 is a diagrammatic view of a target useful with embodiments of the present invention. Important features of target 190 are that a visual target is provided having a known size and geometry such that an image thereof can be processed to calculate x, y, z positions, as well as yaw, pitch and roll. Such six-degree positional calculations have heretofore not been accommodated with wireless substrate-like sensors. Preferably, target 190 has a size of 50 mm×50 mm and includes four circular marks 192 which have a known size, geometry and positional relationship with respect to one another. Careful imaging and processing of target 190 can allow a system to calculate a vector to transform positions from that of the image acquisition system (wireless substrate-like sensor) to the target (semiconductor processing system component or reference marker).

For example, suppose the exact position of a flat surface in three dimensions must be found from a two-dimensional image of the surface taken by a camera. The position of the surface can be described by three vectors illustrated in FIG. 7 as follows: A and B are two vectors in the plane of the surface, which together describe the orientation of the surface. These can be thought of as axes of a local coordinate system on the surface. C is a vector from the image acquisition system to a reference point on the surface, which describes the position of the surface. (In fact C measures from a specific point inside the lens of the image acquisition system; the exact position of this point depends on the design of the image acquisition system.) If the surface has some markings on it, such as indicia 192, a point in the marking pattern can be described in the surface coordinates by two numbers (u, v). The position of this point in three-dimensional space is then described by the vector equation:
P=C+u·A+v·B  EQ. 1
The position (x, y) in the camera's image of this point is determined by the perspective transformation:
x=k·Px/Pz; and y=k·Py/Pz
where k is a constant related to the field of view of the image acquisition system.

The relationship between the position of the mark on the surface and the position of the mark on the image can be obtained by combining these equations:
x·(Cz+u·Az+v·Bz)=k·(Cx+u·Ax+v·Bx); and
y·(Cz+u·Az+v·Bz)=k·(Cy+u·Ay+v·By).

If a known pattern is used, u and v for each mark are known constants. Also, x and y for each mark can be measured from the image, and k can be determined by calibrating the camera.

One method of calibrating the camera is to image a mark at a known position relative to the camera, (Px, Py, Pz). If (x, y) is the position of the mark in the camera image, the camera magnification can be computed by either
k=x*Pz/Px,
or
k=y*Pz/Py.

More accurate values for k can be determined if necessary by making several measurements and using statistical techniques.

If a pattern with four marks, as illustrated in FIG. 6, is used, this results in a system of eight linear equations which can then be solved for these nine unknowns Cx, Cy,Cz,Ax,Ay,Az,Bx,By and Bz.

Once these nine values are known, the position and orientation in space of the surface can be computed. Because there are only eight equations and nine unknowns, one more constraint must be applied to find a unique solution. The lack of uniqueness exists because the same image will result from changing the system by any scaling factor—a large target will look exactly same to the image acquisition system as a small target close up. This can be seen in the equations by noting that multiplying all three vectors A, B, C by a constant does not change these equations. This means that the final constraint cannot be added by simply using five marks to get an additional three linear equations. Instead, a constraint on the size of the system should be used. The easiest constraint to chose is a constraint such as the absolute value |A|=1 which requires that the units used to measure u and v are the same as the units used on the vectors A, B, C.

The solution to these eight linear equations and one non-linear equation can be found for any particular pattern of markings (except for a few special cases such as all four marks in a straight line). The results can then be used in combination with simple image processing techniques to create a computer program which automatically calculates three-dimensional position and orientation of the surface from video images.

In summary, calculation of the position and orientation of target 190 is done by choosing a target with four easily found marks which can be attached to a surface of interest. Then, the method described above is employed and the chosen positions of the marks are used to create a system of eight equations in nine variables. Then, the system is solved to obtain expressions for eight of the nine components of the position vectors in terms of the ninth. For example, solve for A, B, and the x and y components of C in terms of Cz. The following steps are performed by the sensor each time it performs a measurement:

1) digitize an image of the target on the surface;

2) use standard image processing techniques such as blob analysis to find the reference marks in the image;

3) use the expressions described above to obtain eight components of the position vectors assuming the ninth is 1.0;

4) compute the length of A and divide all components by this length to produce correct values for the vectors A, B, C; and

5) optionally convert the results for the orientation vectors A and B to rotation angles and add an offset to C such that the position is reported relative to some reference point other than the lens of the image acquisition system. The method described above is an illustrative solution only, and it is contemplated that other approaches to finding the position of a surface using a two-dimensional image can be used without departing from the spirit and scope of the invention.

Although the present invention has been described with reference to preferred embodiments, workers skilled in the art will recognize that changes may be made in form and detail without departing from the spirit and scope of the invention. Although embodiments of the present invention have been described with respect to acquiring optical images of calibration targets and processing such images to ascertain position and orientation information in at least three degrees of freedom, additional optical features can be provided. For example, the wireless substrate-like sensor in some embodiments, is adapted to recognize characters and/or barcodes.

Ramsey, Craig C., Lassahn, Jeffrey K., Huntzinger, Greg, Gardner, DelRae H.

Patent Priority Assignee Title
10058996, Nov 18 2014 PERSIMMON TECHNOLOGIES, CORP Robot adaptive placement system with end-effector position estimation
10509052, Feb 06 2017 Lam Research Corporation Smart vibration wafer with optional integration with semiconductor processing tool
10537997, May 13 2016 Applied Materials, Inc Sensor based auto-calibration wafer
10580681, Jul 10 2016 YASKAWA AMERICA INC Robotic apparatus and method for transport of a workpiece
10741433, Nov 29 2017 TAIWAN SEMICONDUCTOR MANUFACTURING CO , LTD Systems and methods for wafer pod alignment
11469128, Nov 29 2017 Taiwan Semiconductor Manufacturing Co., Ltd. Systems and methods for wafer pod alignment
11676845, Jun 30 2020 BROOKS AUTOMATION HOLDING, LLC; Brooks Automation US, LLC Automated teach apparatus for robotic systems and method therefor
7629184, Mar 20 2007 Tokyo Electron Limited RFID temperature sensing wafer, system and method
7855549, Dec 03 2002 KLA-Tencor Corporation Integrated process condition sensing wafer and data analysis system
8823933, Sep 29 2006 CyberOptics Corporation Substrate-like particle sensor
9099508, Apr 09 2009 AES MOTOMATION GMBH Method for automatic measurement and for teaching-in of location positions of objects within a substrate processing system by means of sensor carriers and associated sensor carrier
9987747, May 24 2016 SEMES CO., LTD. Stocker for receiving cassettes and method of teaching a stocker robot disposed therein
Patent Priority Assignee Title
1668328,
3815020,
3835264,
3876833,
4033053, Jan 14 1976 Digital segment display to braille converter
4074114, Mar 12 1976 Monarch Marking Systems, Inc. Bar code and method and apparatus for interpreting the same
4119381, Dec 17 1976 CLINICAL DIAGNOSTIC SYSTEMS INC Incubator and radiometric scanner
4180199, Feb 27 1978 ELECTROVERT U S A CORP Mass soldering control system
4633578, Dec 01 1983 Miniature thermal fluid flow sensors and batch methods of making same
4701096, Mar 05 1986 BRUCE TECHNOLOGIES, INC Wafer handling station
4753569, Dec 28 1982 LMI TECHNOLOGIES INC Robot calibration
4791482, Feb 06 1987 Westinghouse Electric Corp. Object locating system
4810996, Oct 28 1986 Patient communication and diagnostic device
4843287, Dec 31 1987 Westinghouse Electric Corp. Path contriving system for look-ahead sensor in a robotic control system
4880384, Feb 03 1989 MURPHY, KEVIN C ; MURPHY, JANICE L Braille teaching apparatus
4891030, Apr 28 1988 Superior Toy & Manufacturing Company, Inc. Toy with lighted playpieces
4918627, Aug 04 1986 FMC Corporation Computer integrated gaging system
5076794, Apr 29 1991 HEWLETT-PACKARD DEVELOPMENT COMPANY, L P Space-saving mounting interconnection between electrical components and a printed circuit board
5175601, Oct 15 1991 OPTON CO , LTD High-speed 3-D surface measurement surface inspection and reverse-CAD system
5232331, Aug 07 1987 Canon Kabushiki Kaisha Automatic article feeding system
5248553, Mar 16 1989 TOYO INK MANUFACTURING CO , LTD Coated molded article
5265957, Aug 11 1992 Texas Instruments Incorporated Wireless temperature calibration device and method
5267143, Oct 12 1984 Sensor Adaptive Machines, Incorporated Vision assisted fixture construction
5276609, Nov 20 1989 3-D amusement and display device
5298363, Jun 17 1991 Eastman Kodak Company Photolithographically patterned fluorescent coating
5301248, Nov 09 1987 Hitachi, Ltd. Method for pattern inspection and apparatus therefor
5321989, Feb 12 1990 ACACIA PATENT ACQUISITION LLC Integratable capacitative pressure sensor and process for its manufacture
5371728, Mar 07 1992 Canon Kabushiki Kaisha Information recording/reproducing apparatus using probe
5382911, Mar 29 1993 International Business Machines Corporation Reaction chamber interelectrode gap monitoring by capacitance measurement
5393706, Jan 07 1993 Texas Instruments Incorporated; Texas Instrument Incorporated Integrated partial sawing process
5435682, Oct 15 1987 ASM America, Inc Chemical vapor desposition system
5442297, Jun 30 1994 SEMILAB SEMICONDUCTOR PHYSICS LABORATORY RT Contactless sheet resistance measurement method and apparatus
5444637, Sep 28 1993 GLOBALFOUNDRIES Inc Programmable semiconductor wafer for sensing, recording and retrieving fabrication process conditions to which the wafer is exposed
5521123, Apr 17 1992 Terumo Kabushiki Kaisha Infrared sensor and method for production thereof
5573728, Jun 04 1992 Gaz De France Device for selective detection of gas
5619027, May 04 1995 Intermec IP Corporation Single width bar code symbology with full character set utilizing robust start/stop characters and error detection scheme
5642293, Jun 03 1996 CamSys, Inc. Method and apparatus for determining surface profile and/or surface strain
5675396, Nov 30 1993 Matsushita Electric Industrial Co., Ltd. Liquid crystal display unit having grounding frame
5680384, Nov 17 1992 Seiko Epson Corporation Laser emission unit, optical head and optical memory device
5721677, Oct 12 1984 Sensor Adaptive Machines, Inc. Vision assisted fixture construction
5726066, Mar 10 1994 LG Electronics Inc. Method for manufacturing an infrared sensor array
5783341, May 25 1994 Canon Kabushiki Kaisha Alignment for layer formation through determination of target values for translation, rotation and magnification
5784282, Jun 11 1993 Bertin & Cie; Sollac Method and apparatus for identifying the position in three dimensions of a movable object such as a sensor or a tool carried by a robot
5786704, Apr 13 1995 Mirae Corporation Metallic tray unit for testing a semiconductor device
5805289, Jul 07 1997 General Electric Company Portable measurement system using image and point measurement devices
5839215, Dec 27 1996 Raised indicia labels
5854880, Oct 12 1984 Sensor Adaptive Machines, Inc. Target based determination of robot and sensor alignment
5956417, Feb 16 1982 LMI TECHNOLOGIES INC Robot vision using target holes, corners and other object features
5962909, Sep 12 1996 Institut National D'Optique Microstructure suspended by a microsupport
5969639, Jul 28 1997 UT-Battelle, LLC Temperature measuring device
5973788, Oct 12 1995 Metronor ASA System for point-by-point measuring of spatial coordinates
5981116, Dec 12 1996 Nikon Corporation Alignment in a projection exposure method
6010009, Oct 13 1995 Empak, Inc. Shipping and transport cassette with kinematic coupling
6011294, Apr 08 1996 OmniVision Technologies, Inc Low cost CCD packaging
6013236, Oct 03 1996 Bridgestone Corporation Wafer
6022811, Dec 28 1990 Mitsubishi Denki Kabushiki Kaisha Method of uniform CVD
6075909, Jun 26 1998 Bell Semiconductor, LLC Optical monitoring system for III-V wafer processing
6106457, Apr 04 1997 Welch Allyn, Inc Compact imaging instrument system
6129278, May 19 1994 Symbol Technologies, LLC Detecting image cell position with subpixel accuracy
6175124, Jun 30 1998 Bell Semiconductor, LLC Method and apparatus for a wafer level system
6184771, May 25 1998 Kabushiki Kaisha Toshiba Sintered body having non-linear resistance characteristics
6206441, Aug 03 1999 Taiwan Semiconductor Manufacturing Company, Ltd. Apparatus and method for transferring wafers by robot
6232615, Mar 31 1998 ASML NETHERLANDS B V Lithographic projection apparatus with improved substrate holder
6244121, Mar 06 1998 Applied Materials, Inc. Sensor device for non-intrusive diagnosis of a semiconductor processing system
6275742, Apr 16 1999 MOOG INC Wafer aligner system
6300974, Feb 28 1997 Commissariat a l'Energie Atomique Process and device for locating an object in space
6323952, Dec 02 1998 Nikon Corporation Flatness measuring apparatus
6325536, Jul 10 1998 SensArray Corporation Integrated wafer temperature sensors
6326228, Mar 25 1996 SHENZHEN XINGUODU TECHNOLOGY CO , LTD Sensor and method of fabrication
6389158, Jul 22 1996 Metronor AS System and method for determining spatial coordinates
6465281, Sep 08 2000 SHENZHEN XINGUODU TECHNOLOGY CO , LTD Method of manufacturing a semiconductor wafer level package
6466325, Jul 08 1999 BAE SYSTEMS PLC Method and apparatus for calibrating positions of a plurality of first light sources on a first part
6468816, Mar 06 1998 Applied Materials, Inc. Method for sensing conditions within a substrate processing system
6476825, May 13 1999 American Patents LLC Hand-held video viewer and remote control device
6480537, Feb 25 1999 HANGER SOLUTIONS, LLC Active techniques for video transmission and playback
6526668, Mar 11 1999 MicroTool, Inc. Electronic level
6532403, Apr 21 2000 MicroTool, Inc Robot alignment system and method
6535650, Jul 21 1998 Intel Corporation Creating high resolution images
6607951, Jun 26 2001 United Microelectronics Corp. Method for fabricating a CMOS image sensor
6625305, Aug 16 1999 HEWLETT-PACKARD DEVELOPMENT COMPANY, L P Image demosaicing method
6628803, Nov 25 1998 PENTAX Corporation Device for calculating positional data of standard points of photogrammetric target
6681151, Dec 15 2000 Cognex Technology and Investment LLC System and method for servoing robots based upon workpieces with fiducial marks using machine vision
6691068, Aug 22 2000 Regents of the University of California Methods and apparatus for obtaining data for process operation, optimization, monitoring, and control
6724930, Feb 04 1999 Olympus Corporation Three-dimensional position and orientation sensing system
6734027, Mar 14 2001 ASM INTERNATIONAL N V Inspection system for process devices for treating substrates, sensor intended for such inspection system, and method for inspecting process devices
6801257, Jan 12 2001 COGNITENS LTD Optical three-dimensional digital imaging and mensuration system for industrial applications
6816755, May 24 2002 ROBOTICVISIONTECH, INC Method and apparatus for single camera 3D vision guided robotics
6852975, Apr 07 2000 Riegl Laser Measurement Systems, GmbH Method for the recording of an object space
6891276, Jan 09 2002 Bridge Semiconductor Corporation Semiconductor package device
6925356, Apr 19 1999 Applied Materials, Inc. Method and apparatus for aligning a cassette
6958768, Oct 20 2000 ASTI Holdings Limited; ASTI, INC CMOS inspection apparatus
6966235, Oct 06 2000 Remote monitoring of critical parameters for calibration of manufacturing equipment and facilities
6985169, Feb 09 1998 Lenovo PC International Image capture system for mobile communications
6990215, Jul 31 2000 GEODETIC SERVICES, INC Photogrammetric measurement system and method
7035913, Sep 28 2001 HEWLETT-PACKARD DEVELOPMENT COMPANY L P System for collection and distribution of calendar information
7059936, Mar 23 2004 CMC MATERIALS LLC Low surface energy CMP pad
7135852, Dec 03 2002 SensArray Corporation Integrated process condition sensing wafer and data analysis system
7149643, Dec 03 2002 SensArray Corporation Integrated process condition sensing wafer and data analysis system
7158857, Apr 19 1999 Applied Materials, Inc. Method and apparatus for aligning a cassette
7180607, Nov 15 2002 Leica Geosystems AG Method and device for calibrating a measuring system
7206080, Jul 30 2001 TOPCON CORPORATION Surface shape measurement apparatus, surface shape measurement method, surface state graphic apparatus
7222789, Oct 17 1997 Hand Held Products, Inc. Bar code reading device having image processing mode
20010034222,
20010050769,
20020006675,
20020006687,
20020028629,
20020078770,
20020092369,
20020101508,
20020148307,
20030001083,
20030112448,
20030127589,
20030160883,
20030209097,
20030223057,
20040158426,
20040202362,
20050017712,
D344302, Nov 24 1992 LEGO A S Element for a toy buidling set
D478494, Sep 20 2002 Replaceable, trimmable grip with pressure sensitive adhesive for tool handle
D490276, Feb 21 2003 Transparent cake shield having candle apertures and handles
DE101065774,
EP583007,
EP1150187,
EP1184805,
EP1253471,
JP1082823,
JP11260706,
JP11307606,
JP163340,
JP2004276151,
JP3214783,
JP6076193,
JP6163340,
JP62054108,
JP7074229,
JP7280644,
JP8233855,
WO12263,
WO70495,
WO165317,
WO188976,
WO217364,
WO229385,
WO247115,
WO3060989,
/
Executed onAssignorAssigneeConveyanceFrameReelDoc
Mar 15 2006CyberOptics Semiconductor, Inc.(assignment on the face of the patent)
Date Maintenance Fee Events
Dec 14 2011M2551: Payment of Maintenance Fee, 4th Yr, Small Entity.
May 19 2016M2552: Payment of Maintenance Fee, 8th Yr, Small Entity.
Jan 30 2020M2553: Payment of Maintenance Fee, 12th Yr, Small Entity.


Date Maintenance Schedule
Nov 25 20114 years fee payment window open
May 25 20126 months grace period start (w surcharge)
Nov 25 2012patent expiry (for year 4)
Nov 25 20142 years to revive unintentionally abandoned end. (for year 4)
Nov 25 20158 years fee payment window open
May 25 20166 months grace period start (w surcharge)
Nov 25 2016patent expiry (for year 8)
Nov 25 20182 years to revive unintentionally abandoned end. (for year 8)
Nov 25 201912 years fee payment window open
May 25 20206 months grace period start (w surcharge)
Nov 25 2020patent expiry (for year 12)
Nov 25 20222 years to revive unintentionally abandoned end. (for year 12)