This specification describes trajectory planning for robotic devices. A robotic navigation system can obtain, for each of multiple time steps, data representing an environment of a robot at the time step. The system generates a series of occupancy maps for the multiple time steps, and uses the series of occupancy maps to determine occupancy predictions for one or more future time steps. Each occupancy prediction can identify predicted locations of obstacles in the environment of the robot at a different one of the future time steps. A planned trajectory can be determined for the robot using the occupancy predictions, and the robot initiates travel along the planned trajectory.
|
1. A computer-implemented method, comprising:
obtaining, for each of a plurality of time steps, data representing an environment of a robot at the time step;
generating a series of occupancy maps that each identify locations of one or more obstacles in the environment of the robot at a different one of the plurality of time steps,
wherein generating the occupancy map for a particular time step of the plurality of time steps comprises selectively adding to the occupancy map a representation of a first obstacle detected in the data representing the environment of the robot for the particular time step, but not adding to the occupancy map a representation of a second obstacle detected in the data representing the environment of the robot for the particular time step,
wherein the representation of the first obstacle is selected to be added to the occupancy map based on the first obstacle being classified as capable of interfering with travel of the robot, and the representation of the second obstacle is selected not to be added to the occupancy map based on the second obstacle being classified as incapable of interfering with travel of the robot,
wherein the first obstacle that is added to the occupancy map is a dynamic obstacle that moves around the environment concurrently with the robot;
generating, using the series of occupancy maps, occupancy predictions for one or more future time steps that follow the plurality of time steps, each occupancy prediction identifying predicted locations of one or more obstacles in the environment of the robot at a different one of the future time steps, wherein the one or more obstacles for which occupancy predictions are generated includes the first obstacle;
determining a planned trajectory for the robot using the occupancy predictions; and
initiating travel of the robot along the planned trajectory.
12. One or more non-transitory computer-readable media having instructions stored thereon that, when executed by one or more processors, cause the one or more processors to perform operations comprising:
obtaining, for each of a plurality of time steps, data representing an environment of a robot at the time step;
generating a series of occupancy maps that each identify locations of one or more obstacles in the environment of the robot at a different one of the plurality of time steps,
wherein generating the occupancy map for a particular time step of the plurality of time steps comprises selectively adding to the occupancy map a representation of a first obstacle detected in the data representing the environment of the robot for the particular time step, but not adding to the occupancy map a representation of a second obstacle detected in the data representing the environment of the robot for the particular time step,
wherein the representation of the first obstacle is selected to be added to the occupancy map based on the first obstacle being classified as capable of interfering with travel of the robot, and the representation of the second obstacle is selected not to be added to the occupancy map based on the second obstacle being classified as incapable of interfering with travel of the robot,
wherein the first obstacle that is added to the occupancy map is a dynamic obstacle that moves around the environment concurrently with the robot;
generating, using the series of occupancy maps, occupancy predictions for one or more future time steps that follow the plurality of time steps, each occupancy prediction identifying predicted locations of one or more obstacles in the environment of the robot at a different one of the future time steps, wherein the one or more obstacles for which occupancy predictions are generated includes the first obstacle;
determining a planned trajectory for the robot using the occupancy predictions; and
initiating travel of the robot along the planned trajectory.
18. A system, comprising:
one or more processors; and
one or more computer-readable media having instructions stored thereon that, when executed by the one or more processors, cause the one or more processors to perform operations comprising:
obtaining, for each of a plurality of time steps, data representing an environment of a robot at the time step;
generating a series of occupancy maps that each identify locations of one or more obstacles in the environment of the robot at a different one of the plurality of time steps,
wherein generating the occupancy map for a particular time step of the plurality of time steps comprises selectively adding to the occupancy map a representation of a first obstacle detected in the data representing the environment of the robot for the particular time step, but not adding to the occupancy map a representation of a second obstacle detected in the data representing the environment of the robot for the particular time step,
wherein the representation of the first obstacle is selected to be added to the occupancy map based on the first obstacle being classified as capable of interfering with travel of the robot, and the representation of the second obstacle is selected not to be added to the occupancy map based on the second obstacle being classified as incapable of interfering with travel of the robot,
wherein the first obstacle that is added to the occupancy map is a dynamic obstacle that moves around the environment concurrently with the robot;
generating, using the series of occupancy maps, occupancy predictions for one or more future time steps that follow the plurality of time steps, each occupancy prediction identifying predicted locations of one or more obstacles in the environment of the robot at a different one of the future time steps, wherein the one or more obstacles for which occupancy predictions are generated includes the first obstacle;
determining a planned trajectory for the robot using the occupancy predictions; and
initiating travel of the robot along the planned trajectory.
2. The computer-implemented method of
3. The computer-implemented method of
4. The computer-implemented method of
the occupancy map for the particular time step comprises a two-dimensional (2D) grid of occupancy cells, each occupancy cell corresponding to a different portion of the environment of the robot; and
the occupancy map for the particular time step indicates, for each occupancy cell in the 2D grid of occupancy cells, whether an obstacle classified as capable of interfering with travel of the robot has been identified as occupying the portion of the environment of the robot that corresponds to the occupancy cell.
5. The computer-implemented method of
6. The computer-implemented method of
7. The computer-implemented method of
8. The computer-implemented method of
9. The computer-implemented method of
determining, based on the occupancy predictions, respective travel costs of the robot traveling to various locations in the environment of the robot; and
generating the planned trajectory by optimizing a function that is based on the respective travel costs of the robot traveling to the various locations in the environment of the robot.
10. The computer-implemented method of
identifying an obstacle that is represented in at least one occupancy map of the series of occupancy maps;
identifying a motion model that corresponds to the identified obstacle that is represented in the at least one occupancy map, wherein the motion model describes motion characteristics of the identified obstacle or motion characteristics of a class of obstacles that corresponds to the identified obstacle; and
applying the motion model to predict locations of the identified obstacle at the one or more future time steps.
11. The computer-implemented method of
the method comprises generating a predicted occupancy map for the first future time step, the predicted occupancy map comprising a two-dimensional (2D) grid of occupancy cells, each occupancy cell corresponding to a different portion of the environment of the robot and indicating a probability that any obstacle will occupy the portion of the environment that corresponds to the occupancy cell at the first future time step.
13. The computer-readable media of
14. The computer-readable media of
15. The computer-readable media of
the occupancy map for the particular time step comprises a two-dimensional (2D) grid of occupancy cells, each occupancy cell corresponding to a different portion of the environment of the robot; and
the occupancy map for the particular time step indicates, for each occupancy cell in the 2D grid of occupancy cells, whether an obstacle classified as capable of interfering with travel of the robot has been identified as occupying the portion of the environment of the robot that corresponds to the occupancy cell.
16. The computer-readable media of
17. The computer-readable media of
19. The computer-implemented method of
|
This specification generally relates to autonomous navigation, such as for mobile robotic devices.
Robots have facilitated automation of tasks in various settings from factories to offices, homes, and other environments. Some robots are autonomous, and are capable of performing tasks with little or no human supervision. Some robots are both mobile and autonomous, and are capable of navigating in an environment autonomously. For example, a mobile autonomous robot may travel in designated areas of an environment to clean, pick up and re-arrange items, or perform other specified tasks. Robots may also include sensors that allow it to sense information about its environment, which can then be processed to map the environment and locate the robot therein.
This specification discloses systems, methods, devices, and other techniques for facilitating navigation of an autonomous device, such as a robot. An important task of autonomous devices is to avoid obstacles as the device moves around an environment, and to minimize risks of collision. The techniques disclosed herein are aimed at planning paths for an autonomous robot that allow the robot to travel to a goal location while maneuvering around obstacles, thereby avoiding collisions and respecting social considerations that, for example, may impose wider boundaries around objects than are required by their physical presence. In some implementations, the autonomous navigational system includes a predictive model that is operable to predict locations of dynamic obstacles that move in an environment concurrently with the robot's travel in the environment. By predicting the locations of moving objects, the system can plan a path for the robot that accounts for these movements and minimizes the likelihood of the robot's travel interfering with the movement of the other objects in the environment.
Some implementations of the subject matter described herein include a computer-implemented method. The method can include obtaining, for each of multiple time steps, data representing an environment of a robot at the time step. A series of occupancy maps can be generated for the multiple time steps, each occupancy map identifying locations of obstacles in the environment of the robot at a different one of the multiple time steps. Using the series of occupancy maps, occupancy predictions can be generated for one or more future time steps that follow the multiple time steps, each occupancy prediction identifying predicted locations of obstacles in the environment of the robot at a different one of the future time steps. A planned trajectory can be determined for the robot using the occupancy predictions, and the robot can initiate travel of the robot along the planned trajectory.
These and other implementations can optionally include one or more of the following features.
For each of the multiple time steps, the data representing the environment of the robot at the time step can be a three-dimensional (3D) representation of the environment based on measurements by one or more sensors of the robot.
The one or more sensors of the robot can include at least one of a light detection and ranging (LIDAR) sensor, a camera, a sound detection and ranging (SONAR) sensor, or a radio detection and ranging (RADAR) sensor.
For each of the multiple time steps: the corresponding occupancy map for the time step can include a two-dimensional (2D) grid of occupancy cells, each occupancy cell corresponding to a different portion of the environment of the robot; and the corresponding occupancy map for the time step can indicate, for each occupancy cell in the 2D grid of occupancy cells, whether any obstacle has been identified as occupying the portion of the environment of the robot that corresponds to the occupancy cell.
For each of the multiple time steps, generating the occupancy map for the time step can include analyzing the data representing the environment of the robot at the time step to distinguish between first objects in the environment of the robot that are not capable of interfering with travel of the robot and second objects in the environment of the robot that are capable of interfering with the travel of the robot. Additionally, for each of the multiple time steps and for each second object in the environment of the robot that is identified as being capable of interfering with the travel of the robot, one or more occupancy cells in the 2D grid of occupancy cells of the occupancy map can be designated that corresponds to the portion of the environment where the second object is located as occupied by an obstacle.
Generating the occupancy predictions can include generating a first occupancy prediction for a first future time step that follows the multiple time steps by processing the series of occupancy maps with an obstacle location predictive model to generate the first occupancy prediction. The first occupancy prediction can identify predicted locations of obstacles in the environment of the robot at the first future time step. Generating the occupancy predictions can further include generating additional occupancy predictions for future time steps after the first future time step by iteratively processing with the obstacle location predictive model, for each particular future time step after the first future time step, at least a preceding occupancy map that is based on the occupancy prediction from the time step that precedes the particular future time step to generate an occupancy prediction for the particular future time step.
The obstacle location predictive model can be a deep neural network. The deep neural network can include one or more convolutional layers and one or more recurrent layers.
Determining the planned trajectory for the robot can include: determining, based on the occupancy predictions, respective travel costs of the robot traveling to various locations in the environment of the robot; and generating the planned trajectory by optimizing a function that is based on the respective travel costs of the robot traveling to the various locations in the environment of the robot.
Generating the occupancy predictions can include identifying an obstacle that is represented in at least one occupancy map of the series of occupancy maps, identifying a motion model that corresponds to the identified obstacle that is represented in the at least one occupancy map, and applying the motion model to predict locations of the identified obstacle at the one or more future time steps. The motion model can describe motion characteristics of the identified obstacle or motion characteristics of a class of obstacles that corresponds to the identified obstacle.
The occupancy predictions can be predicted occupancy maps. For each of the one or more future time steps, the method can include generating a predicted occupancy map for the future time step. The predicted occupancy map can include a two-dimensional (2D) grid of occupancy cells, and each occupancy cell can correspond to a different portion of the environment of the robot and indicates a probability that an obstacle will occupy the portion of the environment that corresponds to the occupancy cell at the future time step.
Some implementations of the subject matter described herein include one or more computer-readable media having instructions stored thereon that, when executed by one or more processors, are operable to cause the one or more processors to perform any of the computer-implemented methods described herein. The computer-readable media may be non-transitory, and can be part of a computing system that includes the one or more processors in one or more computers. The computers can be provided in a single physical system, or may be distributed across multiple locations.
Some implementations of the subject matter described in this specification can, in certain instances, realize one or more of the following advantages. A path (also referred to as a “trajectory”) for an autonomous robot can be planned that protects against the risk of collisions with obstacles in the environment. Additionally, a navigation system for the robot can identify obstacles in the environment and predict how they will move over an upcoming period of time. The navigation system can then plan a path for the robot that avoids the predicted locations of the moving obstacles. The system predicts locations of obstacles using a model that can be trained based on actual observations of obstacles in various environments, without requiring the training data to be manually labeled. Moreover, the model can be continuously updated during operation of a robot based on comparison of predicted obstacle locations and movements to actual obstacle locations and movements.
The details of one or more embodiments are set forth in the drawings and the detailed description below. Additional features and advantages will be apparent to those of ordinary skill in the field to which the invention pertains.
Like references and numerals in the various drawings indicate like elements.
This specification describes systems, methods, devices, and other techniques for planning paths (also referred to herein as ‘trajectories’) for a self-navigating device, such as an autonomous robot, to travel in an environment. In general, autonomous robots are capable of moving about an environment in which they are located without explicit human control. However, to navigate effectively, the robot must be capable of avoiding collisions with obstacles in its environment. Accordingly, the robot may be equipped with sensors that allow it to detect obstacles and maneuver around them. Some obstacles are static, in that they are fixed in place as the robot moves about the environment. Other obstacles are dynamic, in that they can move about the environment even as the robot travels. The techniques described herein may allow the robot to detect obstacles and travel around them. Moreover, for dynamic obstacles, the robot may predict their trajectories and/or locations at future times, and plan a path for the robot to travel that is unlikely to intersect with either the locations of static obstacles in the environment or the predicted trajectories of the dynamic obstacles.
Referring to
To facilitate navigation through the environment 100, the robot 102 can include a planning engine, or can access a remote planning engine, that determines how the robot 112 should travel toward a goal location 107. Planning the robot's travel can broadly occur in two phases: global planning and local planning. Global planning refers to the process of determining a navigable route from a starting or current location of the robot to the goal location 107, i.e., the destination where the robot 102 is tasked to travel. Local planning, in contrast, involves determining precise maneuvers for the robot to travel along the route set by the global planner. By way of example, a global planner may determine a route through a home environment that would take the robot through a living room and a hallway to reach a goal location in a kitchen. The local planner may then execute on this route by constantly determining short paths or maneuvers that allow the robot to travel along a segment of the route while navigating around obstacles in the immediate vicinity of the robot. For instance, while traveling down a hallway, the robot may detect obstacles that require the robot to veer toward the left or right side of the hallway for avoidance. The local planner may repeatedly plan short paths during travel for an upcoming segment of the route that allow the robot to maneuver around such obstacles and ultimately reach the goal location of the route.
In
Turning to
The robot 202 is an autonomous robot that is capable of moving about an environment and performing tasks within the environment. Being autonomous, the robot may respond to human requests and perform actions on behalf of humans, but its navigation is performed substantially independent of human or operator control.
The robot 202 can include one or more sensor subsystems 204. The sensor subsystems 204 allow the robot 202 to collect information about the environment where it is located. This information can be processed and analyzed for various purposes, including to identify objects in the environment and to navigate about the environment. In some implementations, the robot 202 includes a camera system 208. The camera system 208 captures images and/or video of the robot's environment. The camera system 208 can be adapted to capture images and/or video with depth information that indicates distances of objects from the robot in the environment. For example, the camera system 208 may include a stereoscopic camera. In some implementations, the robot 202 includes a LIDAR system 206 that uses light detection and ranging techniques to determine distances of objects from the robot in the environment. In some implementations, the robot 202 includes location sensors 210 such as GPS that allow the robot 202 to ascertain its geographic location and, in some examples, its micro-position within the environment (e.g., a particular location within a residence or office environment). The sensor subsystems 204 generate sensor data 212 characterizing signals captured by all or some of the sensor subsystems 204.
The mapping engine 214 is operable to receive sensor data 212 from the sensor subsystems 204 and to generate a virtual representation of at least a portion of the robot's environment from the sensor data 212.
In some implementations, the mapping engine 214 creates a three-dimensional (3D) representation of the environment based on the sensor data 212. For example, the mapping engine 214 may receive the sensor data 212 that includes measurements from one or more sensors on the robot 202, and may process the measurements to determine where surfaces of obstacles and other objects are located in the environment. The surfaces may be represented using, for example, polygonal meshes, point clouds, point splatting, or any other form of 3D representation. In some implementations, the 3D virtual representation of the environment may represent spaces within the property, instead of surfaces. For example, the virtual representation may include a collection of cells that each represent a particular 3D space or volume of the environment. A suitable resolution for the 3D representation may be selected depending on system requirements. For example, surfaces with more contours may be represented with a higher resolution, e.g., a small polygon size, than surfaces with fewer contours.
In some implementations, the mapping engine 214 is configured to generate a 2D representation of the robot's environment. The 2D representation may be derived from the 3D representation of the environment, or the 2D representation may be generated directly from the sensor data 212 without the intermediate step of creating a 3D representation of the environment. In some implementations, the 2D representation of the environment is an occupancy map. An occupancy map is a map of at least a portion of the environment of the robot 202 that identifies locations of obstacles in the at least the portion of the environment represented by the occupancy map. In some forms, the occupancy map defines a 2D grid of cells, where each cell represents a different portion (e.g., location) of the environment of the robot 202. For each cell, the occupancy map indicates whether an obstacle is located in the corresponding portion of the physical environment represented by the cell. In some implementations, a binary occupancy value (e.g., 1 or 0) is assigned to each cell indicating whether an obstacle is located in the corresponding portion of the physical environment represented by the cell. In other implementations, a floating point occupancy value can be assigned to each cell indicating a probability that an obstacle is located in the corresponding portion of the physical environment represented by the cell. A suitable size and resolution of the occupancy map can be determined based on system requirements. For example, the portion of the environment represented by the occupancy map may extend in one or more directions from the robot 202 (e.g., to the front, back, left, and right of the robot 202) as far as the sensing range of the sensor subsystems 204 permit, or the portion of the environment represented by the occupancy map may extend in one or more directions from the robot 202 by a fixed distance that is less than the full sensing range of the sensor subsystems 204. Generally, the occupancy map represents a portion of the environment 202 that extends at least as far as the local planner 232 may generate a path for the robot 202, which may be a function of the speed of the robot 202 and the frequency with which the local planner 232 updates the path for the robot 202. In some implementations, the occupancy map has relatively fine resolution. For example, each cell in the 2D grid of the occupancy map may represent between a 2×2 millimeter and a 1×1 millimeter region of the physical environment of the robot 202. In some implementations, the occupancy map is a two-dimensional grid having between 100 and 500 rows and columns, and each cell is less than 10×10 mm. In some implementations, the occupancy map has 400 rows, 400 columns, and each cell corresponds to a 0.05 meter×0.05 meter region of the environment, with the robot located at the center of the map.
Referring again to
In some implementations, obstacle location prediction data 226 indicates, for each obstacle represented in an occupancy map 216 at a current or recent time, a probability distribution of the object being located at one or more locations in the environment at one or more future times. For example, the human 104 shown in
The obstacle location prediction engine 218 can include one or more sub-components, including object recognizer 220 and predictive model 222. The predictive model 222 is configured to process one or more occupancy maps 216, and optionally one or more motion models from motion model library 224, to generate the obstacle location prediction data 226. The predictive model 222 can be a machine-learning model such as a deep neural network. Neural networks are machine-learning models that employ one or more layers of nonlinear units to predict an output for a received input. Some neural networks include one or more hidden layers in addition to an output layer. The output of each hidden layer is used as input to the next layer in the network, e.g., the next hidden layer or the output layer. Each layer of the network generates an output from a received input in accordance with current values of a respective set of parameters. In some implementations, the predictive model 222 includes one or more convolutional layers followed by one or more recurrent layers. In some implementations, the recurrent layers are followed by transpose convolutional layers mirroring the convolutional layers at the input.
In some implementations, the obstacle location prediction engine 218 receives a series of occupancy maps 216 from the mapping engine 214. Each occupancy map 216 indicates locations of obstacles in the robot's environment at a different one of a series of time steps (e.g., a current time step and one or more preceding time steps). The occupancy maps 216 can be based on observations of the robot's environment at times corresponding to the series of time steps. For example, the robot 202 and sensor subsystems 204 may periodically provide updated sensor data 212 to the mapping engine 214 at each passing time step. The mapping engine 214 may then process each iteration of sensor data 212 to generate an occupancy map 216 for a next time step. The obstacle location prediction engine 218 processes occupancy maps 216 from a set of the most recent time steps (e.g., a predetermined number n of the most recent time steps) to generate obstacle location predictions for one or more future time steps. The obstacle location prediction for each future time step can be, for example, predicted occupancy maps indicating likelihoods of various locations in the robot's environment being occupied by an obstacle at that future time step. For example, the obstacle location prediction data 226 may specify obstacle location prediction maps indicating predicted likelihoods of obstacle locations 50 ms from the current time step, 100 ms from the current time step, 150 ms from the current time step, and so on until a last predicted time step (e.g., 2 seconds from the current time step). The interval between time steps can be, for example, in the range of 5-milliseconds to 5 seconds, and can preferably be in the range 50-500 milliseconds. A suitable interval can be selected based on system requirements for particular implementations. In some implementations, the intervals between occupancy maps 216 for current and past locations of obstacles may be different than the intervals between the future time steps at which predicted obstacle locations are generated. In other implementations, the intervals between past, current, and future time steps is the same.
In some implementations, the predictive model 222 includes recurrent layers that maintain a hidden state that is updated each time a new occupancy map 216 is processed. As such, the obstacle location prediction for a given future time step can be based on not just the single most recent occupancy map 216, but is also partly a function of occupancy maps 216 from one or more preceding time steps. By analyzing occupancy maps 216 across a series of time steps, the predictive model 222 may detect, for example, an object's past speed and direction to help predict the object's future trajectory and location at one or more future time steps by tracking its motion across the environment through the series of occupancy maps 216.
In some implementations, the predictive model 222 may process preceding obstacle location predictions (e.g., predicted occupancy maps from preceding time steps) to generate obstacle location predictions for time steps further in the future. For example, the predictive model may initially process occupancy maps 216 from time steps n−10 through n, where n is the current time or most recent time step, one at a time and in sequence (thereby updating the hidden state of the recurrent portion of the model 222) to generate a predicted occupancy map for future time step n+1. The predicted occupancy map for future time step n+1 may then be provided as input to the predictive model 222 to generate the next predicted occupancy map for future time step n+2. This feedback loop may continue for a pre-defined number of time steps, or until some other terminating condition is met. In some implementations, the predicted occupancy maps from preceding time steps are modified before being fed back as input to the predictive model 222. The modifications can include, for example, converting granular probability distributions that were initially specified by the predicted occupancy maps to binary occupancy predictions that indicate only the most likely future location of each obstacle in the environment. In other implementations, the modifications can include comparing the probability value for each occupancy cell in the prediction to a threshold value and assigning a binary occupancy prediction value to the cell based on whether the probability value meets the threshold. In other implementations, no modifications may be required to provide the output of the predictive model 222 as a preceding time step as input to the model 222 at a next time step. No modifications may be required because the prediction model 222 is capable of processing granular probability distributions from the preceding predicted occupancy map(s), or the model 222 may already be configured to output predicted occupancy map(s) in the same format as occupancy maps 216 (e.g., with only binary occupancy values assigned to each cell indicating whether or not the respective locations in the environment corresponding to cells in the map are predicted to be occupied by an obstacle at a future time step associated with the map).
In some implementations, the predictive model 222 is further configured to use motion models associated with dynamic obstacles in the environment of the robot 202 to generate obstacle location predictions (e.g., predicted occupancy maps for one or more future time steps). The obstacle location predictions can be provided in obstacle location prediction data 226. In general, a motion model is a model that describes motion (e.g., travel characteristics) of a particular obstacle or type of obstacle. The motion models are stored in motion model library 224, which is a data repository provided on storage devices accessible to one or more computers of the system 200. The library 224 may include respective motion models for any number of obstacles that the robot 202 may encounter in an environment. For example, different motion models may be defined for humans, bicycles, automotive vehicles, pets, and the like. In some implementations, the library 224 maintains motion models only for dynamic obstacles that are actually capable of or likely to move about an environment. In other implementations, the library 224 maintain motion models for both static and dynamic obstacles. The motion models for static obstacles, however, may indicate that the obstacles will be stationary as the robot 202 travels in an environment. Motion models for dynamic obstacles may specify travel characteristics for the obstacle, such as expected velocities, accelerations, trajectories, directional changes, or a combination of these and/or other characteristics.
The predictive model 222 may apply a motion model to predict the location of an obstacle at a future time step. For example, if a person riding a bicycle and a person walking are identified adjacent to each other on a sidewalk, the predictive model 222 may predict that the person riding the bicycle will be located ahead of the person walking at a future time step, as a result of the motion model for the bicycle specifying a higher expected velocity than that of the person walking. In some implementations, the motion model for an obstacle specifies a function that predicts a path or motion of the obstacle based on observed values for one or more parameters of the obstacle. For example, the obstacle location prediction engine 218 may provide the motion model with information indicating a current velocity of the obstacle, a recent velocity of the obstacle at one or more preceding time steps, a current location of the obstacle, a recent location of the obstacle at one or more preceding time steps, a current direction that the obstacle is moving or facing/oriented, a recent direction that the obstacle was moving or facing/oriented at one or more preceding time steps, an indication of whether the obstacle is traveling with other obstacles (e.g., whether a person riding a bicycle and a person walking are traveling together and are unlikely to separate, or whether the person riding the bicycle and the person walking are independent), a pose or posture of the obstacle (e.g., whether a person is sitting and unlikely to move or standing and likely to move), an operational status of the obstacle (e.g., whether a device is powered on and capable of moving or powered off and stationary), or a combination of values for all or some of these parameters and/or other parameters.
In some implementations, the system 200 explicitly provides a library of motion models 224 separately from predictive model 222. The motion models 224 and predictive model 222 can then be used together to generate obstacle location predictions. For example, the obstacle location prediction engine 218 may analyze one or more occupancy maps 216 from the mapping engine 214 to identify obstacles in the environment and determine values for any parameters of the obstacles that can be provided as input to the corresponding motion models for the obstacles. The motion models may then output predicted motion data describing how the obstacles are predicted to move in the environment for one or more future time steps. The predictive model 222 may then process the predicted motion data for each obstacle from the motion models, along with data representing one or more occupancy maps, to generate obstacle location predictions such as obstacle location prediction data 226. In implementations where the motion models are explicitly provided outside of the obstacle location prediction model 222, the motion models and predictive model 222 may be trained either jointly or independently of each other. The motion models may be pre-defined based on historical analysis of the movements of different types (e.g., classes) or instances of obstacles. In some cases, the motion models for various obstacles may be automatically learned and refined over time as additional observations of actual movements of the obstacles are obtained and analyzed. To select appropriate motion models form the library 224, the obstacle location prediction engine 218 can include an object recognizer 220 that recognizes obstacles represented in occupancy maps 216, e.g., based on the shape or pattern of clusters of occupied cells in the maps 216 that correspond to obstacles, based on the locations of the obstacles in the environment, or a combination of these and other factors.
In other implementations, the system 200 may not include explicit motion models. Instead, the obstacle location predictive model 222 may be trained to implicitly predict motion/travel characteristics of different types of obstacles in an environment. For example, the predictive model 222 may be trained by processing data representing a sequence of occupancy maps over a series of time steps to predict the locations of obstacles at a future time step. During the training procedure, the predictive model 222 may learn how different types of obstacles represented in the occupancy maps tend to move. In such cases, the predictive model 222 need not rely on an object recognizer 220 to explicitly identify the types or instances of obstacles represented in occupancy maps 216. The predictive model 222 instead directly generates obstacle location predictions from data representing occupancy maps of the robot's environment. Further detail on training the predictive model 222 is discussed with respect to
Next, the system 200 further includes a planning engine 228. The planning engine 228 includes a global planner 230 and a local planner 232. The global planner 230 is configured to plan a route for an autonomous robot to travel toward a specified goal (target) location. The local planner 232 is configured to plan a short path (trajectory) for the autonomous robot to follow along the planned route. The route for a robot is generally less specific than the path. For instance, while the route may indicate which rooms or regions of an environment the robot should travel through to reach the goal location, the path can indicate precise positions and movements for the robot to take to actually execute the route. The local planner 232, for example, plan maneuvers that ensure the robot avoids collisions with obstacles in the environment. Either planner 230, 232 may use data from one or more sources to generate a route or path. The global planner 230, for instance, may access mapping data representing a map or layout of the robot's environment, as well as semantic data that identifies semantic regions of the environment represented in the map and any constraints that bear on route planning. Based on a starting location, goal location, map, and any applicable semantic data, the global planner can generate a route for the robot. The route may be optimized, for example, to minimize the time or distance for the robot to reach the goal location.
The local planner 232 receives the planned route from global planner 230, along with obstacle prediction data 226 from the obstacle location prediction engine 218, and generates a planned path from the current location of the robot 202 to a farther point along the route. The local planner 232 can constantly update the path as the robot travels, and so it may or may not generate a complete path that terminates at the goal location in a single pass. For example, the local planner 232 may only generate a path for a segment of the route that extends a limited distance or that would guide the robot's travel for a limited amount of time (e.g., the next 1, 2, or 5 seconds). In some implementations, the local planner 232 generates a path by determining particular positions/locations in the environment for the robot 202 to travel and corresponding times for the robot 202 to advance to those locations. For instance, the planner 232 may determine precise locations where the robot 202 should advance to at each time step in a series of time steps representing fixed time intervals (e.g., 100, 200, or 500 milliseconds) for the robot's impending travel.
The local planner 232 can generate a path for the robot along a segment of the route by optimizing a path cost function that minimizes a cost of the robot traveling with respect to obstacles in the robot's environment. For example, if a path for the robot's travel is to be planned over an upcoming three-second period of time, the planner 232 the may analyze obstacle prediction data 226 from the obstacle location prediction engine 218 to identify likelihoods of obstacles being located at various locations of the environment during the upcoming three-second period of time. The various locations of the environment (e.g., each location corresponding to a cell in a predicted occupancy map) may be assigned travel cost values at each of a series of time steps over the upcoming three-second period of time. The travel cost value of a given location and time step in the environment indicates a cost of the robot 202 traveling to that location at that time step. Locations that are occupied by static obstacles such as walls or plants may have high travel cost values over the entire three-second period of time. However, locations where dynamic obstacles are predicted to pass through temporarily may have high travel cost values at times when the obstacles are located at those locations, but may have low travel cost values at times when the obstacles are not located there. Additionally, in some implementations, the local planner 232 may assign higher cost values to areas of the environment surrounding obstacles, thereby providing a “buffer zone” around the obstacles that the robot 202 should not cross. The local planner 232 may then generate a travel path for the robot 202 that minimizes an overall travel cost of the path. By minimizing travel cost, the robot 202 is caused to avoid collisions with obstacles or violating the buffer zones around obstacles, for both static obstacles that have a fixed location and dynamic obstacles that move about the environment at the same time as the robot 202.
The planning engine 228 provides trajectory data 234 as output to a controller 236 of the robot 202. The trajectory data 234 represents the planned path or trajectory for the robot 202. For instance, the trajectory data 234 can include a vector of coordinates representing a series of locations along the path that the robot should advance to at discrete times. The controller 236 is operable to generate control signals 238 to one or more mechanical actuators in the robot 202, thereby causing the robot 202 to move along the planned path. In some implementations, even before the robot 202 has traveled along the entire length of a planned path (e.g., corresponding to a segment of the route), the system 200 may re-analyze the environment and update the planned path on a frequent basis to ensure the robot continues to travel on an optimized trajectory even as conditions in the environment may change over time during the course of the robot's travel.
At stage 402, the robot uses one or more sensor subsystems to scan the environment and generate sensor data that describes measurements by the one or more sensor subsystems. The sensor subsystems can include LIDAR sensors, RADAR sensors, SONAR sensors, image sensors (e.g., of a stereoscopic camera), location sensors, or a combination of these and/or other sensors. For example, when the robot enters a room or other space in its environment, the robot may scan the room to collect information that can be processed to identify where obstacles are located in the room. In some implementations, the re-scans the environment and generates updated sensor data at regular time intervals. That is, at each time step in a series of time steps, the robot can scan its environment and generate an updated instance of sensor data for the time step.
At stage 404, the system generates occupancy maps based on the sensor data obtained at stage 402. In some implementations, the system includes a mapping engine for generating occupancy maps. In particular, for each time step in the series of time steps where the robot provided an updated instance of sensor data, the mapping engine determines an occupancy map for the time step that indicates the current locations of obstacles in the environment at that time step. The occupancy map can include a grid of occupancy cells that each correspond to a small area/location of the environment of the robot, and the grid can span a region of the environment proximate to the robot that is covered by the sensor data. Each occupancy cell can have one of two states as indicated by an occupancy value for the cell: occupied or not-occupied. If the cell is occupied, this signifies that the sensor data indicates the corresponding location of the environment is occupied by an obstacle. If the cell is not-occupied, this signifies that the sensor data indicates the corresponding location of the environment is not occupied by an obstacle. In some instances, occupancy cells representing the same location in an environment may have different states at different time steps. This can occur, for example, if a dynamic obstacle moves through a location of the environment and is present at that location at a first time step but not present at that location at another time step.
At stage 406, the system determines occupancy predictions for one or more future time steps. In some implementations, an obstacle location prediction engine provides a sequence of occupancy maps for a series of time steps to a predictive model. The predictive model processes the sequence of occupancy maps one at a time and then outputs an occupancy prediction for a future time step. The occupancy prediction can be, for example, in the form of a predicted occupancy map that includes a grid of occupancy cells and that indicates, for each occupancy cell, a likelihood (e.g., a probability) that the location of the environment corresponding to the occupancy cell will be occupied by an obstacle at the future time step. In some implementations, for time steps after a first future time step, predicted occupancy map from a preceding time step can be fed back to the predictive model as an input to generate a new predicted occupancy map for the next future time step. For example, a series of occupancy maps up to a current time step based on sensor data can be processed to generate a first predicted occupancy map for a first future time step. An input based on the first predicted occupancy map can then be processed by the predictive model to generate a second predicted occupancy map for a second future time step. These operations can repeat until predicted occupancy maps are generated for a pre-defined number of future time steps.
At stage 408, the system determines travel costs based on the occupancy predictions. In some implementations, for each future time step, a local planning engine determines travel costs for the future time step by generating a travel cost map. The travel cost map can include a grid of cells like the grid of cells contained in the occupancy map for the future time step. Each cell in the travel cost map corresponds to a different area/location of the environment of the robot. The local planning engine can assign a travel cost value to each cell in the travel cost map based at least in part on the likelihood that the area/location of the environment corresponding to the cell or a nearby cell will be occupied by an obstacle at the future time step, as indicated by the predicted occupancy map for the future time step. If there is no possibility, or an insignificant possibility, of the area/location of an environment corresponding to a cell being occupied, the cell may be assigned a low travel cost value. If there is a high probability of the area/location of an environment corresponding to a cell being occupied, the cell may be assigned a high travel cost value. In some implementations, a higher travel cost may be assigned to cells that are more likely to be occupied than to cells than are less likely to be occupied by obstacles in the environment. In some implementations, a higher travel cost may be assigned to cells that are adjacent to or otherwise nearby cells (e.g., within a threshold number of cells) that are likely to be occupied by an obstacle at the time step, so as to provide a “buffer” zone around obstacles that the robot should not travel in.
At stage 410, the local planning engine generates a trajectory for the robot based on the travel costs for the future time steps. In some implementations, the local planning engine solves an optimization function to determine a path that minimizes the cost of traveling through the environment toward a goal location of a route for the robot. In this way, the path is incented to traverse areas/locations of the environment corresponding to cells having lower travel cost values where obstacles are unlikely to be located, while avoiding areas/locations of the environment corresponding to cells having higher travel cost values where obstacles are more likely to be located or that are nearby predicted locations of obstacles. As a result, the robot can avoid collisions with obstacles in the environment and can comply with policies that respect social considerations, such as providing adequate space when traveling around humans, pets, or precious objects. At stage 412, the robot then initiates travel and follows the planned path/trajectory. In some instances, a new planned path/trajectory may be generated before the robot completes travel along the current planned path/trajectory and the robot may then shift to traveling along the new path/trajectory before completion of the preceding path/trajectory.
At stage 506, the system provides the training series of occupancy maps to the predictive model for processing. In general, the predictive model is tasked with predicting generating a predicted occupancy map (or other form of occupancy prediction) for time step n+j based on the series of occupancy maps from time steps n-m through n. The predictive model may have one or more recurrent layers that effectively provide a memory in the form of a hidden state that is updated as each of the series of occupancy maps is processed in sequence. The predictive model outputs data representing a predicted occupancy map as a result of processing the series of occupancy maps. At stage 508, the training system compares the predicted occupancy map to the target occupancy map. The difference between the predicted and target occupancy maps represents an error in the prediction. At stage 510, the system updates the values of the internal parameters of the predictive model based on the error. In some implementations, the values of the model's parameters are updated using machine-learning techniques such as backpropagation with gradient descent.
The device 600 includes a processor 610, a memory 620, a storage device 630, and an input/output device 640. Each of the components 610, 620, 630, and 640 are interconnected using a system bus 650. The processor 610 is capable of processing instructions for execution within the device 600. The processor can be designed using any of a number of architectures. For example, the processor 610 can be a CISC (Complex Instruction Set Computers) processor, a RISC (Reduced Instruction Set Computer) processor, or a MISC (Minimal Instruction Set Computer) processor.
In one implementation, the processor 610 is a single-threaded processor. In another implementation, the processor 610 is a multi-threaded processor. The processor 610 is capable of processing instructions stored in the memory 620 or on the storage device 630 to display graphical information for a user interface on the input/output device 640.
The memory 620 stores information within the device 600. In one implementation, the memory 620 is a computer-readable medium. In one implementation, the memory 620 is a volatile memory unit. In another implementation, the memory 620 is a non-volatile memory unit.
The storage device 630 is capable of providing mass storage for the system 600. In one implementation, the storage device 630 is a computer-readable medium. In various different implementations, the storage device 630 can be a floppy disk device, a hard disk device, an optical disk device, or a tape device.
The input/output device 640 provides input/output operations for the system 400. In one implementation, the input/output device 640 includes a keyboard and/or pointing device. In another implementation, the input/output device 640 includes a display unit for displaying graphical user interfaces.
The features described can be implemented in digital electronic circuitry, or in computer hardware, firmware, software, or in combinations of them. The apparatus can be implemented in a computer program product tangibly embodied in an information carrier, e.g., in a machine-readable storage device for execution by a programmable processor; and method steps can be performed by a programmable processor executing a program of instructions to perform functions of the described implementations by operating on input data and generating output. The described features can be implemented advantageously in one or more computer programs that are executable on a programmable system including at least one programmable processor coupled to receive data and instructions from, and to transmit data and instructions to, a data storage system, at least one input device, and at least one output device. A computer program is a set of instructions that can be used, directly or indirectly, in a computer to perform a certain activity or bring about a certain result. A computer program can be written in any form of programming language, including compiled or interpreted languages, and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment.
Suitable processors for the execution of a program of instructions include, by way of example, both general and special purpose microprocessors, and the sole processor or one of multiple processors of any kind of computer. Generally, a processor will receive instructions and data from a read-only memory or a random access memory or both. The essential elements of a computer are a processor for executing instructions and one or more memories for storing instructions and data. Generally, a computer will also include, or be operatively coupled to communicate with, one or more mass storage devices for storing data files; such devices include magnetic disks, such as internal hard disks and removable disks; magneto-optical disks; and optical disks. Storage devices suitable for tangibly embodying computer program instructions and data include all forms of non-volatile memory, including by way of example semiconductor memory devices, such as EPROM, EEPROM, and flash memory devices; magnetic disks such as internal hard disks and removable disks; magneto-optical disks; and CD-ROM and DVD-ROM disks. The processor and the memory can be supplemented by, or incorporated in, ASICs (application-specific integrated circuits).
To provide for interaction with a user, the features can be implemented on a computer having a display device such as a CRT (cathode ray tube) or LCD (liquid crystal display) monitor for displaying information to the user and a keyboard and a pointing device such as a mouse or a trackball by which the user can provide input to the computer. Additionally, such activities can be implemented via touchscreen flat-panel displays and other appropriate mechanisms.
The features can be implemented in a computer system that includes a back-end component, such as a data server, or that includes a middleware component, such as an application server or an Internet server, or that includes a front-end component, such as a client computer having a graphical user interface or an Internet browser, or any combination of them. The components of the system can be connected by any form or medium of digital data communication such as a communication network. Examples of communication networks include a local area network (“LAN”), a wide area network (“WAN”), peer-to-peer networks (having ad-hoc or static members), grid computing infrastructures, and the Internet.
The computer system can include clients and servers. A client and server are generally remote from each other and typically interact through a network, such as the described one. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
As used in this specification, an “engine,” or “software engine,” refers to a software implemented input/output system that provides an output that is different from the input. An engine can be an encoded block of functionality, such as a library, a platform, a software development kit (“SDK”), or an object. Each engine can be implemented on any appropriate type of computing device, e.g., servers, mobile phones, tablet computers, notebook computers, music players, e-book readers, laptop or desktop computers, PDAs, smart phones, or other stationary or portable devices, that includes one or more processors and computer readable media. Additionally, two or more of the engines may be implemented on the same computing device, or on different computing devices.
While this specification contains many specific implementation details, these should not be construed as limitations on the scope of any inventions or of what can be claimed, but rather as descriptions of features specific to particular implementations of particular inventions. Certain features that are described in this specification in the context of separate implementations can also be implemented in combination in a single implementation. Conversely, various features that are described in the context of a single implementation can also be implemented in multiple implementations separately or in any suitable subcombination. Moreover, although features can be described above as acting in certain combinations and even initially claimed as such, one or more features from a claimed combination can in some cases be excised from the combination, and the claimed combination can be directed to a subcombination or variation of a subcombination.
Similarly, while operations are depicted in the drawings in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all illustrated operations be performed, to achieve desirable results. In certain circumstances, multitasking and parallel processing can be advantageous. Moreover, the separation of various system components in the implementations described above should not be understood as requiring such separation in all implementations, and it should be understood that the described program components and systems can generally be integrated together in a single software product or packaged into multiple software products.
As used throughout this specification, the terms optimum, optimize, and variations thereof refer to improving performance of a system through the use of the techniques discussed herein, and does not imply that the performance obtained is necessarily the most optimum performance that is achievable.
Thus, particular implementations of the subject matter have been described. Other implementations are within the scope of the following claims. In some cases, the actions recited in the claims can be performed in a different order and still achieve desirable results. In addition, the processes depicted in the accompanying figures do not necessarily require the particular order shown, or sequential order, to achieve desirable results. In certain implementations, multitasking and parallel processing can be advantageous.
Patent | Priority | Assignee | Title |
11314254, | Mar 26 2019 | Intel Corporation | Methods and apparatus for dynamically routing robots based on exploratory on-board mapping |
11373412, | Feb 14 2019 | HANWHA AEROSPACE CO , LTD | Obstacle map generating method and apparatus |
11454974, | Jun 29 2020 | Baidu USA LLC | Method, apparatus, device, and storage medium for controlling guide robot |
11465633, | Nov 14 2018 | HUAWEI TECHNOLOGIES CO , LTD | Method and system for generating predicted occupancy grid maps |
11470821, | Dec 07 2017 | AMICRO SEMICONDUCTOR CO., LTD. | Method for monitoring pet by robot based on grid map and chip |
11514812, | Apr 02 2018 | PINEGEM TECHNOLOGIES LLC | Autonomous physical activity assistance systems and methods thereof |
11571814, | Sep 13 2018 | DEXAI ROBOTICS, INC | Determining how to assemble a meal |
11597084, | Sep 13 2018 | DEXAI ROBOTICS, INC | Controlling robot torque and velocity based on context |
11597085, | Sep 13 2018 | DEXAI ROBOTICS, INC | Locating and attaching interchangeable tools in-situ |
11597086, | Sep 13 2018 | DEXAI ROBOTICS, INC | Food-safe, washable interface for exchanging tools |
11597087, | Sep 13 2018 | DEXAI ROBOTICS, INC | User input or voice modification to robot motion plans |
11607801, | Aug 21 2019 | LG Electronics Inc.; LG Electronics Inc | Artificial intelligence robot for managing movement of object using artificial intelligence and method of operating the same |
11607810, | Sep 13 2018 | DEXAI ROBOTICS, INC | Adaptor for food-safe, bin-compatible, washable, tool-changer utensils |
11628566, | Sep 13 2018 | DEXAI ROBOTICS, INC | Manipulating fracturable and deformable materials using articulated manipulators |
11642791, | Oct 08 2019 | NATIONAL UNIVERSITY CORPORATION SHIZUOKA UNIVERSITY; NATIONAL UNIVERSITY CORPORATION IWATE UNIVERSITY; Toyota Jidosha Kabushiki Kaisha | Autonomous mobile robot and control program for autonomous mobile robot |
11648669, | Sep 13 2018 | DEXAI ROBOTICS, INC | One-click robot order |
11673268, | Sep 13 2018 | DEXAI ROBOTICS, INC | Food-safe, washable, thermally-conductive robot cover |
11726492, | Oct 02 2019 | ZOOX, INC | Collision avoidance perception system |
11730646, | Sep 25 2020 | System and method to combine input from multiple sensors on an autonomous wheelchair | |
11731274, | Mar 24 2021 | Ford Global Technologies, LLC | Predictive time horizon robotic motion control |
11780472, | Nov 16 2019 | AURORA OPERATIONS, INC | Systems and methods for generating motion forecast data for a plurality of actors with respect to an autonomous vehicle |
11794347, | Jan 08 2021 | Tata Consultancy Services Limited | Navigation of tele-robot in dynamic environment using in-situ intelligence |
11827214, | Mar 05 2020 | HUAWEI TECHNOLOGIES CO , LTD | Machine-learning based system for path and/or motion planning and method of training the same |
11835343, | Aug 06 2004 | AI Incorporated | Method for constructing a map while performing work |
11872702, | Sep 13 2018 | DEXAI ROBOTICS, INC | Robot interaction with human co-workers |
11967157, | Aug 20 2021 | Samsung Electronics Co., Ltd. | Robot and method for controlling thereof |
11994866, | Oct 02 2019 | ZOOX, INC ; Zoox, Inc. | Collision avoidance perception system |
Patent | Priority | Assignee | Title |
7211980, | Jul 05 2006 | Humatics Corporation | Robotic follow system and method |
8214079, | Mar 30 2007 | Sungkyunkwan University Foundation for Corporate Collaboration | Central information processing system and method for service robot having layered information structure according to recognition and reasoning level |
8396254, | Feb 09 2012 | GOOGLE LLC | Methods and systems for estimating a location of a robot |
8432443, | Feb 27 2008 | Ajou University Industry Cooperation Foundation | Method for object localization using visual images with reference coordinates |
9129148, | Nov 09 2012 | Amazon Technologies, Inc | System, method and apparatus for scene recognition |
9574883, | Mar 24 2015 | GOOGLE LLC | Associating semantic location data with automated environment mapping |
9751212, | May 05 2016 | Toyota Jidosha Kabushiki Kaisha | Adapting object handover from robot to human using perceptual affordances |
20170249504, | |||
20180364731, | |||
20190080604, | |||
20190113927, | |||
20190213438, | |||
20190332121, | |||
20200159227, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Jan 24 2018 | MILLARD, DAVID | X Development LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 045004 | /0092 | |
Jan 26 2018 | X Development LLC | (assignment on the face of the patent) | / | |||
Apr 01 2023 | X Development LLC | GOOGLE LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 064658 | /0001 |
Date | Maintenance Fee Events |
Jan 26 2018 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Nov 12 2024 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Date | Maintenance Schedule |
May 25 2024 | 4 years fee payment window open |
Nov 25 2024 | 6 months grace period start (w surcharge) |
May 25 2025 | patent expiry (for year 4) |
May 25 2027 | 2 years to revive unintentionally abandoned end. (for year 4) |
May 25 2028 | 8 years fee payment window open |
Nov 25 2028 | 6 months grace period start (w surcharge) |
May 25 2029 | patent expiry (for year 8) |
May 25 2031 | 2 years to revive unintentionally abandoned end. (for year 8) |
May 25 2032 | 12 years fee payment window open |
Nov 25 2032 | 6 months grace period start (w surcharge) |
May 25 2033 | patent expiry (for year 12) |
May 25 2035 | 2 years to revive unintentionally abandoned end. (for year 12) |