Nothing Special   »   [go: up one dir, main page]

CN114323035A - Positioning method, device and system - Google Patents

Positioning method, device and system Download PDF

Info

Publication number
CN114323035A
CN114323035A CN202011063252.0A CN202011063252A CN114323035A CN 114323035 A CN114323035 A CN 114323035A CN 202011063252 A CN202011063252 A CN 202011063252A CN 114323035 A CN114323035 A CN 114323035A
Authority
CN
China
Prior art keywords
vehicle
moment
particles
information
feature points
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011063252.0A
Other languages
Chinese (zh)
Inventor
叶爱学
高亚军
陈哲
许光林
温丰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Priority to CN202011063252.0A priority Critical patent/CN114323035A/en
Priority to PCT/CN2021/101620 priority patent/WO2022068274A1/en
Publication of CN114323035A publication Critical patent/CN114323035A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/20Instruments for performing navigational calculations
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/28Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network with correlation of data from several navigational instruments
    • G01C21/30Map- or contour-matching
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/34Route searching; Route guidance

Landscapes

  • Engineering & Computer Science (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Automation & Control Theory (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Navigation (AREA)

Abstract

The application provides a positioning method, a positioning device and a positioning system, after a vehicle is started, sensing information of the vehicle is obtained according to a time interval, when the vehicle is positioned, sensing information corresponding to the current moment is obtained through a vehicle-mounted sensing device, the sensing information obtained at a key moment before is obtained, and therefore the vehicle is positioned according to the obtained sensing information and map information of an area where the vehicle is located. The sensing information is acquired from the time dimension, so that the process of acquiring the sensing information is not influenced by road congestion, the sensing information for positioning can be updated according to the time interval when the vehicle runs, and the positioning accuracy is improved. And moreover, the sensing information is acquired from the time dimension, and when the vehicle runs at a low speed, the environmental information around the vehicle can be acquired in the gap between the vehicles.

Description

Positioning method, device and system
Technical Field
The embodiment of the application relates to the technical field of intelligent driving (intelligent driving), in particular to a positioning method, a positioning device and a positioning system.
Background
Laser radar (lidar) is used as an unmanned driving automobile positioning sensor, and realizes the perception of the vehicle to the surrounding environment and the global positioning (gobal positioning) of the vehicle together with technologies such as image recognition and the like. Among them, the laser radar (lidar) mounting system on a vehicle includes a top laser mounting system and a front security mounting system.
The laser radar (lidar) is arranged at the front part and two sides of the vehicle in a front security installation mode, the installation position is low, the laser radar is easy to block, the sensing range is small, the information for positioning is less, and the positioning is inaccurate. Therefore, in order to improve the accuracy of the positioning, a laser radar (lidar) is controlled to collect environmental information around the vehicle once every time the vehicle travels a preset distance. When the vehicle is positioned, the positioning is carried out according to the environment information around the vehicle which is collected for many times, and the positioning accuracy is higher.
However, when the road is congested, the vehicle runs slowly, and the environmental information around the vehicle cannot be collected in time, so that the environmental information around the vehicle, which is collected for multiple times for positioning, is not updated when the vehicle is positioned, and thus the positioning accuracy is poor.
Disclosure of Invention
The application provides a positioning method, a positioning device and a positioning system, and aims to solve the problem of inaccurate positioning of a vehicle caused by obstruction, road congestion and the like.
In a first aspect, the present application provides a positioning method, which may be implemented by a vehicle, or implemented by a component in the vehicle, such as a processing device, a circuit, a chip, or the like in the vehicle, or implemented by a cloud server in communication with the vehicle through a gateway. The method comprises the following steps:
at the current moment, acquiring first sensing information through a vehicle-mounted sensing device, wherein the first sensing information comprises coordinates of characteristic points acquired by the vehicle-mounted sensing device at the current moment under a vehicle body coordinate system and odometer information of a vehicle at the current moment;
acquiring second sensing information acquired through the vehicle-mounted sensing device at a key moment, wherein the second sensing information comprises coordinates of characteristic points acquired by the vehicle-mounted sensing device at the key moment under a vehicle body coordinate system and odometer information of the vehicle at the key moment, the key moment comprises a first key moment, the first key moment is a moment of acquiring second sensing information according to a time interval, and the time interval between the first key moment and the current moment is less than or equal to a preset time length;
and determining the position of the vehicle at the current moment according to the map information of the driving area of the vehicle, the first sensing information and the second sensing information.
In one possible embodiment, the determining the position of the vehicle at the current time according to the map information of the driving area of the vehicle and the first sensing information and the second sensing information includes:
mapping the coordinates of the feature points acquired at the key moment to the vehicle body coordinate system at the current moment according to the odometer information in the first sensing information and the odometer information in the second sensing information, and acquiring the coordinates of the feature points acquired at the key moment in the vehicle body coordinate system at the current moment;
and determining the position of the vehicle at the current moment according to the coordinates of the feature points acquired at the key moment and the feature points acquired at the current moment under the vehicle body coordinate system at the current moment and the map information.
In a possible implementation manner, the determining the position of the vehicle at the current time according to the coordinates of the feature points acquired at the critical time and the feature points acquired at the current time in the vehicle body coordinate system at the current time and the map information includes:
at the current moment, for each particle in M particles corresponding to the vehicle, obtaining matching degrees of the feature points acquired at the current moment and the key moment and a target object in the map information according to coordinates of the feature points acquired at the key moment and the feature points acquired at the current moment under a vehicle body coordinate system at the current moment and the position of each particle, respectively, wherein the M particles are particles adopted when the vehicle is positioned last time, the position of each particle is obtained according to odometer information when the vehicle is positioned last time, the odometer information at the current moment and the position of each particle when the vehicle is positioned last time, and M is a positive integer;
obtaining K particles according to the matching degree of each particle in the M particles at the current moment and the M particles, wherein K is a positive integer;
and obtaining the position of the vehicle at the current moment according to the positions of the K particles corresponding to the vehicle at the current moment.
In a possible implementation manner, the obtaining, according to the coordinates of the feature points acquired at the key moment and the feature points acquired at the current moment under the vehicle body coordinate system at the current moment and the position of each particle, the matching degrees of the feature points acquired at the current moment and the key moment and the target object in the map information includes:
respectively mapping the coordinates of the feature points acquired at the current moment and the key moment under the vehicle body coordinate system at the current moment to a particle coordinate system with the particles as the origin;
and obtaining the matching degree of a target object to be matched, which is formed by the characteristic points acquired at the current moment and the key moment and mapped to the particle coordinate system, and the corresponding target object in the map information.
In a possible implementation manner, the obtaining K particles according to the matching degree of each of the M particles at the current time and the M particles includes:
obtaining L particles with a matching degree greater than or equal to a preset matching degree in the M particles, wherein L is less than or equal to M and less than or equal to K;
determining the L particles as the K particles; or,
obtaining at least one particle from the L particles;
obtaining the K particles according to the L particles and the at least one particle.
In a possible embodiment, said obtaining said K particles from said L particles and said at least one particle comprises:
according to the matching degree of each particle in the at least one particle, copying each particle at least once;
and determining the L particles and the particles obtained after copying as the K particles.
In a possible implementation manner, the obtaining the position of the vehicle at the current time according to the positions of the K particles corresponding to the vehicle at the current time includes:
and acquiring the average position of the K particles according to the position of each particle in the K particles, and determining the average position of the K particles as the position of the vehicle at the current moment.
In one possible embodiment, the method further comprises:
obtaining an evaluation value of the position of the vehicle at the current moment according to the position of each of the K particles corresponding to the vehicle at the current moment and the position of the vehicle at the current moment, wherein the evaluation value indicates a difference between the position of the vehicle at the current moment and real position information of the vehicle at the current moment.
In a possible implementation manner, the obtaining, according to the odometer information in the second sensing information and the odometer information in the second sensing information, the coordinates of the feature point acquired at the key time are respectively mapped to the vehicle body coordinate system at the current time, and the coordinates of the feature point acquired at the key time in the vehicle body coordinate system at the current time are obtained, including:
determining a coordinate mapping relation between the vehicle body coordinate system corresponding to the current moment and the vehicle body coordinate system corresponding to the first key moment according to the mileage information in the first sensing information and the mileage information in the second sensing information;
and according to the coordinate mapping relation, respectively mapping the coordinates of the characteristic points acquired at the first key moment to the vehicle body coordinate system at the current moment, and acquiring the coordinates of the characteristic points acquired at the first key moment in the vehicle body coordinate system at the current moment.
In a possible implementation manner, the critical time further includes a second critical time, where the second critical time is a time when second sensing information is acquired according to a distance interval, and a travel distance from a position of the second critical time to a position corresponding to the current time is less than or equal to a preset distance.
In one possible embodiment, the determining the position of the vehicle at the current time according to the map information of the driving area of the vehicle and the first sensing information and the second sensing information includes:
and determining the pose of the vehicle at the current moment according to the map information and the first and second sensing information.
In a possible implementation manner, before obtaining matching degrees between feature points acquired at a current time and the key time and a target object in the map information according to coordinates of the feature points acquired at the key time and the feature points acquired at the current time under the vehicle body coordinate system at the current time and a position of each particle, the method further includes:
determining target feature points according to the feature points acquired at each key moment and the feature values of each feature point in the feature points acquired at the current moment;
determining target feature points again according to feature points acquired at each key moment and feature values of the feature points acquired at the current moment except feature points contained in a feature point set, wherein the feature points contained in the feature point set are feature points located in a preset distance range of the last determined target feature point and the last determined target feature point;
when the number of the target characteristic points meets a preset number, N target characteristic points are obtained, wherein N is equal to the preset number;
the obtaining of the matching degree between the feature points acquired at the current moment and the key moment and the target object in the map information according to the coordinates of the feature points acquired at the key moment and the feature points acquired at the current moment under the vehicle body coordinate system at the current moment and the position of each particle respectively includes:
and according to the coordinates of the N target characteristic points in the vehicle body coordinate system at the current moment and the position of each particle, obtaining the matching degree of the characteristic points acquired at the current moment and the key moment and the target object in the map information of the target area.
In one possible embodiment, the characteristic value of the characteristic point is obtained by an evaluation function, and the characteristic value is used to evaluate the stability of the characteristic point.
In one possible embodiment, the target feature point is a feature point having a largest feature value among the feature points.
In one possible embodiment, the origin of the body coordinate system is located at any position on the vehicle.
In a second aspect, the present application provides a positioning device comprising:
the acquisition module is used for acquiring first sensing information through the vehicle-mounted sensing device at the current moment, wherein the first sensing information comprises coordinates of characteristic points acquired by the vehicle-mounted sensing device at the current moment under a vehicle body coordinate system and odometer information of a vehicle at the current moment; the vehicle-mounted sensing device is further used for acquiring second sensing information acquired by the vehicle-mounted sensing device at a key moment, wherein the second sensing information comprises coordinates of characteristic points acquired by the vehicle-mounted sensing device at the key moment under a vehicle body coordinate system and odometer information of the vehicle at the key moment, the key moment comprises a first key moment, the first key moment is a moment when second sensing information is acquired according to a time interval, and the time interval between the first key moment and the current moment is less than or equal to a preset duration;
and the positioning module is used for determining the position of the vehicle at the current moment according to the map information of the driving area of the vehicle, the first sensing information and the second sensing information.
In a possible implementation manner, when the positioning module determines the position of the vehicle at the current time according to the map information of the vehicle driving area and the first sensing information and the second sensing information, the positioning module is specifically configured to:
mapping the coordinates of the feature points acquired at the key moment to the vehicle body coordinate system at the current moment according to the odometer information in the first sensing information and the odometer information in the second sensing information, and acquiring the coordinates of the feature points acquired at the key moment in the vehicle body coordinate system at the current moment;
and determining the position of the vehicle at the current moment according to the coordinates of the feature points acquired at the key moment and the feature points acquired at the current moment under the vehicle body coordinate system at the current moment and the map information.
In a possible implementation manner, when the positioning module determines the position of the vehicle at the current time according to the coordinates of the feature points acquired at the key time and the feature points acquired at the current time in the vehicle body coordinate system at the current time and the map information, the positioning module is specifically configured to:
at the current moment, for each particle in M particles corresponding to the vehicle, obtaining matching degrees of the feature points acquired at the current moment and the key moment and a target object in the map information according to coordinates of the feature points acquired at the key moment and the feature points acquired at the current moment under a vehicle body coordinate system at the current moment and the position of each particle, respectively, wherein the M particles are particles adopted when the vehicle is positioned last time, the position of each particle is obtained according to odometer information when the vehicle is positioned last time, the odometer information at the current moment and the position of each particle when the vehicle is positioned last time, and M is a positive integer;
obtaining K particles according to the matching degree of each particle in the M particles at the current moment and the M particles, wherein K is a positive integer;
and obtaining the position of the vehicle at the current moment according to the positions of the K particles corresponding to the vehicle at the current moment.
In a possible implementation manner, when the positioning module obtains the matching degrees between the feature points acquired at the current time and the key time and the target object in the map information according to the coordinates of the feature points acquired at the key time and the feature points acquired at the current time under the vehicle body coordinate system at the current time and the position of each particle, the positioning module is specifically configured to:
respectively mapping the coordinates of the feature points acquired at the current moment and the key moment under the vehicle body coordinate system at the current moment to a particle coordinate system with the particles as the origin;
and obtaining the matching degree of a target object to be matched, which is formed by the characteristic points acquired at the current moment and the key moment and mapped to the particle coordinate system, and the corresponding target object in the map information.
In a possible implementation manner, when the positioning module obtains K particles according to the matching degree of each of the M particles at the current time and the M particles, the positioning module is specifically configured to:
obtaining L particles with a matching degree greater than or equal to a preset matching degree in the M particles, wherein L is less than or equal to M and less than or equal to K;
determining the L particles as the K particles; or,
obtaining at least one particle from the L particles;
obtaining the K particles according to the L particles and the at least one particle.
In a possible implementation manner, when the positioning module obtains the K particles according to the L particles and the at least one particle, the positioning module is specifically configured to:
according to the matching degree of each particle in the at least one particle, copying each particle at least once;
and determining the L particles and the particles obtained after copying as the K particles.
In a possible implementation manner, when the location module obtains the location of the vehicle at the current time according to the locations of the K particles corresponding to the vehicle at the current time, the location module is specifically configured to:
and acquiring the average position of the K particles according to the position of each particle in the K particles, and determining the average position of the K particles as the position of the vehicle at the current moment.
In a possible implementation, the positioning module is further configured to:
obtaining an evaluation value of the position of the vehicle at the current moment according to the position of each of the K particles corresponding to the vehicle at the current moment and the position of the vehicle at the current moment, wherein the evaluation value indicates a difference between the position of the vehicle at the current moment and real position information of the vehicle at the current moment.
In a possible implementation manner, the positioning module, according to the odometer information in the second sensing information and the odometer information in the second sensing information, respectively maps the coordinates of the feature point acquired at the key moment to the vehicle body coordinate system at the current moment, and when obtaining the coordinates of the feature point acquired at the key moment in the vehicle body coordinate system at the current moment, is specifically configured to:
determining a coordinate mapping relation between the vehicle body coordinate system corresponding to the current moment and the vehicle body coordinate system corresponding to the first key moment according to the mileage information in the first sensing information and the mileage information in the second sensing information;
and according to the coordinate mapping relation, respectively mapping the coordinates of the characteristic points acquired at the first key moment to the vehicle body coordinate system at the current moment, and acquiring the coordinates of the characteristic points acquired at the first key moment in the vehicle body coordinate system at the current moment.
In a possible implementation manner, the critical time further includes a second critical time, where the second critical time is a time when second sensing information is acquired according to a distance interval, and a travel distance from a position of the second critical time to a position corresponding to the current time is less than or equal to a preset distance.
In a possible implementation manner, when the positioning module determines the position of the vehicle at the current time according to the map information of the vehicle driving area and the first sensing information and the second sensing information, the positioning module is specifically configured to:
and determining the pose of the vehicle at the current moment according to the map information and the first and second sensing information.
In a possible implementation manner, before the positioning module obtains the matching degrees between the feature points acquired at the current time and the target object in the map information according to the coordinates of the feature points acquired at the key time and the feature points acquired at the current time under the vehicle body coordinate system at the current time and the position of each particle, the positioning module is further configured to:
determining target feature points according to the feature points acquired at each key moment and the feature values of each feature point in the feature points acquired at the current moment;
determining target feature points again according to feature points acquired at each key moment and feature values of the feature points acquired at the current moment except feature points contained in a feature point set, wherein the feature points contained in the feature point set are feature points located in a preset distance range of the last determined target feature point and the last determined target feature point;
when the number of the target characteristic points meets a preset number, N target characteristic points are obtained, wherein N is equal to the preset number;
the positioning module is specifically configured to, when obtaining matching degrees of the feature points acquired at the current time and the key time and a target object in the map information according to the feature points acquired at the key time and the coordinates of the feature points acquired at the current time under the vehicle body coordinate system at the current time and the position of each particle, specifically:
and according to the coordinates of the N target characteristic points in the vehicle body coordinate system at the current moment and the position of each particle, obtaining the matching degree of the characteristic points acquired at the current moment and the key moment and the target object in the map information of the target area.
In one possible embodiment, the characteristic value of the characteristic point is obtained by an evaluation function, and the characteristic value is used to evaluate the stability of the characteristic point.
In one possible embodiment, the target feature point is a feature point having a largest feature value among the feature points.
In one possible embodiment, the origin of the body coordinate system is located at any position on the vehicle.
In a third aspect, the present application provides a positioning device, comprising: a memory and at least one processor;
the memory to store program instructions;
the processor is configured to call the program instructions in the memory to execute the positioning method according to any one of the first aspect.
In a fourth aspect, the present application provides a positioning system comprising a vehicle and an onboard sensing device;
the vehicle-mounted sensing device is mounted on the vehicle;
the vehicle for executing the positioning method according to any one of the first aspect; or,
the vehicle-mounted sensing device is used for executing the positioning method of any one of the first aspect.
In a fifth aspect, the present application provides a non-transitory computer-readable storage medium having stored thereon computer instructions for causing the computer to perform the positioning method of any one of the first aspect.
In a sixth aspect, the present application provides a program product comprising a computer program stored in a readable storage medium, from which the computer program can be read by at least one processor of a positioning apparatus, the execution of which by the at least one processor causes the positioning apparatus to implement the positioning method according to any one of the first aspect.
The application provides a positioning method, a positioning device and a positioning system, wherein after a vehicle is started, sensing information of the vehicle is obtained according to a time interval, when the vehicle is positioned, first sensing information corresponding to the current moment and second sensing information corresponding to at least one key moment are obtained, and the vehicle is positioned according to the first sensing information, the at least one second sensing information and map information. The sensing information is acquired from the time dimension, so that the process of acquiring the sensing information is not influenced by road congestion, the sensing information for positioning can be updated according to time intervals when the vehicle runs, and the positioning accuracy is improved. And moreover, the sensing information is acquired from the time dimension, when the vehicle runs at a low speed, the environmental information around the vehicle can be acquired in the gap between the vehicles, and at the moment, the acquired environmental information around the vehicle is clearer due to the fact that the vehicle speed is lower, and the positioning accuracy is further improved.
Drawings
Fig. 1 is a schematic view of an application scenario provided in an embodiment of the present application;
fig. 2 is a flowchart of a positioning method according to an embodiment of the present application
FIG. 3 is a schematic view of a vehicle body coordinate system provided in an embodiment of the present application;
FIG. 4 is a schematic illustration of vehicle positioning provided by an embodiment of the present application;
FIG. 5 is a schematic illustration of vehicle positioning provided by another embodiment of the present application;
FIG. 6 is a schematic illustration of vehicle positioning provided by another embodiment of the present application;
fig. 7a is a schematic view of ambient environment information when a vehicle faces a first direction according to an embodiment of the present application;
fig. 7b is a schematic view of the surrounding environment information when the vehicle faces a second direction according to an embodiment of the present application;
fig. 8 is a flowchart of a positioning method according to another embodiment of the present application;
fig. 9 is a flowchart of a positioning method according to another embodiment of the present application;
FIG. 10 is a schematic diagram of particle initialization according to an embodiment of the present application;
FIG. 11 is a schematic diagram of the peripheral environment information under the vehicle coordinate system and the particle coordinate system, respectively, according to an embodiment of the present disclosure;
fig. 12 is a schematic structural diagram of a positioning device according to an embodiment of the present application;
fig. 13 is a schematic structural diagram of a positioning device according to an embodiment of the present application;
fig. 14 is a schematic structural diagram of a positioning system according to an embodiment of the present application.
Detailed Description
To facilitate understanding of the embodiments of the present application, concepts related to the embodiments of the present application will be described first.
At the current moment: and after the vehicle is started, positioning time is carried out.
The key moment is as follows: after the vehicle is started, the sensor device scans the surrounding environment information of the vehicle, and the scanned surrounding environment information of the vehicle is used for positioning.
The characteristic points are as follows: in the point cloud corresponding to the vehicle surrounding environment scanned by the sensing device, the points of the characteristics of the target object in the surrounding environment can be reflected, so that the target object in the surrounding environment can be identified through the characteristic points.
Fig. 1 is a schematic view of an application scenario provided in an embodiment of the present application. In the intelligent driving mode of the vehicle, after the vehicle is started, for example, the vehicle is running or is temporarily in a parking state due to congestion, waiting for traffic lights and the like, the information of the surrounding environment of the vehicle is sensed through the laser radar, and the vehicle is positioned. Specifically, the method comprises the following steps: and setting a preset distance, and transmitting a laser signal to the periphery of the vehicle by the laser radar on the vehicle after the vehicle runs the preset distance every time. The target object in the vehicle surroundings is scanned with the laser beam and displayed as a point Cloud (clouds of points). And extracting the mark information around the vehicle by acquiring the point cloud each time. And during positioning, matching the mark information around the vehicle and the map information which are acquired for many times to realize the positioning of the vehicle.
However, when the vehicle runs on a congested road, the vehicle runs slowly, so that the collected surrounding environment information cannot be updated in time when the vehicle is positioned, and positioning is inaccurate.
For example, as shown in fig. 1, when a vehicle travels on a congested road, environment information around the vehicle is collected every 100 meters, for example, the vehicle collects environment information around the vehicle at a point a, a point B, a point C, and a point D, respectively, and the vehicle travels to a point E, and when positioning is required at the point E because a travel distance between the point D and the point E is less than 100 meters, positioning is performed by collecting environment information around the vehicle according to the point a, the point B, the point C, and the point D, and at this time, the position where the vehicle is positioned is located near the point D due to an error.
Due to road congestion, when the vehicle runs slowly and needs to be positioned when the vehicle runs to the point F, the running distance between the point D and the point F is still less than 100 meters, so that the environmental information around the vehicle, which is acquired by the vehicle, is still acquired at the point A, the point B, the point C and the point D. Therefore, during positioning, the environmental information around the vehicle is still collected according to the points a, B, C and D for positioning, which results in a large difference between the position of the vehicle obtained by positioning and the actual position of the vehicle, and inaccurate positioning.
It should be noted that the above examples are only used for exemplarily illustrating application scenarios to which the embodiments of the present application may be applied, and are not to be construed as limitations on the application scenarios of the embodiments of the present application.
Therefore, to solve the problems existing in the prior art, the present application proposes: the time dimension has continuity, so that the vehicle-mounted device, such as a laser radar, can be controlled in the time dimension to continuously acquire the environmental information around the vehicle, and the position of the vehicle at the current moment is positioned through the continuously acquired environmental information around the vehicle. According to the method and the device, the environmental information around the vehicle is continuously acquired in the time dimension, and the time dimension is not affected by the conditions of road congestion and the like. Therefore, after the vehicle is started, the collected environmental information around the vehicle can be updated in time, and the positioning accuracy is further improved.
It should be understood that, in practical application, the technical scheme of this application except being applied to vehicle location also can be applied to the aircraft location, for example, in the location scene of unmanned aerial vehicle, improves the accuracy of location. The present application is described with reference to vehicle positioning, however, this should not be construed as limiting the scope of the application.
The following describes the technical solutions of the present application and how to solve the above technical problems with specific embodiments. The following several specific embodiments may be combined with each other, and details of the same or similar concepts or processes may not be repeated in some embodiments. Embodiments of the present application will be described below with reference to the accompanying drawings.
Fig. 2 is a flowchart of a positioning method according to an embodiment of the present application. The execution subject of this embodiment is, for example, a vehicle, and may also be implemented by a component in the vehicle, such as a processing device, a circuit, a chip, and the like in the vehicle, or a cloud server that communicates with the vehicle through a gateway, which is not limited in this application. As shown in fig. 2, the method of the embodiment of the present application includes:
s201, at the current moment, first sensing information is acquired through the vehicle-mounted sensing device.
The first sensing information comprises the coordinates of the characteristic points acquired by the vehicle-mounted sensing device at the current moment under the vehicle body coordinate system and the odometer information of the vehicle at the current moment.
In this step, optionally, the vehicle body coordinate system may be a planar coordinate system, such as a planar rectangular coordinate system, a polar coordinate system, or a spatial coordinate system, such as a spatial rectangular coordinate system, a cylindrical coordinate system, a spherical coordinate system, which is not limited in this application. The present embodiment is described by taking a spatial rectangular coordinate system as an example.
Alternatively, the body coordinate system is a rectangular spatial coordinate system, and the origin of the body coordinate system may be located at any position on the vehicle, for example, as shown in fig. 3, the geometric center of the vehicle is the origin, the driving direction is an X-axis, the direction perpendicular to the driving direction on the horizontal plane is a Y-axis, and the direction perpendicular to the driving direction on the vertical plane is a Z-axis.
The origin of the body coordinate system is determined with the vehicle as a reference object, but since the position of the vehicle changes during the travel of the vehicle, the origin of the body coordinate system corresponding to the vehicle at different positions differs.
After the vehicle is started, environmental information around the vehicle and odometer information of the vehicle are acquired through the vehicle-mounted sensing device. The vehicle-mounted sensing device is, for example, a system integrated by multiple sensors (sensors), and may include, for example, a laser radar for scanning the surrounding environment of the vehicle to obtain the environmental information around the vehicle, and an inertia detection unit for obtaining the odometer information of the vehicle.
It should be noted that, in some possible designs, the point cloud of the environmental information around the vehicle may also be obtained by other sensors, such as an ultrasonic sensor.
At the current moment, a laser radar in the vehicle-mounted sensing device emits laser outwards to obtain point cloud corresponding to the surrounding environment of the vehicle, and the inertia detection unit obtains odometer information at the current moment, so that the vehicle obtains sensing information at the current moment, namely first sensing information, through the point cloud corresponding to the surrounding environment of the vehicle at the current moment and the odometer information. The odometer information may include, for example: steering information of the vehicle after starting, the number of turns of the wheels, etc.
Optionally, when the vehicle-mounted sensing device acquires the point cloud and the odometer information corresponding to the vehicle surrounding environment at the current time, the acquired point cloud and the odometer information corresponding to the vehicle surrounding environment may be directly sent to the vehicle processing device, and the vehicle processing device processes the point cloud and the odometer information corresponding to the vehicle surrounding environment to acquire the first sensing information. For example, the vehicle processing device extracts at least one feature point from the point cloud corresponding to the vehicle surrounding environment, maps the coordinates of the feature point in the vehicle-mounted sensing device coordinate system to the vehicle body coordinate system, and determines the coordinates of the at least one feature point in the vehicle body coordinate system.
Or, the vehicle-mounted sensing device performs preliminary processing on the point cloud and the odometer information corresponding to the acquired vehicle surrounding environment, for example, conversion of a data format, extraction of feature points in the point cloud, and the like, and sends the point cloud corresponding to the vehicle surrounding environment after the preliminary processing and the odometer information acquired by the inertia detection unit at the current time to the processing device of the vehicle. The vehicle processing device acquires first sensing information according to the point cloud corresponding to the vehicle surrounding environment after the primary processing and the inertia detection unit.
And then, or the vehicle-mounted sensing device acquires first sensing information according to the acquired point cloud corresponding to the vehicle surrounding environment at the current moment and the acquired odometer information, and then sends the first sensing information to a processing device of the vehicle. Therefore, when the vehicle is positioned, the vehicle processing device can directly use the first sensing information, the processing amount of the vehicle processing device is reduced, and the positioning speed of the vehicle processing device is accelerated.
S202, second sensing information acquired through the vehicle-mounted sensing device at the key moment is acquired.
The second sensing information comprises coordinates of the characteristic points acquired by the vehicle-mounted sensing device at the key moment under the vehicle body coordinate system and odometer information of the vehicle at the key moment.
The key time comprises a first key time, the first key time is the time for acquiring the second sensing information according to the time interval, and the time interval between the key time and the current time is less than or equal to the preset time length.
In this step, in the prior art, by setting the preset distance, when the vehicle travels the preset distance, the vehicle-mounted sensing device acquires the environmental information and the odometer information around the vehicle, so that the vehicle processing device acquires the sensing information, and performs positioning according to the sensing information acquired for multiple times, but the road congestion causes the vehicle to travel slowly, and the updating of the sensing information is slow. In addition, due to road congestion, the environmental information around the vehicle is shielded by obstacles such as surrounding vehicles and the like, the sensing range is small, and in a section of congested road, the sensing information is acquired according to the preset distance, and only once sensing information is acquired when the vehicle runs the preset distance, so that less effective information in the sensing information is caused, and the positioning accuracy is influenced. The effective information is information which can improve the positioning accuracy in the perception information.
Therefore, in the present application, after the vehicle is started, the vehicle-mounted sensing device continuously acquires the point cloud and the odometer information corresponding to the vehicle surrounding environment in the time dimension, for example, a time period of acquisition is preset, and when the time period of acquisition is reached, the vehicle-mounted sensing device acquires the point cloud and the odometer information corresponding to the vehicle surrounding environment once. Or adjusting the time interval of the point cloud and the odometer information corresponding to the vehicle surrounding environment obtained by the vehicle-mounted sensing device twice in a neighboring mode according to the running speed of the vehicle. Correspondingly, the vehicle processing device acquires the sensing information once corresponding to each time the vehicle-mounted sensing device acquires the point cloud corresponding to the surrounding environment of the vehicle and the odometer information.
At the current moment, the vehicle processing device acquires sensing information corresponding to at least one key moment, namely second sensing information. The key moment is the moment when the second sensing information is acquired according to the time interval. The time interval between each key moment and the current moment is less than or equal to the preset time length.
For example, if the preset time period is 5 minutes, the time when the sensing device collects the environmental information and the odometer information around the vehicle before the current time and within 5 minutes from the current time is the critical time. As shown in fig. 4, after the vehicle is started, the sensing device collects environmental information and odometer information around the vehicle every 30S intervals, so that the vehicle obtains sensing information every 30S intervals. If the current time is 10:00, 9: after 55, the sensing information before 10:00 is the sensing information corresponding to the key moment. Wherein the vehicle is at 9:55, the vehicle arrives at the position 1 in fig. 4 in a time-sharing manner, the time interval between the moment when the vehicle arrives at the position 2 in fig. 4 and the moment when the last sensing device collects the environmental information and the odometer information around the vehicle is the key moment is 30s, and at the position 2 in fig. 4, the sensing device collects the environmental information and the odometer information around the vehicle, and the vehicle acquires the sensing information. Thus, in the vehicle at 10:00 reaches position 3 in fig. 4, 9: after 55, the key moment before 10:00 is the moment marked "it" in fig. 4.
Optionally, after the vehicle processing device obtains the sensing information each time, the sensing information is stored, and when the second sensing information is obtained at the current time, the second sensing information may be obtained from the stored sensing information according to the current time and the time when the sensing device corresponding to the stored sensing information acquires the environmental information and the odometer information around the vehicle.
Optionally, when the second sensing information is obtained at the current time, the second sensing information corresponding to each key time may be obtained from the key time list. For example, after the vehicle processing device obtains the sensing information each time, the sensing information is stored, and according to the key time corresponding to the sensing information and the key time corresponding to each stored sensing information, the sensing information in the key list, in which the time interval between the key times corresponding to the sensing information is longer than the preset time, is deleted, and the key time list is updated.
Optionally, the critical time further includes a second critical time, where the second critical time is a time when the second sensing information is acquired according to the distance interval, and a driving distance from a position of the second critical time to a position corresponding to the current time is less than or equal to a preset distance.
After the vehicle is started, the vehicle-mounted sensing device can continuously acquire the corresponding point cloud and the odometer information of the surrounding environment of the vehicle according to the distance interval. For example, the acquired distance interval is preset to be 100 meters, and the vehicle-mounted sensing device acquires point cloud and odometer information corresponding to the surrounding environment of the vehicle every time the vehicle runs for 100 meters. Correspondingly, the vehicle processing device acquires the sensing information once corresponding to the point cloud and the odometer information corresponding to the surrounding environment of the vehicle acquired by the vehicle-mounted sensing device each time.
In this way, at the current time, the vehicle processing device obtains the second sensing information corresponding to at least one second key time, for example, as shown in fig. 5, after the vehicle is started, the vehicle processing device obtains the sensing information every 100 meters of the vehicle, and the positions of the vehicle when the vehicle-mounted sensing device obtains the point cloud and the odometer information corresponding to the surrounding environment of the vehicle respectively correspond to S1-S7 in the graph. The preset distance is 500 meters, when the vehicle runs 50 meters from S7 to S8 in FIG. 5, the vehicle is positioned, the vehicle at the current moment is located at S8 in FIG. 5, and S3-S7 in FIG. 5 correspond to a second critical moment.
After the vehicle is started, the vehicle-mounted sensing device may continuously obtain the point cloud and the odometer information corresponding to the surrounding environment of the vehicle according to the time interval and the distance interval, for example, on the basis of fig. 4 and 5, the vehicle-mounted sensor collects the surrounding environment information and the odometer information of the vehicle every 30s, so that the vehicle obtains the sensing information every 30s, and the preset duration is 5 minutes. And the vehicle-mounted sensor collects the environmental information and the odometer information around the vehicle once every 100 meters of the vehicle, so that the vehicle obtains the sensing information once every 100 meters of the vehicle, and the preset distance is 500 meters. As shown in fig. 6, ". o" indicates the time when the point cloud corresponding to the surrounding environment of the vehicle and the odometer information are acquired according to the time interval, and "●" indicates the time when the point cloud corresponding to the surrounding environment of the vehicle and the odometer information are acquired according to the distance interval.
Optionally, in practical application, the time when the vehicle-mounted sensing device acquires the point cloud and the odometer information corresponding to the vehicle surrounding environment according to the time interval may be the same as the time when the vehicle-mounted sensing device acquires the point cloud and the odometer information corresponding to the vehicle surrounding environment according to the distance interval, and in fig. 6, "□" indicates the time when the vehicle-mounted sensing device continuously acquires the point cloud and the odometer information corresponding to the vehicle surrounding environment according to the time interval and the distance interval.
Corresponding to fig. 6, when the vehicle travels to position 1 in fig. 6 (i.e., position 1 in fig. 4) at 9:55:00, and to position 2 in fig. 6 (i.e., position 2 in fig. 4), the sensing device collects environmental information and odometer information of the surroundings of the vehicle, and the vehicle is driven at 10: 55:00 reaches S8 (i.e., position 3 in fig. 4), at the current time, the times corresponding to the first key time and the second key time are as shown in fig. 6, and the key time is the union of the first key time and the second key time.
It should be noted that the time critical point mentioned in the following may be a first time critical point, a second time critical point, or a union of the first time critical point and the second time critical point.
And S203, determining the position of the vehicle at the current moment according to the map information of the vehicle driving area, the first sensing information and the second sensing information.
In this step, after the first sensing information and the at least one second sensing information are acquired, the mark information of the surrounding environment of the vehicle at the current moment is acquired according to the first sensing information and the at least one second sensing information, and the mark information is compared with the map information of the driving area of the vehicle to acquire the position of the vehicle at the current moment.
Alternatively, the map information of the vehicle travel area is map information acquired from a position at the time of previous positioning, and for example, the map information of the vehicle travel area is map information acquired from a position at the time of last positioning.
When the vehicle is started, the position of the vehicle is initialized, for example, the initial position of the vehicle is located by GPS or is input. After the position of the vehicle is initialized, the map information of the current area where the vehicle is located is obtained according to the initial position of the vehicle.
After the vehicle is started, the position of the vehicle changes, and the map of the area where the vehicle is located also changes, so that after the vehicle is positioned, the map information needs to be updated according to the positioned position, and thus, compared with the map information before updating, the updated map information has higher matching degree with the area where the vehicle is located. Therefore, the vehicle can be positioned more accurately based on the updated map information at the next positioning.
Optionally, in S203, the pose of the vehicle at the current time is determined according to the map information of the vehicle driving area, the first sensing information and the second sensing information.
Specifically, for the same target object, the vehicle has different postures at the same position, for example, the vehicle has different orientations, and the vehicle body coordinate systems thereof are different, when the target object is scanned by the vehicle-mounted sensing device, for example, a laser radar, and the same laser beam is emitted to the target object, the position of the target object in the space is fixed, so that the coordinates of the characteristic point obtained by the laser beam are different in different vehicle body coordinate systems. Therefore, according to the map information of the driving area of the vehicle and the first sensing information and the at least one second sensing information, not only the position of the vehicle at the current moment but also the posture of the position of the vehicle can be determined.
For example, as shown in fig. 7a and 7b, for the same billboard at the roadside, the billboard information collected from different positions of the vehicle at the same position. When the vehicle scans the billboard by the lidar in the first direction, the position of at least one characteristic point of the billboard under the first body coordinate system according to the collected information is shown in fig. 7 a. When the vehicle scans the billboard by the lidar toward the second direction, the position of at least one characteristic point of the billboard under the second vehicle body coordinate system according to the collected information is shown in fig. 7 b.
When the advertising board under the first vehicle body coordinate system and the advertising board under the second vehicle body coordinate system are respectively matched with the advertising boards in the map information, the first vehicle body coordinate system and the second vehicle body coordinate system can be respectively determined, and the posture of the vehicle volume at the position is determined according to the first vehicle body coordinate system and the second vehicle body coordinate system.
In the embodiment, after the vehicle is started, the sensing information of the vehicle is acquired according to the time interval, when the vehicle is positioned, the first sensing information corresponding to the current moment and the second sensing information corresponding to at least one key moment are acquired, and the vehicle is positioned according to the first sensing information, the at least one second sensing information and the map information. The sensing information is acquired from the time dimension, so that the process of acquiring the sensing information is not influenced by road congestion, the sensing information for positioning can be updated according to time intervals when the vehicle runs, and the positioning accuracy is improved. And moreover, the sensing information is acquired from the time dimension, when the vehicle runs at a low speed, the environmental information around the vehicle can be acquired in the gap between the vehicles, and at the moment, the acquired environmental information around the vehicle is clearer due to the fact that the vehicle speed is lower, and the positioning accuracy is further improved.
Fig. 8 is a flowchart of a positioning method according to another embodiment of the present application. On the basis of the embodiment shown in fig. 2, as shown in fig. 8, the method of the embodiment of the present application includes:
s801, acquiring first sensing information through the vehicle-mounted sensing device at the current moment.
In this step, S201 may be referred to as an implementation manner of S801, and details are not described herein.
S802, second sensing information acquired through the vehicle-mounted sensing device at the key moment is acquired.
In this step, S202 may be referred to as an implementation manner of S802, and details are not described here.
And S803, respectively mapping the coordinates of the feature points acquired at the key moment to the vehicle body coordinate system at the current moment according to the odometer information in the first sensing information and the odometer information in the second sensing information, and acquiring the coordinates of the feature points acquired at the key moment in the vehicle body coordinate system at the current moment.
In this step, when the vehicle is powered on, the position of the vehicle when the vehicle is started is acquired and a vehicle-mounted sensing device, such as an inertia detection unit, is started. After the vehicle is started, the odometer information during the running of the vehicle, such as the steering information of the vehicle after the start, the number of turns of the wheels, the speed of the vehicle, etc., is recorded by the inertia detection unit.
Therefore, the relative position information of the position where the vehicle is located when the odometry information is recorded relative to the power-on position of the vehicle can be determined according to the odometry information of the vehicle.
Therefore, the travel distance and the rotation angle of the vehicle from the position at t1 to the position at t2 can be determined from the odometer information at t1 and the odometer information at t 2. That is, the relative position information of the position at t2 of the vehicle with respect to the position at t1 of the vehicle is determined by the odometer information at t1 and the odometer information at t 2.
Therefore, the coordinates of the at least one characteristic point acquired at the critical moment can be respectively mapped to the vehicle body coordinate system at the current moment according to the odometer information in the first sensing information and the odometer information in the second sensing information, and the coordinates of the at least one characteristic point acquired at the critical moment in the vehicle body coordinate system at the current moment can be obtained.
Optionally, a specific implementation manner of S803 is:
and S8031, determining a coordinate mapping relation between the vehicle body coordinate system corresponding to the current moment and the vehicle body coordinate system corresponding to the key moment according to the mileage information in the first sensing information and the mileage information in the second sensing information.
In this step, the vehicle turns left from the position where the key moment is located to the position where the current moment is located, and the vehicle turns left from the position where the key moment is located to the position where the current moment is located and is recorded in the odometer information of the vehicle. Therefore, the turning direction of the vehicle from the position of the key moment to the position of the current moment and the driving distance of the vehicle from the position of the key moment to the position of the current moment can be obtained through the odometer information of the key moment and the odometer information of the current moment.
Therefore, by using the odometer information at the current time and the odometer information at the critical time, a mapping matrix of the body coordinate system of the vehicle at t1 and the body coordinate system of the vehicle at the critical time can be calculated, and a formula such as formula one is calculated:
Tkey×Tkey→cur=Tcurformula one
Wherein, TkeyA coordinate matrix T representing a body coordinate system of the vehicle relative to a powered-on position of the vehicle at any one of the critical timescurA coordinate matrix T representing a body coordinate system of the vehicle at the present time with respect to a powered-on position of the vehiclekey→curAnd the mapping matrix represents the body coordinate system of the vehicle at the key moment relative to the body coordinate system of the vehicle at the current moment.
Therefore, the coordinate mapping relationship between the vehicle body coordinate system corresponding to the current time and the vehicle body coordinate system corresponding to the critical time is determined based on the odometer information at the critical time and the odometer information at the current time.
S8032, respectively mapping the coordinates of the feature points acquired at the first key moment to the vehicle body coordinate system at the current moment according to the coordinate mapping relation, and acquiring the coordinates of the feature points acquired at the first key moment in the vehicle body coordinate system at the current moment.
In this step, the mapping matrix is known according to formula one
Figure BDA0002713024620000121
For one feature point of one key moment, the coordinates mapped to the current moment under the vehicle body coordinate system are obtained through a formula II:
Pcur=Tkey→cur×Pkeyformula two
Wherein, PcurThe coordinate, P, of the feature point of one of the key moments in the body coordinate system of the current momentkeyAnd the coordinates of the characteristic point of one key moment in the vehicle body coordinate system of the key moment are represented.
S804, determining the position of the vehicle at the current moment according to the coordinates of the feature points acquired at each key moment and the coordinates of the feature points acquired at the current moment under the vehicle body coordinate system at the current moment and the map information.
In this step, for the sensing information at any key moment, the coordinates of at least one feature point acquired at the key moment are mapped to the vehicle body coordinate system at the current moment according to the corresponding mapping matrix, so that the position of the vehicle at the current moment is determined according to the feature points (including the feature points acquired by mapping at least one feature point acquired at each key moment to the vehicle body coordinate system at the current moment and the at least one feature point acquired at the current moment) and the map information in the vehicle body coordinate system at the current moment.
The coordinates of at least one characteristic point acquired at the key moment are mapped to the vehicle body coordinate system at the current moment according to the corresponding mapping matrix, so that the number of the characteristic points at the current moment can be increased. Therefore, the contour of the target object around the vehicle is extracted through the characteristic points and is more clear and closer to the real contour, so that the matching degree between the contour of the target object extracted through the characteristic points and the contour of the target object in the map information is higher when positioning is carried out according to the map information, and the positioning accuracy is improved.
Fig. 9 is a flowchart of a positioning method according to another embodiment of the present application. On the basis of the embodiment shown in fig. 8, the method of the embodiment of the present application includes:
and S901, acquiring first sensing information through the vehicle-mounted sensing device at the current moment.
In this step, S201 may be referred to as an implementation manner of S901, which is not described herein again.
And S902, acquiring second sensing information acquired by the vehicle-mounted sensing device at the key moment.
In this step, S202 may be referred to as an implementation manner of S902, and details are not described herein.
And S903, mapping the coordinates of the feature points acquired at the key moment to the vehicle body coordinate system at the current moment according to the odometer information in the second sensing information and the odometer information in the second sensing information, and acquiring the coordinates of the feature points acquired at the key moment in the vehicle body coordinate system at the current moment.
In this step, reference may be made to S803 for implementation of S903, which is not described herein again.
And S904, aiming at each particle in M particles corresponding to the vehicle at the current moment, obtaining the matching degree of the feature points acquired at the current moment and the key moment and the target object in the map information according to the coordinates of the feature points acquired at the key moment and the feature points acquired at the current moment under the vehicle body coordinate system at the current moment and the position of each particle.
The method comprises the steps that M particles are used for positioning a vehicle last time, the position of each particle is obtained according to odometer information obtained when the vehicle is positioned last time, odometer information at the current moment and the position of each particle obtained when the vehicle is positioned last time, and M is a positive integer.
In this step, a framework for particle filtering may be used when positioning the vehicle. The method specifically comprises the following steps:
as shown in fig. 10, an initial position of the vehicle is acquired immediately after the vehicle starts, and some particles, i.e., initialization particles, are arranged around the vehicle, for example, the positions of the initialization particles satisfy a normal distribution. Wherein each particle is used for simulating a vehicle, the position of the particle simulates the position of the vehicle, and the position of each particle is recorded. At this time, the position of the particle is the initialized position
After the vehicle starts, the motion process of each particle is consistent with the motion process of the vehicle. The movement information of the vehicle traveling from one location to another, such as the distance traveled by the vehicle, the direction of the turn during travel, etc., may be determined from the odometer information of the vehicle. Therefore, the movement process of the particles can be predicted from the odometry information of the vehicle, thereby predicting the positions of the particles.
And for M particles at the current moment, wherein the M particles are the particles adopted in the last positioning of the vehicle, and the position of each particle at the current moment is determined according to the position of each particle in the M particles in the last positioning of the vehicle, the odometry information of the vehicle in the last positioning and the odometry information at the current moment.
At the current moment, according to the feature points (including the feature points acquired by mapping at least one feature point acquired at each key moment on the vehicle body coordinate system at the current moment and at least one feature point acquired at the current moment) in the vehicle body coordinate system at the current moment, the environmental information around the vehicle at the current moment can be determined, that is, according to the feature points in the vehicle body coordinate system at the current moment, the mark information around the vehicle, such as the target object to be matched around the vehicle and the position of the target object to be matched relative to the vehicle, can be extracted.
The characteristic points under the vehicle body coordinate system at the current moment are actually characteristic points acquired by the vehicle at the real position through a laser radar, so that the target object to be matched around the vehicle, which is obtained according to the characteristic points under the vehicle body coordinate system at the current moment, is actually the target object to be matched observed at the real position of the vehicle, and the coordinates of the target object to be matched under the vehicle body coordinate system are actually the relative position of the target object to be matched with the real position of the vehicle. Therefore, under the optimal condition, the target object to be matched extracted from the feature points in the vehicle body coordinate system at the current moment is completely matched with the target object in the map information.
Therefore, since each particle is a real position of the simulated vehicle, and the odometer information recorded in the vehicle does not faithfully reflect the real movement process of the vehicle, for example, the driving distance of the vehicle obtained from the odometer information is different from the actual driving distance of the vehicle. Therefore, when the position of the particle at the present time is determined from the odometer information, there is a difference between the position of the particle and the true position of the vehicle.
Therefore, for each particle, according to the feature point under the vehicle body coordinate system at the current moment and the position of the particle at the current moment, the difference between the position of the particle and the real position of the vehicle can be determined through the matching degree of the target object to be matched, which is extracted from the feature point under the vehicle body coordinate system at the current moment, and the target object in the map information.
Optionally, a specific implementation manner of S904 is as follows:
s9041, respectively mapping coordinates of the feature points acquired at the current moment and the key moment in a body coordinate system at the current moment to a particle coordinate system with the particles as origins.
Specifically, each particle is selected one by one, for each particle, the particle coordinate system with the particle as the origin of coordinates replaces the vehicle body coordinate system at the current moment, the feature point under the vehicle body coordinate system at the current moment is mapped into the particle coordinate system, and the coordinate of the feature point under the vehicle body coordinate system at the current moment mapped into the particle coordinate system is obtained.
The particle coordinate system may be a planar coordinate system, such as a planar rectangular coordinate system, a polar coordinate system, or a spatial coordinate system, such as a spatial rectangular coordinate system, a cylindrical coordinate system, or a spherical coordinate system, which is not limited in this application. The present embodiment is described by taking a spatial rectangular coordinate system as an example.
Optionally, the vehicle body coordinate system and the particle coordinate system may be the same type of coordinate system, for example, both the vehicle body coordinate system and the particle coordinate system are rectangular spatial coordinate systems, or the vehicle body coordinate system and the particle coordinate system may be different types of coordinate systems, for example, the vehicle body coordinate system is a rectangular spatial coordinate system, and both the particle coordinate systems are spherical coordinate systems, which is not limited in this embodiment of the present application.
And S9042, obtaining the matching degree of a target object to be matched, which is formed by the feature points acquired at the current moment and the key moment and mapped to the particle coordinate system, and the corresponding target object in the map information.
Specifically, according to feature points acquired at the current moment and at least one key moment mapped to a particle coordinate system, target objects to be matched around a vehicle at the current moment are extracted, the positions of the particles in the map information are determined in the map information corresponding to the positions of the particles, the positions of the particles in the map information are used as coordinate origin points, the target objects to be matched are matched with corresponding target objects in the map information, and the matching degree of the target objects to be matched and the corresponding target objects in the map information is determined according to the coincidence degree of the target objects to be matched and the corresponding target objects in the map information.
Since the motion of the particle is consistent with the motion of the vehicle, if the position of the particle is closer to the real position of the vehicle, the matching degree between the target object to be matched, which is composed of the feature points mapped to the current time of the particle coordinate system in the vehicle body coordinate system, and the corresponding target object in the map information is higher.
As shown in fig. 11, the vehicle coordinate system and the particle coordinate system are a spatial rectangular coordinate system (the spatial rectangular coordinate system is not shown in fig. 11), the real position of the vehicle in the map information is point O, the position of one of the particles is point O1, and the billboard and the position of the roadside are shown by a solid line in fig. 11, which are formed from the feature points in the vehicle body coordinate system at the present time. The billboard shown by the solid line is overlapped with the billboard in the map information.
After the feature points in the vehicle body coordinate system at the current time are mapped into the particle coordinate system, the billboard and the positions formed by the feature points are shown by dotted lines in fig. 11.
If the position of the particle is the true position of the vehicle, the billboard position shown by the solid line coincides with the billboard position shown by the dashed line in FIG. 11. The matching degree of the particles is determined according to the distance between the position of the billboard shown by the solid line and the position of the billboard shown by the dotted line, and the smaller the distance, the higher the matching degree.
Optionally, at least one feature point acquired at each key moment is mapped to the vehicle body coordinate system at the current moment, so that more feature points are acquired in the vehicle body coordinate system at the current moment. Moreover, when the environmental information around the vehicle is collected at different key moments, repeated characteristic points can be collected, so that the calculated amount is increased, and the positioning efficiency is influenced. Therefore, before S904, the method further includes:
s1001, determining a target feature point according to at least one feature point acquired at each key moment and a feature value of each feature point in the feature points acquired at the current moment.
In this step, stable and representative target feature points are selected from the feature points in the vehicle body coordinate system at the current time (including the feature points acquired by mapping at least one feature point acquired at each key time on the vehicle body coordinate system at the current time and at least one feature point acquired at the current time), and the target feature points are stored. On the basis of reducing the number of the feature points, the environmental information around the vehicle extracted by the target feature point is made to approach the environmental information around the vehicle extracted by the feature point in the vehicle body coordinate system at the present time.
Optionally, the feature value of each feature point in the feature points in the vehicle body coordinate system at the current time is calculated according to the evaluation function, and one or more feature points are selected as the target feature points according to the feature value of each feature point.
Optionally, the feature value of each feature point in the feature points in the vehicle body coordinate system at the current time is calculated according to the evaluation function, and one or more feature points with the largest feature value are selected as the target feature points according to the size of the feature values.
S1002, determining the target feature points again according to the feature points acquired at each key moment and the feature values of the feature points except the feature points included in the feature point set in the feature points acquired at the current moment.
The feature points included in the feature point set are the feature points located in the preset distance range of the target feature point determined last time and the target feature point determined last time.
In this step, after the target feature point is determined in S1001, since the target feature point has stability and representativeness, in order to reduce the number of feature points, feature points within a preset distance range from the target feature point and the target feature point may be deleted from the feature points in the vehicle body coordinate system at the current time, so as to obtain a feature point set.
Then, S1001 and S1002 are repeated, the target feature point is determined again from the feature point set, and a new feature point set is obtained based on the determined target feature point.
S1003, when the number of the target characteristic points meets the preset number, N target characteristic points are obtained, wherein N is equal to the preset number.
In this step, when the number of the target feature points meets a preset number, the acquisition of new target feature points is stopped, and the vehicle is positioned according to the currently acquired target feature points.
S905, obtaining K particles according to the matching degree of each particle in the M particles at the current moment and the M particles.
Wherein K is a positive integer.
In this step, although the movement process of the particles is consistent with the movement process of the vehicle, the odometer information recorded in the vehicle does not faithfully reflect the real movement process of the vehicle, for example, the driving distance of the vehicle obtained according to the odometer information is different from the actual driving distance of the vehicle. Therefore, when the position of the particle at the current time is determined from the odometer information, the position of the particle with respect to the vehicle at the current time is different from the position of the particle with respect to the vehicle at the last positioning. Therefore, the matching degree of the particles varies with the movement of the vehicle. That is, the matching degree of the particle having the high matching degree at the previous time may be lowered.
Therefore, it is necessary to resample the particles according to the matching degree of each of the M particles, i.e., to re-determine the K particles for vehicle positioning.
Optionally, the matching degree of the particle represents the degree of closeness of the position of the particle to the real position of the vehicle, and the higher the matching degree is, the closer the position of the particle is to the real position of the vehicle. Therefore, the particles with the higher M particles are selected and determined as the K particles, for example, a preset matching degree is set, L particles with a matching degree greater than or equal to the preset matching degree among the M particles are obtained, and the particles are determined as the K particles.
Optionally, after L particles with a matching degree greater than or equal to a preset matching degree are obtained from the M particles, at least one particle may be selected from the L particles, and K particles are obtained according to the L particles and the selected at least one particle.
Optionally, after at least one particle is selected from the L particles, the particle may be copied at least once according to the matching degree of each particle in the at least one particle, that is, the number of the particles is increased, for example, the higher the matching degree of the particle is, the more the number of the particle is after the increase is, and K particles are obtained according to the L particles and the increased particles.
And S906, obtaining the position of the vehicle at the current moment according to the positions of the K particles corresponding to the vehicle at the current moment.
In this step, since the position of each particle simulates the real position of the vehicle, the position of the vehicle at the current time can be determined according to the positions of the K particles at the current time.
Optionally, the average positions of the K particles are obtained according to the positions of the K particles, and the average position of the K particles is determined as the position of the vehicle at the current moment.
Optionally, the weighted average position of the K particles is obtained according to the matching degree of each particle of the positions of the K particles, and the weighted average position of the K particles is determined as the position of the vehicle at the current moment.
And when the vehicle is positioned next time, replacing the M particles in the S1004 by the K particles determined at the current moment, and repeating the steps to position the vehicle.
Optionally, according to the positions of the K particles, the obtained position of the vehicle at the current time is not the actual position of the vehicle at the current time. Therefore, if the difference between the obtained position of the vehicle at the current time and the actual position of the vehicle at the current time is large according to the positions of the K particles, a potential safety hazard exists. Therefore, the method of the embodiment of the present application further includes:
and S907, obtaining an evaluation value of the position of the vehicle at the current time according to the position of each particle in the K particles corresponding to the vehicle at the current time and the position of the vehicle at the current time.
Wherein the evaluation value represents a difference between the position of the vehicle at the present time and the true position information of the vehicle at the present time.
In this step, the distribution variance of the K particles is calculated from the position of each of the K particles corresponding to the vehicle at the current time and the position of the vehicle at the current time, and the evaluation value of the position of the vehicle at the current time is obtained from the distribution variance. For example, the distribution variance is determined as an evaluation value of the position of the vehicle at the current time. The larger the distribution variance, the larger the difference between the position of the vehicle at the present time and the true position of the vehicle is. Therefore, it is possible to determine whether to remotely control the vehicle, or to activate another positioning device to position the vehicle, for example, according to the evaluation value.
Fig. 12 is a schematic structural diagram of a positioning device according to an embodiment of the present application. As shown in fig. 12, the positioning device may be a vehicle as described above, or may be a component of a vehicle (e.g., an integrated circuit, a chip, etc.). The locating device may also be a server, such as a cloud server, or a component of a server (e.g., an integrated circuit, a chip, etc.). The positioning device may also be another communication module, which is used to implement the method in the embodiment of the method of the present application and the above-mentioned optional embodiments. The positioning device may include: an acquisition module 1201 and a positioning module 1202.
An obtaining module 1201, configured to perform:
for specific reference to detailed description in the method example, details of any optional embodiment of S201 and S202 and any optional embodiment of S201 and S202 in the embodiment shown in fig. 2, and details of any optional embodiment of S801 and S802 and any optional embodiment of S801 and S802 in the embodiment shown in fig. 8 are not described herein again.
A positioning module 1202 for performing:
s203 and any optional embodiment thereof in the embodiment shown in fig. 2, any optional embodiment of S803, S804 and S803, S804 in the embodiment shown in fig. 8, and any optional embodiment of S903 to S907 and S903 to S907 in the embodiment shown in fig. 9. For details, reference is made to the detailed description in the method example, which is not repeated herein.
The obtaining module 1201 is configured to obtain first sensing information at a current time through a vehicle-mounted sensing device, where the first sensing information includes coordinates of a feature point acquired by the vehicle-mounted sensing device at the current time in a vehicle body coordinate system and odometer information of a vehicle at the current time.
The obtaining module 1201 is further configured to obtain second sensing information obtained by the vehicle-mounted sensing device at the key moment, where the second sensing information includes coordinates of a feature point, collected by the vehicle-mounted sensing device at the key moment, in the vehicle body coordinate system and odometer information of the vehicle at the key moment. The key time comprises a first key time, the first key time is the time for acquiring the second sensing information according to the time interval, and the time interval between the first key time and the current time is less than or equal to the preset time length.
The positioning module 1202 is configured to determine a position of the vehicle at the current time according to the map information of the driving area of the vehicle, the first sensing information, and the second sensing information.
It should be understood that the positioning device in the embodiments of the present application may be implemented by software, for example, a computer program or instructions having the above-mentioned functions, and the corresponding computer program or instructions may be stored in a memory inside the terminal, and the corresponding computer program or instructions inside the memory may be read by the processor to implement the above-mentioned functions.
Alternatively, the positioning apparatus in the embodiment of the present application may also be implemented by hardware. The obtaining module 1201 and the positioning module 1202 are processors (e.g., NPUs, GPUs, processors in a system chip).
Alternatively, the positioning device in the embodiment of the present application may also be implemented by a combination of a processor and a software module.
Optionally, the positioning module 1202 is further configured to map, according to the odometer information in the first sensing information and the odometer information in the second sensing information, the coordinates of the feature points acquired at the key moment to the vehicle body coordinate system at the current moment, and obtain the coordinates of the feature points acquired at the key moment in the vehicle body coordinate system at the current moment.
And determining the position of the vehicle at the current moment according to the coordinates of the characteristic points acquired at the key moment and the characteristic points acquired at the current moment under the vehicle body coordinate system at the current moment and the map information.
Optionally, the positioning module 1202 is further configured to, at the current time, obtain, for each particle in M particles corresponding to the vehicle, a matching degree between the feature point acquired at the current time and the target object in the map information according to the feature point acquired at the key time and the coordinate of the feature point acquired at the current time in the vehicle body coordinate system at the current time and the position of each particle, where the M particles are particles used when the vehicle is last positioned, the position of each particle is obtained according to the odometer information when the vehicle is last positioned, the odometer information at the current time, and the position of each particle when the vehicle is last positioned, and M is a positive integer;
obtaining K particles according to the matching degree of each particle in the M particles at the current moment and the M particles, wherein K is a positive integer;
and obtaining the position of the vehicle at the current moment according to the positions of the K particles corresponding to the vehicle at the current moment.
Optionally, the positioning module 1202 is further configured to map coordinates of the feature points acquired at the current time and the key time in the vehicle body coordinate system at the current time to a particle coordinate system with each particle as an origin;
and obtaining the matching degree of the target object to be matched, which is formed by the characteristic points acquired at the current moment and the key moment and mapped to each particle coordinate system, and the corresponding target object in the map information.
Optionally, the positioning module 1202 is further configured to obtain L particles, of the M particles, whose matching degree is greater than or equal to a preset matching degree, where L is less than or equal to M and less than or equal to K;
determining L particles as K particles; or,
obtaining at least one particle from the L particles;
k particles are obtained from the L particles and the at least one particle.
Optionally, the positioning module 1202 is further configured to copy each particle at least once according to the matching degree of each particle in at least one particle;
the L particles and the particles obtained after the replication were determined as K particles.
Optionally, the positioning module 1202 is further configured to obtain an average position of the K particles according to a position of each particle of the K particles, and determine the average position of the K particles as a position of the vehicle at the current time.
Optionally, the positioning module 1202 is further configured to obtain an evaluation value of the position of the vehicle at the current time according to the position of each particle of the K particles corresponding to the vehicle at the current time and the position of the vehicle at the current time. Wherein the evaluation value indicates a difference between the position of the vehicle at the current time and actual position information of the vehicle at the current time.
Optionally, the positioning module 1202 is further configured to determine, according to the mileage information in the first sensing information and the mileage information in the second sensing information, a coordinate mapping relationship between the vehicle body coordinate system corresponding to the current time and the vehicle body coordinate system corresponding to the first key time;
and respectively mapping the coordinates of the characteristic points acquired at the first key moment to the vehicle body coordinate system at the current moment according to the coordinate mapping relation, and acquiring the coordinates of the characteristic points acquired at the first key moment in the vehicle body coordinate system at the current moment.
Optionally, the critical time further includes a second critical time, where the second critical time is a time when the second sensing information is acquired according to the distance interval, and a driving distance from a position of the second critical time to a position corresponding to the current time is less than or equal to a preset distance.
Optionally, the positioning module 1202 is further configured to determine the pose of the vehicle at the current time according to the map information and the first sensing information and the second sensing information.
Optionally, the positioning module 1202 is further configured to determine a target feature point according to the feature points acquired at each key time and a feature value of each feature point in the feature points acquired at the current time;
determining the target characteristic points again according to the characteristic points acquired at each key moment and the characteristic values of the characteristic points, except the characteristic points contained in the characteristic point set, in the characteristic points acquired at the current moment, wherein the characteristic points contained in the characteristic point set are the characteristic points located in the preset distance range of the target characteristic points determined last time and the target characteristic points determined last time;
when the number of the target characteristic points meets the preset number, N target characteristic points are obtained, wherein N is equal to the preset number;
correspondingly, the positioning module 1202 is further configured to obtain matching degrees between the feature points acquired at the current time and the key time and the target object in the map information of the target area according to the coordinates of the N target feature points in the vehicle body coordinate system at the current time and the position of each particle.
Optionally, the feature value of the feature point is obtained through an evaluation function, and the feature value is used for evaluating the stability of the feature point.
Optionally, the target feature point is a feature point with a maximum feature value among the feature points.
Optionally, the origin of the vehicle coordinate system is located at any position on the vehicle.
The apparatus of this embodiment may be configured to implement the technical solution in the foregoing method embodiment, and the implementation principle and the technical effect are similar, which are not described herein again.
Fig. 13 is a schematic structural diagram of a positioning device according to an embodiment of the present application. The positioning device of the present embodiment may be the vehicle as described above, or may be a server. The positioning device may be used to implement the method described in the above method embodiment, and specific reference may be made to the description in the above method embodiment.
The positioning device may include one or more processors 1301, and the processors 1301 may also be referred to as processing units, which may implement certain control functions. The processor 1301 may be a general purpose processor, a special purpose processor, or the like.
Optionally, the processor 1301 may also store instructions and/or data 1303, and the instructions and/or data 1303 may be executed by the processor, so that the positioning apparatus executes the method described in the foregoing method embodiment.
Optionally, the positioning apparatus may include one or more memories 1302, on which instructions 1304 may be stored, and the instructions may be executed on the processor, so that the positioning apparatus performs the method described in the above method embodiments. Optionally, the memory may further store data therein. Optionally, instructions and/or data may also be stored in the processor. The processor and the memory may be provided separately or may be integrated together. For example, the correspondence described in the above method embodiments may be stored in a memory or in a processor.
The processor and transceiver described in this embodiment may be fabricated using various IC process technologies, such as Complementary Metal Oxide Semiconductor (CMOS), N-type metal oxide semiconductor (NMOS), P-type metal oxide semiconductor (PMOS), Bipolar Junction Transistor (BJT), Bipolar CMOS (bicmos), silicon germanium (SiGe), gallium arsenide (GaAs), and the like.
It should be understood that the processor in the embodiments of the present application may be an integrated circuit chip having signal processing capability. In implementation, the steps of the above method embodiments may be performed by integrated logic circuits of hardware in a processor or instructions in the form of software. The processor may be a general purpose processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or other programmable logic device, discrete gate or transistor logic device, or discrete hardware components.
It will be appreciated that the memory in the embodiments of the subject application can be either volatile memory or nonvolatile memory, or can include both volatile and nonvolatile memory. The non-volatile memory may be a read-only memory (ROM), a Programmable ROM (PROM), an Erasable PROM (EPROM), an electrically Erasable EPROM (EEPROM), or a flash memory. Volatile memory can be Random Access Memory (RAM), which acts as external cache memory. By way of example, but not limitation, many forms of RAM are available, such as Static Random Access Memory (SRAM), Dynamic Random Access Memory (DRAM), Synchronous Dynamic Random Access Memory (SDRAM), double data rate SDRAM, enhanced SDRAM, SLDRAM, Synchronous Link DRAM (SLDRAM), and direct rambus RAM (DR RAM). It should be noted that the memory of the systems and methods described herein is intended to comprise, without being limited to, these and any other suitable types of memory.
The scope of the positioning device described in the embodiment of the present application is not limited thereto, and the structure of the positioning device may not be limited by fig. 13. The positioning means described in the embodiments of the present application may be a stand-alone device or may be part of a larger device.
Fig. 14 is a schematic structural diagram of a positioning system according to an embodiment of the present application. As shown in fig. 14, the positioning system includes: a vehicle 1401 and an onboard sensing device 1402.
The in-vehicle sensor device 1402 is mounted on the vehicle 1401, and communicates with the vehicle 1401.
Optionally, the vehicle-mounted sensing device 1402 is configured to collect environmental information and odometer information around the vehicle 1401, and send the environmental information and the odometer information around the vehicle 1401 to the vehicle 1401.
A vehicle 1401 for carrying out the method described in the above method embodiments.
Optionally, the vehicle-mounted sensing device 1402 is configured to collect environmental information and odometer information around the vehicle 1401, send the environmental information and the odometer information around the vehicle 1401 to the vehicle 1401, and perform the method described in the foregoing method embodiment to locate the vehicle 1401.
The system of this embodiment may be configured to implement the technical solution in any of the method embodiments, and the implementation principle and the technical effect are similar, which are not described herein again.
The present application also provides a computer-readable medium having stored thereon a computer program which, when executed by a computer, implements the method as shown in any of the method embodiments described above.
The present application also provides a computer program product which, when executed by a computer, implements the method shown in any of the above method embodiments.
Where the above embodiments are implemented using software, they may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. When loaded and executed on a computer, cause the processes or functions described in accordance with the embodiments of the application to occur, in whole or in part. The computer may be a general purpose computer, a special purpose computer, a network of computers, or other programmable device. The computer instructions may be stored on a computer readable storage medium or transmitted from one computer readable storage medium to another, for example, from one website, computer, server, or data center to another website, computer, server, or data center via wire (e.g., coaxial cable, fiber optic, Digital Subscriber Line (DSL)) or wireless (e.g., infrared, wireless, microwave, etc.). The computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device, such as a server, a data center, etc., that incorporates one or more of the available media. The usable medium may be a magnetic medium (e.g., a floppy disk, a hard disk, a magnetic tape), an optical medium (e.g., a Digital Video Disk (DVD)), or a semiconductor medium (e.g., a Solid State Disk (SSD)), among others.

Claims (33)

1. A method of positioning, comprising:
at the current moment, acquiring first sensing information through a vehicle-mounted sensing device, wherein the first sensing information comprises coordinates of characteristic points acquired by the vehicle-mounted sensing device at the current moment under a vehicle body coordinate system and odometer information of a vehicle at the current moment;
acquiring second sensing information acquired through the vehicle-mounted sensing device at a key moment, wherein the second sensing information comprises coordinates of characteristic points acquired by the vehicle-mounted sensing device at the key moment under a vehicle body coordinate system and odometer information of the vehicle at the key moment, the key moment comprises a first key moment, the first key moment is a moment of acquiring second sensing information according to a time interval, and the time interval between the first key moment and the current moment is less than or equal to a preset time length;
and determining the position of the vehicle at the current moment according to the map information of the driving area of the vehicle, the first sensing information and the second sensing information.
2. The method of claim 1, wherein determining the position of the vehicle at the current time based on the map information of the vehicle travel area and the first and second sensory information comprises:
mapping the coordinates of the feature points acquired at the key moment to the vehicle body coordinate system at the current moment according to the odometer information in the first sensing information and the odometer information in the second sensing information, and acquiring the coordinates of the feature points acquired at the key moment in the vehicle body coordinate system at the current moment;
and determining the position of the vehicle at the current moment according to the coordinates of the feature points acquired at the key moment and the feature points acquired at the current moment under the vehicle body coordinate system at the current moment and the map information.
3. The method according to claim 2, wherein the determining the position of the vehicle at the current time according to the coordinates of the feature points acquired at the critical time and the feature points acquired at the current time under the vehicle body coordinate system at the current time and the map information respectively comprises:
at the current moment, for each particle in M particles corresponding to the vehicle, obtaining matching degrees of the feature points acquired at the current moment and the key moment and a target object in the map information according to coordinates of the feature points acquired at the key moment and the feature points acquired at the current moment under a vehicle body coordinate system at the current moment and the position of each particle, respectively, wherein the M particles are particles adopted when the vehicle is positioned last time, the position of each particle is obtained according to odometer information when the vehicle is positioned last time, the odometer information at the current moment and the position of each particle when the vehicle is positioned last time, and M is a positive integer;
obtaining K particles according to the matching degree of each particle in the M particles at the current moment and the M particles, wherein K is a positive integer;
and obtaining the position of the vehicle at the current moment according to the positions of the K particles corresponding to the vehicle at the current moment.
4. The method according to claim 3, wherein the obtaining of the matching degree between the feature points acquired at the current time and the target object in the map information according to the coordinates of the feature points acquired at the key time and the feature points acquired at the current time under the vehicle coordinate system at the current time and the position of each particle respectively comprises:
respectively mapping the coordinates of the feature points acquired at the current moment and the key moment under the vehicle body coordinate system at the current moment to a particle coordinate system with the particles as the origin;
and obtaining the matching degree of a target object to be matched, which is formed by the characteristic points acquired at the current moment and the key moment and mapped to the particle coordinate system, and the corresponding target object in the map information.
5. The method according to claim 3, wherein obtaining K particles according to the matching degree of each of the M particles at the current time and the M particles comprises:
obtaining L particles with a matching degree greater than or equal to a preset matching degree in the M particles, wherein L is less than or equal to M and less than or equal to K;
determining the L particles as the K particles; or,
obtaining at least one particle from the L particles;
obtaining the K particles according to the L particles and the at least one particle.
6. The method of claim 5, wherein obtaining the K particles from the L particles and the at least one particle comprises:
according to the matching degree of each particle in the at least one particle, copying each particle at least once;
and determining the L particles and the particles obtained after copying as the K particles.
7. The method according to any one of claims 3-6, wherein the obtaining the position of the vehicle at the current time according to the positions of the K particles corresponding to the vehicle at the current time comprises:
and acquiring the average position of the K particles according to the position of each particle in the K particles, and determining the average position of the K particles as the position of the vehicle at the current moment.
8. The method according to any one of claims 3-7, further comprising:
obtaining an evaluation value of the position of the vehicle at the current moment according to the position of each of the K particles corresponding to the vehicle at the current moment and the position of the vehicle at the current moment, wherein the evaluation value indicates a difference between the position of the vehicle at the current moment and real position information of the vehicle at the current moment.
9. The method according to any one of claims 2 to 8, wherein the obtaining coordinates of the feature points acquired at the key time in the vehicle body coordinate system at the current time by mapping the coordinates of the feature points acquired at the key time to the vehicle body coordinate system at the current time according to the odometer information in the second sensing information and the odometer information in the second sensing information comprises:
determining a coordinate mapping relation between the vehicle body coordinate system corresponding to the current moment and the vehicle body coordinate system corresponding to the first key moment according to the mileage information in the first sensing information and the mileage information in the second sensing information;
and according to the coordinate mapping relation, respectively mapping the coordinates of the characteristic points acquired at the first key moment to the vehicle body coordinate system at the current moment, and acquiring the coordinates of the characteristic points acquired at the first key moment in the vehicle body coordinate system at the current moment.
10. The method according to any one of claims 1 to 9, wherein the critical time further includes a second critical time, the second critical time is a time when second sensing information is acquired according to a distance interval, and a travel distance from a position of the second critical time to a position corresponding to the current time is less than or equal to a preset distance.
11. The method according to any one of claims 1 to 10, wherein the determining the position of the vehicle at the current time based on the map information of the vehicle travel area and the first and second sensing information comprises:
and determining the pose of the vehicle at the current moment according to the map information and the first and second sensing information.
12. The method according to claim 3, wherein before obtaining the matching degree between the feature points acquired at the current time and the target object in the map information according to the coordinates of the feature points acquired at the key time and the feature points acquired at the current time under the vehicle coordinate system at the current time and the position of each particle, the method further comprises:
determining target feature points according to the feature points acquired at each key moment and the feature values of each feature point in the feature points acquired at the current moment;
determining target feature points again according to feature points acquired at each key moment and feature values of the feature points acquired at the current moment except feature points contained in a feature point set, wherein the feature points contained in the feature point set are feature points located in a preset distance range of the last determined target feature point and the last determined target feature point;
when the number of the target characteristic points meets a preset number, N target characteristic points are obtained, wherein N is equal to the preset number;
the obtaining of the matching degree between the feature points acquired at the current moment and the key moment and the target object in the map information according to the coordinates of the feature points acquired at the key moment and the feature points acquired at the current moment under the vehicle body coordinate system at the current moment and the position of each particle respectively includes:
and according to the coordinates of the N target characteristic points in the vehicle body coordinate system at the current moment and the position of each particle, obtaining the matching degree of the characteristic points acquired at the current moment and the key moment and the target object in the map information of the target area.
13. The method according to claim 12, wherein the feature value of the feature point is obtained by an evaluation function, and the feature value is used to evaluate the stability of the feature point.
14. The method according to claim 12 or 13, wherein the target feature point is a feature point having a largest feature value among the feature points.
15. The method according to any of claims 1-14, wherein the origin of the body coordinate system is located at any position on the vehicle.
16. A positioning device, comprising:
the acquisition module is used for acquiring first sensing information through the vehicle-mounted sensing device at the current moment, wherein the first sensing information comprises coordinates of characteristic points acquired by the vehicle-mounted sensing device at the current moment under a vehicle body coordinate system and odometer information of a vehicle at the current moment;
the acquisition module is further configured to acquire second sensing information acquired by the vehicle-mounted sensing device at a key moment, where the second sensing information includes coordinates of feature points acquired by the vehicle-mounted sensing device at the key moment in a vehicle body coordinate system and odometer information of the vehicle at the key moment, the key moment includes a first key moment, the first key moment is a moment at which second sensing information is acquired according to a time interval, and a time interval between the first key moment and the current moment is less than or equal to a preset duration;
and the positioning module is used for determining the position of the vehicle at the current moment according to the map information of the driving area of the vehicle, the first sensing information and the second sensing information.
17. The apparatus according to claim 16, wherein the positioning module, when determining the position of the vehicle at the current time according to the map information of the driving area of the vehicle and the first and second sensing information, is specifically configured to:
mapping the coordinates of the feature points acquired at the key moment to the vehicle body coordinate system at the current moment according to the odometer information in the first sensing information and the odometer information in the second sensing information, and acquiring the coordinates of the feature points acquired at the key moment in the vehicle body coordinate system at the current moment;
and determining the position of the vehicle at the current moment according to the coordinates of the feature points acquired at the key moment and the feature points acquired at the current moment under the vehicle body coordinate system at the current moment and the map information.
18. The apparatus according to claim 17, wherein the positioning module, when determining the position of the vehicle at the current time according to the coordinates of the feature points acquired at the key time and the feature points acquired at the current time in the body coordinate system at the current time and the map information, is specifically configured to:
at the current moment, for each particle in M particles corresponding to the vehicle, obtaining matching degrees of the feature points acquired at the current moment and the key moment and a target object in the map information according to coordinates of the feature points acquired at the key moment and the feature points acquired at the current moment under a vehicle body coordinate system at the current moment and the position of each particle, respectively, wherein the M particles are particles adopted when the vehicle is positioned last time, the position of each particle is obtained according to odometer information when the vehicle is positioned last time, the odometer information at the current moment and the position of each particle when the vehicle is positioned last time, and M is a positive integer;
obtaining K particles according to the matching degree of each particle in the M particles at the current moment and the M particles, wherein K is a positive integer;
and obtaining the position of the vehicle at the current moment according to the positions of the K particles corresponding to the vehicle at the current moment.
19. The apparatus according to claim 18, wherein the positioning module, when obtaining the matching degrees between the feature points acquired at the current time and the key time and the target object in the map information according to the coordinates of the feature points acquired at the key time and the feature points acquired at the current time in the vehicle coordinate system at the current time and the position of each particle, is specifically configured to:
respectively mapping the coordinates of the feature points acquired at the current moment and the key moment under the vehicle body coordinate system at the current moment to a particle coordinate system with the particles as the origin;
and obtaining the matching degree of a target object to be matched, which is formed by the characteristic points acquired at the current moment and the key moment and mapped to the particle coordinate system, and the corresponding target object in the map information.
20. The apparatus according to claim 18, wherein the positioning module is specifically configured to, when obtaining K particles according to the matching degree of each of the M particles at the current time and the M particles:
obtaining L particles with a matching degree greater than or equal to a preset matching degree in the M particles, wherein L is less than or equal to M and less than or equal to K;
determining the L particles as the K particles; or,
obtaining at least one particle from the L particles;
obtaining the K particles according to the L particles and the at least one particle.
21. The apparatus according to claim 20, wherein the positioning module, when obtaining the K particles according to the L particles and the at least one particle, is specifically configured to:
according to the matching degree of each particle in the at least one particle, copying each particle at least once;
and determining the L particles and the particles obtained after copying as the K particles.
22. The apparatus according to any one of claims 18 to 21, wherein when the location of the vehicle at the current time is obtained according to the locations of the K particles corresponding to the vehicle at the current time by the location module, the apparatus is specifically configured to:
and acquiring the average position of the K particles according to the position of each particle in the K particles, and determining the average position of the K particles as the position of the vehicle at the current moment.
23. The apparatus of any one of claims 18-22, wherein the positioning module is further configured to:
obtaining an evaluation value of the position of the vehicle at the current moment according to the position of each of the K particles corresponding to the vehicle at the current moment and the position of the vehicle at the current moment, wherein the evaluation value indicates a difference between the position of the vehicle at the current moment and real position information of the vehicle at the current moment.
24. The apparatus according to any one of claims 17 to 23, wherein the positioning module is configured to map, according to the odometer information in the second sensing information and the odometer information in the second sensing information, the coordinates of the feature point acquired at the key time to the coordinates of the vehicle body at the current time, and when obtaining the coordinates of the feature point acquired at the key time in the vehicle body coordinate system at the current time, specifically:
determining a coordinate mapping relation between the vehicle body coordinate system corresponding to the current moment and the vehicle body coordinate system corresponding to the first key moment according to the mileage information in the first sensing information and the mileage information in the second sensing information;
and according to the coordinate mapping relation, respectively mapping the coordinates of the characteristic points acquired at the first key moment to the vehicle body coordinate system at the current moment, and acquiring the coordinates of the characteristic points acquired at the first key moment in the vehicle body coordinate system at the current moment.
25. The apparatus according to any one of claims 16 to 24, wherein the critical time further includes a second critical time, the second critical time being a time at which second sensing information is acquired according to a distance interval, wherein a travel distance of the vehicle from a position of the second critical time to a position corresponding to the current time is less than or equal to a preset distance.
26. The apparatus according to any one of claims 16 to 25, wherein the positioning module is configured to, when determining the position of the vehicle at the current time according to the map information of the driving area of the vehicle and the first and second sensing information, specifically:
and determining the pose of the vehicle at the current moment according to the map information and the first and second sensing information.
27. The apparatus according to claim 18, wherein the positioning module is further configured to, before obtaining matching degrees between feature points acquired at a current time and the key time and the target object in the map information according to coordinates of the feature points acquired at the key time and the feature points acquired at the current time in the vehicle coordinate system at the current time and the position of each particle, respectively:
determining target feature points according to the feature points acquired at each key moment and the feature values of each feature point in the feature points acquired at the current moment;
determining target feature points again according to feature points acquired at each key moment and feature values of the feature points acquired at the current moment except feature points contained in a feature point set, wherein the feature points contained in the feature point set are feature points located in a preset distance range of the last determined target feature point and the last determined target feature point;
when the number of the target characteristic points meets a preset number, N target characteristic points are obtained, wherein N is equal to the preset number;
the positioning module is specifically configured to, when obtaining matching degrees of the feature points acquired at the current time and the key time and a target object in the map information according to the feature points acquired at the key time and the coordinates of the feature points acquired at the current time under the vehicle body coordinate system at the current time and the position of each particle, specifically:
and according to the coordinates of the N target characteristic points in the vehicle body coordinate system at the current moment and the position of each particle, obtaining the matching degree of the characteristic points acquired at the current moment and the key moment and the target object in the map information of the target area.
28. The apparatus according to claim 27, wherein the feature value of the feature point is obtained by an evaluation function, and the feature value is used for evaluating stability of the feature point.
29. The apparatus according to claim 27 or 28, wherein the target feature point is a feature point having a largest feature value among the feature points.
30. The apparatus of any one of claims 16-29, wherein the origin of the body coordinate system is located at any position on the vehicle.
31. A positioning device, comprising: a memory and at least one processor;
the memory to store program instructions;
the processor, configured to invoke program instructions in the memory, to perform the positioning method according to any one of claims 1 to 15.
32. A positioning system, comprising: a vehicle and an onboard sensing device;
the vehicle-mounted sensing device is mounted on the vehicle;
the vehicle for performing the positioning method according to any one of claims 1-15; or,
the vehicle-mounted sensing device is used for executing the positioning method according to any one of claims 1 to 16.
33. A readable storage medium, characterized in that the readable storage medium has stored thereon a computer program; the computer program, when executed, implements a positioning method as claimed in any one of claims 1-15.
CN202011063252.0A 2020-09-30 2020-09-30 Positioning method, device and system Pending CN114323035A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202011063252.0A CN114323035A (en) 2020-09-30 2020-09-30 Positioning method, device and system
PCT/CN2021/101620 WO2022068274A1 (en) 2020-09-30 2021-06-22 Positioning method, apparatus and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011063252.0A CN114323035A (en) 2020-09-30 2020-09-30 Positioning method, device and system

Publications (1)

Publication Number Publication Date
CN114323035A true CN114323035A (en) 2022-04-12

Family

ID=80949680

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011063252.0A Pending CN114323035A (en) 2020-09-30 2020-09-30 Positioning method, device and system

Country Status (2)

Country Link
CN (1) CN114323035A (en)
WO (1) WO2022068274A1 (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114659531B (en) * 2022-05-16 2022-09-23 苏州挚途科技有限公司 Map positioning method and device of vehicle and electronic equipment
CN114719867B (en) * 2022-05-24 2022-09-02 北京捷升通达信息技术有限公司 Vehicle navigation method and system based on sensor
CN115847427B (en) * 2023-02-07 2024-07-16 成都秦川物联网科技股份有限公司 Dual-identification cooperative robot industrial Internet of things monitoring system and control method thereof

Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004117017A (en) * 2002-09-24 2004-04-15 Mitsubishi Electric Corp Target position determination method for sensor bias error estimation
JP2017181476A (en) * 2016-03-31 2017-10-05 株式会社デンソーアイティーラボラトリ Vehicle location detection device, vehicle location detection method and vehicle location detection-purpose computer program
CN108007451A (en) * 2017-11-10 2018-05-08 未来机器人(深圳)有限公司 Detection method, device, computer equipment and the storage medium of cargo carrying device pose
CN109186625A (en) * 2018-10-24 2019-01-11 北京奥特贝睿科技有限公司 Intelligent vehicle carries out pinpoint method and system using mixing sampling filter
CN109932713A (en) * 2019-03-04 2019-06-25 北京旷视科技有限公司 Localization method, device, computer equipment, readable storage medium storing program for executing and robot
CN110673115A (en) * 2019-09-25 2020-01-10 杭州飞步科技有限公司 Combined calibration method, device, equipment and medium for radar and integrated navigation system
JP2020038360A (en) * 2018-08-31 2020-03-12 株式会社デンソー Vehicle-side device, method, and storage medium
CN111006655A (en) * 2019-10-21 2020-04-14 南京理工大学 Multi-scene autonomous navigation positioning method for airport inspection robot
CN111158035A (en) * 2019-12-31 2020-05-15 广东科学技术职业学院 Unmanned vehicle positioning method and unmanned vehicle
CN111220154A (en) * 2020-01-22 2020-06-02 北京百度网讯科技有限公司 Vehicle positioning method, device, equipment and medium
CN111240331A (en) * 2020-01-17 2020-06-05 仲恺农业工程学院 Intelligent trolley positioning and navigation method and system based on laser radar and odometer SLAM
CN111351493A (en) * 2018-12-24 2020-06-30 上海欧菲智能车联科技有限公司 Positioning method and system
CN111539305A (en) * 2020-04-20 2020-08-14 肇庆小鹏汽车有限公司 Map construction method and system, vehicle and storage medium
CN111595333A (en) * 2020-04-26 2020-08-28 武汉理工大学 Modularized unmanned vehicle positioning method and system based on visual inertial laser data fusion
CN111666797A (en) * 2019-03-08 2020-09-15 深圳市速腾聚创科技有限公司 Vehicle positioning method and device and computer equipment

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106123890A (en) * 2016-06-14 2016-11-16 中国科学院合肥物质科学研究院 A kind of robot localization method of Fusion
JP6654977B2 (en) * 2016-07-04 2020-02-26 株式会社Soken Own vehicle position specifying device and own vehicle position specifying method
CN110398765B (en) * 2018-04-25 2022-02-01 北京京东乾石科技有限公司 Positioning method and device and unmanned equipment
CN109556611B (en) * 2018-11-30 2020-11-10 广州高新兴机器人有限公司 Fusion positioning method based on graph optimization and particle filtering
CN110530368B (en) * 2019-08-22 2021-06-15 浙江华睿科技有限公司 Robot positioning method and equipment

Patent Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004117017A (en) * 2002-09-24 2004-04-15 Mitsubishi Electric Corp Target position determination method for sensor bias error estimation
JP2017181476A (en) * 2016-03-31 2017-10-05 株式会社デンソーアイティーラボラトリ Vehicle location detection device, vehicle location detection method and vehicle location detection-purpose computer program
CN108007451A (en) * 2017-11-10 2018-05-08 未来机器人(深圳)有限公司 Detection method, device, computer equipment and the storage medium of cargo carrying device pose
JP2020038360A (en) * 2018-08-31 2020-03-12 株式会社デンソー Vehicle-side device, method, and storage medium
CN109186625A (en) * 2018-10-24 2019-01-11 北京奥特贝睿科技有限公司 Intelligent vehicle carries out pinpoint method and system using mixing sampling filter
CN111351493A (en) * 2018-12-24 2020-06-30 上海欧菲智能车联科技有限公司 Positioning method and system
CN109932713A (en) * 2019-03-04 2019-06-25 北京旷视科技有限公司 Localization method, device, computer equipment, readable storage medium storing program for executing and robot
CN111666797A (en) * 2019-03-08 2020-09-15 深圳市速腾聚创科技有限公司 Vehicle positioning method and device and computer equipment
CN110673115A (en) * 2019-09-25 2020-01-10 杭州飞步科技有限公司 Combined calibration method, device, equipment and medium for radar and integrated navigation system
CN111006655A (en) * 2019-10-21 2020-04-14 南京理工大学 Multi-scene autonomous navigation positioning method for airport inspection robot
CN111158035A (en) * 2019-12-31 2020-05-15 广东科学技术职业学院 Unmanned vehicle positioning method and unmanned vehicle
CN111240331A (en) * 2020-01-17 2020-06-05 仲恺农业工程学院 Intelligent trolley positioning and navigation method and system based on laser radar and odometer SLAM
CN111220154A (en) * 2020-01-22 2020-06-02 北京百度网讯科技有限公司 Vehicle positioning method, device, equipment and medium
CN111539305A (en) * 2020-04-20 2020-08-14 肇庆小鹏汽车有限公司 Map construction method and system, vehicle and storage medium
CN111595333A (en) * 2020-04-26 2020-08-28 武汉理工大学 Modularized unmanned vehicle positioning method and system based on visual inertial laser data fusion

Also Published As

Publication number Publication date
WO2022068274A1 (en) 2022-04-07

Similar Documents

Publication Publication Date Title
CN111448478B (en) System and method for correcting high-definition maps based on obstacle detection
CN114323035A (en) Positioning method, device and system
US20200309538A1 (en) System for producing and/or updating a digital model of a digital map
JP5404861B2 (en) Stationary object map generator
CN112136021B (en) System and method for constructing landmark-based high definition map
US20210323577A1 (en) Methods and systems for managing an automated driving system of a vehicle
JP2008065087A (en) Apparatus for creating stationary object map
KR20210111181A (en) Method, apparatus, device and medium for detecting environmental change
US12085403B2 (en) Vehicle localisation
WO2024012211A1 (en) Autonomous-driving environmental perception method, medium and vehicle
US10839522B2 (en) Adaptive data collecting and processing system and methods
CN115205311B (en) Image processing method, device, vehicle, medium and chip
CN110390252B (en) Obstacle detection method and device based on prior map information and storage medium
CN117470258A (en) Map construction method, device, equipment and medium
US11908095B2 (en) 2-D image reconstruction in a 3-D simulation
CN113312403B (en) Map acquisition method and device, electronic equipment and storage medium
CN115359332A (en) Data fusion method and device based on vehicle-road cooperation, electronic equipment and system
US11410432B2 (en) Methods and systems for displaying animal encounter warnings in vehicles
JP7232727B2 (en) Map data management device and map data management method
CN113874681B (en) Evaluation method and system for point cloud map quality
CN112074871A (en) High-definition map management system and method
EP4431974A1 (en) Generating representation of a surrounding environment of a vehicle
US20240233390A9 (en) Identification of unknown traffic objects
US20230024799A1 (en) Method, system and computer program product for the automated locating of a vehicle
CN118113028A (en) Information service method, apparatus, device, storage medium, and program product for vehicle

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination