[go: up one dir, main page]

CN111366917B - Method, device and equipment for detecting travelable area and computer readable storage medium - Google Patents

Method, device and equipment for detecting travelable area and computer readable storage medium Download PDF

Info

Publication number
CN111366917B
CN111366917B CN202010175382.7A CN202010175382A CN111366917B CN 111366917 B CN111366917 B CN 111366917B CN 202010175382 A CN202010175382 A CN 202010175382A CN 111366917 B CN111366917 B CN 111366917B
Authority
CN
China
Prior art keywords
target image
image
grid
projection
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010175382.7A
Other languages
Chinese (zh)
Other versions
CN111366917A (en
Inventor
唐逸之
韩承志
谭日成
王智
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Baidu Netcom Science and Technology Co Ltd
Original Assignee
Beijing Baidu Netcom Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baidu Netcom Science and Technology Co Ltd filed Critical Beijing Baidu Netcom Science and Technology Co Ltd
Priority to CN202010175382.7A priority Critical patent/CN111366917B/en
Publication of CN111366917A publication Critical patent/CN111366917A/en
Application granted granted Critical
Publication of CN111366917B publication Critical patent/CN111366917B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S11/00Systems for determining distance or velocity not using reflection or reradiation
    • G01S11/12Systems for determining distance or velocity not using reflection or reradiation using electromagnetic waves other than radio waves
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/10Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration
    • G01C21/12Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning
    • G01C21/16Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning by integrating acceleration or speed, i.e. inertial navigation
    • G01C21/165Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning by integrating acceleration or speed, i.e. inertial navigation combined with non-inertial navigation instruments
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/16Anti-collision systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Electromagnetism (AREA)
  • Automation & Control Theory (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The application discloses a method, a device and equipment for detecting a drivable area and a computer readable storage medium, and relates to the technical field of autonomous parking. The specific implementation scheme is as follows: acquiring a matching image and a target image acquired by shooting a surrounding area of a vehicle with a vehicle-mounted monocular camera, wherein the matching image is a previous frame image continuous with the target image; then, performing three-dimensional reconstruction on the target image according to the matched image to obtain target three-dimensional point cloud data corresponding to the target image, so that the reconstruction of the three-dimensional point cloud data of the continuous frame image acquired by the monocular camera is realized; and then, according to the target three-dimensional point cloud data, determining the nearest barrier point in the shooting area of the target image, and according to each nearest barrier point, determining the travelable area in the shooting area of the target image, without acquiring a three-dimensional image or a large number of image samples for machine learning, thereby reducing the dependence on an image acquisition mode, reducing the processing difficulty and improving the accuracy and reliability of detection.

Description

可行驶区域检测方法、装置、设备及计算机可读存储介质Driving area detection method, apparatus, device, and computer-readable storage medium

技术领域technical field

本申请实施例涉及图像处理技术领域,具体涉及一种自主泊车技术。The embodiments of the present application relate to the technical field of image processing, and in particular, to an autonomous parking technology.

背景技术Background technique

为了避免车身与障碍物碰撞或者超出道路边界,车辆在自动驾驶和自主泊车过程中都需要进行可行驶区域检测。In order to avoid the collision between the body and the obstacle or beyond the road boundary, the vehicle needs to detect the drivable area during the process of autonomous driving and autonomous parking.

目前主要是基于有监督深度学习的图像检测或利用三维相机采集三维点云进行可行驶区域检测。但基于有监督深度学习的图像检测需要大量人工标注成本,且在有限数据集上训练的模型难以解决泛化问题。而三维相机结构复杂,制造难度大、检测成本较高。At present, it is mainly based on supervised deep learning image detection or using 3D cameras to collect 3D point clouds for drivable area detection. However, image detection based on supervised deep learning requires a lot of manual annotation costs, and models trained on limited datasets are difficult to solve the generalization problem. The 3D camera has a complex structure, is difficult to manufacture, and has a high detection cost.

可见,现有的可行驶区域检测方法可靠性不够高。It can be seen that the reliability of the existing drivable area detection method is not high enough.

发明内容SUMMARY OF THE INVENTION

本申请的目的是提供一种可行驶区域检测方法、装置、设备及计算机可读存储介质,提高了可行驶区域检测的可靠性。The purpose of the present application is to provide a drivable area detection method, apparatus, device and computer-readable storage medium, which improve the reliability of drivable area detection.

根据本申请的第一方面,提供一种可行驶区域检测方法,包括:According to a first aspect of the present application, a method for detecting a drivable area is provided, comprising:

获取以车载单目相机对车辆周围区域拍摄所采集的匹配图像和目标图像,其中,所述匹配图像是与所述目标图像连续的前一帧图像;Obtaining a matching image and a target image captured by the vehicle-mounted monocular camera on the surrounding area of the vehicle, wherein the matching image is a previous frame image that is continuous with the target image;

根据所述匹配图像对所述目标图像进行三维重建,得到所述目标图像对应的目标三维点云数据;Perform 3D reconstruction on the target image according to the matching image to obtain target 3D point cloud data corresponding to the target image;

根据所述目标三维点云数据,确定所述目标图像的拍摄区域中最近障碍物点;According to the target three-dimensional point cloud data, determine the nearest obstacle point in the shooting area of the target image;

根据各所述最近障碍物点,确定所述目标图像的拍摄区域中的可行驶区域。According to each of the nearest obstacle points, a drivable area in the shooting area of the target image is determined.

本申请实施例通过对单目相机采集的连续帧图像,重建三维点云数据,并提取最近障碍物点,得到可行驶区域,降低了可行驶区域检测对于图像采集方式的依赖,降低了处理难度、提高了检测的准确性和可靠性。In the embodiment of the present application, the 3D point cloud data is reconstructed from the continuous frame images collected by the monocular camera, and the nearest obstacle point is extracted to obtain the drivable area, which reduces the dependence of the drivable area detection on the image acquisition method, and reduces the processing difficulty , Improve the accuracy and reliability of detection.

在一些实施例中,所述根据所述匹配图像对所述目标图像进行三维重建,得到所述目标图像对应的目标三维点云数据,包括:In some embodiments, performing 3D reconstruction on the target image according to the matching image to obtain target 3D point cloud data corresponding to the target image includes:

将所述匹配图像按照所述目标图像的视角投影至预设的多个投影面上,得到多个投影图像,其中,每个所述投影面对应一相对于相机原点的深度;Projecting the matching image onto a plurality of preset projection surfaces according to the viewing angle of the target image to obtain a plurality of projection images, wherein each of the projection surfaces corresponds to a depth relative to the origin of the camera;

根据所述目标图像中像素与所述多个投影图像中相应像素的匹配代价,确定所述目标图像中像素的估计深度;Determine the estimated depth of the pixels in the target image according to the matching cost of the pixels in the target image and the corresponding pixels in the plurality of projection images;

根据所述目标图像中像素的估计深度,获取所述目标图像对应的目标三维点云数据。Acquire target three-dimensional point cloud data corresponding to the target image according to the estimated depth of the pixels in the target image.

本申请实施例利用匹配图像在不同深度投影面上的投影图像,以代价匹配对目标图像实现深度恢复,转化得到目标三维点云数据,提高了对单目图像进行三维重建的准确性和可靠性,进而提高了可行驶区域检测的准确性和可靠性。The embodiment of the present application utilizes the projected images of the matching images on projection surfaces of different depths to achieve depth recovery of the target image at cost matching, and transforms to obtain the target 3D point cloud data, which improves the accuracy and reliability of the 3D reconstruction of the monocular image. , thereby improving the accuracy and reliability of drivable area detection.

在一些实施例中,所述投影面包括:N1个竖直投影平面;In some embodiments, the projection planes include: N1 vertical projection planes;

所述N1个竖直投影平面平行于相机正对面,且所述相机原点到所述N1个竖直投影平面的距离成反比例等差分布,其中,N1为大于1的整数。The N1 vertical projection planes are parallel to the front face of the camera, and the distances from the camera origin to the N1 vertical projection planes are distributed in inverse proportion, wherein N1 is an integer greater than 1.

本申请实施例通过竖直投影平面实现对相机前方区域的深度恢复,提高在弯道等复杂环境下深度恢复的准确性,进而提高对目标图像三维重建的准确性和可靠性。The embodiment of the present application realizes the depth recovery of the area in front of the camera through the vertical projection plane, improves the accuracy of depth recovery in complex environments such as curves, and further improves the accuracy and reliability of the three-dimensional reconstruction of the target image.

在一些实施例中,所述投影面还包括:N2个水平投影平面和/或N3个投影球面;In some embodiments, the projection surface further includes: N2 horizontal projection planes and/or N3 projection spherical surfaces;

其中,所述N2个水平投影平面平行于相机正下方地面,且所述N2个水平投影平面在以所述地面为对称中心的地面分布范围内均匀排列,其中,N2为大于1的整数;Wherein, the N2 horizontal projection planes are parallel to the ground directly below the camera, and the N2 horizontal projection planes are evenly arranged within the ground distribution range with the ground as the center of symmetry, where N2 is an integer greater than 1;

所述N3个投影球面为以所述相机原点为球心的同心球面,且所述N3个投影球面的半径成反比例等差分布,其中,N3为大于1的整数。The N3 projection spheres are concentric spheres with the origin of the camera as the sphere center, and the radii of the N3 projection spheres are inversely proportional and equally distributed, wherein N3 is an integer greater than 1.

本申请实施例通过水平投影平面恢复目标图像中地面区域的深度,通过投影球面引入更多法向采样,提高深度恢复的准确性和可靠性。对于目标图像中既不在水平面上又不在竖直面上的点,通过结合投影球面可以增加法向采样能够提高这些点的深度恢复的准确性。另外,引入平行的投影球面还能够对视角大于180度的鱼眼的目标图像提供有利的投影面,提高深度恢复的准确性和可靠性。In the embodiment of the present application, the depth of the ground area in the target image is recovered by the horizontal projection plane, and more normal samples are introduced by the projection sphere, so as to improve the accuracy and reliability of the depth recovery. For the points in the target image that are neither on the horizontal nor vertical planes, the depth recovery accuracy of these points can be improved by increasing the normal sampling by combining with the projection sphere. In addition, the introduction of a parallel projection sphere can also provide a favorable projection surface for the fisheye target image with a viewing angle greater than 180 degrees, and improve the accuracy and reliability of depth recovery.

在一些实施例中,所述根据所述目标图像中像素与所述多个投影图像中相应像素的匹配代价,确定所述目标图像中像素的估计深度,包括:In some embodiments, the determining the estimated depth of the pixels in the target image according to the matching cost of the pixels in the target image and the corresponding pixels in the plurality of projection images includes:

获取所述目标图像中像素的目标像素窗口特征;Obtain the target pixel window feature of the pixel in the target image;

获取所述多个投影图像中相应像素的投影像素窗口特征;acquiring projected pixel window features of corresponding pixels in the plurality of projected images;

根据所述目标像素窗口特征和所述投影像素窗口特征,获取所述目标图像中像素与各所述投影图像中相应像素的匹配代价;According to the target pixel window feature and the projection pixel window feature, obtain the matching cost of the pixel in the target image and the corresponding pixel in each of the projection images;

将所述匹配代价最小的所述相应像素对应的深度,作为所述目标图像中像素的估计深度,其中,所述相应像素对应的深度,是所述相应像素所在投影面对应的深度。The depth corresponding to the corresponding pixel with the smallest matching cost is taken as the estimated depth of the pixel in the target image, wherein the depth corresponding to the corresponding pixel is the depth corresponding to the projection surface where the corresponding pixel is located.

本申请实施例通过将匹配代价最小的相应像素所对应的深度,作为目标图像中像素的估计深度,提高了对目标图像中像素深度恢复的准确性。The embodiment of the present application improves the accuracy of restoring the depth of pixels in the target image by using the depth corresponding to the corresponding pixel with the smallest matching cost as the estimated depth of the pixel in the target image.

在一些实施例中,在所述根据所述目标图像中像素与所述多个投影图像中相应像素的匹配代价,确定所述目标图像中像素的估计深度之前,还包括:In some embodiments, before the determining the estimated depth of the pixels in the target image according to the matching cost of the pixels in the target image and the corresponding pixels in the plurality of projection images, the method further includes:

根据所述目标图像与所述匹配图像的相机相对位姿,确定所述匹配图像中与所述目标图像中像素一一对应的相应像素;According to the camera relative poses of the target image and the matching image, determine the corresponding pixels in the matching image that correspond to the pixels in the target image one-to-one;

根据所述匹配图像中的所述相应像素,确定各所述投影图像中与所述目标图像中像素一一对应的相应像素。According to the corresponding pixels in the matching image, the corresponding pixels in each of the projection images corresponding to the pixels in the target image one-to-one are determined.

本申请实施例通过相机相对位姿定位匹配图像和目标图像中相对应的像素,提高了对目标图像深度恢复的准确性和可靠性。The embodiment of the present application improves the accuracy and reliability of the depth recovery of the target image by matching the corresponding pixels in the image and the target image through the relative pose positioning of the camera.

在一些实施例中,在所述根据所述目标图像与所述匹配图像的相机相对位姿,确定所述匹配图像中与所述目标图像中像素一一对应的相应像素之前,还包括:In some embodiments, before determining the corresponding pixels in the matching image that correspond one-to-one with the pixels in the target image according to the relative poses of the cameras of the target image and the matching image, the method further includes:

采集车辆后轮的轮速计数据和车载惯性测量单元IMU数据,其中,所述车辆后轮的轮速计数据指示了所述车载单目相机的水平运动距离,所述车载IMU数据指示了所述车载单目相机的水平运动方向;Collect wheel speedometer data of the rear wheel of the vehicle and on-board inertial measurement unit IMU data, wherein the wheel speedometer data of the rear wheel of the vehicle indicates the horizontal movement distance of the vehicle-mounted monocular camera, and the vehicle-mounted IMU data indicates all Describe the horizontal movement direction of the vehicle-mounted monocular camera;

根据所述车辆后轮的轮速计数据和所述车载IMU数据,确定所述车载单目相机的相机位姿数据;Determine the camera pose data of the vehicle-mounted monocular camera according to the wheel speedometer data of the rear wheel of the vehicle and the vehicle-mounted IMU data;

根据所述车载单目相机的相机位姿数据,确定所述目标图像与所述匹配图像的相机相对位姿。According to the camera pose data of the vehicle-mounted monocular camera, the relative camera poses of the target image and the matching image are determined.

本申请实施例由于后轮与车身之间在水平方向无相对转动,后轮的轮速可以直接表征车身的移速,结合车辆后轮的轮速计数据和车载IMU数据得到相机位姿,提高了相机相对位姿的可靠性,进而提高可行驶区域检测的准确性和可靠性。In the embodiment of the present application, since there is no relative rotation between the rear wheel and the body in the horizontal direction, the wheel speed of the rear wheel can directly represent the moving speed of the body, and the camera pose is obtained by combining the wheel speedometer data of the rear wheel of the vehicle and the data of the on-board IMU, which improves the The reliability of the relative pose of the camera is improved, thereby improving the accuracy and reliability of the drivable area detection.

在一些实施例中,所述根据所述目标三维点云数据,确定所述目标图像的拍摄区域中最近障碍物点,包括:In some embodiments, determining the closest obstacle point in the shooting area of the target image according to the target three-dimensional point cloud data includes:

根据所述目标三维点云数据和对所述目标图像的拍摄区域水平切分的极坐标栅格网络,确定所述极坐标栅格网络中各栅格中包含障碍物点的数量,其中,所述障碍物点为对地高度大于预设障碍物高度阈值的目标三维点;Determine the number of obstacle points included in each grid in the polar coordinate grid network according to the target three-dimensional point cloud data and the polar coordinate grid network that horizontally divides the shooting area of the target image, wherein the The obstacle point is a target three-dimensional point whose height above the ground is greater than the preset obstacle height threshold;

根据所述极坐标栅格网络的各扇形分区中的最近障碍物栅格,确定所述目标图像的拍摄区域中各方向的最近障碍物点,其中,所述最近障碍物栅格是所述扇形分区中与所述相机原点径向距离最近、且所包含障碍物点的数量大于预设数量阈值的栅格。According to the nearest obstacle grid in each sector partition of the polar coordinate grid network, determine the nearest obstacle point in each direction in the shooting area of the target image, wherein the nearest obstacle grid is the sector The grid in the partition with the closest radial distance to the origin of the camera and containing the number of obstacle points greater than the preset number threshold.

本申请实施例利用对相机原点建立的极坐标栅格网络对目标图像的拍摄区域进行空间划分,将各方向扇形分区中可能是障碍物且对相机原点径向距离最小的栅格提取出来,用于确定相机原点各方向的最近障碍物点,提高了最近障碍物点的准确性,进而提高可行驶区域检测的准确性和可靠性。In this embodiment of the present application, the polar coordinate grid network established for the camera origin is used to spatially divide the shooting area of the target image, and the grid that may be an obstacle and the smallest radial distance from the camera origin in the fan-shaped partitions in each direction is extracted, and the It is used to determine the nearest obstacle points in all directions of the origin of the camera, which improves the accuracy of the nearest obstacle points, thereby improving the accuracy and reliability of the drivable area detection.

在一些实施例中,所述根据所述极坐标栅格网络的各扇形分区中的最近障碍物栅格,确定所述目标图像的拍摄区域中各方向的最近障碍物点,包括:In some embodiments, determining the nearest obstacle points in each direction in the shooting area of the target image according to the nearest obstacle grid in each sector of the polar coordinate grid network includes:

获取所述最近障碍物栅格中所包含障碍物点的平均位置点;Obtain the average position point of the obstacle points contained in the nearest obstacle grid;

将各所述最近障碍物栅格对应的平均位置点,作为所述目标图像的拍摄区域中最近障碍物点。The average position point corresponding to each of the nearest obstacle grids is taken as the nearest obstacle point in the shooting area of the target image.

本申请实施例进一步优化最近障碍物点的位置,提高障碍物点的准确性。The embodiment of the present application further optimizes the position of the nearest obstacle point, and improves the accuracy of the obstacle point.

在一些实施例中,所述根据各所述最近障碍物点,确定所述目标图像的拍摄区域中的可行驶区域,包括:In some embodiments, the determining a drivable area in the shooting area of the target image according to each of the nearest obstacle points includes:

在对所述目标图像的拍摄区域水平切分的均匀分割网络中,根据所述均匀分割网络中各网格单元相对于所述最近障碍物点的位置,确定各网格单元的加权值;In the uniform segmentation network for horizontally segmenting the shooting area of the target image, the weighted value of each grid unit is determined according to the position of each grid unit in the uniform segmentation network relative to the nearest obstacle point;

根据各所述网格单元的初始权值以及所述加权值,确定各所述网格单元的新权值,其中,所述新权值大于或等于最小权阈值、且小于或等于最大权阈值,所述网格单元的初始权值是0或者是在对前一帧图像确定可行驶区域时确定的新权值;According to the initial weight of each grid unit and the weight value, a new weight value of each grid unit is determined, wherein the new weight value is greater than or equal to the minimum weight threshold value and less than or equal to the maximum weight threshold value , the initial weight of the grid unit is 0 or a new weight determined when the drivable area is determined for the previous frame of image;

根据所述新权值指示的第一类网格单元或第二类网格单元,确定所述目标图像的拍摄区域中的可行驶区域,其中,所述新权值指示的第一类网格单元与所述相机原点之间有所述最近障碍物点,所述新权值指示的第二类网格单元与所述相机原点之间没有所述最近障碍物点。Determine a drivable area in the shooting area of the target image according to the first-type grid unit or the second-type grid unit indicated by the new weight, wherein the first-type grid indicated by the new weight There is the closest obstacle point between the unit and the camera origin, and there is no closest obstacle point between the second type of grid unit indicated by the new weight and the camera origin.

本申请实施例通过计算目标图像拍摄区域在均匀分割网络中各网格单元的加权值,对网格单元更新权值,随着连续图像帧的不断更新,不断加权更新网格单元的新权值,不仅能够平滑噪声还能够起到弥补单帧最近障碍物点漏检的问题,提高了对可行驶区域检测的可靠性。In the embodiment of the present application, the weighted value of each grid unit in the uniform segmentation network of the target image shooting area is calculated, and the weighted value of the grid unit is updated. , which can not only smooth the noise, but also make up for the problem of missing detection of the nearest obstacle point in a single frame, and improve the reliability of the detection of the drivable area.

在一些实施例中,所述根据所述均匀分割网络中各网格单元相对于所述最近障碍物点的位置,确定各网格单元的加权值,包括:In some embodiments, determining the weighted value of each grid unit according to the position of each grid unit in the uniformly divided network relative to the nearest obstacle point includes:

若所述均匀分割网络中网格单元对相机原点的径向距离,减去所述网格单元方向上的最近障碍物点对相机原点的径向距离的差值,大于或等于距离上限阈值,则所述网格单元的加权值为第一值;If the radial distance between the grid unit in the uniform segmentation network and the camera origin, minus the difference between the radial distance between the nearest obstacle point in the direction of the grid unit and the camera origin, is greater than or equal to the upper limit of the distance threshold, then the weighted value of the grid unit is the first value;

若所述最近障碍物点对相机原点的径向距离,减去所述最近障碍物点方向上网格单元对相机原点的径向距离的差值,小于或等于距离下限阈值,则所述网格单元的加权值为第二值;If the radial distance between the closest obstacle point and the camera origin, minus the difference between the radial distance between the grid unit and the camera origin in the direction of the closest obstacle point, is less than or equal to the lower threshold of the distance, then the grid The weighted value of the unit is the second value;

若所述最近障碍物点对相机原点的径向距离,减去所述最近障碍物点方向上网格单元对相机原点的径向距离的差值,小于所述距离上限阈值且大于所述距离下限阈值,则所述网格单元的加权值为第三值,其中,所述第三值是根据所述差值和预设的平滑连续函数确定的值;If the radial distance between the closest obstacle point and the camera origin, minus the difference between the radial distance between the grid unit and the camera origin in the direction of the closest obstacle point, is less than the distance upper threshold and greater than the distance lower limit threshold, the weighted value of the grid unit is a third value, wherein the third value is a value determined according to the difference and a preset smooth continuous function;

其中,所述距离上限阈值是所述距离下限阈值的相反数,所述第一值是所述第二值的相反数,所述第三值的绝对值小于所述距离上限阈值的绝对值或所述距离下限阈值的绝对值。Wherein, the distance upper threshold is the inverse of the distance lower threshold, the first value is the inverse of the second value, and the absolute value of the third value is less than the absolute value of the distance upper threshold or The absolute value of the distance lower threshold.

本申请实施例实现了对各网格单元加权值的确定,并通过平滑连续函数对差值在距离上限阈值和距离下限阈值之间的网格单元平滑过渡,进一步降低多帧融合加权累积过程中的噪声,提高了对网格单元新权值确定的可靠性,进而提高了对可行驶区域检测的可靠性。The embodiment of the present application realizes the determination of the weighted value of each grid unit, and uses a smooth continuous function to smoothly transition the grid units whose difference is between the upper threshold of the distance and the lower threshold of the distance, further reducing the weighted accumulation process of multi-frame fusion. , which improves the reliability of the determination of the new weights of the grid cells, and further improves the reliability of the detection of the drivable area.

在一些实施例中,所述匹配图像和目标图像都是鱼眼图像。In some embodiments, the matching image and the target image are both fisheye images.

本申请实施例通过采用鱼眼图像,实现了增大水平视角(能够超过180度),扩大图像拍摄区域视野范围,提高可行驶区域检测的可靠性。By adopting the fisheye image, the embodiment of the present application achieves an increase in the horizontal viewing angle (which can exceed 180 degrees), enlarges the field of view of the image shooting area, and improves the reliability of the detection of the drivable area.

根据本申请的第二方面,提供一种可行驶区域检测装置,包括:According to a second aspect of the present application, there is provided a drivable area detection device, comprising:

图像采集模块,用于获取以车载单目相机对车辆周围区域拍摄所采集的匹配图像和目标图像,其中,所述匹配图像是所述目标图像的前一帧图像;an image acquisition module, configured to acquire a matching image and a target image captured by the vehicle-mounted monocular camera on the surrounding area of the vehicle, wherein the matching image is a previous frame of the target image;

第一处理模块,用于根据所述匹配图像对所述目标图像进行三维重建,得到所述目标图像对应的目标三维点云数据;a first processing module, configured to perform three-dimensional reconstruction on the target image according to the matching image, and obtain target three-dimensional point cloud data corresponding to the target image;

第二处理模块,用于根据所述目标三维点云数据,确定所述目标图像的拍摄区域中最近障碍物点;a second processing module, configured to determine the nearest obstacle point in the shooting area of the target image according to the target three-dimensional point cloud data;

第三处理模块,用于根据各所述最近障碍物点,确定所述目标图像的拍摄区域中的可行驶区域。The third processing module is configured to determine a drivable area in the shooting area of the target image according to each of the nearest obstacle points.

根据本申请的第三方面,提供一种电子设备,包括:According to a third aspect of the present application, an electronic device is provided, comprising:

至少一个处理器;以及at least one processor; and

与所述至少一个处理器通信连接的存储器;其中,a memory communicatively coupled to the at least one processor; wherein,

所述存储器存储有可被所述至少一个处理器执行的指令,所述指令被所述至少一个处理器执行,以使所述至少一个处理器能够执行本申请第一方面及第一方面任一实施例所述的可行驶区域检测方法。The memory stores instructions executable by the at least one processor, the instructions being executed by the at least one processor to enable the at least one processor to perform any of the first and first aspects of the present application The drivable area detection method described in the embodiment.

根据本申请的第四方面,提供一种存储有计算机指令的非瞬时计算机可读存储介质,所述计算机指令用于使所述计算机执行本申请第一方面及第一方面任一实施例所述的可行驶区域检测方法。According to a fourth aspect of the present application, there is provided a non-transitory computer-readable storage medium storing computer instructions, the computer instructions being used to cause the computer to execute the first aspect of the present application and any embodiment of the first aspect. drivable area detection method.

上述申请中的一个实施例具有如下优点或有益效果:通过获取以车载单目相机对车辆周围区域拍摄所采集的匹配图像和目标图像,其中,所述匹配图像是与所述目标图像连续的前一帧图像;然后根据所述匹配图像对所述目标图像进行三维重建,得到所述目标图像对应的目标三维点云数据,从而实现对单目相机采集的连续帧图像重建三维点云数据;接着根据所述目标三维点云数据,确定所述目标图像的拍摄区域中最近障碍物点,并根据各所述最近障碍物点,确定所述目标图像的拍摄区域中的可行驶区域,无需采集三维图像或大量图像样本进行机器学习,降低了对于图像采集方式的依赖,降低了处理难度、提高了检测的准确性和可靠性。An embodiment in the above application has the following advantages or beneficial effects: by acquiring a matching image and a target image captured by shooting the surrounding area of the vehicle with a vehicle-mounted monocular camera, wherein the matching image is a preceding image that is continuous with the target image. One frame of image; then carry out three-dimensional reconstruction of the target image according to the matching image to obtain the target three-dimensional point cloud data corresponding to the target image, so as to realize the reconstruction of three-dimensional point cloud data for the continuous frame images collected by the monocular camera; then Determine the nearest obstacle point in the shooting area of the target image according to the target three-dimensional point cloud data, and determine the drivable area in the shooting area of the target image according to each of the nearest obstacle points, without collecting three-dimensional Machine learning is performed on images or a large number of image samples, which reduces the dependence on the image acquisition method, reduces the processing difficulty, and improves the accuracy and reliability of detection.

上述可选方式所具有的其他效果将在下文中结合具体实施例加以说明。Other effects of the above-mentioned optional manners will be described below with reference to specific embodiments.

附图说明Description of drawings

附图用于更好地理解本方案,不构成对本申请的限定。其中:The accompanying drawings are used for better understanding of the present solution, and do not constitute a limitation to the present application. in:

图1是本申请实施例提供的一种自主泊车的应用场景示意图;1 is a schematic diagram of an application scenario of autonomous parking provided by an embodiment of the present application;

图2是本申请实施例提供的一种可行驶区域检测方法流程示意图;2 is a schematic flowchart of a method for detecting a drivable area provided by an embodiment of the present application;

图3是本申请实施例提供的一种图2中步骤S102的方法流程示意图;FIG. 3 is a schematic flowchart of the method of step S102 in FIG. 2 provided by an embodiment of the present application;

图4是本申请实施例提供的一种在多个投影面上投影匹配深度的示意图;4 is a schematic diagram of projecting matching depths on multiple projection surfaces provided by an embodiment of the present application;

图5是本申请实施例提供的一种竖直投影平面分布的俯视图;5 is a top view of a vertical projection plane distribution provided by an embodiment of the present application;

图6是本申请实施例提供的一种多类型投影平面分布的侧视图;6 is a side view of a multi-type projection plane distribution provided by an embodiment of the present application;

图7是本申请实施例提供的一种极坐标栅格网络的俯视图;7 is a top view of a polar coordinate grid network provided by an embodiment of the present application;

图8是本申请实施例提供的一种目标图像的拍摄区域中最近障碍物点示意图;8 is a schematic diagram of the nearest obstacle point in a shooting area of a target image provided by an embodiment of the present application;

图9是本申请实施例提供的一种均匀分割网络俯视图;9 is a top view of a uniformly divided network provided by an embodiment of the present application;

图10是本申请实施例提供的一种可行驶区域示意图;10 is a schematic diagram of a drivable area provided by an embodiment of the present application;

图11是本申请实施例提供的一种可行驶区域检测装置结构示意图;11 is a schematic structural diagram of a driveable area detection device provided by an embodiment of the present application;

图12是根据本申请实施例的可行驶区域检测方法的电子设备的框图。FIG. 12 is a block diagram of an electronic device of a drivable area detection method according to an embodiment of the present application.

具体实施方式Detailed ways

以下结合附图对本申请的示范性实施例做出说明,其中包括本申请实施例的各种细节以助于理解,应当将它们认为仅仅是示范性的。因此,本领域普通技术人员应当认识到,可以对这里描述的实施例做出各种改变和修改,而不会背离本申请的范围和精神。同样,为了清楚和简明,以下的描述中省略了对公知功能和结构的描述。Exemplary embodiments of the present application are described below with reference to the accompanying drawings, which include various details of the embodiments of the present application to facilitate understanding, and should be considered as exemplary only. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the present application. Also, descriptions of well-known functions and constructions are omitted from the following description for clarity and conciseness.

在车辆自动驾驶或自主泊车的场景中,车辆需要对车辆周围的可行驶区域进行检测,以规划安全可行的行驶路径或泊车路径,实现自动驾驶中对障碍物的自动规避或者自动泊车入库。例如是在车库进行自主泊车过程中,需要检测前进或后退方向上的可行驶区域,再根据可行驶区域控制车辆进入车位。参见图1,是本申请实施例提供的一种自主泊车的应用场景示意图。在图1所示的泊车场景中,车辆倒车进入车位时,需要避开空置车位左侧的其他车辆以及右侧的石柱。首先需要采集车辆倒车后方的物体三维信息,识别出其他车辆和石柱的位置,从而准确地获取到车辆倒车时的可行驶区域。In the scenario of autonomous driving or autonomous parking, the vehicle needs to detect the drivable area around the vehicle in order to plan a safe and feasible driving path or parking path, and realize the automatic avoidance of obstacles or automatic parking in autonomous driving. Inventory. For example, in the process of autonomous parking in a garage, it is necessary to detect the drivable area in the forward or backward direction, and then control the vehicle to enter the parking space according to the drivable area. Referring to FIG. 1 , it is a schematic diagram of an application scenario of autonomous parking provided by an embodiment of the present application. In the parking scene shown in Figure 1, when the vehicle reverses into the parking space, it needs to avoid other vehicles on the left side of the vacant parking space and the stone pillars on the right side. First, it is necessary to collect the three-dimensional information of the object behind the vehicle when it is reversing, and identify the positions of other vehicles and stone pillars, so as to accurately obtain the drivable area when the vehicle is reversing.

现有的可行驶区域检测方法中,通过有监督深度学习的图像检测需要大量的图片样本进行学习,遇到新的障碍物或复杂环境可能存在识别不准确的风险。而利用三维相机采集车辆后方三维点云则需要给车辆配置结构复杂的三维相机,车辆行驶中可能存在检测不可靠的问题。In the existing drivable area detection methods, image detection through supervised deep learning requires a large number of image samples for learning, and there may be a risk of inaccurate recognition when encountering new obstacles or complex environments. Using a 3D camera to collect a 3D point cloud behind the vehicle requires configuring a 3D camera with a complex structure for the vehicle, which may cause unreliable detection while the vehicle is running.

本申请通过提供一种可行驶区域检测方法、装置、设备及计算机可读存储介质,利用车载单目相机实现对可行驶区域的检测,降低检测的难度,提高检测的准确性和可靠性。The present application provides a drivable area detection method, device, equipment and computer-readable storage medium, which utilizes a vehicle-mounted monocular camera to detect the drivable area, reduces the difficulty of detection, and improves the accuracy and reliability of detection.

参见图2,是本申请实施例提供的一种可行驶区域检测方法流程示意图,图2所示方法的执行主体可以是软件和/或硬件的可行驶区域检测装置,具体例如可以是各种类型的终端、车载检测系统、云端等之一或多者的结合。图2所示的方法包括步骤S101至步骤S104,具体如下:Referring to FIG. 2 , it is a schematic flowchart of a drivable area detection method provided by an embodiment of the present application. The execution body of the method shown in FIG. 2 may be a software and/or hardware drivable area detection device, for example, various types of drivable area detection devices may be used. The combination of one or more of the terminal, vehicle detection system, cloud, etc. The method shown in FIG. 2 includes steps S101 to S104, and the details are as follows:

S101,获取以车载单目相机对车辆周围区域拍摄所采集的匹配图像和目标图像,其中,所述匹配图像是与所述目标图像连续的前一帧图像。S101: Acquire a matching image and a target image captured by a vehicle-mounted monocular camera on a surrounding area of a vehicle, where the matching image is a previous frame image that is continuous with the target image.

车载单目相机可以是安装在车辆前方或者后方。车载单目相机拍摄车辆周围区域,例如可以是对车辆前方区域进行拍摄,也可以是对车辆后方区域进行拍摄。或者,可以根据车辆的前进或后退动作,选择采集车辆前方区域图像或车辆后方区域图像。Vehicle monocular cameras can be installed in the front or rear of the vehicle. The vehicle-mounted monocular camera shoots the surrounding area of the vehicle, for example, the area in front of the vehicle may be photographed, or the area behind the vehicle may be photographed. Alternatively, the image of the area in front of the vehicle or the image of the area behind the vehicle can be selected to be collected according to the forward or reverse motion of the vehicle.

在车载单目相机采集的连续帧图像中确定目标图像,并将目标图像的前一帧作为匹配图像。这里的目标图像例如可以是车载单目相机实时采集的当前帧图像,但也可以是历史帧图像,本实施例不做限定。The target image is determined from the consecutive frame images collected by the vehicle-mounted monocular camera, and the previous frame of the target image is used as the matching image. The target image here may be, for example, a current frame image collected in real time by a vehicle-mounted monocular camera, but may also be a historical frame image, which is not limited in this embodiment.

其中,车载单目相机可以是采用非广角镜头、广角镜头或者超广角镜头的相机。鱼眼镜头是一种超大视场、大孔径的光学成像系统,一般采用两块或三块负弯月形透镜作为前光组,将物方超大视场压缩至常规镜头要求的视场范围。采用鱼眼镜头的相机拍摄图像,视角例如可达到220°或230°。The vehicle-mounted monocular camera may be a camera using a non-wide-angle lens, a wide-angle lens, or an ultra-wide-angle lens. Fisheye lens is an optical imaging system with large field of view and large aperture. Generally, two or three negative meniscus lenses are used as the front light group to compress the large field of view on the object side to the field of view required by conventional lenses. A camera with a fisheye lens captures images with a viewing angle of, for example, 220° or 230°.

在一些实施例中,车载单目相机可以是采用鱼眼镜头的相机,其采集的匹配图像和目标图像可以都是鱼眼图像。本实施例可以通过采用鱼眼图像,增大水平视角(例如超过180°),扩大图像拍摄区域视野范围,提高可行驶区域检测的可靠性。In some embodiments, the vehicle-mounted monocular camera may be a camera using a fisheye lens, and the matched image and the target image collected by the camera may both be fisheye images. In this embodiment, by using a fisheye image, the horizontal viewing angle (for example, exceeding 180°) can be increased, the field of view of the image shooting area can be expanded, and the reliability of the detection of the drivable area can be improved.

S102,根据所述匹配图像对所述目标图像进行三维重建,得到所述目标图像对应的目标三维点云数据。S102: Perform three-dimensional reconstruction on the target image according to the matching image, to obtain target three-dimensional point cloud data corresponding to the target image.

目标图像是二维图像,但与其前一帧图像之间可以得到相机的位姿改变,进而可以通过投影进对目标图像生成深度图像,进而实现三维重建。步骤S102的实现方式有多种,下面结合图3和具体实施例进行举例说明。参见图3,是本申请实施例提供的一种图2中步骤S102的方法流程示意图。图3所示方法具体包括步骤S201至S203,具体如下:The target image is a two-dimensional image, but the pose change of the camera can be obtained between it and the previous frame image, and then a depth image can be generated by projecting into the target image, thereby realizing three-dimensional reconstruction. There are various implementations of step S102, which are described below with reference to FIG. 3 and specific embodiments. Referring to FIG. 3 , it is a schematic flowchart of the method of step S102 in FIG. 2 provided by an embodiment of the present application. The method shown in FIG. 3 specifically includes steps S201 to S203, which are as follows:

S201,将所述匹配图像按照所述目标图像的视角投影至预设的多个投影面上,得到多个投影图像,其中,每个所述投影面对应一相对于相机原点的深度。S201 , project the matching image onto a plurality of preset projection surfaces according to the viewing angle of the target image to obtain a plurality of projection images, wherein each of the projection surfaces corresponds to a depth relative to the origin of the camera.

参见图4,是本申请实施例提供的一种在多个投影面上投影匹配深度的示意图。如图所示,按照目标图像的视角,将匹配图像投影到多个投影面上,得到具有各种深度的投影图像。可以理解为,对相机预先设置有多个深度的投影面。深度可以理解为是与相机原点所在垂面的距离。不同深度的多个投影面,可以理解为是空间中设置有多个投影面,且每个投影面与相机原点径向距离(也是与相机原点所在垂面的距离)不相同,由此每个投影面对应了一个深度。当匹配图像被投影到投影面上后,投影面上的投影图像就具有了与该投影面对应的深度,即单个投影图像中所有像素的深度都是其所述投影面的深度。Referring to FIG. 4 , it is a schematic diagram of projecting matching depths on multiple projection surfaces according to an embodiment of the present application. As shown in the figure, according to the viewing angle of the target image, the matching image is projected onto multiple projection surfaces to obtain projected images with various depths. It can be understood that the camera is preset with projection surfaces with multiple depths. Depth can be understood as the distance from the vertical plane where the camera origin is located. Multiple projection surfaces at different depths can be understood as multiple projection surfaces in space, and each projection surface has a different radial distance from the camera origin (also the distance from the vertical plane where the camera origin is located), so each The projection surface corresponds to a depth. After the matching image is projected onto the projection surface, the projection image on the projection surface has a depth corresponding to the projection surface, that is, the depth of all pixels in a single projection image is the depth of the projection surface.

在一些实施例中,上述投影面例如包括N1个竖直投影平面,其中,N1为大于1的整数。具体地,N1个竖直投影平面平行于相机正对面,且所述相机原点到所述N1个竖直投影平面的距离成反比例等差分布。在图1所示场景中,加入车辆的车载单目相机正对前方车库的墙壁,那么,N1个竖直投影平面可以理解为是N1个与前方车库墙壁平行的空间平面。相机所拍摄图像,近处物体图像分辨率通常大于远处物体图像分辨率,可以理解为远景像素所表示的实际尺寸大于近景像素所表示的实际尺寸。因此,为了提高深度恢复的可靠性,靠近相机原点的竖直投影平面分布密度,大于远离相机原点的竖直投影平面分布密度。另外,障碍物越靠近车辆,可能产生的影响或者危险就越大,在越靠近车辆的区域可以提高深度恢复的精度,以对靠近车辆的区域生成更精确的深度数据。例如N1个竖直投影平面的分布密度可以满足:靠近相机原点的分布密度大于远离相机原点的分布密度。参见图5,是本申请实施例提供的一种竖直投影平面分布的俯视图。图5所示相机原点的前方拍摄区域布置有64个竖直投影平面(图中未全部示出),且相机原点到各竖直投影平面的距离成反比例等差分布,例如依次是:20/64米,20/63米,20/62米,....,20/3米,20/2米,20米。其中,越靠近相机原点,竖直投影平面分布越密集,所恢复深度的精度越大。本实施例通过竖直投影平面实现对相机前方区域的深度恢复,提高在弯道等复杂环境下深度恢复的准确性,进而提高对目标图像三维重建的准确性和可靠性。In some embodiments, the above-mentioned projection planes include, for example, N1 vertical projection planes, where N1 is an integer greater than 1. Specifically, the N1 vertical projection planes are parallel to the front face of the camera, and the distances from the camera origin to the N1 vertical projection planes are distributed in inverse proportion and equally. In the scene shown in Figure 1, the on-board monocular camera added to the vehicle is facing the wall of the front garage, then the N1 vertical projection planes can be understood as N1 space planes parallel to the front garage wall. In the image captured by the camera, the resolution of the near object image is usually larger than that of the distant object image, which can be understood as the actual size represented by the far-field pixels is larger than the actual size represented by the near-field pixels. Therefore, in order to improve the reliability of depth recovery, the distribution density of the vertical projection plane close to the camera origin is greater than the distribution density of the vertical projection plane far from the camera origin. In addition, the closer the obstacle is to the vehicle, the greater the possible impact or danger. In the area closer to the vehicle, the accuracy of depth recovery can be improved to generate more accurate depth data for the area close to the vehicle. For example, the distribution densities of the N1 vertical projection planes may satisfy: the distribution density near the camera origin is greater than the distribution density far from the camera origin. Referring to FIG. 5 , it is a top view of a vertical projection plane distribution provided by an embodiment of the present application. 64 vertical projection planes (not all shown in the figure) are arranged in the front shooting area of the camera origin shown in Figure 5, and the distances from the camera origin to each vertical projection plane are distributed in inverse proportion, for example: 20/ 64 meters, 20/63 meters, 20/62 meters, ...., 20/3 meters, 20/2 meters, 20 meters. Among them, the closer to the camera origin, the denser the distribution of the vertical projection plane, and the greater the accuracy of the recovered depth. This embodiment realizes the depth recovery of the area in front of the camera through the vertical projection plane, improves the accuracy of depth recovery in complex environments such as curves, and further improves the accuracy and reliability of the three-dimensional reconstruction of the target image.

在另一些实施例中,在上述竖直投影平面的基础上,还可以引入N2个水平投影平面和/或N3个投影球面作为投影面。其中,水平投影平面可以用于对地面的深度恢复,投影球面可以用于对鱼眼图像中畸变图像的深度恢复。In other embodiments, on the basis of the above vertical projection plane, N2 horizontal projection planes and/or N3 projection spheres may also be introduced as projection surfaces. Among them, the horizontal projection plane can be used for the depth recovery of the ground, and the projection sphere can be used for the depth recovery of the distorted image in the fisheye image.

参见图6,是本申请实施例提供的一种多类型投影平面分布的侧视图。图6中示意出了多个相互平行的竖直投影平面、相互平行的水平投影平面以及同球心的投影球面。N2个水平投影平面平行于相机正下方地面,且所述N2个水平投影平面在以所述地面为对称中心的地面分布范围内均匀排列,其中,N2为大于1的整数。以图6所示为例,相机标定后,可以确定出相机原点正下方地面所在平面位置,以4个水平投影平面均匀分布在相机正下方地面附近-5cm到5cm的地面分布范围内,用于对接近地面的点的深度进行恢复。其中,水平投影平面的个数也可以是8个。本实施例通过水平投影平面,可以恢复目标图像中地面区域像素的深度,提高对地面图像区域深度恢复的准确性和可靠性。Referring to FIG. 6 , it is a side view of a multi-type projection plane distribution provided by an embodiment of the present application. FIG. 6 illustrates a plurality of mutually parallel vertical projection planes, mutually parallel horizontal projection planes and concentric projection spheres. The N2 horizontal projection planes are parallel to the ground directly below the camera, and the N2 horizontal projection planes are evenly arranged within a ground distribution range with the ground as the center of symmetry, where N2 is an integer greater than 1. Taking Figure 6 as an example, after the camera is calibrated, the plane position of the ground directly below the camera origin can be determined, and four horizontal projection planes are evenly distributed in the ground distribution range of -5cm to 5cm near the ground directly below the camera. The depth of the point close to the ground is recovered. The number of horizontal projection planes may also be 8. In this embodiment, through the horizontal projection plane, the depth of the pixels in the ground area in the target image can be recovered, and the accuracy and reliability of the depth recovery of the ground image area can be improved.

上述实施例中,N3个投影球面为以所述相机原点为球心的同心球面,且所述N3个投影球面的半径成反比例等差分布,其中,N3为大于1的整数。继续参见图6,以相机原点为球心,64个半径从0.5m到32m按反比例等差分布,形成投影球面。设置投影球面能够在投影平面的基础上引入更多法向采样,尤其对于目标图像中既不在水平面上又不在竖直面上的像素,通过结合投影球面增加法向采样能够提高这些像素的深度恢复的准确性。另外,在上述目标图像和匹配图像都是鱼眼图像的实施例中,引入平行的N3个投影球面,还能够对视角大于180°的鱼眼的目标图像提供有利的投影面,提高深度恢复的准确性和可靠性。In the above embodiment, the N3 projection spheres are concentric spheres with the camera origin as the sphere center, and the radii of the N3 projection spheres are inversely proportional and equally distributed, wherein N3 is an integer greater than 1. Continuing to refer to Figure 6, taking the camera origin as the center of the sphere, the 64 radii from 0.5m to 32m are equally distributed in inverse proportion to form a projection sphere. Setting the projection sphere can introduce more normal sampling on the basis of the projection plane, especially for pixels in the target image that are neither on the horizontal nor vertical plane, by adding normal sampling in combination with the projection sphere, the depth recovery of these pixels can be improved. accuracy. In addition, in the above-mentioned embodiment in which the target image and the matching image are both fisheye images, the introduction of N3 parallel projection spheres can also provide favorable projection surfaces for fisheye target images with a viewing angle greater than 180°, improving depth recovery efficiency. Accuracy and reliability.

S202,根据所述目标图像中像素与所述多个投影图像中相应像素的匹配代价,确定所述目标图像中像素的估计深度。S202: Determine the estimated depth of the pixels in the target image according to the matching cost of the pixels in the target image and the corresponding pixels in the multiple projection images.

继续参见图4,目标图像在各投影图像中都具有相应像素,匹配代价(match cost)越小,则像素特征越相关,由此将匹配代价最小的相应像素所在深度作为目标图像中像素的估计深度,实现对目标图像的深度恢复,可以得到目标图像对应的深度图。Continuing to refer to Figure 4, the target image has corresponding pixels in each projection image. The smaller the matching cost, the more relevant the pixel features are. Therefore, the depth of the corresponding pixel with the smallest matching cost is used as the estimation of the pixel in the target image. Depth, to achieve depth recovery of the target image, the depth map corresponding to the target image can be obtained.

在一些实施例中,在进行匹配代价计算之前,可以先根据所述目标图像与所述匹配图像的相机相对位姿,确定所述匹配图像中与所述目标图像中像素一一对应的相应像素。具体可以采用现有的各种追踪算法实现相应像素的追踪,在此不做限定。在匹配图像中的相应像素确定后,可以根据匹配图像中的所述相应像素,确定各所述投影图像中与所述目标图像中像素一一对应的相应像素。参见图4中每个投影图像都具有与目标图像中像素的相应像素。本实施例通过相机相对位姿定位匹配图像和目标图像中相对应的像素,提高了对目标图像深度恢复的准确性和可靠性。In some embodiments, before performing the matching cost calculation, the corresponding pixels in the matching image that correspond one-to-one with the pixels in the target image may be determined according to the relative camera poses of the target image and the matching image. . Specifically, various existing tracking algorithms can be used to implement the tracking of the corresponding pixels, which is not limited here. After the corresponding pixels in the matching image are determined, the corresponding pixels in each of the projection images that correspond one-to-one with the pixels in the target image may be determined according to the corresponding pixels in the matching image. See Figure 4. Each projected image has a corresponding pixel to a pixel in the target image. This embodiment improves the accuracy and reliability of the depth recovery of the target image by matching the corresponding pixels in the image and the target image through the relative pose positioning of the camera.

车载单目相机、车辆的车身以及车载惯性测量单元(Inertial measurementunit,IMU)之间的位置关系预先标定。根据车身的移动方向和距离,就能确定车载单目相机的相机原点位置、相机拍摄区域、视角等信息。The positional relationship between the on-board monocular camera, the body of the vehicle, and the on-board inertial measurement unit (IMU) is pre-calibrated. According to the moving direction and distance of the vehicle body, information such as the camera origin position, camera shooting area, and angle of view of the vehicle-mounted monocular camera can be determined.

在根据所述目标图像与所述匹配图像的相机相对位姿,确定所述匹配图像中与所述目标图像中像素一一对应的相应像素之前,可以先确定出匹配图像和目标图像之间的相机相对位姿。例如,先采集车辆后轮的轮速计数据和车载IMU数据,其中,所述车辆后轮的轮速计数据指示了所述车载单目相机的水平运动距离,所述车载IMU数据指示了所述车载单目相机的水平运动方向;根据所述车辆后轮的轮速计数据和所述车载IMU数据,可以确定所述车载单目相机的相机位姿数据。车载单目相机拍摄的每一帧图像,都有对应的相机位姿。由此可以根据所述车载单目相机的相机位姿数据,确定匹配图像和目标图像之间的相机相对位姿。本实施中,由于车辆后轮与车身之间在水平方向无相对转动,后轮的轮速可以直接表征车身的移速,结合车辆后轮的轮速计数据和车载IMU数据得到相机位姿,提高了相机位姿的可靠性,进而提高可行驶区域检测的准确性和可靠性。Before determining the corresponding pixels in the matching image that correspond one-to-one with the pixels in the target image according to the relative poses of the cameras of the target image and the matching image, the relationship between the matching image and the target image may be determined first. The relative pose of the camera. For example, first collect wheel speedometer data and onboard IMU data of the rear wheel of the vehicle, wherein the wheel speedometer data of the rear wheel of the vehicle indicates the horizontal movement distance of the onboard monocular camera, and the onboard IMU data indicates all The horizontal movement direction of the vehicle-mounted monocular camera; the camera pose data of the vehicle-mounted monocular camera can be determined according to the wheel speedometer data of the rear wheels of the vehicle and the vehicle-mounted IMU data. Each frame of image captured by the vehicle-mounted monocular camera has a corresponding camera pose. Thereby, the relative pose of the camera between the matched image and the target image can be determined according to the camera pose data of the vehicle-mounted monocular camera. In this implementation, since there is no relative rotation between the rear wheels of the vehicle and the vehicle body in the horizontal direction, the wheel speed of the rear wheels can directly represent the moving speed of the vehicle body. The reliability of the camera pose is improved, thereby improving the accuracy and reliability of the drivable area detection.

步骤S202中,具体的实现方式例如可以是先获取所述目标图像中像素的目标像素窗口特征,以及所述多个投影图像中相应像素的投影像素窗口特征。这里的窗口特征例如是以预设大小的采样窗口在目标图像和投影图像上滑动,以窗口内像素特征的均值作为窗口中心像素的窗口特征。该采样窗口的大小可以是7*7、5*5,也可以是1*1,在此不做限定。目标像素窗口特征例如是以目标图像中像素为中心的采样窗口内像素的灰度均值。匹配像素窗口特征例如是以相应像素为中心的采样窗口内像素的灰度均值。然后,根据所述目标像素窗口特征和所述投影像素窗口特征,获取所述目标图像中像素与各所述投影图像中相应像素的匹配代价。例如,对目标图像中像素和投影图像中相应像素以7*7窗口采样得到的灰度均值误差,作为相应像素与目标图像中像素的匹配代价。得到各投影图像中相应像素对目标图像中像素的匹配代价后,可以将所述匹配代价最小的所述相应像素对应的深度,作为所述目标图像中像素的估计深度,其中,所述相应像素对应的深度,是所述相应像素所在投影面对应的深度。本实施例通过将匹配代价最小的相应像素所对应的深度,作为目标图像中像素的估计深度,提高了对目标图像中像素深度恢复的准确性。In step S202, a specific implementation manner may be, for example, to first acquire target pixel window characteristics of pixels in the target image and projection pixel window characteristics of corresponding pixels in the multiple projection images. The window feature here is, for example, sliding a sampling window of a preset size on the target image and the projection image, and taking the mean value of the pixel features in the window as the window feature of the center pixel of the window. The size of the sampling window may be 7*7, 5*5, or 1*1, which is not limited here. The target pixel window feature is, for example, the average gray value of the pixels in the sampling window centered on the pixel in the target image. The matching pixel window feature is, for example, the gray mean value of the pixels in the sampling window centered on the corresponding pixel. Then, according to the feature of the target pixel window and the feature of the projected pixel window, the matching cost of the pixel in the target image and the corresponding pixel in each of the projected images is obtained. For example, the gray mean error obtained by sampling the pixels in the target image and the corresponding pixels in the projection image with a 7*7 window is used as the matching cost between the corresponding pixels and the pixels in the target image. After obtaining the matching cost of the corresponding pixel in each projection image to the pixel in the target image, the depth corresponding to the corresponding pixel with the smallest matching cost can be used as the estimated depth of the pixel in the target image, wherein the corresponding pixel The corresponding depth is the depth corresponding to the projection surface where the corresponding pixel is located. This embodiment improves the accuracy of restoring the depth of pixels in the target image by using the depth corresponding to the corresponding pixel with the smallest matching cost as the estimated depth of the pixel in the target image.

S203,根据所述目标图像中像素的估计深度,获取所述目标图像对应的目标三维点云数据。S203: Acquire target three-dimensional point cloud data corresponding to the target image according to the estimated depth of the pixels in the target image.

目标图像中各像素确定估计深度的过程可以并行执行。在目标图像中像素都确定出估计深度后,得到目标图像对应的深度图像。进而将深度图像结合目标图像中各像素的像素位置,可以得到各像素的三维信息,得到目标图像对应的目标三维点云数据。The process of determining the estimated depth for each pixel in the target image can be performed in parallel. After all the pixels in the target image have determined the estimated depth, a depth image corresponding to the target image is obtained. Then, combining the depth image with the pixel position of each pixel in the target image, the three-dimensional information of each pixel can be obtained, and the target three-dimensional point cloud data corresponding to the target image can be obtained.

图2所示三维重建的实施例,通过利用匹配图像在不同深度投影面上的投影图像,以代价匹配对目标图像实现深度恢复,转化得到目标三维点云数据,提高了对单目图像进行三维重建的准确性和可靠性,进而提高了可行驶区域检测的准确性和可靠性。In the embodiment of 3D reconstruction shown in FIG. 2 , by using the projection images of the matching images on the projection surfaces of different depths, the depth recovery of the target image is achieved by matching the cost, and the target 3D point cloud data is obtained by transformation, which improves the 3D reconstruction of the monocular image. The accuracy and reliability of the reconstruction, in turn, improve the accuracy and reliability of the drivable area detection.

S103,根据所述目标三维点云数据,确定所述目标图像的拍摄区域中最近障碍物点。S103, according to the target three-dimensional point cloud data, determine the nearest obstacle point in the shooting area of the target image.

目标三维点云数据体现了目标图像的拍摄区域中的三维立体信息,由此可以过滤出障碍物的三维点,并根据相对于相机原点的径向距离再次过滤出用于确定可行驶区域边界的最近障碍物点。The target 3D point cloud data reflects the 3D stereo information in the shooting area of the target image, so that the 3D points of obstacles can be filtered out, and the 3D points used to determine the boundary of the drivable area can be filtered again according to the radial distance relative to the origin of the camera. The nearest obstacle point.

在一些实施例中,可以根据所述目标三维点云数据和对所述目标图像的拍摄区域水平切分的极坐标栅格网络,确定所述极坐标栅格网络中各栅格中包含障碍物点的数量,然后根据所述极坐标栅格网络的各扇形分区中的最近障碍物栅格,确定所述目标图像的拍摄区域中各方向的最近障碍物点。In some embodiments, it may be determined that each grid in the polar coordinate grid network contains obstacles according to the target three-dimensional point cloud data and the polar coordinate grid network that horizontally divides the shooting area of the target image The number of points is determined, and then the nearest obstacle points in each direction in the shooting area of the target image are determined according to the nearest obstacle grids in each sector of the polar coordinate grid network.

具体地,首先可以利用极坐标栅格网络对相机拍摄区域进行空间划分,以便对其中最近障碍物点进行提取。例如,可以根据所述目标三维点云数据和所述目标图像对应的相机原点位置,确定预设的极坐标栅格网络的各栅格中的目标三维点。相机原点位置例如是预先标定的、用于指示相机原点相对于地面高度、相对于车身位置的信息。参见图7,是本申请实施例提供的一种极坐标栅格网络的俯视图。如图7所示,所述极坐标栅格网络是对目标图像的拍摄区域,以扇形排布的第一类切割面和正对相机且平行排布的第二类切割面叠加分割形成的分割网络,所述第一类切割面的交线与相机原点的对地垂线共线。第一类切割面和第二类切割面都是与地面垂直的平面。参见图7,多个第一类切割面对目标图像的拍摄区域的分割,在俯视图上例如可以形成以相机原点为中心,将水平方向175°划分为128个扇面分区。在此基础上,多个第二类切割面与相机正对面平行分布进行叠加分割。在一些实施例中,越靠近车辆的像素分辨率越高,目标三维点越密集,而且越靠近车辆的障碍物可能产生越大的影响,对靠近相机原点的分割密度应大于远离相机原点的分割密度。靠近相机原点的第二类切割面分布密度大于远离相机原点的第二类切割面的分布密度。例如,第二类切割面按与相机原点的径向距离从0.5m到32m成反比例等差划分出63段形成栅格。Specifically, firstly, a polar coordinate grid network can be used to spatially divide the area captured by the camera, so as to extract the nearest obstacle points therein. For example, the target three-dimensional point in each grid of the preset polar coordinate grid network may be determined according to the target three-dimensional point cloud data and the camera origin position corresponding to the target image. The camera origin position is, for example, pre-calibrated information used to indicate the position of the camera origin relative to the ground height and relative to the vehicle body. Referring to FIG. 7 , it is a top view of a polar coordinate grid network provided by an embodiment of the present application. As shown in Figure 7, the polar coordinate grid network is a segmentation network formed by superimposing and segmenting the first type of cutting planes arranged in a fan shape and the second type of cutting planes facing the camera and arranged in parallel for the shooting area of the target image. , the intersection line of the first type of cutting plane is collinear with the vertical line to the ground of the origin of the camera. Both the first type of cutting surface and the second type of cutting surface are planes perpendicular to the ground. Referring to FIG. 7 , the division of the shooting area of the target image by a plurality of first-type cutting planes can be formed, for example, in a plan view, with the camera origin as the center, dividing the horizontal direction 175° into 128 sector divisions. On this basis, multiple second-type cutting planes are distributed in parallel with the front face of the camera for superposition and segmentation. In some embodiments, the closer to the vehicle the higher the pixel resolution, the denser the target 3D points, and the closer the vehicle to the obstacle may have a greater impact, the density of the segmentation close to the camera origin should be greater than the segmentation far from the camera origin density. The distribution density of the second type of cut surfaces close to the camera origin is greater than that of the second type of cut surfaces far from the camera origin. For example, the second type of cutting plane is divided into 63 segments to form a grid according to the radial distance from the camera origin from 0.5m to 32m in inverse proportion.

继续参见图7,确定预设的极坐标栅格网络的各栅格中的目标三维点后,可以确定各所述栅格中包含障碍物点的数量,其中,所述障碍物点为对地高度大于预设障碍物高度阈值的目标三维点。预设障碍物高度阈值例如可以是4cm、5cm、6cm等车辆可以直接越过的限额高度,例如车辆的底盘高度。然后,在所述第一类切割面分割出的扇形分区中,确定最近障碍物栅格,其中,所述最近障碍物栅格是所述扇形分区中与所述相机原点径向距离最近、且所包含障碍物点的数量大于预设数量阈值的栅格。例如,在以原点为中心向图7所示极坐标栅格网络各方向搜索障碍物点数量大于预设数量阈值的第一个栅格,作为该方向上的最近障碍物栅格。最后可以根据所述最近障碍物栅格,确定所述目标图像的拍摄区域中最近障碍物点。参见图8,是本申请实施例提供的一种目标图像的拍摄区域中最近障碍物点示意图。在图8所示的场景中,车位两侧的其他车辆和石柱都在最近障碍物点以外,在最近障碍物点圈定的相机侧区域中,可以确定出可行驶区域。本实施例利用对相机原点建立的极坐标栅格网络对目标图像的拍摄区域进行空间划分,将各方向扇形分区中可能是障碍物且对相机原点径向距离最小的栅格提取出来,用于确定相机原点各方向的最近障碍物点,提高了最近障碍物点的准确性,进而提高可行驶区域检测的准确性和可靠性。Continuing to refer to FIG. 7 , after determining the target three-dimensional point in each grid of the preset polar coordinate grid network, the number of obstacle points contained in each grid can be determined, wherein the obstacle point is the ground The target 3D point whose height is greater than the preset obstacle height threshold. The preset obstacle height threshold may be, for example, a limit height that the vehicle can directly cross, such as 4 cm, 5 cm, 6 cm, etc., for example, the chassis height of the vehicle. Then, in the sector partition segmented by the first type of cutting plane, a nearest obstacle grid is determined, wherein the nearest obstacle grid is the radial distance closest to the camera origin in the sector partition, and A raster that contains more than a preset number of obstacle points. For example, search for the first grid with the number of obstacle points greater than the preset number threshold in each direction of the polar coordinate grid network shown in FIG. 7 from the origin as the center, as the nearest obstacle grid in this direction. Finally, the nearest obstacle point in the shooting area of the target image may be determined according to the nearest obstacle grid. Referring to FIG. 8 , it is a schematic diagram of a nearest obstacle point in a shooting area of a target image provided by an embodiment of the present application. In the scene shown in Figure 8, other vehicles and stone pillars on both sides of the parking space are outside the nearest obstacle point. In the area on the camera side delineated by the nearest obstacle point, the drivable area can be determined. In this embodiment, the polar coordinate grid network established for the camera origin is used to spatially divide the shooting area of the target image, and the grids that may be obstacles and the smallest radial distance from the camera origin in the fan-shaped partitions in each direction are extracted for use in The nearest obstacle points in all directions of the origin of the camera are determined, which improves the accuracy of the nearest obstacle points, thereby improving the accuracy and reliability of the drivable area detection.

最近障碍物点可以是最近障碍物栅格的中心点或者是根据最近障碍物栅格中目标三维点确定的点。在一些实施例中,可以是获取所述最近障碍物栅格中所包含障碍物点的平均位置点;将各所述最近障碍物栅格对应的平均位置点,作为所述目标图像的拍摄区域中最近障碍物点。本实施例进一步优化最近障碍物点的位置,提高障碍物点的准确性。The closest obstacle point can be the center point of the closest obstacle grid or a point determined from the target 3D point in the closest obstacle grid. In some embodiments, the average position point of the obstacle points included in the nearest obstacle grid may be obtained; the average position point corresponding to each of the nearest obstacle grids may be used as the shooting area of the target image The nearest obstacle point in the middle. This embodiment further optimizes the position of the nearest obstacle point and improves the accuracy of the obstacle point.

S104,根据各所述最近障碍物点,确定所述目标图像的拍摄区域中的可行驶区域。S104: Determine a drivable area in the shooting area of the target image according to each of the nearest obstacle points.

在最近障碍物点确定后,可以以最近障碍物点作为可行驶区域的边界。而为了提高可行驶区域的可靠性,还可以结合均匀分割网络景进行多帧融合处理,对可行驶区域的边界进行持续优化。After the nearest obstacle point is determined, the nearest obstacle point can be used as the boundary of the drivable area. In order to improve the reliability of the drivable area, multi-frame fusion processing can also be performed in combination with the evenly divided network scene, and the boundary of the drivable area can be continuously optimized.

在一些实施例中,可以根据所述最近障碍物点和所述目标图像对应的相机原点位置,获取包含所述最近障碍物点的均匀分割网络。其中,所述均匀分割网络是对目标图像的拍摄区域在水平方向以均匀方形网格分割形成的分割网络。参见图9,是本申请实施例提供的一种均匀分割网络俯视图。如图9所示,将水平空间按0.1m*0.1m的网格均匀划分出均匀分割网络,其中的方形网格实际上是对水平方向进行均匀分割的立体网格,每个网格单元实际上是横截面为方形的棱柱体区域。In some embodiments, a uniform segmentation network including the closest obstacle point may be obtained according to the closest obstacle point and the camera origin position corresponding to the target image. Wherein, the uniform segmentation network is a segmentation network formed by dividing the shooting area of the target image with a uniform square grid in the horizontal direction. Referring to FIG. 9 , it is a plan view of a uniformly divided network provided by an embodiment of the present application. As shown in Figure 9, the horizontal space is evenly divided into a uniformly divided network according to a grid of 0.1m*0.1m. The square grid is actually a three-dimensional grid that uniformly divides the horizontal direction. Above is a prismatic region with a square cross-section.

本实施例通过均匀分割网络俯视图对空间进行均匀分割,在对所述目标图像的拍摄区域水平切分的均匀分割网络中,可以根据所述均匀分割网络中各网格单元相对于所述最近障碍物点的位置,确定各网格单元的加权值。具体地,可以根据所述均匀分割网络中各网格单元对相机原点的径向距离,以及所述网格单元方向上的最近障碍物点对相机原点的径向距离,确定各所述网格单元的加权值。参见图9所示,每个箭头代表一个径向方向,对每个径向方向上的网格单元确定各自的加权值。该加权值的计算是根据网格单元对相机原点的径向距离,和其网格单元所在方向上障碍物点对相机原点的径向距离来确定的。具体在后续实施例中对加权值的计算进行举例说明。在每个网格单元得到加权值时,形成了以最近障碍物点为截断面的截断有向距离场,可以根据相对于截断面的有向距离,对截断有效距离场中各网格单元进行权值计算。得到各网格单元的加权值后,根据各所述网格单元的初始权值以及所述加权值,确定各所述网格单元的新权值,其中,所述新权值大于或等于最小权阈值、且小于或等于最大权阈值,所述网格单元的初始权值是0或者是在对前一帧图像确定可行驶区域时确定的新权值。应当理解地,最大权阈值和最小权阈值是用于对新权值进行限定,避免新权值的无限增大。例如,假设多个连续帧中,车库墙壁对应的网格单元的加权值维持取+1,那么其新权值增大到10(最大权阈值)后不再增大。In this embodiment, the space is evenly divided by evenly dividing the top view of the network. In the evenly dividing network that horizontally divides the shooting area of the target image, each grid unit in the evenly dividing network can be divided according to the relative position of each grid unit to the nearest obstacle. The position of the object point determines the weighted value of each grid cell. Specifically, each grid unit may be determined according to the radial distance from each grid unit in the uniform segmentation network to the camera origin, and the radial distance from the nearest obstacle point in the direction of the grid unit to the camera origin. The weighted value of the unit. Referring to FIG. 9 , each arrow represents a radial direction, and a respective weighting value is determined for the grid cells in each radial direction. The calculation of the weighted value is determined according to the radial distance of the grid unit to the camera origin, and the radial distance of the obstacle point to the camera origin in the direction of the grid unit. Specifically, the calculation of the weighted value will be illustrated in the following embodiments. When each grid cell obtains a weighted value, a truncated directional distance field with the nearest obstacle point as the truncated surface is formed. According to the directional distance relative to the truncated surface, each grid cell in the truncated effective distance field can be calculated. Weight calculation. After the weighted value of each grid unit is obtained, a new weight of each grid unit is determined according to the initial weight of each grid unit and the weighted value, wherein the new weight is greater than or equal to the minimum value The weight threshold is less than or equal to the maximum weight threshold, and the initial weight of the grid unit is 0 or a new weight determined when the drivable area is determined for the previous frame image. It should be understood that the maximum weight threshold and the minimum weight threshold are used to limit the new weight to avoid an infinite increase of the new weight. For example, assuming that in multiple consecutive frames, the weight value of the grid unit corresponding to the garage wall remains +1, then its new weight value will not increase after it increases to 10 (the maximum weight threshold).

然后,可以根据所述新权值指示的第一类网格单元或第二类网格单元,确定所述目标图像的拍摄区域中的可行驶区域,其中,所述新权值指示的第一类网格单元与所述相机原点之间有所述最近障碍物点,所述新权值指示的第二类网格单元与所述相机原点之间没有所述最近障碍物点。可以理解为,根据新权值的数值可以对网格单元实现分类,确定各网格单元是第一类网格单元还是第二类网格单元。例如,可以根据所述网格单元的新权值,确定所述网格单元是第一类网格单元或者第二类网格单元,其中,所述第一类网格单元与所述相机原点之间有所述最近障碍物点,所述第二类网格单元与所述相机原点之间没有所述最近障碍物点。第二类网格单元可以理解为是在最近障碍物点和相机原点之间的网格单元。最后,根据所述第二类网格单元,确定所述目标图像的拍摄区域中的可行驶区域,其中,与所述第一类网格单元相邻的所述第二类网格单元是所述可行驶区域的边界。应理解地,第二类网格单元合并后可以形成目标图像的拍摄区域中的可行驶区域。Then, a drivable area in the shooting area of the target image may be determined according to the first type of grid unit or the second type of grid unit indicated by the new weight, wherein the first type of grid unit indicated by the new weight There is the closest obstacle point between the grid-like unit and the camera origin, and there is no closest obstacle point between the second grid unit indicated by the new weight and the camera origin. It can be understood that the grid cells can be classified according to the value of the new weight, and it is determined whether each grid cell is a first-type grid cell or a second-type grid cell. For example, according to the new weight of the grid unit, it may be determined that the grid unit is a first-type grid unit or a second-type grid unit, wherein the first-type grid unit and the camera origin There is the closest obstacle point in between, and there is no closest obstacle point between the second type of grid cell and the camera origin. The second type of grid cell can be understood as the grid cell between the nearest obstacle point and the camera origin. Finally, a drivable area in the shooting area of the target image is determined according to the second type of grid unit, wherein the second type of grid unit adjacent to the first type of grid unit is the the boundary of the drivable area. It should be understood that the second type of grid cells can form a drivable area in the shooting area of the target image after being merged.

上述实施例中,在根据各所述网格单元的初始权值以及所述加权值,确定各所述网格单元的新权值之前,还可以先确定所述目标图像对应的各网格单元的初始权值。目标图像各网格单元的初始权值是根据匹配图像和目标图像的相机相对位姿,将匹配图像对应的截断有向距离场旋转平移变换至目标图像对应的截断有向距离场中,确定匹配图像对应的网格单元与目标图像对应的网格单元之间的对应关系。而在匹配图像对应的网格单元所确定的新权值,就是目标图像对应的网格单元的初始权值。In the above-mentioned embodiment, before determining the new weight value of each grid unit according to the initial weight value and the weight value of each grid unit, each grid unit corresponding to the target image may also be determined first. the initial weight of . The initial weight of each grid unit of the target image is based on the camera relative pose of the matching image and the target image, and the truncated directional distance field corresponding to the matching image is rotated and translated into the truncated directional distance field corresponding to the target image to determine the matching. Correspondence between the grid unit corresponding to the image and the grid unit corresponding to the target image. The new weight determined in the grid unit corresponding to the matching image is the initial weight of the grid unit corresponding to the target image.

假如目标图像是连续帧图像的首帧图像,则初始权值为0。If the target image is the first frame image of consecutive frame images, the initial weight is 0.

上述实施例通过计算目标图像拍摄区域在均匀分割网络中各网格单元的加权值,对网格单元更新权值,随着连续图像帧的不断更新,不断加权更新网格单元的新权值,不仅能够平滑噪声还能够起到弥补单帧最近障碍物点漏检的问题,提高了对可行驶区域检测的可靠性。In the above-mentioned embodiment, by calculating the weighted value of each grid unit in the uniform segmentation network of the target image shooting area, the weighted value of the grid unit is updated, and the new weighted value of the grid unit is continuously updated with the continuous updating of continuous image frames, It can not only smooth the noise, but also make up for the problem of missed detection of the nearest obstacle point in a single frame, and improve the reliability of the detection of the drivable area.

在上述实施例中,确定各所述网格单元的加权值的实现方式,例如可以是:In the above embodiment, the implementation manner of determining the weighted value of each grid unit may be, for example:

若所述均匀分割网络中网格单元对相机原点的径向距离,减去所述网格单元方向上的最近障碍物点对相机原点的径向距离的差值,大于或等于距离上限阈值,则所述网格单元的加权值为第一值。If the radial distance between the grid unit in the uniform segmentation network and the camera origin, minus the difference between the radial distance between the nearest obstacle point in the direction of the grid unit and the camera origin, is greater than or equal to the upper limit of the distance threshold, Then the weighted value of the grid unit is the first value.

若所述最近障碍物点对相机原点的径向距离,减去所述最近障碍物点方向上网格单元对相机原点的径向距离的差值,小于或等于距离下限阈值,则所述网格单元的加权值为第二值。If the radial distance between the closest obstacle point and the camera origin, minus the difference between the radial distance between the grid unit and the camera origin in the direction of the closest obstacle point, is less than or equal to the lower threshold of the distance, then the grid The weighted value of the cell is the second value.

若所述最近障碍物点对相机原点的径向距离,减去所述最近障碍物点方向上网格单元对相机原点的径向距离的差值,小于所述距离上限阈值且大于所述距离下限阈值,则所述网格单元的加权值为第三值,其中,所述第三值是根据所述差值和预设的平滑连续函数确定的值。If the radial distance between the closest obstacle point and the camera origin, minus the difference between the radial distance between the grid unit and the camera origin in the direction of the closest obstacle point, is less than the distance upper threshold and greater than the distance lower limit the threshold value, the weighted value of the grid unit is a third value, where the third value is a value determined according to the difference value and a preset smooth continuous function.

其中,所述距离上限阈值是所述距离下限阈值的相反数,所述第一值是所述第二值的相反数,所述第三值的绝对值小于所述距离上限阈值的绝对值或所述距离下限阈值的绝对值。Wherein, the distance upper threshold is the inverse of the distance lower threshold, the first value is the inverse of the second value, and the absolute value of the third value is less than the absolute value of the distance upper threshold or The absolute value of the distance lower threshold.

作为一种示例,可以是根据下列公式一,确定指定方向上的网格单元的加权值f(c):As an example, the weighted value f(c) of the grid cells in the specified direction can be determined according to the following formula 1:

Figure BDA0002410636520000171
Figure BDA0002410636520000171

其中,c是网格单元对相机原点的径向距离,d是该网格方向上最近障碍物点对相机原点的径向距离。公式一中,距离上限阈值是0.3,距离下线阈值是-0.3,第一值是1,第二值是-1,平滑连续函数是预设的三角函数。where c is the radial distance from the grid unit to the camera origin, and d is the radial distance from the closest obstacle point in the grid direction to the camera origin. In formula 1, the upper threshold of the distance is 0.3, the lower threshold of the distance is -0.3, the first value is 1, the second value is -1, and the smooth continuous function is a preset trigonometric function.

参见图10,是本申请实施例提供的一种可行驶区域示意图。图10中阴影区域为可行驶区域。如图10所示的可行驶区域边界是通过多帧图像的加权值累计加权更新得到的结果,是对图8所示最近障碍物点所标记位置的进一步优化。图10所示的可行驶区域边界融合了目标图像的最近障碍物所确定的网格单元加权值,以及之前连续帧累计得到的初始权值,具有较高的可靠性。Referring to FIG. 10 , it is a schematic diagram of a drivable area provided by an embodiment of the present application. The shaded area in Figure 10 is the drivable area. The drivable area boundary shown in FIG. 10 is the result obtained by the cumulative weighted update of the weighted values of the multi-frame images, which is a further optimization of the position marked by the nearest obstacle point shown in FIG. 8 . The drivable area boundary shown in Figure 10 combines the grid cell weights determined by the nearest obstacles in the target image and the initial weights accumulated in the previous consecutive frames, which has high reliability.

上述实施例实现了对各网格单元加权值的确定,并通过平滑连续函数对差值在距离上限阈值和距离下限阈值之间的网格单元平滑过渡,进一步降低多帧融合加权累积过程中的噪声,提高了对网格单元新权值确定的可靠性,进而提高了对可行驶区域检测的可靠性。The above embodiment realizes the determination of the weighted value of each grid unit, and the smooth transition of the grid unit with the difference between the distance upper threshold and the distance lower threshold through a smooth continuous function, further reduces the multi-frame fusion weighted accumulation process. noise, which improves the reliability of the determination of the new weights of the grid cells, and further improves the reliability of the detection of the drivable area.

图1所示实施例通过获取以车载单目相机对车辆周围区域拍摄所采集的匹配图像和目标图像,其中,所述匹配图像是与所述目标图像连续的前一帧图像;然后根据所述匹配图像对所述目标图像进行三维重建,得到所述目标图像对应的目标三维点云数据,从而实现对单目相机采集的连续帧图像重建三维点云数据;接着根据所述目标三维点云数据,确定所述目标图像的拍摄区域中最近障碍物点,并根据各所述最近障碍物点,确定所述目标图像的拍摄区域中的可行驶区域,无需采集三维图像或大量图像样本进行机器学习,降低了对于图像采集方式的依赖,降低了处理难度、提高了检测的准确性和可靠性。The embodiment shown in FIG. 1 acquires a matching image and a target image obtained by shooting the surrounding area of the vehicle with a vehicle-mounted monocular camera, wherein the matching image is the previous frame image that is continuous with the target image; and then according to the Performing three-dimensional reconstruction on the target image by matching the image to obtain the target three-dimensional point cloud data corresponding to the target image, so as to reconstruct the three-dimensional point cloud data of the continuous frame images collected by the monocular camera; then according to the target three-dimensional point cloud data , determine the nearest obstacle point in the shooting area of the target image, and determine the drivable area in the shooting area of the target image according to each of the nearest obstacle points, without collecting three-dimensional images or a large number of image samples for machine learning , reducing the dependence on the image acquisition method, reducing the processing difficulty, and improving the accuracy and reliability of detection.

参见图11,是本申请实施例提供的一种可行驶区域检测装置结构示意图。如图11所示的可行驶区域检测装置30包括:Referring to FIG. 11 , it is a schematic structural diagram of a drivable area detection device provided by an embodiment of the present application. The drivable area detection device 30 shown in FIG. 11 includes:

图像采集模块31,用于获取以车载单目相机对车辆周围区域拍摄所采集的匹配图像和目标图像,其中,所述匹配图像是所述目标图像的前一帧图像。The image acquisition module 31 is configured to acquire a matching image and a target image captured by the vehicle-mounted monocular camera on the surrounding area of the vehicle, wherein the matching image is a previous frame image of the target image.

第一处理模块32,用于根据所述匹配图像对所述目标图像进行三维重建,得到所述目标图像对应的目标三维点云数据。The first processing module 32 is configured to perform three-dimensional reconstruction on the target image according to the matching image, and obtain target three-dimensional point cloud data corresponding to the target image.

第二处理模块33,用于根据所述目标三维点云数据,确定所述目标图像的拍摄区域中最近障碍物点。The second processing module 33 is configured to determine the nearest obstacle point in the shooting area of the target image according to the target three-dimensional point cloud data.

第三处理模块34,用于根据各所述最近障碍物点,确定所述目标图像的拍摄区域中的可行驶区域。The third processing module 34 is configured to determine a drivable area in the shooting area of the target image according to each of the nearest obstacle points.

本实施例提供的可行驶区域检测装置,通过获取以车载单目相机对车辆周围区域拍摄所采集的匹配图像和目标图像,其中,所述匹配图像是与所述目标图像连续的前一帧图像;然后根据所述匹配图像对所述目标图像进行三维重建,得到所述目标图像对应的目标三维点云数据,从而实现对单目相机采集的连续帧图像重建三维点云数据;接着根据所述目标三维点云数据,确定所述目标图像的拍摄区域中最近障碍物点,并根据各所述最近障碍物点,确定所述目标图像的拍摄区域中的可行驶区域,无需采集三维图像或大量图像样本进行机器学习,降低了对于图像采集方式的依赖,降低了处理难度、提高了检测的准确性和可靠性。The drivable area detection device provided by this embodiment acquires a matching image and a target image collected by shooting the surrounding area of the vehicle with a vehicle-mounted monocular camera, wherein the matching image is a previous frame image that is continuous with the target image Then carry out three-dimensional reconstruction of the target image according to the matching image, and obtain the target three-dimensional point cloud data corresponding to the target image, thereby realizing the reconstruction of three-dimensional point cloud data for the continuous frame images collected by the monocular camera; then according to the Target three-dimensional point cloud data, determine the nearest obstacle point in the shooting area of the target image, and determine the drivable area in the shooting area of the target image according to each of the nearest obstacle points, without collecting three-dimensional images or a large number of Machine learning is performed on image samples, which reduces the dependence on the image acquisition method, reduces the processing difficulty, and improves the accuracy and reliability of detection.

在一些实施例中,第一处理模块32,具体用于将所述匹配图像按照所述目标图像的视角投影至预设的多个投影面上,得到多个投影图像,其中,每个所述投影面对应一相对于相机原点的深度;根据所述目标图像中像素与所述多个投影图像中相应像素的匹配代价,确定所述目标图像中像素的估计深度;根据所述目标图像中像素的估计深度,获取所述目标图像对应的目标三维点云数据。In some embodiments, the first processing module 32 is specifically configured to project the matching image onto a plurality of preset projection surfaces according to the viewing angle of the target image to obtain a plurality of projection images, wherein each of the The projection surface corresponds to a depth relative to the origin of the camera; according to the matching cost of the pixels in the target image and the corresponding pixels in the plurality of projection images, the estimated depth of the pixels in the target image is determined; The estimated depth of the pixel is obtained, and the target 3D point cloud data corresponding to the target image is obtained.

本实施例利用匹配图像在不同深度投影面上的投影图像,以代价匹配对目标图像实现深度恢复,转化得到目标三维点云数据,提高了对单目图像进行三维重建的准确性和可靠性,进而提高了可行驶区域检测的准确性和可靠性。In this embodiment, the projected images of the matched images on the projection surfaces of different depths are used to achieve depth recovery of the target image at cost matching, and the target three-dimensional point cloud data is obtained by transformation, which improves the accuracy and reliability of the three-dimensional reconstruction of the monocular image. Thus, the accuracy and reliability of the drivable area detection are improved.

在一些实施例中,所述投影面包括:N1个竖直投影平面;所述N1个竖直投影平面平行于相机正对面,且所述相机原点到所述N1个竖直投影平面的距离成反比例等差分布,其中,N1为大于1的整数。本实施例通过竖直投影平面实现对相机前方区域的深度恢复,提高在弯道等复杂环境下深度恢复的准确性,进而提高对目标图像三维重建的准确性和可靠性。In some embodiments, the projection planes include: N1 vertical projection planes; the N1 vertical projection planes are parallel to the front of the camera, and the distances from the camera origin to the N1 vertical projection planes are equal to Inversely proportional equal difference distribution, where N1 is an integer greater than 1. This embodiment realizes the depth recovery of the area in front of the camera through the vertical projection plane, improves the accuracy of depth recovery in complex environments such as curves, and further improves the accuracy and reliability of the three-dimensional reconstruction of the target image.

在一些实施例中,所述投影面还包括:N2个水平投影平面和/或N3个投影球面;其中,所述N2个水平投影平面平行于相机正下方地面,且所述N2个水平投影平面在以所述地面为对称中心的地面分布范围内均匀排列,其中,N2为大于1的整数;所述N3个投影球面为以所述相机原点为球心的同心球面,且所述N3个投影球面的半径成反比例等差分布,其中,N3为大于1的整数。本实施例通过水平投影平面恢复目标图像中地面区域的深度,通过投影球面引入更多法向采样,提高深度恢复的准确性和可靠性。对于目标图像中既不在水平面上又不在竖直面上的点,通过结合投影球面可以增加法向采样能够提高这些点的深度恢复的准确性。另外,引入平行的投影球面还能够对视角大于180度的鱼眼的目标图像提供有利的投影面,提高深度恢复的准确性和可靠性。In some embodiments, the projection surface further includes: N2 horizontal projection planes and/or N3 projection spherical surfaces; wherein the N2 horizontal projection planes are parallel to the ground directly below the camera, and the N2 horizontal projection planes Evenly arranged in the ground distribution range with the ground as the center of symmetry, where N2 is an integer greater than 1; the N3 projection spheres are concentric spheres with the camera origin as the center of the sphere, and the N3 projection spheres are The radius of the sphere is inversely proportional and equally distributed, where N3 is an integer greater than 1. In this embodiment, the depth of the ground area in the target image is recovered through the horizontal projection plane, and more normal samples are introduced through the projection sphere to improve the accuracy and reliability of depth recovery. For the points in the target image that are neither on the horizontal nor vertical planes, the depth recovery accuracy of these points can be improved by increasing the normal sampling by combining with the projection sphere. In addition, the introduction of a parallel projection sphere can also provide a favorable projection surface for the fisheye target image with a viewing angle greater than 180 degrees, and improve the accuracy and reliability of depth recovery.

在一些实施例中,第一处理模块32,具体用于获取所述目标图像中像素的目标像素窗口特征;获取所述多个投影图像中相应像素的投影像素窗口特征;根据所述目标像素窗口特征和所述投影像素窗口特征,获取所述目标图像中像素与各所述投影图像中相应像素的匹配代价;将所述匹配代价最小的所述相应像素对应的深度,作为所述目标图像中像素的估计深度,其中,所述相应像素对应的深度,是所述相应像素所在投影面对应的深度。In some embodiments, the first processing module 32 is specifically configured to acquire target pixel window characteristics of pixels in the target image; acquire projection pixel window characteristics of corresponding pixels in the multiple projection images; feature and the feature of the projected pixel window, obtain the matching cost between the pixel in the target image and the corresponding pixel in each of the projected images; take the depth corresponding to the corresponding pixel with the smallest matching cost as the depth of the corresponding pixel in the target image. The estimated depth of the pixel, wherein the depth corresponding to the corresponding pixel is the depth corresponding to the projection surface where the corresponding pixel is located.

本实施例通过将匹配代价最小的相应像素所对应的深度,作为目标图像中像素的估计深度,提高了对目标图像中像素深度恢复的准确性。This embodiment improves the accuracy of restoring the depth of pixels in the target image by using the depth corresponding to the corresponding pixel with the smallest matching cost as the estimated depth of the pixel in the target image.

在一些实施例中,第一处理模块32,在所述根据所述目标图像中像素与所述多个投影图像中相应像素的匹配代价,确定所述目标图像中像素的估计深度之前,还用于根据所述目标图像与所述匹配图像的相机相对位姿,确定所述匹配图像中与所述目标图像中像素一一对应的相应像素;根据所述匹配图像中的所述相应像素,确定各所述投影图像中与所述目标图像中像素一一对应的相应像素。In some embodiments, the first processing module 32, before determining the estimated depth of the pixels in the target image according to the matching cost of the pixels in the target image and the corresponding pixels in the plurality of projection images, further use according to the camera relative poses of the target image and the matching image, to determine the corresponding pixels in the matching image that correspond to the pixels in the target image one-to-one; according to the corresponding pixels in the matching image, determine The corresponding pixels in each of the projection images correspond to the pixels in the target image one-to-one.

本实施例通过相机相对位姿定位匹配图像和目标图像中相对应的像素,提高了对目标图像深度恢复的准确性和可靠性。This embodiment improves the accuracy and reliability of the depth recovery of the target image by matching the corresponding pixels in the image and the target image through the relative pose positioning of the camera.

在一些实施例中,第一处理模块32,在所述根据所述目标图像与所述匹配图像的相机相对位姿,确定所述匹配图像中与所述目标图像中像素一一对应的相应像素之前,还用于采集车辆后轮的轮速计数据和车载惯性测量单元IMU数据,其中,所述车辆后轮的轮速计数据指示了所述车载单目相机的水平运动距离,所述车载IMU数据指示了所述车载单目相机的水平运动方向;根据所述车辆后轮的轮速计数据和所述车载IMU数据,确定所述车载单目相机的相机位姿数据;根据所述车载单目相机的相机位姿数据,确定所述目标图像与所述匹配图像的相机相对位姿。In some embodiments, the first processing module 32 determines, according to the camera relative poses of the target image and the matching image, corresponding pixels in the matching image that correspond to pixels in the target image one-to-one Before, it is also used to collect the wheel speedometer data of the rear wheel of the vehicle and the IMU data of the vehicle-mounted inertial measurement unit, wherein the wheel speedometer data of the rear wheel of the vehicle indicates the horizontal movement distance of the vehicle-mounted monocular camera, and the vehicle-mounted The IMU data indicates the horizontal movement direction of the vehicle-mounted monocular camera; the camera pose data of the vehicle-mounted monocular camera is determined according to the wheel speedometer data of the rear wheels of the vehicle and the vehicle-mounted IMU data; according to the vehicle-mounted monocular camera The camera pose data of the monocular camera is used to determine the relative camera pose of the target image and the matching image.

本实施例由于后轮与车身之间在水平方向无相对转动,后轮的轮速可以直接表征车身的移速,结合车辆后轮的轮速计数据和车载IMU数据得到相机位姿,提高了相机位姿的可靠性,进而提高可行驶区域检测的准确性和可靠性。In this embodiment, since there is no relative rotation between the rear wheel and the body in the horizontal direction, the wheel speed of the rear wheel can directly represent the moving speed of the body, and the camera pose is obtained by combining the wheel speedometer data of the rear wheel of the vehicle and the data of the vehicle IMU, which improves the performance of the camera. The reliability of the camera pose, thereby improving the accuracy and reliability of the drivable area detection.

在一些实施例中,第二处理模块33,用于根据所述目标三维点云数据和对所述目标图像的拍摄区域水平切分的极坐标栅格网络,确定所述极坐标栅格网络中各栅格中包含障碍物点的数量,其中,所述障碍物点为对地高度大于预设障碍物高度阈值的目标三维点;根据所述极坐标栅格网络的各扇形分区中的最近障碍物栅格,确定所述目标图像的拍摄区域中各方向的最近障碍物点,其中,所述最近障碍物栅格是所述扇形分区中与所述相机原点径向距离最近、且所包含障碍物点的数量大于预设数量阈值的栅格。具体地,例如是根据所述目标三维点云数据和所述目标图像对应的相机原点位置,确定预设的极坐标栅格网络的各栅格中的目标三维点,其中,所述极坐标栅格网络是对目标图像的拍摄区域,以扇形排布的第一类切割面和正对相机且平行排布的第二类切割面叠加分割形成的分割网络,所述第一类切割面的交线与相机原点的对地垂线共线;确定各所述栅格中包含障碍物点的数量,其中,所述障碍物点为对地高度大于预设障碍物高度阈值的目标三维点;在所述第一类切割面分割出的扇形分区中,确定最近障碍物栅格,其中,所述最近障碍物栅格是所述扇形分区中与所述相机原点径向距离最近、且所包含障碍物点的数量大于预设数量阈值的栅格;根据所述最近障碍物栅格,确定所述目标图像的拍摄区域中最近障碍物点。In some embodiments, the second processing module 33 is configured to, according to the target three-dimensional point cloud data and the polar coordinate grid network horizontally segmenting the shooting area of the target image, determine the The number of obstacle points contained in each grid, wherein the obstacle point is the target three-dimensional point whose height above the ground is greater than the preset obstacle height threshold; according to the nearest obstacle in each sector partition of the polar coordinate grid network The object grid is used to determine the nearest obstacle points in each direction in the shooting area of the target image, wherein the nearest obstacle grid is the radial distance closest to the camera origin in the fan-shaped partition and contains obstacles A raster whose number of object points is greater than the preset number threshold. Specifically, for example, according to the target 3D point cloud data and the camera origin position corresponding to the target image, the target 3D point in each grid of the preset polar coordinate grid network is determined, wherein the polar coordinate grid The grid network is a segmentation network formed by superimposing and segmenting the first type of cutting planes arranged in a fan shape and the second type of cutting planes facing the camera and arranged in parallel for the shooting area of the target image. The intersection of the first type of cutting planes Collinear with the vertical line to the ground of the camera origin; determine the number of obstacle points contained in each of the grids, wherein the obstacle points are the target three-dimensional points whose height to the ground is greater than the preset obstacle height threshold; In the sector partition divided by the first type of cutting plane, determine the nearest obstacle grid, wherein the nearest obstacle grid is the radial distance closest to the camera origin in the sector partition, and the obstacles included A grid with a number of points greater than a preset number threshold; and according to the nearest obstacle grid, determine the nearest obstacle point in the shooting area of the target image.

本实施例利用对相机原点建立的极坐标栅格网络对目标图像的拍摄区域进行空间划分,将各方向扇形分区中可能是障碍物且对相机原点径向距离最小的栅格提取出来,用于确定相机原点各方向的最近障碍物点,提高了最近障碍物点的准确性,进而提高可行驶区域检测的准确性和可靠性。In this embodiment, the polar coordinate grid network established for the camera origin is used to spatially divide the shooting area of the target image, and the grids that may be obstacles and the smallest radial distance from the camera origin in the fan-shaped partitions in each direction are extracted for use in The nearest obstacle points in all directions of the origin of the camera are determined, which improves the accuracy of the nearest obstacle points, thereby improving the accuracy and reliability of the drivable area detection.

在一些实施例中,第二处理模块33,用于获取所述最近障碍物栅格中所包含障碍物点的平均位置点;将各所述最近障碍物栅格对应的平均位置点,作为所述目标图像的拍摄区域中最近障碍物点。In some embodiments, the second processing module 33 is configured to obtain the average position point of the obstacle points included in the nearest obstacle grid; the average position point corresponding to each of the nearest obstacle grids is used as the The nearest obstacle point in the shooting area of the target image.

本实施例进一步优化最近障碍物点的位置,提高障碍物点的准确性。This embodiment further optimizes the position of the nearest obstacle point and improves the accuracy of the obstacle point.

在一些实施例中,第三处理模块34,用于在对所述目标图像的拍摄区域水平切分的均匀分割网络中,根据所述均匀分割网络中各网格单元相对于所述最近障碍物点的位置,确定各网格单元的加权值;根据各所述网格单元的初始权值以及所述加权值,确定各所述网格单元的新权值,其中,所述新权值大于或等于最小权阈值、且小于或等于最大权阈值,所述网格单元的初始权值是0或者是在对前一帧图像确定可行驶区域时确定的新权值;根据所述新权值指示的第一类网格单元或第二类网格单元,确定所述目标图像的拍摄区域中的可行驶区域,其中,所述新权值指示的第一类网格单元与所述相机原点之间有所述最近障碍物点,所述新权值指示的第二类网格单元与所述相机原点之间没有所述最近障碍物点。例如,具体可以是根据所述最近障碍物点和所述目标图像对应的相机原点位置,获取包含所述最近障碍物点的均匀分割网络,其中,所述均匀分割网络是对目标图像的拍摄区域在水平方向以均匀方形网格分割形成的分割网络;根据所述均匀分割网络中各网格单元对相机原点的径向距离,以及所述网格单元方向上的最近障碍物点对相机原点的径向距离,确定各所述网格单元的加权值;根据各所述网格单元的初始权值以及所述加权值,确定各所述网格单元的新权值,其中,所述新权值大于或等于最小权阈值、且小于或等于最大权阈值,所述网格单元的初始权值是0或者是在对前一帧图像确定可行驶区域时确定的新权值;根据所述网格单元的新权值,确定所述网格单元是第一类网格单元或者第二类网格单元,其中,所述第一类网格单元与所述相机原点之间有所述最近障碍物点,所述第二类网格单元与所述相机原点之间没有所述最近障碍物点;根据所述第二类网格单元,确定所述目标图像的拍摄区域中的可行驶区域,其中,与所述第一类网格单元相邻的所述第二类网格单元是所述可行驶区域的边界。In some embodiments, the third processing module 34 is configured to, in the uniform segmentation network for horizontally segmenting the shooting area of the target image, according to the relative position of each grid unit in the uniform segmentation network to the nearest obstacle The weighted value of each grid unit is determined according to the position of the point, and the new weight of each grid unit is determined according to the initial weight of each grid unit and the weighted value, wherein the new weight is greater than or equal to the minimum weight threshold, and less than or equal to the maximum weight threshold, the initial weight of the grid unit is 0 or a new weight determined when the drivable area is determined for the previous frame image; according to the new weight The indicated first-type grid unit or the second-type grid unit determines a drivable area in the shooting area of the target image, wherein the first-type grid unit indicated by the new weight is related to the camera origin There is the closest obstacle point in between, and there is no closest obstacle point between the second type of grid unit indicated by the new weight and the camera origin. For example, specifically, according to the closest obstacle point and the camera origin position corresponding to the target image, obtain a uniform segmentation network including the closest obstacle point, where the uniform segmentation network is the shooting area of the target image A segmentation network formed by dividing a uniform square grid in the horizontal direction; according to the radial distance of each grid unit in the uniform segmentation network to the camera origin, and the distance between the nearest obstacle point in the grid unit direction to the camera origin The radial distance is used to determine the weighted value of each of the grid units; according to the initial weight of each of the grid units and the weighted value, the new weight of each of the grid units is determined, wherein the new weight is The value is greater than or equal to the minimum weight threshold and less than or equal to the maximum weight threshold, the initial weight of the grid unit is 0 or a new weight determined when the drivable area is determined for the previous frame image; according to the grid The new weight of the grid unit, which determines whether the grid unit is a first-type grid unit or a second-type grid unit, wherein there is the closest obstacle between the first-type grid unit and the camera origin object point, there is no nearest obstacle point between the second type grid unit and the camera origin; according to the second type grid unit, determine the drivable area in the shooting area of the target image, Wherein, the grid unit of the second type adjacent to the grid unit of the first type is the boundary of the drivable area.

本实施例通过计算目标图像拍摄区域在均匀分割网络中各网格单元的加权值,对网格单元更新权值,随着连续图像帧的不断更新,不断加权更新网格单元的新权值,不仅能够平滑噪声还能够起到弥补单帧最近障碍物点漏检的问题,提高了对可行驶区域检测的可靠性。In this embodiment, by calculating the weighted value of each grid unit in the uniform segmentation network of the target image shooting area, the weight value of the grid unit is updated. It can not only smooth the noise, but also make up for the problem of missed detection of the nearest obstacle point in a single frame, and improve the reliability of the detection of the drivable area.

在一些实施例中,第三处理模块34,具体用于若所述均匀分割网络中网格单元对相机原点的径向距离,减去所述网格单元方向上的最近障碍物点对相机原点的径向距离的差值,大于或等于距离上限阈值,则所述网格单元的加权值为第一值;若所述最近障碍物点对相机原点的径向距离,减去所述最近障碍物点方向上网格单元对相机原点的径向距离的差值,小于或等于距离下限阈值,则所述网格单元的加权值为第二值;若所述最近障碍物点对相机原点的径向距离,减去所述最近障碍物点方向上网格单元对相机原点的径向距离的差值,小于所述距离上限阈值且大于所述距离下限阈值,则所述网格单元的加权值为第三值,其中,所述第三值是根据所述差值和预设的平滑连续函数确定的值;其中,所述距离上限阈值是所述距离下限阈值的相反数,所述第一值是所述第二值的相反数,所述第三值的绝对值小于所述距离上限阈值的绝对值或所述距离下限阈值的绝对值。In some embodiments, the third processing module 34 is specifically configured to subtract the closest obstacle point in the direction of the grid unit to the camera origin if the radial distance between the grid unit in the uniform segmentation network and the camera origin is If the difference between the radial distances is greater than or equal to the distance upper limit threshold, the weighted value of the grid unit is the first value; if the radial distance between the nearest obstacle point and the camera origin is subtracted from the nearest obstacle The difference between the radial distance between the grid unit and the camera origin in the direction of the object point is less than or equal to the lower threshold of the distance, then the weight of the grid unit is the second value; if the distance between the closest obstacle point and the camera origin is To the distance, minus the difference between the radial distance between the grid unit and the camera origin in the direction of the nearest obstacle point, if it is less than the upper threshold of the distance and greater than the lower threshold of the distance, then the weighted value of the grid unit is a third value, wherein the third value is a value determined according to the difference value and a preset smooth continuous function; wherein the distance upper threshold is the inverse of the distance lower threshold, and the first value is the inverse of the second value, and the absolute value of the third value is smaller than the absolute value of the distance upper threshold value or the absolute value of the distance lower threshold value.

本实施例实现了对各网格单元加权值的确定,并通过平滑连续函数对差值在距离上限阈值和距离下限阈值之间的网格单元平滑过渡,进一步降低多帧融合加权累积过程中的噪声,提高了对网格单元新权值确定的可靠性,进而提高了对可行驶区域检测的可靠性。This embodiment realizes the determination of the weighted value of each grid unit, and uses a smooth continuous function to smoothly transition the grid units whose difference is between the upper and lower distance thresholds, thereby further reducing the multi-frame fusion weighted accumulation process. noise, which improves the reliability of the determination of the new weights of the grid cells, and further improves the reliability of the detection of the drivable area.

在一些实施例中,所述匹配图像和目标图像都是鱼眼图像。本申请实施例通过采用鱼眼图像,实现了增大水平视角,扩大图像拍摄区域视野范围,提高可行驶区域检测的可靠性。In some embodiments, the matching image and the target image are both fisheye images. By adopting the fisheye image, the embodiment of the present application realizes the increase of the horizontal viewing angle, the expansion of the field of view of the image shooting area, and the improvement of the reliability of the detection of the drivable area.

根据本申请的实施例,本申请还提供了一种电子设备和一种可读存储介质。According to the embodiments of the present application, the present application further provides an electronic device and a readable storage medium.

参见图12,是根据本申请实施例的可行驶区域检测方法的电子设备的框图。电子设备旨在表示各种形式的数字计算机,诸如,膝上型计算机、台式计算机、工作台、个人数字助理、服务器、刀片式服务器、大型计算机、和其它适合的计算机。电子设备还可以表示各种形式的移动装置,诸如,个人数字处理、蜂窝电话、智能电话、可穿戴设备和其它类似的计算装置。本文所示的部件、它们的连接和关系、以及它们的功能仅仅作为示例,并且不意在限制本文中描述的和/或者要求的本申请的实现。Referring to FIG. 12 , it is a block diagram of an electronic device of a drivable area detection method according to an embodiment of the present application. Electronic devices are intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframe computers, and other suitable computers. Electronic devices may also represent various forms of mobile devices, such as personal digital processors, cellular phones, smart phones, wearable devices, and other similar computing devices. The components shown herein, their connections and relationships, and their functions are by way of example only, and are not intended to limit implementations of the application described and/or claimed herein.

如图12所示,该电子设备包括:一个或多个处理器1201、存储器1202,以及用于连接各部件的接口,包括高速接口和低速接口。各个部件利用不同的总线互相连接,并且可以被安装在公共主板上或者根据需要以其它方式安装。处理器可以对在电子设备内执行的指令进行处理,包括存储在存储器中或者存储器上以在外部输入/输出装置(诸如,耦合至接口的显示设备)上显示GUI的图形信息的指令。在其它实施方式中,若需要,可以将多个处理器和/或多条总线与多个存储器和多个存储器一起使用。同样,可以连接多个电子设备,各个设备提供部分必要的操作(例如,作为服务器阵列、一组刀片式服务器、或者多处理器系统)。图12中以一个处理器1201为例。As shown in FIG. 12, the electronic device includes: one or more processors 1201, a memory 1202, and interfaces for connecting various components, including a high-speed interface and a low-speed interface. The various components are interconnected using different buses and may be mounted on a common motherboard or otherwise as desired. The processor may process instructions executed within the electronic device, including instructions stored in or on memory to display graphical information of the GUI on an external input/output device, such as a display device coupled to the interface. In other embodiments, multiple processors and/or multiple buses may be used with multiple memories and multiple memories, if desired. Likewise, multiple electronic devices may be connected, each providing some of the necessary operations (eg, as a server array, a group of blade servers, or a multiprocessor system). In FIG. 12, a processor 1201 is used as an example.

存储器1202即为本申请所提供的非瞬时计算机可读存储介质。其中,所述存储器存储有可由至少一个处理器执行的指令,以使所述至少一个处理器执行本申请所提供的可行驶区域检测方法。本申请的非瞬时计算机可读存储介质存储计算机指令,该计算机指令用于使计算机执行本申请所提供的可行驶区域检测方法。The memory 1202 is the non-transitory computer-readable storage medium provided by the present application. Wherein, the memory stores instructions executable by at least one processor, so that the at least one processor executes the drivable area detection method provided by the present application. The non-transitory computer-readable storage medium of the present application stores computer instructions for causing the computer to execute the drivable area detection method provided by the present application.

存储器1202作为一种非瞬时计算机可读存储介质,可用于存储非瞬时软件程序、非瞬时计算机可执行程序以及模块,如本申请实施例中的可行驶区域检测方法对应的程序指令/模块(例如,附图11所示的图像采集模块31、第一处理模块32、第二处理模块33和第三处理模块34)。处理器1201通过运行存储在存储器1202中的非瞬时软件程序、指令以及模块,从而执行服务器的各种功能应用以及数据处理,即实现上述方法实施例中的可行驶区域检测的方法。As a non-transitory computer-readable storage medium, the memory 1202 can be used to store non-transitory software programs, non-transitory computer-executable programs and modules, such as program instructions/modules corresponding to the drivable area detection method in the embodiments of the present application (for example, , the image acquisition module 31, the first processing module 32, the second processing module 33 and the third processing module 34 shown in FIG. 11). The processor 1201 executes various functional applications and data processing of the server by running the non-transitory software programs, instructions and modules stored in the memory 1202, that is, implementing the method for detecting a drivable area in the above method embodiments.

存储器1202可以包括存储程序区和存储数据区,其中,存储程序区可存储操作系统、至少一个功能所需要的应用程序;存储数据区可存储根据可行驶区域检测的电子设备的使用所创建的数据等。此外,存储器1202可以包括高速随机存取存储器,还可以包括非瞬时存储器,例如至少一个磁盘存储器件、闪存器件、或其他非瞬时固态存储器件。在一些实施例中,存储器1202可选包括相对于处理器1201远程设置的存储器,这些远程存储器可以通过网络连接至可行驶区域检测的电子设备。上述网络的实例包括但不限于互联网、企业内部网、局域网、移动通信网及其组合。The memory 1202 may include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application program required for at least one function; the storage data area may store data created according to the use of the electronic device detected in the drivable area Wait. Additionally, memory 1202 may include high-speed random access memory, and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid-state storage device. In some embodiments, the memory 1202 may optionally include memory located remotely relative to the processor 1201, and these remote memories may be connected to the electronic device for drivable area detection via a network. Examples of such networks include, but are not limited to, the Internet, an intranet, a local area network, a mobile communication network, and combinations thereof.

可行驶区域检测的方法的电子设备还可以包括:输入装置1203和输出装置1204。处理器1201、存储器1202、输入装置1203和输出装置1204可以通过总线或者其他方式连接,图12中以通过总线连接为例。The electronic device of the method for detecting a drivable area may further include: an input device 1203 and an output device 1204 . The processor 1201 , the memory 1202 , the input device 1203 and the output device 1204 may be connected by a bus or in other ways, and the connection by a bus is taken as an example in FIG. 12 .

输入装置1203可接收输入的数字或字符信息,以及产生与可行驶区域检测的电子设备的用户设置以及功能控制有关的键信号输入,例如触摸屏、小键盘、鼠标、轨迹板、触摸板、指示杆、一个或者多个鼠标按钮、轨迹球、操纵杆等输入装置。输出装置1204可以包括显示设备、辅助照明装置(例如,LED)和触觉反馈装置(例如,振动电机)等。该显示设备可以包括但不限于,液晶显示器(LCD)、发光二极管(LED)显示器和等离子体显示器。在一些实施方式中,显示设备可以是触摸屏。The input device 1203 can receive input numerical or character information, and generate key signal input related to user settings and function control of the electronic equipment for driving area detection, such as touch screen, keypad, mouse, trackpad, touchpad, pointing stick , one or more mouse buttons, trackballs, joysticks and other input devices. Output devices 1204 may include display devices, auxiliary lighting devices (eg, LEDs), haptic feedback devices (eg, vibration motors), and the like. The display device may include, but is not limited to, a liquid crystal display (LCD), a light emitting diode (LED) display, and a plasma display. In some implementations, the display device may be a touch screen.

此处描述的系统和技术的各种实施方式可以在数字电子电路系统、集成电路系统、专用ASIC(专用集成电路)、计算机硬件、固件、软件、和/或它们的组合中实现。这些各种实施方式可以包括:实施在一个或者多个计算机程序中,该一个或者多个计算机程序可在包括至少一个可编程处理器的可编程系统上执行和/或解释,该可编程处理器可以是专用或者通用可编程处理器,可以从存储系统、至少一个输入装置、和至少一个输出装置接收数据和指令,并且将数据和指令传输至该存储系统、该至少一个输入装置、和该至少一个输出装置。Various implementations of the systems and techniques described herein can be implemented in digital electronic circuitry, integrated circuit systems, application specific ASICs (application specific integrated circuits), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include being implemented in one or more computer programs executable and/or interpretable on a programmable system including at least one programmable processor that The processor, which may be a special purpose or general-purpose programmable processor, may receive data and instructions from a storage system, at least one input device, and at least one output device, and transmit data and instructions to the storage system, the at least one input device, and the at least one output device an output device.

这些计算程序(也称作程序、软件、软件应用、或者代码)包括可编程处理器的机器指令,并且可以利用高级过程和/或面向对象的编程语言、和/或汇编/机器语言来实施这些计算程序。如本文使用的,术语“机器可读介质”和“计算机可读介质”指的是用于将机器指令和/或数据提供给可编程处理器的任何计算机程序产品、设备、和/或装置(例如,磁盘、光盘、存储器、可编程逻辑装置(PLD)),包括,接收作为机器可读信号的机器指令的机器可读介质。术语“机器可读信号”指的是用于将机器指令和/或数据提供给可编程处理器的任何信号。These computational programs (also referred to as programs, software, software applications, or codes) include machine instructions for programmable processors, and may be implemented using high-level procedural and/or object-oriented programming languages, and/or assembly/machine languages calculation program. As used herein, the terms "machine-readable medium" and "computer-readable medium" refer to any computer program product, apparatus, and/or apparatus for providing machine instructions and/or data to a programmable processor ( For example, magnetic disks, optical disks, memories, programmable logic devices (PLDs), including machine-readable media that receive machine instructions as machine-readable signals. The term "machine-readable signal" refers to any signal used to provide machine instructions and/or data to a programmable processor.

为了提供与用户的交互,可以在计算机上实施此处描述的系统和技术,该计算机具有:用于向用户显示信息的显示装置(例如,CRT(阴极射线管)或者LCD(液晶显示器)监视器);以及键盘和指向装置(例如,鼠标或者轨迹球),用户可以通过该键盘和该指向装置来将输入提供给计算机。其它种类的装置还可以用于提供与用户的交互;例如,提供给用户的反馈可以是任何形式的传感反馈(例如,视觉反馈、听觉反馈、或者触觉反馈);并且可以用任何形式(包括声输入、语音输入或者、触觉输入)来接收来自用户的输入。To provide interaction with a user, the systems and techniques described herein may be implemented on a computer having a display device (eg, a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to the user ); and a keyboard and pointing device (eg, a mouse or trackball) through which a user can provide input to the computer. Other kinds of devices can also be used to provide interaction with the user; for example, the feedback provided to the user can be any form of sensory feedback (eg, visual feedback, auditory feedback, or tactile feedback); and can be in any form (including acoustic input, voice input, or tactile input) to receive input from the user.

可以将此处描述的系统和技术实施在包括后台部件的计算系统(例如,作为数据服务器)、或者包括中间件部件的计算系统(例如,应用服务器)、或者包括前端部件的计算系统(例如,具有图形用户界面或者网络浏览器的用户计算机,用户可以通过该图形用户界面或者该网络浏览器来与此处描述的系统和技术的实施方式交互)、或者包括这种后台部件、中间件部件、或者前端部件的任何组合的计算系统中。可以通过任何形式或者介质的数字数据通信(例如,通信网络)来将系统的部件相互连接。通信网络的示例包括:局域网(LAN)、广域网(WAN)和互联网。The systems and techniques described herein may be implemented on a computing system that includes back-end components (eg, as a data server), or a computing system that includes middleware components (eg, an application server), or a computing system that includes front-end components (eg, a user's computer having a graphical user interface or web browser through which a user may interact with implementations of the systems and techniques described herein), or including such backend components, middleware components, Or any combination of front-end components in a computing system. The components of the system may be interconnected by any form or medium of digital data communication (eg, a communication network). Examples of communication networks include: Local Area Networks (LANs), Wide Area Networks (WANs), and the Internet.

计算机系统可以包括客户端和服务器。客户端和服务器一般远离彼此并且通常通过通信网络进行交互。通过在相应的计算机上运行并且彼此具有客户端-服务器关系的计算机程序来产生客户端和服务器的关系。A computer system can include clients and servers. Clients and servers are generally remote from each other and usually interact through a communication network. The relationship of client and server arises by computer programs running on the respective computers and having a client-server relationship to each other.

应该理解,可以使用上面所示的各种形式的流程,重新排序、增加或删除步骤。例如,本发申请中记载的各步骤可以并行地执行也可以顺序地执行也可以不同的次序执行,只要能够实现本申请公开的技术方案所期望的结果,本文在此不进行限制。It should be understood that steps may be reordered, added or deleted using the various forms of flow shown above. For example, the steps described in the present application can be performed in parallel, sequentially or in different orders, and as long as the desired results of the technical solutions disclosed in the present application can be achieved, no limitation is imposed herein.

上述具体实施方式,并不构成对本申请保护范围的限制。本领域技术人员应该明白的是,根据设计要求和其他因素,可以进行各种修改、组合、子组合和替代。任何在本申请的精神和原则之内所作的修改、等同替换和改进等,均应包含在本申请保护范围之内。The above-mentioned specific embodiments do not constitute a limitation on the protection scope of the present application. It should be understood by those skilled in the art that various modifications, combinations, sub-combinations and substitutions may occur depending on design requirements and other factors. Any modifications, equivalent replacements and improvements made within the spirit and principles of this application shall be included within the protection scope of this application.

Claims (18)

1. A travelable region detection method, comprising:
acquiring a matching image and a target image acquired by shooting a surrounding area of a vehicle by using a vehicle-mounted monocular camera, wherein the matching image is a previous frame image continuous with the target image;
performing three-dimensional reconstruction on the target image according to the matching image to obtain target three-dimensional point cloud data corresponding to the target image;
determining the nearest barrier point in the shooting area of the target image according to the target three-dimensional point cloud data;
in a uniform segmentation network for horizontally segmenting the shooting area of the target image, determining the weighted value of each grid unit according to the radial distance of each grid unit in the uniform segmentation network to the origin of the camera and the radial distance of the nearest barrier point in the direction of the grid unit to the origin of the camera;
determining a new weight of each grid cell according to the initial weight of each grid cell and the weighted value, wherein the new weight is greater than or equal to a minimum weight threshold and less than or equal to a maximum weight threshold, and the initial weight of each grid cell is 0 or is a new weight determined when a travelable area is determined for a previous frame of image;
and determining a travelable area in the shooting area of the target image according to the first type of grid cells or the second type of grid cells indicated by the new weight, wherein the nearest barrier point is arranged between the first type of grid cells indicated by the new weight and the camera origin, and the nearest barrier point is not arranged between the second type of grid cells indicated by the new weight and the camera origin.
2. The method of claim 1, wherein the three-dimensional reconstruction of the target image according to the matching image to obtain target three-dimensional point cloud data corresponding to the target image comprises:
projecting the matched image onto a plurality of preset projection surfaces according to the visual angle of the target image to obtain a plurality of projection images, wherein each projection surface corresponds to a depth relative to the origin of the camera;
determining the estimated depth of the pixels in the target image according to the matching cost of the pixels in the target image and the corresponding pixels in the plurality of projection images;
and acquiring target three-dimensional point cloud data corresponding to the target image according to the estimated depth of the pixels in the target image.
3. The method of claim 2, wherein the projection surface comprises: n1 vertical projection planes;
the N1 vertical projection planes are parallel to the right opposite of the camera, and the distances from the camera origin to the N1 vertical projection planes are in an inversely proportional equal difference distribution, wherein N1 is an integer greater than 1.
4. The method of claim 3, wherein the projection surface further comprises: n2 horizontal projection planes and/or N3 projection spheres;
the N2 horizontal projection planes are parallel to the ground right below the camera, and the N2 horizontal projection planes are uniformly arranged in the ground distribution range taking the ground as a symmetry center, wherein N2 is an integer larger than 1;
the N3 projection spherical surfaces are concentric spherical surfaces taking the camera origin as the sphere center, and the radiuses of the N3 projection spherical surfaces are in inverse proportion equal difference distribution, wherein N3 is an integer larger than 1.
5. The method of any of claims 2 to 4, wherein determining the estimated depth of the pixel in the target image based on the cost of matching the pixel in the target image to the corresponding pixel in the plurality of projection images comprises:
acquiring target pixel window characteristics of pixels in the target image;
acquiring projection pixel window characteristics of corresponding pixels in the plurality of projection images;
according to the target pixel window characteristic and the projection pixel window characteristic, obtaining the matching cost of the pixel in the target image and the corresponding pixel in each projection image;
and taking the depth corresponding to the corresponding pixel with the minimum matching cost as the estimated depth of the pixel in the target image, wherein the depth corresponding to the corresponding pixel is the depth corresponding to the projection plane where the corresponding pixel is located.
6. The method according to claim 1, wherein the determining a nearest obstacle point in a shooting area of the target image according to the target three-dimensional point cloud data comprises:
determining the number of barrier points contained in each grid in the polar coordinate grid network according to the target three-dimensional point cloud data and the polar coordinate grid network which horizontally divides the shooting area of the target image, wherein the barrier points are target three-dimensional points with the ground height larger than a preset barrier height threshold;
and determining the nearest barrier points in each direction in the shooting area of the target image according to the nearest barrier grid in each fan-shaped partition of the polar coordinate grid network, wherein the nearest barrier grid is the grid which has the closest radial distance with the origin of the camera in the fan-shaped partition and contains barrier points with the number larger than a preset number threshold.
7. The method of claim 6, wherein determining nearest obstacle points in each direction in the capture area of the target image from the nearest obstacle grids in each sector of the polar grid mesh network comprises:
obtaining the average position point of the obstacle points contained in the nearest obstacle grid;
and taking the average position point corresponding to each nearest obstacle grid as the nearest obstacle point in the shooting area of the target image.
8. The method of claim 1, 2, 3, 4, 6 or 7, wherein the matching image and the target image are both fisheye images.
9. A travelable region detection apparatus, characterized by comprising:
the system comprises an image acquisition module, a matching image acquisition module and a target image acquisition module, wherein the image acquisition module is used for acquiring the matching image and the target image acquired by shooting the surrounding area of a vehicle by using a vehicle-mounted monocular camera, and the matching image is the previous frame image of the target image;
the first processing module is used for carrying out three-dimensional reconstruction on the target image according to the matching image to obtain target three-dimensional point cloud data corresponding to the target image;
the second processing module is used for determining a nearest barrier point in a shooting area of the target image according to the target three-dimensional point cloud data;
the third processing module is used for determining the weighted value of each grid unit in an evenly-divided network for horizontally dividing the shooting area of the target image according to the radial distance of each grid unit in the evenly-divided network to the camera origin and the radial distance of the nearest barrier point in the direction of the grid unit to the camera origin; determining a new weight of each grid cell according to the initial weight of each grid cell and the weighted value, wherein the new weight is greater than or equal to a minimum weight threshold and less than or equal to a maximum weight threshold, and the initial weight of each grid cell is 0 or is a new weight determined when a travelable area is determined for a previous frame of image; and determining a travelable area in the shooting area of the target image according to the first type of grid cells or the second type of grid cells indicated by the new weight, wherein the nearest barrier point is arranged between the first type of grid cells indicated by the new weight and the origin of the camera, and the nearest barrier point is not arranged between the second type of grid cells indicated by the new weight and the origin of the camera.
10. The apparatus according to claim 9, wherein the first processing module is specifically configured to project the matching image onto a plurality of preset projection surfaces according to the view angle of the target image, so as to obtain a plurality of projection images, where each projection surface corresponds to a depth relative to an origin of a camera; determining the estimated depth of the pixels in the target image according to the matching cost of the pixels in the target image and the corresponding pixels in the plurality of projection images; and acquiring target three-dimensional point cloud data corresponding to the target image according to the estimated depth of the pixels in the target image.
11. The apparatus of claim 10, wherein the projection surface comprises: n1 vertical projection planes; the N1 vertical projection planes are parallel to the right opposite of the camera, and the distances from the camera origin to the N1 vertical projection planes are in an inversely proportional equal difference distribution, wherein N1 is an integer greater than 1.
12. The apparatus of claim 11, wherein the plane of projection further comprises: n2 horizontal projection planes and/or N3 projection spheres; the N2 horizontal projection planes are parallel to the ground right below the camera, and the N2 horizontal projection planes are uniformly arranged in the ground distribution range taking the ground as a symmetry center, wherein N2 is an integer larger than 1; the N3 projection spherical surfaces are concentric spherical surfaces taking the camera origin as the sphere center, and the radiuses of the N3 projection spherical surfaces are in inverse proportion equal difference distribution, wherein N3 is an integer larger than 1.
13. The apparatus according to any one of claims 10 to 12, wherein the first processing module is specifically configured to obtain a target pixel window characteristic of a pixel in the target image; acquiring projection pixel window characteristics of corresponding pixels in the plurality of projection images; according to the target pixel window characteristic and the projection pixel window characteristic, obtaining the matching cost of the pixel in the target image and the corresponding pixel in each projection image; and taking the depth corresponding to the corresponding pixel with the minimum matching cost as the estimated depth of the pixel in the target image, wherein the depth corresponding to the corresponding pixel is the depth corresponding to the projection plane where the corresponding pixel is located.
14. The apparatus according to claim 9, wherein the second processing module is configured to determine, according to the target three-dimensional point cloud data and a polar grid network horizontally dividing a shooting area of the target image, a number of obstacle points included in each grid in the polar grid network, where the obstacle points are target three-dimensional points whose height to ground is greater than a preset obstacle height threshold; and determining the nearest barrier points in each direction in the shooting area of the target image according to the nearest barrier grid in each fan-shaped partition of the polar coordinate grid network, wherein the nearest barrier grid is the grid which has the closest radial distance with the origin of the camera in the fan-shaped partition and contains barrier points with the number larger than a preset number threshold.
15. The apparatus according to claim 14, wherein the second processing module 33 is configured to obtain an average position point of obstacle points included in the nearest obstacle grid; and taking the average position point corresponding to each nearest obstacle grid as the nearest obstacle point in the shooting area of the target image.
16. The apparatus of claim 9, 10, 11, 12, 14 or 15, wherein the matching image and the target image are both fisheye images.
17. An electronic device, comprising:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the travelable region detection method of any of claims 1-8.
18. A non-transitory computer-readable storage medium storing computer instructions for causing a computer to execute the travelable region detection method of any of claims 1 to 8.
CN202010175382.7A 2020-03-13 2020-03-13 Method, device and equipment for detecting travelable area and computer readable storage medium Active CN111366917B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010175382.7A CN111366917B (en) 2020-03-13 2020-03-13 Method, device and equipment for detecting travelable area and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010175382.7A CN111366917B (en) 2020-03-13 2020-03-13 Method, device and equipment for detecting travelable area and computer readable storage medium

Publications (2)

Publication Number Publication Date
CN111366917A CN111366917A (en) 2020-07-03
CN111366917B true CN111366917B (en) 2022-07-15

Family

ID=71208941

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010175382.7A Active CN111366917B (en) 2020-03-13 2020-03-13 Method, device and equipment for detecting travelable area and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN111366917B (en)

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112289070B (en) * 2020-10-20 2021-12-07 广州小鹏自动驾驶科技有限公司 Parking space detection method and device, vehicle and readable medium
CN113409268B (en) * 2021-06-18 2023-04-18 广东工业大学 Method and device for detecting passable area based on monocular camera and storage medium
CN114049486B (en) * 2021-11-26 2025-03-14 珠海格力智能装备有限公司 3D point cloud target detection method and system based on monocular structured light
CN116740671A (en) * 2022-03-01 2023-09-12 长沙智能驾驶研究院有限公司 Obstacle detection method, device, equipment and computer storage medium
CN114663754B (en) * 2022-03-04 2025-01-10 深圳鹏行智能研究有限公司 Detection method, device, multi-legged robot and storage medium
CN114690779A (en) * 2022-04-12 2022-07-01 武汉湾流科技股份有限公司 Positioning method and device based on robot vision recognition
CN115406457A (en) * 2022-08-30 2022-11-29 重庆长安汽车股份有限公司 A drivable region detection method, system, device and storage medium
CN115661238B (en) * 2022-12-06 2023-03-07 广汽埃安新能源汽车股份有限公司 Method and device for generating travelable region, electronic equipment and computer readable medium
CN116778444B (en) * 2023-05-30 2025-11-25 嘉兴聚速电子技术有限公司 Environmental free space recognition method based on 4D millimeter-wave radar point cloud in autonomous driving
CN117218244B (en) * 2023-11-07 2024-02-13 武汉博润通文化科技股份有限公司 Intelligent 3D animation model generation method based on image recognition

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102903096A (en) * 2012-07-04 2013-01-30 北京航空航天大学 Monocular video based object depth extraction method
CN103955920A (en) * 2014-04-14 2014-07-30 桂林电子科技大学 Binocular vision obstacle detection method based on three-dimensional point cloud segmentation
CN107064955A (en) * 2017-04-19 2017-08-18 北京汽车集团有限公司 barrier clustering method and device
CN108038905A (en) * 2017-12-25 2018-05-15 北京航空航天大学 A kind of Object reconstruction method based on super-pixel
CN109949355A (en) * 2019-03-14 2019-06-28 大连民族大学 A Method of Establishing Half-sector Isometric Line Model in Monocular Vision Pedestrian Distance Estimation
CN110244321A (en) * 2019-04-22 2019-09-17 武汉理工大学 A detection method of road passable area based on 3D lidar
CN112243518A (en) * 2019-08-29 2021-01-19 深圳市大疆创新科技有限公司 Method and device for acquiring depth map and computer storage medium

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4256812B2 (en) * 2004-04-26 2009-04-22 三菱重工業株式会社 Obstacle avoidance method for moving body and moving body
ES2675363T3 (en) * 2014-06-05 2018-07-10 Softbank Robotics Europe Procedure for constructing a probability map between the absence and the presence of obstacles for an autonomous robot
US9933264B2 (en) * 2015-04-06 2018-04-03 Hrl Laboratories, Llc System and method for achieving fast and reliable time-to-contact estimation using vision and range sensor data for autonomous navigation
CN104950883A (en) * 2015-05-14 2015-09-30 西安电子科技大学 Mobile robot route planning method based on distance grid map
US10635913B2 (en) * 2016-10-17 2020-04-28 Mediatek Inc. Path planning method and related navigation device
US10296012B2 (en) * 2016-12-21 2019-05-21 X Development Llc Pre-computation of kinematically feasible roadmaps
KR102275310B1 (en) * 2017-04-20 2021-07-12 현대자동차주식회사 Mtehod of detecting obstacle around vehicle

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102903096A (en) * 2012-07-04 2013-01-30 北京航空航天大学 Monocular video based object depth extraction method
CN103955920A (en) * 2014-04-14 2014-07-30 桂林电子科技大学 Binocular vision obstacle detection method based on three-dimensional point cloud segmentation
CN107064955A (en) * 2017-04-19 2017-08-18 北京汽车集团有限公司 barrier clustering method and device
CN108038905A (en) * 2017-12-25 2018-05-15 北京航空航天大学 A kind of Object reconstruction method based on super-pixel
CN109949355A (en) * 2019-03-14 2019-06-28 大连民族大学 A Method of Establishing Half-sector Isometric Line Model in Monocular Vision Pedestrian Distance Estimation
CN110244321A (en) * 2019-04-22 2019-09-17 武汉理工大学 A detection method of road passable area based on 3D lidar
CN112243518A (en) * 2019-08-29 2021-01-19 深圳市大疆创新科技有限公司 Method and device for acquiring depth map and computer storage medium

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
一种快速多视图立体匹配方法;赵洪田;《现代计算机》;20180131;全文 *
基于视差的单目视觉障碍检测;王辉等;《万方数据》;20050627;正文第237-239页 *

Also Published As

Publication number Publication date
CN111366917A (en) 2020-07-03

Similar Documents

Publication Publication Date Title
CN111366917B (en) Method, device and equipment for detecting travelable area and computer readable storage medium
CN111968229B (en) High-precision map making method and device
CN112241718B (en) Vehicle information detection method, detection model training method and device
US11694445B2 (en) Obstacle three-dimensional position acquisition method and apparatus for roadside computing device
CN106599108B (en) A method for constructing a multi-modal environment map in a three-dimensional environment
US8199977B2 (en) System and method for extraction of features from a 3-D point cloud
CN114359744A (en) Depth estimation method based on fusion of laser radar and event camera
CN111612760A (en) Method and apparatus for detecting obstacles
WO2021218123A1 (en) Method and device for detecting vehicle pose
CN111274343A (en) A vehicle positioning method, device, electronic device and storage medium
CN111721281B (en) Position identification method and device and electronic equipment
CN112184914B (en) Method and device for determining three-dimensional position of target object and road side equipment
CN113359782A (en) Unmanned aerial vehicle autonomous addressing landing method integrating LIDAR point cloud and image data
CN111340922A (en) Method and electronic device for localization and map construction
CN115147809B (en) Obstacle detection method, device, equipment and storage medium
CN111784579B (en) Mapping method and device
CN112509126A (en) Method, device, equipment and storage medium for detecting three-dimensional object
CN112528932A (en) Method and device for optimizing position information, road side equipment and cloud control platform
CN113763468B (en) Positioning method, device, system and storage medium
CN115790621A (en) High-precision map updating method and device and electronic equipment
WO2022246812A1 (en) Positioning method and apparatus, electronic device, and storage medium
CN111783611A (en) Unmanned vehicle positioning method, device, unmanned vehicle and storage medium
Cigla et al. Gaussian mixture models for temporal depth fusion
CN115294234B (en) Image generation method and device, electronic equipment and storage medium
CN115019167B (en) Fusion positioning method, system, equipment and storage medium based on mobile terminal

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant