Lane Detection 1

Lane Detection 1

O2f

  • One to Few

  • 去除手工制作的组件,以实现端到端的检测

    • 训练样本的设计
    • 非最大值抑制(NMS),用于删除重复的预测
  • 端到端:输入是原始数据,输出是最后的结果。

    • 其输入数据是未经任何人为加工的原始样本形式,后续则是堆叠在输入层上的众多操作层。这些操作层整体可以看作一个复杂的函数 Fcnn,最终的损失函数由数据损失(data loss)和模型参数的正则化损失(regularization loss)共同组成,模型深度的训练则是在最终损失驱动下对模型进行参数更新并将误差反向传播至网络各层。模型的训练可以简单抽象为从原始数据向最终目标的直接拟合,而中间的这些部件起到了将原始数据 映射为特征 随后在映射为 样本标记 的作用
  • 原来输入端不是直接的原始数据,而是在原始数据中提取的特征,这一点在图像问题上尤为突出,因为图像像素数太多,数据维度高,会产生维度灾难,所以原来一个思路是手工提取图像的一些关键特征

BEV-LaneDet

  • 首先,引入了 Virutal Camera,统一了安装在不同车辆上的相机的内 / 外参数,以保证相机之间空间关系的一致性。由于统一的视觉空间,它可以有效地促进网络的学习过程
    • 通过 homography 矩阵\(P_{i,j}\)将当前摄像机的图像投射到虚拟像机的视图中
  • 其次,提出了一个简单而有效的三维车道表示法,称为关键点表示法(Key-Points Representation), 这个模块更适合于表示复杂多样的三维车道结构
    • 道路坐标 z=0 平面划分 s1*s2 个网格(0.5m*0.5m)
    • 直接预测具有相同分辨率的四个 head,包括置信度 confidence、用于聚类的 embedding、gird 中心到车道在 y 方向的偏移 offset 以及每个 grid 的平均高度 height
  • 最后,本文提出了一个轻量级和芯片友好的空间转换模块,名为空间转换金字塔(Spatial Transformation Pyramid),用于将多尺度的前视特征转换成 BEV 特征
    • 空间转化模块对视觉 3D 任务尤为关键,基于深度的方法和基于 transformer 的方法在计算量上开销很大,而且在部署到自动驾驶芯片上也不太方便。
    • 即基于 MLP 的视图转化模块(VRM)
      • 低分辨率的特征更适合在 VRM 中进行空间转换。低分辨率的特征包含更多的全局信息,需要较少的映射参数,比较容易学习。
      • 受 FPN 的启发,本文设计了一个基于 VRM 的空间变换金字塔。通过实验比较,本文最终分别用输入图像的 1/64 分辨率特征 S64 和 1/32 分辨率特征 S32 进行变换,然后将变换结果串联起来,一起送往提取 BEV 的特征。

LANE DETECTION BASED ON HOUGH TRANSFORM AND ENDPOINTS CLASSIFICATION

  • based on the Gaussian distribution

  • After Hough transform, a method to classify the endpoints that belong to different lane lines and a way to show the lane lines are provided

Image preprocessing

  • Image enhancement
  • Edge extraction

Lane detection

  • Hough transform
    • For the left lane lines, we set \(\theta\) range from 25° to 75° , to find 6 Hough peaks that at least 0.7 times the maximum Hough value. Then lines which have 7 pixels in length are chosen from the 6 lines and all endpoints of the line are recorded.
    • 分别选取左侧和右侧最大的 6 个峰值,然后选取长度大于 7 的线段,记录为车道线上的点 Endpoints of the lines
  • Endpoints classification
    • As there may be more than two endpoints for one lane line, classification is strongly needed.
    • Lines that have the value of \(\rho\) less than 80 pixels are regarded as the same line.
    • 2 条直线判断距离是否小于 80 像素,小于则认为是同一条直线(双实线等问题)
  • linear fitting

Lane Detection: A Survey with New Results

  • Review recent visual-based lane detection datasets and methods.
  • Introduce a new dataset with more detailed annotations for HD map modeling.

Lane Line Detection Optimization Algorithm based on Improved Hough Transform and R-least Squares with Dual Removal

  • Lane line edge extraction based on sobel operator with large law
  • OTSU 阈值分割算法, 类间方差最大化。
    • 像素灰度 i 概率为 ni/n (n1, n2 ... nL)
    • 分出两个灰度区间,使得分割的最大类间方差最大,灰度差距和最显著 OTSU
      Edge detection results of Sobel operator based on OTSU algorithm
  • In this study, an improved algorithm based on traditional Hough transform and least square method is proposed for lane line detection. In this paper, the road image is preprocessed to obtain the edge points of the lane line, and then two methods of lane line fitting are optimized to improve the accuracy and real-time performance of lane line detection. In the actual detection, a detection algorithm combining improved Hough transform and r-lsdr method is designed, and Kalman filter is used to track the lane line.
  • Hough transformation and improvement
    • Constraints of polar angle size and direction
    • Constraint of polar diameter
    • Use the Soble operator twice to get the edge of single pixel width.
    • Find the two directions with the maximum number of edge points as rough directions of lane lines.
  • Least Square Method and Optimization
    • The sample points with too large errors are eliminated by a certain threshold.
    • Remove the sample points with the largest error, until the proportion of the number of remaining points in the original sample quantity is equal to the threshold value R.