DefinePK

DefinePK hosts the largest index of Pakistani journals, research articles, news headlines, and videos. It also offers chapter-level book search.

Visual Perception Deep Drive Model for Self-Driving Car


Article Information

Title: Visual Perception Deep Drive Model for Self-Driving Car

Authors: Waleed Razzaq, Usman Arif, Zia Mohi U Din

Journal: Pakistan Journal of Scientific Research (PJOSR)

HEC Recognition History
Category From To
Y 2023-07-01 2024-09-30
Y 1900-01-01 2005-06-30

Publisher: Pakistan Association for the Advancement of Science

Country: Pakistan

Year: 2021

Volume: 1

Issue: 1

Language: English

Keywords: Convolutional Neural Network Simulated EnvironmentRADARLiDAR

Categories

Abstract

Self driving cars are the need of future technology, there are many companies that are trying to perfect this particular project but there are still some deficiencies there. Most of the companies are using Expensive sensors like RADAR and LiDAR to get the idea of environment, which are very hard to use and need a lot of processing power. Our project focuses on using only visual aid to drive a car, particularly following the lane of the road. We trained a model using Convolutional Neural Network (CNN), in a simulated environment and tested the model in the same environment.


Research Objective

To develop a cost-effective visual perception model for self-driving cars that focuses on lane following using only camera input and Convolutional Neural Networks (CNNs).


Methodology

The study proposes a model that utilizes a monocular camera as the sole sensor for environmental perception. A Convolutional Neural Network (CNN), specifically AlexNet, was trained in a simulated environment (Audacity simulator) using 120x120 grayscale images. The model was trained on data collected directly from the simulator. The action space for the car was defined as [left, forward, right]. The trained model was then tested within the same simulated environment.

Methodology Flowchart
                        graph TD;
    A["Collect Data from Simulator"] --> B["Preprocess Images: 120x120 grayscale"];
    B --> C["Train AlexNet CNN Model"];
    C --> D["Model Achieves Target Accuracy?"];
    D -- Yes --> E["Test Model in Simulator"];
    D -- No --> C;
    E --> F["Predict Car Action: Left/Forward/Right"];
    F --> G["Execute Action in Simulator"];
    G --> H["Evaluate Performance"];                    

Discussion

The paper argues that relying solely on visual input, specifically camera data, is a more cost-effective approach compared to using expensive sensors like RADAR and LiDAR for self-driving car perception. The use of a simulated environment for training and testing simplifies the process and allows for efficient data collection and model validation. The AlexNet architecture was chosen for its established performance in image recognition tasks.


Key Findings

The trained CNN model achieved 96.75 percent accuracy with a 0.082 percent loss rate after approximately 13 hours of training. The model was capable of detecting road lanes, identifying the middle line, and predicting the appropriate action (left, forward, or right) for the car to follow the lane.


Conclusion

The presented visual perception deep drive model, trained using AlexNet CNN in a simulated environment, successfully demonstrated the ability to follow road lanes for a self-driving car. This approach offers a viable and cost-effective alternative to sensor-heavy systems.


Fact Check

1. Training Time: The model took approximately 13 hours to train. (Confirmed in "Model Training" section).
2. Accuracy: The model achieved 96.75 percent accuracy. (Confirmed in "Model Training" section).
3. Image Resolution: The input image resolution was reduced to 120x120 pixels. (Confirmed in "Preprocessing" section).


Mind Map

Loading PDF...

Loading Statistics...