Week 5 - Establishing the MobileNet Training Pipeline
October 28, 2025
A goal: line follower replication with Mobilenet
This week marks significant technical advancements in our autonomous driving research project, focusing on the development and refinement of our training pipeline for imitation learning systems.
The primary objective of our current work is to replicate the research conducted by A. Moncalvillo [1], with a strategic modification to implement an end-to-end training approach using MobileNet architecture. While complete implementation remains in progress, we have achieved substantial technical milestones that demonstrate meaningful progress toward this goal.
We have successfully configured and operationalized a Logitech G-driving steering wheel system within the CARLA simulation environment. Initial driving tests have been conducted, though further calibration is required to optimize steering sensitivity parameters. Ensuring driver comfort and natural control response is paramount for collecting high-quality expert demonstration data. The current test bench configuration is illustrated in Figure 1 below.
With this established setup, we have successfully recorded our first expert driving demonstration video, providing the foundational dataset for subsequent training iterations.
Concurrently, we have initiated comprehensive dataset analysis procedures aimed at achieving optimal class balance and representation. During this preliminary phase, we generated initial distribution histograms for the critical control parameters—"throttle," "steer" and "speed"—across a sample dataset of 100 images, as shown in Figure 2.
A significant paradigm shift has been implemented in our frame acquisition methodology for training data generation. We have transitioned from the previously used image.save_to_disk(filename) function, which failed to guarantee consistent temporal intervals between frames, to a more robust approach utilizing CARLA's logging and replay capabilities.
Our new pipeline employs the client.start_recorder(LOG_FILE_NAME) function to capture simulation states during manual driving sessions, followed by asynchronous frame generation using client.replay_file(LOG_FILE_NAME, start_time, duration, 0). This advanced approach ensures precise temporal consistency across extracted frames, significantly improving dataset quality for temporal learning models.
Reference:
[1] A. Moncalvillo González, "Seguimiento de carril por visión y conducción autónoma con Aprendizaje por Imitación (Vision-based Lane Keeping and Autonomous Driving using Imitation Learning)," Master's thesis, Universidad Rey Juan Carlos, Madrid, Spain, 2024.