Amir Khosravian


Postgraduate Research Student (Ph.D. Student)
Monday to Friday 9:00 - 17:00

About

My research project

My qualifications

2018
BSc
Mechanical Engineering
Qom University of Technology
2020
MSc
Mechanical Engineering - Automotive Engineering
Design of Vehicle Dynamic System
Iran University of Science and Technology

Research

Research interests

Teaching

Publications

Amir Khosravian, Abdollah Amirkhani, Hossein Kashiani, Masoud Masih-Tehrani (2021) Generalizing state-of-the-art object detectors for autonomous vehicles in unseen environments

In scene understanding for autonomous vehicles (AVs), models trained on the available datasets fail to generalize well to the complex, real-world scenarios with higher dynamics. In this work, we attempt to handle the distribution mismatch by employing the generative adversarial network (GAN) and weather modeling to strengthen the intra-domain data. We also alleviate the fragility of our trained models against natural distortions with state-of-the-art augmentation approaches. Finally, we assess our method for cross-domain object detection through CARLA simulation. Our experiments demonstrate that: (1) Augmenting training class with even limited intra-domain data captured from the adverse weather conditions boosts the generalization of the two kinds of object detectors; (2) Exploiting GANs and weather modeling to elaborately simulate the adverse, intra-domain weather conditions manages to surmount the adverse data scarcity issue for intra-domain object detection; (3) A combination of Augmix and style augmentations not only can promote the robustness of our trained models against different natural distortions but also can boost their performance in the cross-domain object detection; (4) Training GANs for unsupervised image-to-image translation by means of the existing, large-scale datasets outside of our training domain is found beneficial to alleviate image-based and instance-based domain shifts.

Abdollah Amirkhani, Amir Khosravian, Masoud Masih-Tehrani, Hossein Kashiani (2021) Robust Semantic Segmentation With Multi-Teacher Knowledge Distillation

Recent studies have recently exploited knowledge distillation (KD) technique to address time-consuming annotation task in semantic segmentation, through which one teacher trained on a single dataset could be leveraged for annotating unlabeled data. However, in this context, knowledge capacity is restricted, and knowledge variety is rare in different conditions, such as cross-model KD, in which the single teacher KD prohibits the student model from distilling information using cross-domain context. In this study, we aim to train a robust, lightweight student under the supervision of several expert teachers, which provide better instructive guidance compared to a single student-teacher learning framework. To be more specific, we first train five distinct convolutional neural networks (CNNs) as teachers for semantic segmentation on several datasets. To this end, several state-of-the-art augmentation transformations have also been utilized in training phase of our teachers. The impacts of such training scenarios are then assessed in terms of student robustness and accuracy. As the main contribution of this paper, our proposed multi-teacher KD paradigm endows the student with the ability to amalgamate and capture a variety of knowledge illustrations from different sources. Results demonstrated that our method outperforms the existing studies on both clean and corrupted data in the semantic segmentation task while benefiting from our proposed score weight system. Experiments validate that our multi-teacher framework results in an improvement of 9% up to 32.18% compared to the single-teacher paradigm. Moreover, it is demonstrated that our paradigm surpasses previous supervised real-time studies in the semantic segmentation challenge.

Amir Khosravian, Abdollah Amirkhani, Masoud Masih-Tehrani (2022) Enhancing the robustness of the convolutional neural networks for traffic sign detection

The detection of traffic signs in clean and noise-free images has been investigated by numerous researchers; however, very few of these works have focused on noisy environments. While in the real world, for different reasons (e.g. the speed and acceleration of a vehicle and the roughness around it), the input images of the convolutional neural networks (CNNs) could be extremely noisy. Contrary to other research works, in this paper, we investigate the robustness of the deep learning models against the synthetically modeled noises in the detection of small objects. To this end, the state-of-the-art architectures of Faster-RCNN Resnet101, R-FCN Resnet101, and Faster-RCNN Inception Resnet V2 are trained by means of the Tsinghua-Tencent 100K database, and the performances of the trained models on noisy data are evaluated. After verifying the robustness of these models, different training scenarios (1 – Modeling various climatic conditions, 2 – Style randomization, and 3 – Augmix augmentation) are used to enhance the model robustness. The findings indicate that these scenarios result in up to 13.09%, 12%, and 13.61% gains in the mentioned three networks by means of the mPC metric. They also result in 11.74%, 8.89%, and 7.27% gains in the rPC metric, demonstrating that improvement in robustness does not lead to performance drop on the clean data.

Amir Khosravian, Abdollah Amirkhani, Masoud Masih-Tehrani, Alireza Yazdanijoo (2023) Multi-domain autonomous driving dataset: Towards enhancing the generalization of the convolutional neural networks in new environments

In this paper, a large-scale dataset called the Iran Autonomous Driving Dataset (IADD) is presented, aiming to improve the generalization capability of the deep networks outside of their training domains. The IADD focuses on 2D object detection and contains more than 97,000 annotated images, covering six common object classes in the field of autonomous vehicles. To improve the generalization of the models, a wide variety of driving conditions and domains, including the city and suburban road settings, adverse weather conditions, and various traffic flows, are presented in the IADD images. The results of exhaustive evaluations conducted on several state-of-the-art convolutional neural networks reveal that not only the trained architectures have performed successfully on test data of the IADD, but also they have upheld high precision in the assessments of generalization capability. In order to challenge the models, broad range of simulations have been performed in the CARLA software environment; which due to the synthetic nature of the simulated images, severe domain shifts have been observed between the CARLA and the IADD. Also, the cross-domain evaluation results have confirmed the efficacy of the IADD in enhancing the generalization ability of the deep learning models. The dataset is available in: https://github.com/ahv1373/IADD.

Amir Khosravian, Masoud Masih-Tehrani, Abdollah Amirkhani, Salman Ebrahimi-Nejad (2024) Robust Autonomous Vehicle Control by Leveraging Multi-Stage MPC and Quantized CNN in HIL Framework

This study presents a novel path planning and path-following system designed explicitly for real-world driving conditions in autonomous vehicles. The pipeline considers uncertainties in the dynamic parameters of the vehicle, which are common in real-world scenarios. To address these uncertainties, a robust multi-stage model predictive control approach is developed to generate optimal control outputs. A customized convolutional neural network, Nano-FastSCNN, quantized to half-precision, is employed for accurately segmenting left and right lane boundaries. To ensure that our visual planner is robust against adverse weather conditions, we have rigorously evaluated its generalization capabilities under challenging weather types. The centerline of these lanes is then selected as the desired path and transformed into the camera reference frame by leveraging the camera projection matrix. The controller utilizes this reference trajectory to guide the vehicle along the desired trajectory. To demonstrate real-time processing capabilities, the entire pipeline, including the vision system module and controller, is implemented in a hardware-in-the-loop setup utilizing three NVIDIA Jetson devices to demonstrate the scalability of our architecture. Comprehensive evaluations demonstrate that the pipeline achieves an inference speed of 15.6 FPS, even on the less powerful NVIDIA Jetson Nano. Moreover, the system successfully alleviates the detrimental effects of uncertainties and reliably follows the desired path.