2,675 research outputs found
A New Vehicle Localization Scheme Based on Combined Optical Camera Communication and Photogrammetry
The demand for autonomous vehicles is increasing gradually owing to their
enormous potential benefits. However, several challenges, such as vehicle
localization, are involved in the development of autonomous vehicles. A simple
and secure algorithm for vehicle positioning is proposed herein without
massively modifying the existing transportation infrastructure. For vehicle
localization, vehicles on the road are classified into two categories: host
vehicles (HVs) are the ones used to estimate other vehicles' positions and
forwarding vehicles (FVs) are the ones that move in front of the HVs. The FV
transmits modulated data from the tail (or back) light, and the camera of the
HV receives that signal using optical camera communication (OCC). In addition,
the streetlight (SL) data are considered to ensure the position accuracy of the
HV. Determining the HV position minimizes the relative position variation
between the HV and FV. Using photogrammetry, the distance between FV or SL and
the camera of the HV is calculated by measuring the occupied image area on the
image sensor. Comparing the change in distance between HV and SLs with the
change in distance between HV and FV, the positions of FVs are determined. The
performance of the proposed technique is analyzed, and the results indicate a
significant improvement in performance. The experimental distance measurement
validated the feasibility of the proposed scheme
Deep Learning Framework for Wireless Systems: Applications to Optical Wireless Communications
Optical wireless communication (OWC) is a promising technology for future
wireless communications owing to its potentials for cost-effective network
deployment and high data rate. There are several implementation issues in the
OWC which have not been encountered in radio frequency wireless communications.
First, practical OWC transmitters need an illumination control on color,
intensity, and luminance, etc., which poses complicated modulation design
challenges. Furthermore, signal-dependent properties of optical channels raise
non-trivial challenges both in modulation and demodulation of the optical
signals. To tackle such difficulties, deep learning (DL) technologies can be
applied for optical wireless transceiver design. This article addresses recent
efforts on DL-based OWC system designs. A DL framework for emerging image
sensor communication is proposed and its feasibility is verified by simulation.
Finally, technical challenges and implementation issues for the DL-based
optical wireless technology are discussed.Comment: To appear in IEEE Communications Magazine, Special Issue on
Applications of Artificial Intelligence in Wireless Communication
Computational Imaging and Artificial Intelligence: The Next Revolution of Mobile Vision
Signal capture stands in the forefront to perceive and understand the
environment and thus imaging plays the pivotal role in mobile vision. Recent
explosive progresses in Artificial Intelligence (AI) have shown great potential
to develop advanced mobile platforms with new imaging devices. Traditional
imaging systems based on the "capturing images first and processing afterwards"
mechanism cannot meet this unprecedented demand. Differently, Computational
Imaging (CI) systems are designed to capture high-dimensional data in an
encoded manner to provide more information for mobile vision systems.Thanks to
AI, CI can now be used in real systems by integrating deep learning algorithms
into the mobile vision platform to achieve the closed loop of intelligent
acquisition, processing and decision making, thus leading to the next
revolution of mobile vision.Starting from the history of mobile vision using
digital cameras, this work first introduces the advances of CI in diverse
applications and then conducts a comprehensive review of current research
topics combining CI and AI. Motivated by the fact that most existing studies
only loosely connect CI and AI (usually using AI to improve the performance of
CI and only limited works have deeply connected them), in this work, we propose
a framework to deeply integrate CI and AI by using the example of self-driving
vehicles with high-speed communication, edge computing and traffic planning.
Finally, we outlook the future of CI plus AI by investigating new materials,
brain science and new computing techniques to shed light on new directions of
mobile vision systems
Federated Learning for Connected and Automated Vehicles: A Survey of Existing Approaches and Challenges
Machine learning (ML) is widely used for key tasks in Connected and Automated
Vehicles (CAV), including perception, planning, and control. However, its
reliance on vehicular data for model training presents significant challenges
related to in-vehicle user privacy and communication overhead generated by
massive data volumes. Federated learning (FL) is a decentralized ML approach
that enables multiple vehicles to collaboratively develop models, broadening
learning from various driving environments, enhancing overall performance, and
simultaneously securing local vehicle data privacy and security. This survey
paper presents a review of the advancements made in the application of FL for
CAV (FL4CAV). First, centralized and decentralized frameworks of FL are
analyzed, highlighting their key characteristics and methodologies. Second,
diverse data sources, models, and data security techniques relevant to FL in
CAVs are reviewed, emphasizing their significance in ensuring privacy and
confidentiality. Third, specific and important applications of FL are explored,
providing insight into the base models and datasets employed for each
application. Finally, existing challenges for FL4CAV are listed and potential
directions for future work are discussed to further enhance the effectiveness
and efficiency of FL in the context of CAV
Perception Intelligence Integrated Vehicle-to-Vehicle Optical Camera Communication.
Ubiquitous usage of cameras and LEDs in modern road and aerial vehicles open up endless opportunities for novel applications in intelligent machine navigation, communication, and networking. To this end, in this thesis work, we hypothesize the benefit of dual-mode usage of vehicular built-in cameras through novel machine perception capabilities combined with optical camera communication (OCC). Current key conception of understanding a line-of-sight (LOS) scenery is from the aspect of object, event, and road situation detection. However, the idea of blending the non-line-of-sight (NLOS) information with the LOS information to achieve a see-through vision virtually is new. This improves the assistive driving performance by enabling a machine to see beyond occlusion. Another aspect of OCC in the vehicular setup is to understand the nature of mobility and its impact on the optical communication channel quality. The research questions gathered from both the car-car mobility modelling, and evaluating a working setup of OCC communication channel can also be inherited to aerial vehicular situations like drone-drone OCC. The aim of this thesis is to answer the research questions along these new application domains, particularly, (i) how to enable a virtual see-through perception in the car assisting system that alerts the human driver about the visible and invisible critical driving events to help drive more safely, (ii) how transmitter-receiver cars behaves while in the mobility and the overall channel performance of OCC in motion modality, (iii) how to help rescue lost Unmanned Aerial Vehicles (UAVs) through coordinated localization with fusion of OCC and WiFi, (iv) how to model and simulate an in-field drone swarm operation experience to design and validate UAV coordinated localization for group of positioning distressed drones. In this regard, in this thesis, we present the end-to-end system design, proposed novel algorithms to solve the challenges in applying such a system, and evaluation results through experimentation and/or simulation
- …