Jove
Visualize
Contact Us
JoVE
x logofacebook logolinkedin logoyoutube logo
ABOUT JoVE
OverviewLeadershipBlogJoVE Help Center
AUTHORS
Publishing ProcessEditorial BoardScope & PoliciesPeer ReviewFAQSubmit
LIBRARIANS
TestimonialsSubscriptionsAccessResourcesLibrary Advisory BoardFAQ
RESEARCH
JoVE JournalMethods CollectionsJoVE Encyclopedia of ExperimentsArchive
EDUCATION
JoVE CoreJoVE BusinessJoVE Science EducationJoVE Lab ManualFaculty Resource CenterFaculty Site
Terms & Conditions of Use
Privacy Policy
Policies

Related Concept Videos

Association Areas of the Cortex01:21

Association Areas of the Cortex

4.9K
Association areas are regions of the cerebral cortex that do not have a specific sensory or motor function. Instead, they integrate and interpret information from various sources to enable higher cognitive processes such as memory, learning, and decision-making. Some key association areas include the following:
Prefrontal Association Area: This area is located in the frontal lobe and is involved in planning, decision-making, and moderating social behavior. It connects with primary motor areas,...
4.9K
Depth Perception and Spatial Vision01:15

Depth Perception and Spatial Vision

523
Depth perception is the ability to perceive objects three-dimensionally. It relies on two types of cues: binocular and monocular. Binocular cues depend on the combination of images from both eyes and how the eyes work together. Since the eyes are in slightly different positions, each eye captures a slightly different image. This disparity between images, known as binocular disparity, helps the brain interpret depth. When the brain compares these images, it determines the distance to an object.
523
  1. Home
  2. Research Domains
  3. Engineering
  4. Communications Engineering
  5. Data Communications
  6. Bi-att3ddet: Attention-based Bi-directional Fusion For Multi-modal 3d Object Detection

Bi-Att3DDet: Attention-Based Bi-Directional Fusion for Multi-Modal 3D Object Detection

Xu Gao1,2, Yaqian Zhao1,2, Yanan Wang1,2

  • 1School of Computer and Artificial Intelligence, Zhengzhou University, Zhengzhou 450001, China.

Sensors (Basel, Switzerland)
|February 13, 2025

Related Experiment Videos

Author Spotlight: Enhancement of Salient Object Detection for Smart Grid Applications
03:31

Author Spotlight: Enhancement of Salient Object Detection for Smart Grid Applications

Published on: December 15, 2023

455
Author Spotlight: An Efficient and Robust Software for Automated Fusion of Multiple Preclinical Imaging Modalities
07:13

Author Spotlight: An Efficient and Robust Software for Automated Fusion of Multiple Preclinical Imaging Modalities

Published on: October 27, 2023

1.0K
A Methodology for Capturing Joint Visual Attention Using Mobile Eye-Trackers
12:39

A Methodology for Capturing Joint Visual Attention Using Mobile Eye-Trackers

Published on: January 18, 2020

7.6K

View abstract on PubMed

Summary
This summary is machine-generated.

This study introduces Bi-Att3DDet, a novel network for multi-modal 3D object detection in autonomous driving. It enhances feature fusion between LiDAR and RGB data, improving detection accuracy and utilizing complementary sensor information effectively.

Area of Science:

  • Computer Vision
  • Autonomous Driving Systems
  • Sensor Fusion

Background:

  • Multi-modal 3D object detection is crucial for autonomous driving.
  • Existing methods struggle with effective fusion of LiDAR and RGB image features.
  • Inadequate capture of structural information in Region of Interest (RoI) features limits performance.

Purpose of the Study:

  • To propose Bi-Att3DDet, a multi-modal sensor fusion network for improved 3D object detection.
  • To enhance the utilization of complementary information between depth and semantic texture features.
  • To better capture structural information within RoI features for more accurate detections.

Main Methods:

  • Developed a Self-Attentive RoI Feature Extraction module (SARoIFE) using self-attention mechanisms.
Keywords:
3D object detectionattention mechanismautonomous drivingmulti-modal sensor fusion

Related Experiment Videos

Author Spotlight: Enhancement of Salient Object Detection for Smart Grid Applications
03:31

Author Spotlight: Enhancement of Salient Object Detection for Smart Grid Applications

Published on: December 15, 2023

455
Author Spotlight: An Efficient and Robust Software for Automated Fusion of Multiple Preclinical Imaging Modalities
07:13

Author Spotlight: An Efficient and Robust Software for Automated Fusion of Multiple Preclinical Imaging Modalities

Published on: October 27, 2023

1.0K
A Methodology for Capturing Joint Visual Attention Using Mobile Eye-Trackers
12:39

A Methodology for Capturing Joint Visual Attention Using Mobile Eye-Trackers

Published on: January 18, 2020

7.6K
  • Implemented a Feature Bidirectional Interactive Fusion module (FBIF) for LiDAR and pseudo RoI features.
  • Employed comprehensive experiments on the KITTI dataset to validate the proposed method.
  • Main Results:

    • Achieved a 1.55% improvement on the hard difficulty level.
    • Secured a 0.19% improvement in mean Average Precision (mAP) on the test dataset.
    • Demonstrated effective fusion of complementary sensor information and improved RoI feature extraction.

    Conclusions:

    • Bi-Att3DDet significantly enhances multi-modal 3D object detection performance.
    • The proposed SARoIFE and FBIF modules effectively address limitations in feature fusion and structural information capture.
    • The method shows strong potential for real-world autonomous driving applications.