Sensor Fusion for Autonomous Outdoor Navigation Using Neural Networks
Tech. Report, CMU-RI-TR-95-05, Robotics Institute, Carnegie Mellon University, 1995
Abstract
For many navigation tasks, a single sensing modality is sufficiently rich to accomplish the desired motion control goals; for practical autonomous outdoor navigation, a single sensing modality is a crippling limitation on what tasks can be undertaken. In the research detailed in this paper, we open the door for a whole new suite of real-time autonomous navigation tasks previously unattainable. Using neural networks, including a neural network paradigm particularly well suited to sensor fusion, and Carnegie Mellon University's HMMWV off-road vehicle, we have successfully performed simulated and real-world navigation tasks that required the use of multiple sensing modalities.
BibTeX
@techreport{Davis-1995-13826,author = {Ian Davis},
title = {Sensor Fusion for Autonomous Outdoor Navigation Using Neural Networks},
year = {1995},
month = {January},
institute = {Carnegie Mellon University},
address = {Pittsburgh, PA},
number = {CMU-RI-TR-95-05},
}
Copyright notice: This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. These works may not be reposted without the explicit permission of the copyright holder.