This article presents two approaches based on deep learning for achieving autonomous navigation and subsequently, conducts an in-depth comparative analysis to elaborate their respective performance. Both methods leverage neural networks to facilitate navigation. The first approach employs a deep convolutional neural network to identify suitable areas for forward movement, while the second method utilizes the drone's front camera to capture images, which are then processed by a generative network to generate corresponding depth images. Importantly, both approaches rely solely on an RGB camera, resulting in cost savings for the robot and reduced energy consumption. The training data used for classifying collision-free and appropriate paths were gathered from simulated environments closely resembling real indoor settings. The proposed model calculates the probability associated with each segment of the input image, enabling the selection of the most viable area for drone flight. Incorporating generative networks as a depth estimation module enhances flight accuracy, making navigation in uncharted environments more straightforward. The accuracy of depth estimation was evaluated using standard criteria, achieving an accuracy rate of $84 \%$. This result underscores the effectiveness of the proposed method.
Yongzhi YangKenneth G. RicksFei Hu
Rudolph Joshua CandareRolyn C. Daguil
Yang ZhangQiang LüLin HuicanJianye Ma
Tzu-Ling HsiehZih-Syuan JhanNai-Jui YehChang-Yu ChenCheng-Ta Chuang