New! Sign up for our free email newsletter.
Science News
from research organizations

New data processing module makes deep neural networks smarter

Date:
September 16, 2020
Source:
North Carolina State University
Summary:
Artificial intelligence researchers have improved the performance of deep neural networks by combining feature normalization and feature attention modules into a single module that they call attentive normalization. The hybrid module improves the accuracy of the system significantly, while using negligible extra computational power.
Share:
FULL STORY

Artificial intelligence researchers at North Carolina State University have improved the performance of deep neural networks by combining feature normalization and feature attention modules into a single module that they call attentive normalization (AN). The hybrid module improves the accuracy of the system significantly, while using negligible extra computational power.

"Feature normalization is a crucial element of training deep neural networks, and feature attention is equally important for helping networks highlight which features learned from raw data are most important for accomplishing a given task," says Tianfu Wu, corresponding author of a paper on the work and an assistant professor of electrical and computer engineering at NC State. "But they have mostly been treated separately. We found that combining them made them more efficient and effective."

To test their AN module, the researchers plugged it into four of the most widely used neural network architectures: ResNets, DenseNets, MobileNetsV2 and AOGNets. They then tested the networks against two industry standard benchmarks: the ImageNet-1000 classification benchmark and the MS-COCO 2017 object detection and instance segmentation benchmark.

"We found that AN improved performance for all four architectures on both benchmarks," Wu says. "For example, top-1 accuracy in the ImageNet-1000 improved by between 0.5% and 2.7%. And Average Precision (AP) accuracy increased by up to 1.8% for bounding box and 2.2% for semantic mask in MS-COCO.

"Another advantage of AN is that it facilitates better transfer learning between different domains," Wu says. "For example, from image classification in ImageNet to object detection and semantic segmentation in MS-COCO. This is illustrated by the performance improvement in the MS-COCO benchmark, which was obtained by fine-tuning ImageNet-pretrained deep neural networks in MS-COCO, a common workflow in state-of-the-art computer vision.

"We have released the source code and hope our AN will lead to better integrative design of deep neural networks."


Story Source:

Materials provided by North Carolina State University. Note: Content may be edited for style and length.


Cite This Page:

North Carolina State University. "New data processing module makes deep neural networks smarter." ScienceDaily. ScienceDaily, 16 September 2020. <www.sciencedaily.com/releases/2020/09/200916131103.htm>.
North Carolina State University. (2020, September 16). New data processing module makes deep neural networks smarter. ScienceDaily. Retrieved December 20, 2024 from www.sciencedaily.com/releases/2020/09/200916131103.htm
North Carolina State University. "New data processing module makes deep neural networks smarter." ScienceDaily. www.sciencedaily.com/releases/2020/09/200916131103.htm (accessed December 20, 2024).

Explore More

from ScienceDaily

RELATED STORIES