Detail publikačního výsledku

Evolutionary Approximation and Neural Architecture Search

PIŇOS, M.; MRÁZEK, V.; SEKANINA, L.

Originální název

Evolutionary Approximation and Neural Architecture Search

Anglický název

Evolutionary Approximation and Neural Architecture Search

Druh

Článek WoS

Originální abstrakt

Automated neural architecture search (NAS) methods are now employed to routinely deliver high-quality neural network architectures for various challenging data sets and reduce the designers effort. The NAS methods utilizing multi-objective evolutionary algorithms are especially useful when the objective is not only to minimize the network error but also to reduce the number of parameters (weights) or power consumption of the inference phase. We propose a multi-objective NAS method based on Cartesian genetic programming for evolving convolutional neural networks (CNN). The method allows approximate operations to be used in CNNs to reduce the power consumption of a target hardware implementation. During the NAS process, a suitable CNN architecture is evolved together with selecting approximate multipliers to deliver the best trade-offs between accuracy, network size, and power consumption. The most suitable 8 x N-bit approximate multipliers are automatically selected from a library of approximate multipliers. Evolved CNNs are compared with CNNs developed by other NAS methods on the CIFAR-10 and SVHN benchmark problems.

Anglický abstrakt

Automated neural architecture search (NAS) methods are now employed to routinely deliver high-quality neural network architectures for various challenging data sets and reduce the designers effort. The NAS methods utilizing multi-objective evolutionary algorithms are especially useful when the objective is not only to minimize the network error but also to reduce the number of parameters (weights) or power consumption of the inference phase. We propose a multi-objective NAS method based on Cartesian genetic programming for evolving convolutional neural networks (CNN). The method allows approximate operations to be used in CNNs to reduce the power consumption of a target hardware implementation. During the NAS process, a suitable CNN architecture is evolved together with selecting approximate multipliers to deliver the best trade-offs between accuracy, network size, and power consumption. The most suitable 8 x N-bit approximate multipliers are automatically selected from a library of approximate multipliers. Evolved CNNs are compared with CNNs developed by other NAS methods on the CIFAR-10 and SVHN benchmark problems.

Klíčová slova

Approximatecomputing, Convolutional neural network, Cartesian genetic programming, Neuroevolution,Energy efficiency

Klíčová slova v angličtině

Approximatecomputing, Convolutional neural network, Cartesian genetic programming, Neuroevolution,Energy efficiency

Autoři

PIŇOS, M.; MRÁZEK, V.; SEKANINA, L.

Rok RIV

2023

Vydáno

10.06.2022

ISSN

1389-2576

Periodikum

Genetic Programming and Evolvable Machines

Svazek

23

Číslo

3

Stát

Spojené státy americké

Strany od

351

Strany do

374

Strany počet

24

URL

BibTex

@article{BUT179451,
  author="Michal {Piňos} and Vojtěch {Mrázek} and Lukáš {Sekanina}",
  title="Evolutionary Approximation and Neural Architecture Search",
  journal="Genetic Programming and Evolvable Machines",
  year="2022",
  volume="23",
  number="3",
  pages="351--374",
  doi="10.1007/s10710-022-09441-z",
  issn="1389-2576",
  url="https://link.springer.com/article/10.1007/s10710-022-09441-z"
}