dc.contributor.author |
Ferhat, Hiba-El-Batoul |
|
dc.contributor.other |
Berrani, Sid-Ahmed, Directeur de thèse |
|
dc.contributor.other |
Eshaghian, Jason, Directeur de thèse |
|
dc.date.accessioned |
2022-09-13T14:15:36Z |
|
dc.date.available |
2022-09-13T14:15:36Z |
|
dc.date.issued |
2022 |
|
dc.identifier.other |
EP00413 |
|
dc.identifier.uri |
http://repository.enp.edu.dz/jspui/handle/123456789/10546 |
|
dc.description |
Mémoire de Projet de Fin d’Études : Électronique : Alger, École Nationale Polytechnique : 2022 |
fr_FR |
dc.description.abstract |
Spiking Neural Networks offer a new approach to tackle the efficiency issues of conventional artificial intelligence (AI), but there is a big performance gap between Spiking Neural Networks and Artificial Neural Networks. This work contributes to the optimization of Spiking Neural Networks in order to reduce this gap. The proposed methods are focused on architecture optimization and the development of a new spike-based loss functions, to prevent the dead neuron issue. The proposed techniques are implemented using PyTorch and SnnTorch frameworks, and benchmarked using two different data sets. The evaluation of the three approaches has been done on two different data sets, using six different loss functions, and using accuracy as a main metric. Hence, new horizons of applications are opened up, as well as many possible optimization techniques. |
fr_FR |
dc.language.iso |
en |
fr_FR |
dc.subject |
Spiking neural networks |
fr_FR |
dc.subject |
Backpropagation |
fr_FR |
dc.subject |
Gradient descent |
fr_FR |
dc.subject |
Neuromorphic computing |
fr_FR |
dc.title |
Spiking neural networks optimization |
fr_FR |
dc.type |
Thesis |
fr_FR |