Filter löschen
Filter löschen

Why my deep learning network is producing NaN outputs?

39 Ansichten (letzte 30 Tage)
BIPIN SAMUEL
BIPIN SAMUEL am 28 Jun. 2024 um 1:26
Kommentiert: BIPIN SAMUEL am 1 Jul. 2024 um 12:17
I have a deep CNN network with 39 layers. While training the network is producing NaN output so the loss function is also producing NaN values. My network also has some custom layers which uses 'dlconv' where the weights are intialized using 'initializeGlorot' function. Why the network is producing 'NaN' outputs? Is there any way to solve this? Kindly request to provide suggestions.
Thanking,
Bipin

Antworten (1)

Prasanna
Prasanna am 28 Jun. 2024 um 5:03
Hi Bipin,
When a deep CNN network produces NaN outputs during training, it typically indicates that there is an issue with the numerical stability of the network. Some reasons for NaN outputs can be:
  • Improper weight initializations
  • High Learning Rate
  • Batch normalizations.
To debug and solve the issue, you can try lowering the learning rate, gradient clipping and check and debug custom layers and their corresponding weight initializations and loss functions. Temporarily, you can use a smaller version of your network to see if the issue persists. This can help isolate whether the problem is due to the network size or specific layers.
For more help on network producing NaN outputs, refer the following links:
Hope this helps.
  1 Kommentar
BIPIN SAMUEL
BIPIN SAMUEL am 1 Jul. 2024 um 12:17
Thank you @Prasanna, the initial layer that I was using the network was custom layer and it was generating NaN while initializing the weight values, but it was working properly for the inner layers. When I replace the initial layer with the in-built matlab layers the network is working. So, I will check with that initial layers.

Melden Sie sich an, um zu kommentieren.

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by