gpt4 book ai didi

machine-learning - Tensorflow:损失变为 'NaN'

转载 作者:行者123 更新时间:2023-11-30 09:29:11 25 4
gpt4 key购买 nike

我正在使用 Tensorflow 对 CPU 进行 CIFAR-10 训练。前几轮比赛中,输球看起来还不错。但在步骤 10210 之后,损失发生变化并最终变为 NaN。

我的网络模型是来自他们网站的 CIFAR-10 CNN 模型。这是我的设置,

image_size = 32
num_channels = 3
num_classes = 10
num_batches_to_run = 50000
batch_size = 128
eval_batch_size = 64
initial_learning_rate = 0.1
learning_rate_decay_factor = 0.1
num_epochs_per_decay = 350.0
moving_average_decay = 0.9999

结果如下所示。

2017-05-12 21:53:05.125242: step 10210, loss = 4.99 (124.9 examples/sec; 1.025 sec/batch)
2017-05-12 21:53:13.960001: step 10220, loss = 7.55 (139.5 examples/sec; 0.918 sec/batch)
2017-05-12 21:53:23.491228: step 10230, loss = 6.63 (149.5 examples/sec; 0.856 sec/batch)
2017-05-12 21:53:33.355805: step 10240, loss = 8.08 (113.3 examples/sec; 1.129 sec/batch)
2017-05-12 21:53:43.007007: step 10250, loss = 7.18 (126.7 examples/sec; 1.010 sec/batch)
2017-05-12 21:53:52.650118: step 10260, loss = 16.61 (138.0 examples/sec; 0.928 sec/batch)
2017-05-12 21:54:02.537279: step 10270, loss = 9.60 (137.6 examples/sec; 0.930 sec/batch)
2017-05-12 21:54:12.390117: step 10280, loss = 46526.25 (145.5 examples/sec; 0.880 sec/batch)
2017-05-12 21:54:22.060741: step 10290, loss = 133479743509972411931057146822656.00 (130.4 examples/sec; 0.982 sec/batch)
2017-05-12 21:54:31.691058: step 10300, loss = nan (115.8 examples/sec; 1.105 sec/batch)

关于 NaN 损失有什么想法吗?

最佳答案

当你的学习率太高时,这种情况在实践中经常发生,我倾向于从 0.001 开始,然后从那里开始,0.1 在大多数数据集上都处于非常高的位置,特别是如果你没有将你的损失除以你的损失批量大小。

关于machine-learning - Tensorflow:损失变为 'NaN',我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/43948571/

25 4 0
Copyright 2021 - 2024 cfsdn All Rights Reserved 蜀ICP备2022000587号
广告合作:1813099741@qq.com 6ren.com