- android - 多次调用 OnPrimaryClipChangedListener
- android - 无法更新 RecyclerView 中的 TextView 字段
- android.database.CursorIndexOutOfBoundsException : Index 0 requested, 光标大小为 0
- android - 使用 AppCompat 时,我们是否需要明确指定其 UI 组件(Spinner、EditText)颜色
我在 Windows 10 上使用 Keras 2.07、Python 3.5、Tensorflow 1.3.0
我正在使用全连接网络进行视频编码来测试论文帧内预测中使用的架构。
我希望将其用于我自己的数据。
我使用了测试数据,我认为它会很快收敛。
我在模型中使用的学习率为 0.1。那么我的val损失并没有减少。
有人可以查看或尝试此代码吗?我是否假设错误、编码错误或不耐烦?谢谢
def multi_input_model():
input1_ = Input(shape=(4,20,1), name='input1')
input2_ = Input(shape=(16,4,1), name='input2')
x1 = Flatten()(input1_)
x2 = Flatten()(input2_)
x = Concatenate()([x1, x2])
for i in range(3):
#x = Dropout(0.3)(x)
x = Dense(1024,
kernel_initializer = RandomNormal(stddev=std),
use_bias=True,
bias_initializer = 'zeros',
activation='relu'
)(x)
#x = Dropout(0.3)(x)
x = Dense(64,
kernel_initializer = RandomNormal(stddev=std),
use_bias=True,
bias_initializer = 'zeros',
)(x)
output_ = Reshape((8,8,1), name='output')(x)
model = Model(inputs=[input1_, input2_], outputs=[output_])
model.summary()
return model
sgd = SGD(lr = 0.1, momentum=0.9)
model = multi_input_model()
model.compile(optimizer=sgd,
loss='mean_squared_error',
metrics=[mse]
)
模型摘要:
Using TensorFlow backend.
__________________________________________________________________________________________________
Layer (type) Output Shape Param # Connected to
==================================================================================================
input1 (InputLayer) (None, 4, 20, 1) 0
__________________________________________________________________________________________________
input2 (InputLayer) (None, 16, 4, 1) 0
__________________________________________________________________________________________________
flatten_1 (Flatten) (None, 80) 0 input1[0][0]
__________________________________________________________________________________________________
flatten_2 (Flatten) (None, 64) 0 input2[0][0]
__________________________________________________________________________________________________
concatenate_1 (Concatenate) (None, 144) 0 flatten_1[0][0]
flatten_2[0][0]
__________________________________________________________________________________________________
dense_1 (Dense) (None, 1024) 148480 concatenate_1[0][0]
__________________________________________________________________________________________________
dense_2 (Dense) (None, 1024) 1049600 dense_1[0][0]
__________________________________________________________________________________________________
dense_3 (Dense) (None, 1024) 1049600 dense_2[0][0]
__________________________________________________________________________________________________
dense_4 (Dense) (None, 64) 65600 dense_3[0][0]
__________________________________________________________________________________________________
output (Reshape) (None, 8, 8, 1) 0 dense_4[0][0]
==================================================================================================
Total params: 2,313,280
Trainable params: 2,313,280
Non-trainable params: 0
__________________________________________________________________________________________________
训练回溯
Train on 985373 samples, validate on 246344 samples
Epoch 1/100
985373/985373 [==============================] - 7s 7us/step - loss: 0.0054 - mse: 353.9386 - val_loss: 0.0087 - val_mse: 566.5364
Lr : 0.08
Epoch 2/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0044 - mse: 288.5897 - val_loss: 0.0082 - val_mse: 534.4153
Lr : 0.08
Epoch 3/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0042 - mse: 270.7345 - val_loss: 0.0080 - val_mse: 517.2601
Lr : 0.08
Epoch 4/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0040 - mse: 259.6213 - val_loss: 0.0078 - val_mse: 504.8340
Lr : 0.08
Epoch 5/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0039 - mse: 251.3669 - val_loss: 0.0076 - val_mse: 495.0704
Lr : 0.08
Epoch 6/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0038 - mse: 244.7449 - val_loss: 0.0075 - val_mse: 486.6413
Lr : 0.08
Epoch 7/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0037 - mse: 239.2320 - val_loss: 0.0074 - val_mse: 480.2631
Lr : 0.08
Epoch 8/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0036 - mse: 234.5702 - val_loss: 0.0073 - val_mse: 473.3974
Lr : 0.08
Epoch 9/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0035 - mse: 230.5504 - val_loss: 0.0072 - val_mse: 468.4981
Lr : 0.08
Epoch 10/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0035 - mse: 227.0740 - val_loss: 0.0071 - val_mse: 463.1125
Lr : 0.08
Epoch 11/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0034 - mse: 224.0050 - val_loss: 0.0071 - val_mse: 459.0103
Lr : 0.08
Epoch 12/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0034 - mse: 221.3146 - val_loss: 0.0070 - val_mse: 455.0988
Lr : 0.08
Epoch 13/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0034 - mse: 218.9012 - val_loss: 0.0069 - val_mse: 451.7273
Lr : 0.08
Epoch 14/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0033 - mse: 216.6984 - val_loss: 0.0069 - val_mse: 448.4461
Lr : 0.08
Epoch 15/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0033 - mse: 214.7047 - val_loss: 0.0069 - val_mse: 446.2414
Lr : 0.08
Epoch 16/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0033 - mse: 212.8751 - val_loss: 0.0068 - val_mse: 443.6305
Lr : 0.08
Epoch 17/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0032 - mse: 211.1775 - val_loss: 0.0068 - val_mse: 441.6251
Lr : 0.08
Epoch 18/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0032 - mse: 209.5941 - val_loss: 0.0068 - val_mse: 439.3194
Lr : 0.08
Epoch 19/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0032 - mse: 208.1229 - val_loss: 0.0067 - val_mse: 436.9524
Lr : 0.08
Epoch 20/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0032 - mse: 206.7133 - val_loss: 0.0067 - val_mse: 435.1670
Lr : 0.08
Epoch 21/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0032 - mse: 205.4114 - val_loss: 0.0067 - val_mse: 432.8971
Lr : 0.08
Epoch 22/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0031 - mse: 204.1449 - val_loss: 0.0066 - val_mse: 431.2357
Lr : 0.08
Epoch 23/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0031 - mse: 202.9737 - val_loss: 0.0066 - val_mse: 430.8359
Lr : 0.08
Epoch 24/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0031 - mse: 201.8537 - val_loss: 0.0066 - val_mse: 429.2760
Lr : 0.08
Epoch 25/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0031 - mse: 200.7699 - val_loss: 0.0066 - val_mse: 428.2598
Lr : 0.08
Epoch 26/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0031 - mse: 199.7556 - val_loss: 0.0066 - val_mse: 426.2759
Lr : 0.08
Epoch 27/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0031 - mse: 198.7569 - val_loss: 0.0065 - val_mse: 425.0138
Lr : 0.08
Epoch 28/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0030 - mse: 197.8049 - val_loss: 0.0065 - val_mse: 423.5451
Lr : 0.08
Epoch 29/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0030 - mse: 196.8937 - val_loss: 0.0065 - val_mse: 422.2164
Lr : 0.08
Epoch 30/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0030 - mse: 196.0350 - val_loss: 0.0065 - val_mse: 421.1507
Lr : 0.08
Epoch 31/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0030 - mse: 195.1807 - val_loss: 0.0065 - val_mse: 420.2791
Lr : 0.08
Epoch 32/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0030 - mse: 194.3681 - val_loss: 0.0065 - val_mse: 420.0271
Lr : 0.08
Epoch 33/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0030 - mse: 193.5861 - val_loss: 0.0064 - val_mse: 418.1573
Lr : 0.08
Epoch 34/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0030 - mse: 192.8048 - val_loss: 0.0064 - val_mse: 417.7967
Lr : 0.08
Epoch 35/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0030 - mse: 192.0708 - val_loss: 0.0064 - val_mse: 416.3381
Lr : 0.08
Epoch 36/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0029 - mse: 191.3404 - val_loss: 0.0064 - val_mse: 416.3695
Lr : 0.08
Epoch 37/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0029 - mse: 190.6411 - val_loss: 0.0064 - val_mse: 415.9791
Lr : 0.08
Epoch 38/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0029 - mse: 189.9464 - val_loss: 0.0064 - val_mse: 414.0931
Lr : 0.08
Epoch 39/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0029 - mse: 189.2725 - val_loss: 0.0064 - val_mse: 413.8717
Lr : 0.08
Epoch 40/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0029 - mse: 188.6250 - val_loss: 0.0064 - val_mse: 413.0042
Lr : 0.08
Epoch 41/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0029 - mse: 187.9844 - val_loss: 0.0064 - val_mse: 413.0950
Lr : 0.08
Epoch 42/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0029 - mse: 187.3532 - val_loss: 0.0063 - val_mse: 412.4408
Lr : 0.08
Epoch 43/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0029 - mse: 186.7512 - val_loss: 0.0063 - val_mse: 411.1885
Lr : 0.08
Epoch 44/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0029 - mse: 186.1417 - val_loss: 0.0063 - val_mse: 410.7527
Lr : 0.08
Epoch 45/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0029 - mse: 185.5806 - val_loss: 0.0063 - val_mse: 409.3184
Lr : 0.08
Epoch 46/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0028 - mse: 185.0026 - val_loss: 0.0063 - val_mse: 410.3592
Lr : 0.08
Epoch 47/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0028 - mse: 184.4485 - val_loss: 0.0063 - val_mse: 409.0613
Lr : 0.08
Epoch 48/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0028 - mse: 183.8902 - val_loss: 0.0063 - val_mse: 409.2569
Lr : 0.08
Epoch 49/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0028 - mse: 183.3491 - val_loss: 0.0063 - val_mse: 408.1287
Lr : 0.08
Epoch 50/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0028 - mse: 182.8373 - val_loss: 0.0063 - val_mse: 407.0794
Lr : 0.08
Epoch 51/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0028 - mse: 182.3068 - val_loss: 0.0063 - val_mse: 407.4011
Lr : 0.08
Epoch 52/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0028 - mse: 181.7958 - val_loss: 0.0062 - val_mse: 405.9963
Lr : 0.08
Epoch 53/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0028 - mse: 181.2954 - val_loss: 0.0062 - val_mse: 406.3628
Lr : 0.08
Epoch 54/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0028 - mse: 180.8150 - val_loss: 0.0062 - val_mse: 405.7962
Lr : 0.08
Epoch 55/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0028 - mse: 180.3107 - val_loss: 0.0062 - val_mse: 405.0873
Lr : 0.08
Epoch 56/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0028 - mse: 179.8412 - val_loss: 0.0062 - val_mse: 405.1387
Lr : 0.08
Epoch 57/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0028 - mse: 179.3717 - val_loss: 0.0062 - val_mse: 404.4157
Lr : 0.08
Epoch 58/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0028 - mse: 178.9016 - val_loss: 0.0062 - val_mse: 403.5622
Lr : 0.08
Epoch 59/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0027 - mse: 178.4647 - val_loss: 0.0062 - val_mse: 403.3207
Lr : 0.08
Epoch 60/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0027 - mse: 177.9997 - val_loss: 0.0062 - val_mse: 403.5156
Lr : 0.08
Epoch 61/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0027 - mse: 177.5587 - val_loss: 0.0062 - val_mse: 403.2921
Lr : 0.08
Epoch 62/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0027 - mse: 177.1332 - val_loss: 0.0062 - val_mse: 402.8525
Lr : 0.08
Epoch 63/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0027 - mse: 176.6831 - val_loss: 0.0062 - val_mse: 402.3887
Lr : 0.08
Epoch 64/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0027 - mse: 176.2728 - val_loss: 0.0062 - val_mse: 401.6309
Lr : 0.08
Epoch 65/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0027 - mse: 175.8403 - val_loss: 0.0062 - val_mse: 401.4650
Lr : 0.08
Epoch 66/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0027 - mse: 175.4337 - val_loss: 0.0062 - val_mse: 401.6886
Lr : 0.08
Epoch 67/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0027 - mse: 175.0153 - val_loss: 0.0062 - val_mse: 401.1379
Lr : 0.08
Epoch 68/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0027 - mse: 174.6212 - val_loss: 0.0062 - val_mse: 401.4452
Lr : 0.08
Epoch 69/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0027 - mse: 174.2141 - val_loss: 0.0062 - val_mse: 401.5032
Lr : 0.08
Epoch 70/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0027 - mse: 173.8265 - val_loss: 0.0062 - val_mse: 400.2583
Lr : 0.08
Epoch 71/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0027 - mse: 173.4480 - val_loss: 0.0062 - val_mse: 399.9907
Lr : 0.08
Epoch 72/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0027 - mse: 173.0551 - val_loss: 0.0062 - val_mse: 400.5084
Lr : 0.08
Epoch 73/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0027 - mse: 172.6990 - val_loss: 0.0062 - val_mse: 400.7729
Lr : 0.08
Epoch 74/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 172.3093 - val_loss: 0.0061 - val_mse: 398.8733
Lr : 0.08
Epoch 75/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 171.9405 - val_loss: 0.0061 - val_mse: 399.6324
Lr : 0.08
Epoch 76/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 171.5539 - val_loss: 0.0062 - val_mse: 400.0404
Lr : 0.08
Epoch 77/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 171.2163 - val_loss: 0.0061 - val_mse: 399.0176
Lr : 0.08
Epoch 78/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 170.8428 - val_loss: 0.0061 - val_mse: 398.2502
Lr : 0.08
Epoch 79/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 170.5107 - val_loss: 0.0061 - val_mse: 399.4756
Lr : 0.08
Epoch 80/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 170.1418 - val_loss: 0.0061 - val_mse: 398.6772
Lr : 0.08
Epoch 81/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 169.7998 - val_loss: 0.0061 - val_mse: 398.6140
Lr : 0.08
Epoch 82/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 169.4597 - val_loss: 0.0061 - val_mse: 398.5729
Lr : 0.08
Epoch 83/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 169.1229 - val_loss: 0.0061 - val_mse: 397.8155
Lr : 0.08
Epoch 84/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 168.7856 - val_loss: 0.0061 - val_mse: 397.5096
Lr : 0.08
Epoch 85/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 168.4588 - val_loss: 0.0061 - val_mse: 397.9902
Lr : 0.08
Epoch 86/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 168.1268 - val_loss: 0.0061 - val_mse: 397.3359
Lr : 0.08
Epoch 87/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 167.8185 - val_loss: 0.0061 - val_mse: 397.9455
Lr : 0.08
Epoch 88/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 167.4879 - val_loss: 0.0061 - val_mse: 397.7877
Lr : 0.08
Epoch 89/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 167.1769 - val_loss: 0.0061 - val_mse: 397.3274
Lr : 0.08
Epoch 90/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 166.8585 - val_loss: 0.0061 - val_mse: 397.2146
Lr : 0.08
Epoch 91/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 166.5457 - val_loss: 0.0061 - val_mse: 397.8896
Lr : 0.08
Epoch 92/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 166.2354 - val_loss: 0.0061 - val_mse: 396.7620
Lr : 0.08
Epoch 93/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0026 - mse: 165.9282 - val_loss: 0.0061 - val_mse: 397.0715
Lr : 0.08
Epoch 94/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0025 - mse: 165.6314 - val_loss: 0.0061 - val_mse: 397.2557
Lr : 0.08
Epoch 95/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0025 - mse: 165.3255 - val_loss: 0.0061 - val_mse: 396.8092
Lr : 0.08
Epoch 96/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0025 - mse: 165.0327 - val_loss: 0.0061 - val_mse: 396.3043
Lr : 0.08
Epoch 97/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0025 - mse: 164.7327 - val_loss: 0.0061 - val_mse: 397.1202
Lr : 0.08
Epoch 98/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0025 - mse: 164.4545 - val_loss: 0.0061 - val_mse: 397.8727
Lr : 0.08
Epoch 99/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0025 - mse: 164.1428 - val_loss: 0.0061 - val_mse: 397.3806
Lr : 0.08
Epoch 100/100
985373/985373 [==============================] - 6s 6us/step - loss: 0.0025 - mse: 163.8922 - val_loss: 0.0061 - val_mse: 396.7912
Lr : 0.08
最佳答案
您的代码工作正常。从损失图中,您可以看到模型收敛得很好。您的问题是您的训练损失和验证损失之间存在很大差距。这有时被称为“低偏差,高方差”。
解决方案包括:
有很多资源可以为您提供想法,例如:
关于python - val损失没有减少,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/56209602/
我是pytorch的新手。请问添加'loss.item()'有什么区别?以下2部分代码: for epoch in range(epochs): trainingloss =0 for
我有一个包含 4 列的 MySQL 表,如下所示。 TransactionID | Item | Amount | Date ------------------------------------
我目前正在使用 cocos2d、Box2D 和 Objective-C 为 iPad 和 iPhone 制作游戏。 每次更新都会发生很多事情,很多事情必须解决。 我最近将我的很多代码重构为几个小方法,
我一直在关注 Mixed Precision Guide .因此,我正在设置: keras.mixed_precision.set_global_policy(mixed_precision) 像这样
double lnumber = Math.pow(2, 1000); 打印 1.0715086071862673E301 我尝试过的事情 我尝试使用 BigDecimal 类来扩展这个数字: St
我正在尝试创建一个神经网络来近似函数(正弦、余弦、自定义...),但我在格式上遇到困难,我不想使用输入标签,而是使用输入输出。我该如何更改它? 我正在关注this tutorial import te
我有一个具有 260,000 行和 35 列的“单热编码”(全一和零)数据矩阵。我正在使用 Keras 训练一个简单的神经网络来预测一个连续变量。制作网络的代码如下: model = Sequenti
什么是像素级 softmax 损失?在我的理解中,这只是一个交叉熵损失,但我没有找到公式。有人能帮我吗?最好有pytorch代码。 最佳答案 您可以阅读 here所有相关内容(那里还有一个指向源代码的
我正在训练一个 CNN 架构来使用 PyTorch 解决回归问题,其中我的输出是一个 20 个值的张量。我计划使用 RMSE 作为模型的损失函数,并尝试使用 PyTorch 的 nn.MSELoss(
在每个时代结束时,我得到例如以下输出: Epoch 1/25 2018-08-06 14:54:12.555511: 2/2 [==============================] - 86
我正在使用 Keras 2.0.2 功能 API (Tensorflow 1.0.1) 来实现一个网络,该网络接受多个输入并产生两个输出 a 和 b。我需要使用 cosine_proximity 损失
我正在尝试设置很少层的神经网络,这将解决简单的回归问题,这应该是f(x) = 0,1x 或 f(x) = 10x 所有代码如下所示(数据生成和神经网络) 4 个带有 ReLu 的全连接层 损失函数 R
我正在研究在 PyTorch 中使用带有梯度惩罚的 Wasserstein GAN,但始终得到大的、正的生成器损失,并且随着时间的推移而增加。 我从 Caogang's implementation
我正在尝试在 TensorFlow 中实现最大利润损失。这个想法是我有一些积极的例子,我对一些消极的例子进行了采样,并想计算类似的东西 其中 B 是我的批处理大小,N 是我要使用的负样本数。 我是 t
我正在尝试预测一个连续值(第一次使用神经网络)。我已经标准化了输入数据。我不明白为什么我会收到 loss: nan从第一个纪元开始的输出。 我阅读并尝试了以前对同一问题的回答中的许多建议,但没有一个对
我目前正在学习神经网络,并尝试训练 MLP 以使用 Python 中的反向传播来学习 XOR。该网络有两个隐藏层(使用 Sigmoid 激活)和一个输出层(也是 Sigmoid)。 网络(大约 20,
尝试在 keras 中自定义损失函数(平滑 L1 损失),如下所示 ValueError: Shape must be rank 0 but is rank 5 for 'cond/Switch' (
我试图在 tensorflow 中为门牌号图像创建一个卷积神经网络 http://ufldl.stanford.edu/housenumbers/ 当我运行我的代码时,我在第一步中得到了 nan 的成
我正在尝试使用我在 Keras 示例( https://github.com/keras-team/keras/blob/master/examples/variational_autoencoder
我试图了解 CTC 损失如何用于语音识别以及如何在 Keras 中实现它。 我认为我理解的内容(如果我错了,请纠正我!)总体而言,CTC 损失被添加到经典网络之上,以便逐个元素(对于文本或语音而言逐个
我是一名优秀的程序员,十分优秀!