gpt4 book ai didi

python - 在keras中实现自定义目标函数

转载 作者:行者123 更新时间:2023-11-30 09:46:26 25 4
gpt4 key购买 nike

我正在尝试实现我自己的成本函数,特别是以下函数:

enter image description here

现在我知道这个问题已经在这个网站上被问过好几次了,我读到的答案通常如下所示:

def custom_objective(y_true, y_pred):
....
return L

人们似乎总是使用y_truey_pred,然后说你只需要编译模型model.compile(loss=custom_objective) 并从那里开始。没有人真正在代码中的某处提到 y_true=somethingy_pred=something。这是我必须在模型中指定的内容吗?

我的代码

不确定我是否正确使用 .predict() 从模型训练时获取运行预测:

params = {'lr': 0.0001,
'batch_size': 30,
'epochs': 400,
'dropout': 0.2,
'optimizer': 'adam',
'losses': 'avg_partial_likelihood',
'activation':'relu',
'last_activation': 'linear'}

def model(x_train, y_train, x_val, y_val):

l2_reg = 0.4
kernel_init ='he_uniform'
bias_init ='he_uniform'
layers=[20, 20, 1]

model = Sequential()

# layer 1
model.add(Dense(layers[0], input_dim=x_train.shape[1],
W_regularizer=l2(l2_reg),
kernel_initializer=kernel_init,
bias_initializer=bias_init))


model.add(BatchNormalization(axis=-1, momentum=momentum, center=True))

model.add(Activation(params['activation']))

model.add(Dropout(params['dropout']))

# layer 2+
for layer in range(0, len(layers)-1):

model.add(Dense(layers[layer+1], W_regularizer=l2(l2_reg),
kernel_initializer=kernel_init,
bias_initializer=bias_init))


model.add(BatchNormalization(axis=-1, momentum=momentum, center=True))

model.add(Activation(params['activation']))

model.add(Dropout(params['dropout']))

# Last layer
model.add(Dense(layers[-1], activation=params['last_activation'],
kernel_initializer=kernel_init,
bias_initializer=bias_init))

model.compile(loss=params['losses'],
optimizer=keras.optimizers.adam(lr=params['lr']),
metrics=['accuracy'])

history = model.fit(x_train, y_train,
validation_data=[x_val, y_val],
batch_size=params['batch_size'],
epochs=params['epochs'],
verbose=1)

y_pred = model.predict(x_train, batch_size=params['batch_size'])

history_dict = history.history

model_output = {'model':model,
'history_dict':history_dict,
'log_risk':y_pred}

return model_output

然后创建模型:

model(x_train, y_train, x_val, y_val)

到目前为止我的目标函数

'log_risk' 将是 y_true 并且 x_train 将用于计算 y_pred:

def avg_partial_likelihood(x_train, log_risk):



from lifelines import CoxPHFitter

cph = CoxPHFitter()

cph.fit(x_train, duration_col='survival_fu_combine', event_col='death',
show_progress=False)

# obtain exp(hx)

cph_output = pd.DataFrame(cph.summary).T

# summing hazard ratio

hazard_ratio_sum = cph_output.iloc[1,].sum()

# -log(sum(exp(hxj)))

neg_log_sum = -np.log(hazard_ratio_sum)

# sum of positive events (death==1)

sum_noncensored_events = (x_train.death==1).sum()

# neg_likelihood

neg_likelihood = -(log_risk + neg_log_sum)/sum_noncensored_events

return neg_likelihood

如果我尝试运行就会出错

  AttributeError                            Traceback (most recent call last)
<ipython-input-26-cf0236299ad5> in <module>()
----> 1 model(x_train, y_train, x_val, y_val)

<ipython-input-25-d0f9409c831a> in model(x_train, y_train, x_val, y_val)
45 model.compile(loss=avg_partial_likelihood,
46 optimizer=keras.optimizers.adam(lr=params['lr']),
---> 47 metrics=['accuracy'])
48
49 history = model.fit(x_train, y_train,

~\Anaconda3\lib\site-packages\keras\engine\training.py in compile(self, optimizer, loss, metrics, loss_weights, sample_weight_mode, weighted_metrics, target_tensors, **kwargs)
331 with K.name_scope(self.output_names[i] + '_loss'):
332 output_loss = weighted_loss(y_true, y_pred,
--> 333 sample_weight, mask)
334 if len(self.outputs) > 1:
335 self.metrics_tensors.append(output_loss)

~\Anaconda3\lib\site-packages\keras\engine\training_utils.py in weighted(y_true, y_pred, weights, mask)
401 """
402 # score_array has ndim >= 2
--> 403 score_array = fn(y_true, y_pred)
404 if mask is not None:
405 # Cast the mask to floatX to avoid float64 upcasting in Theano

<ipython-input-23-ed57799a1f9d> in avg_partial_likelihood(x_train, log_risk)
27
28 cph.fit(x_train, duration_col='survival_fu_combine', event_col='death',
---> 29 show_progress=False)
30
31 # obtain exp(hx)

~\Anaconda3\lib\site-packages\lifelines\fitters\coxph_fitter.py in fit(self, df, duration_col, event_col, show_progress, initial_beta, strata, step_size, weights_col)
90 """
91
---> 92 df = df.copy()
93
94 # Sort on time

AttributeError: 'Tensor' object has no attribute 'copy'

最佳答案

No one really mentions that somewhere in the code that y_true=something and y_pred=something ...

他们没有提及它,因为您不需要这样做!实际上,在每次传递结束时(即一批前向传播),Keras 使用该传递的真实标签和模型预测来提供 y_true 和 y_pred。因此,您根本不需要在模型中定义 y_truey_pred。只需使用后端函数定义损失函数(即 from keras import backend as K),一切都会正常工作(并且永远不要在损失函数中使用 numpy)。要了解更多信息,请查看 built-in loss functions在 Keras 中,看看它们是如何实现的。和here是可用后端函数的(可能不完整)列表。

关于python - 在keras中实现自定义目标函数,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/51953917/

25 4 0
Copyright 2021 - 2024 cfsdn All Rights Reserved 蜀ICP备2022000587号
广告合作:1813099741@qq.com 6ren.com