gpt4 book ai didi

python - Tensorflow中的 block 激活函数实现

转载 作者:太空宇宙 更新时间:2023-11-03 21:29:29 25 4
gpt4 key购买 nike

我试图重现一个 DNN,其中使用了名为 BlockRelu 的 block 激活函数。其定义为

BlockRelu

我尝试根据一些关于自定义激活函数的示例代码来编写这个函数,但这些函数都是标量函数,但BlockRelu将 block 作为一个整体来处理。由于 numpy 数组和张量的差异,这里不能使用 numpy 函数。我想知道是否有人可以帮忙。谢谢。这是我的代码:

import tensorflow as tf
import numpy as np
from tensorflow.python.framework import ops

def block_relu(x):
for i in range(x.shape[0]):
if x[i] > 0:
return x
return x * 0


def grad_block_relu(x):
for i in range(x.shape[0]):
if x[i] > 0:
return np.ones(x.shape[0])
return x * 0


# transferring a common function into a numpy function, not needed here
'''
block_relu_np = np.vectorize(block_relu)
grad_block_relu_np = np.vectorize(grad_block_relu)
'''
# numpy uses float64 but tensorflow uses float32
block_relu_np32 = lambda x: block_relu(x).astype(np.float32)
grad_block_relu_np32 = lambda x: grad_block_relu(x).astype(np.float32)


def grad_block_relu_tf(x, name=None):
with ops.name_scope(name, "grad_block_relu_tf", [x]) as name:
y = tf.py_func(grad_block_relu_np32, [x], [tf.float32], False, name)
return y[0]


def my_py_func(func, inp, Tout, stateful=False, name=None, my_grad_func=None):
# a unique name is required to avoid duplicates:
random_name = "PyFuncGrad" + str(np.random.randint(0, 1E+8))
tf.RegisterGradient(random_name)(my_grad_func)
g = tf.get_default_graph()
with g.gradient_override_map({"PyFunc": random_name, "PyFuncStateless": random_name}):
return tf.py_func(func, inp, Tout, stateful=stateful, name=name)


# The gradient function we need to pass to the above my_py_func function takes a special form:
# It needs to take in (an operation, the previous gradients before the operation)
# and propagate(i.e., return) the gradients backward after the operation.
def _block_relu_grad(op, pre_grad):
x = op.inputs[0]
cur_grad = grad_block_relu(x)
next_grad = pre_grad * cur_grad
return next_grad


def block_relu_tf(x, name=None):
with ops.name_scope(name, "block_relu_tf", [x]) as name:
y = my_py_func(block_relu_np32, [x], [tf.float32], stateful=False, name=name, my_grad_func=_block_relu_grad)
return y[0]


with tf.Session() as sess:
x = tf.constant([-0.3, 0.005, 0.08, 0.12])
y = block_relu_tf(x)
tf.global_variables_initializer().run()
print(x.eval())
print(y.eval())
print(tf.gradients(y, [x])[0].eval())

会报错:TypeError:不允许使用 tf.Tensor 作为 Python bool 。使用 if t is not None: 而不是 if t: 来测试张量是否已定义,并使用 TensorFlow 操作(例如 tf.cond)来执行以该值为条件的子图张量的。

最佳答案

我很确定您可以使用标准 Tensorflow 函数来实现它:

# input: x
y = tf.scalar_mul( tf.sign( tf.reduce_max( tf.nn.relu(x))), x)

关于python - Tensorflow中的 block 激活函数实现,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/53614880/

25 4 0
Copyright 2021 - 2024 cfsdn All Rights Reserved 蜀ICP备2022000587号
广告合作:1813099741@qq.com 6ren.com