Tensorflow 2.1.0 - 函数构建代码之外的操作正在传递一个 "Graph"张量

标签 tensorflow machine-learning deep-learning tf.keras tensorflow2.x

我正在尝试实现最近的一篇论文。此实现的一部分涉及从 tf 1.14 迁移到 tf 2.1.0。该代码适用于 tf 1.14,但不再有效。

注意:如果我禁用急切执行 tf.compat.v1.disable_eager_execution()然后代码按预期工作。

这是解决方案吗?我之前在 中做了很多模型TF 2.x 并且永远不必禁用急切执行来实现正常功能。

我已将问题提炼为一个非常简短的要点,以显示正在发生的事情。

链接和代码,然后是详细的错误信息

链接到要点 -- https://gist.github.com/darien-schettler/fd5b25626e9eb5b1330cce670bf9cc17

代码

# version 2.1.0
import tensorflow as tf

# version 1.18.1
import numpy as np


# ######## DEFINE CUSTOM FUNCTION FOR TF LAMBDA LAYER  ######## #
def resize_like(input_tensor, ref_tensor):
    """ Resize an image tensor to the same size/shape as a reference image tensor

    Args:
        input_tensor : (image tensor) Input image tensor that will be resized
        ref_tensor   : (image tensor) Reference image tensor that we want to resize the input tensor to.

    Returns:
        reshaped tensor
    """
    reshaped_tensor = tf.image.resize(images=input_tensor,
                                      size=tf.shape(ref_tensor)[1:3],
                                      method=tf.image.ResizeMethod.NEAREST_NEIGHBOR,
                                      preserve_aspect_ratio=False,
                                      antialias=False,
                                      name=None)
    return reshaped_tensor
# ############################################################# #

# ############ DEFINE MODEL USING TF.KERAS FN API  ############ #

# INPUTS
model_input_1 = tf.keras.layers.Input(shape=(160,160,3))
model_input_2 = tf.keras.layers.Input(shape=(160,160,3))

# OUTPUTS
model_output_1 = tf.keras.layers.Conv2D(filters=64, 
                                        kernel_size=(1, 1), 
                                        use_bias=False,
                                        kernel_initializer='he_normal',
                                        name='conv_name_base')(model_input_1)
model_output_2 = tf.keras.layers.Lambda(function=resize_like,
                                        arguments={'ref_tensor': model_output_1})(model_input_2)

# MODEL
model = tf.keras.models.Model(inputs=[model_input_1, model_input_2],
                                      outputs=model_output_2,
                                      name="test_model")

# ############################################################# #

# ######### TRY TO UTILIZE PREDICT WITH DUMMY INPUT  ########## #

dummy_input = [np.ones((1,160,160,3)), np.zeros((1,160,160,3))]
model.predict(x=dummy_input) # >>>>ERROR OCCURS HERE<<<<

# ############################################################# #

完整错误

>>> model.predict(x=dummy_input) # >>>>ERROR OCCURS HERE<<<<

Traceback (most recent call last):
  File "/Users/<username>/.virtualenvs/<venv-name>/lib/python3.7/site-packages/tensorflow_core/python/eager/execute.py", line 61, in quick_execute
    num_outputs)
TypeError: An op outside of the function building code is being passed
a "Graph" tensor. It is possible to have Graph tensors
leak out of the function building context by including a
tf.init_scope in your function building code.
For example, the following function will fail:
  @tf.function
  def has_init_scope():
    my_constant = tf.constant(1.)
    with tf.init_scope():
      added = my_constant * 2
The graph tensor has name: conv_name_base_1/Identity:0

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<stdin>", line 1, in <module>
  File "/Users/<user-name>/.virtualenvs/<venv-name>/lib/python3.7/site-packages/tensorflow_core/python/keras/engine/training.py", line 1013, in predict
    use_multiprocessing=use_multiprocessing)
  File "/Users/<user-name>/.virtualenvs/<venv-name>/lib/python3.7/site-packages/tensorflow_core/python/keras/engine/training_v2.py", line 498, in predict
    workers=workers, use_multiprocessing=use_multiprocessing, **kwargs)
  File "/Users/<user-name>/.virtualenvs/<venv-name>/lib/python3.7/site-packages/tensorflow_core/python/keras/engine/training_v2.py", line 475, in _model_iteration
    total_epochs=1)
  File "/Users/<user-name>/.virtualenvs/<venv-name>/lib/python3.7/site-packages/tensorflow_core/python/keras/engine/training_v2.py", line 128, in run_one_epoch
    batch_outs = execution_function(iterator)
  File "/Users/<user-name>/.virtualenvs/<venv-name>/lib/python3.7/site-packages/tensorflow_core/python/keras/engine/training_v2_utils.py", line 98, in execution_function
    distributed_function(input_fn))
  File "/Users/<user-name>/.virtualenvs/<venv-name>/lib/python3.7/site-packages/tensorflow_core/python/eager/def_function.py", line 568, in __call__
    result = self._call(*args, **kwds)
  File "/Users/<user-name>/.virtualenvs/<venv-name>/lib/python3.7/site-packages/tensorflow_core/python/eager/def_function.py", line 638, in _call
    return self._concrete_stateful_fn._filtered_call(canon_args, canon_kwds)  # pylint: disable=protected-access
  File "/Users/<user-name>/.virtualenvs/<venv-name>/lib/python3.7/site-packages/tensorflow_core/python/eager/function.py", line 1611, in _filtered_call
    self.captured_inputs)
  File "/Users/<user-name>/.virtualenvs/<venv-name>/lib/python3.7/site-packages/tensorflow_core/python/eager/function.py", line 1692, in _call_flat
    ctx, args, cancellation_manager=cancellation_manager))
  File "/Users/<user-name>/.virtualenvs/<venv-name>/lib/python3.7/site-packages/tensorflow_core/python/eager/function.py", line 545, in call
    ctx=ctx)
  File "/Users/<user-name>/.virtualenvs/<venv-name>/lib/python3.7/site-packages/tensorflow_core/python/eager/execute.py", line 75, in quick_execute
    "tensors, but found {}".format(keras_symbolic_tensors))
tensorflow.python.eager.core._SymbolicException: Inputs to eager execution function cannot be Keras symbolic tensors, but found [<tf.Tensor 'conv_name_base_1/Identity:0' shape=(None, 160, 160, 64) dtype=float32>]

我想到的一个潜在解决方案是用自定义层替换 Lambda 层……这似乎也解决了这个问题。不过,不确定围绕这一点的最佳实践是什么。代码如下。

# version 2.1.0
import tensorflow as tf

# version 1.18.1
import numpy as np


# ######## DEFINE CUSTOM LAYER DIRECTLY BY SUBCLASSING  ######## #
class ResizeLike(tf.keras.layers.Layer):
    """ tf.keras layer to resize a tensor to the reference tensor shape.

    Attributes:
        keras.layers.Layer: Base layer class.
            This is the class from which all layers inherit.
                -   A layer is a class implementing common neural networks
                    operations, such as convolution, batch norm, etc.
                -   These operations require managing weights,
                    losses, updates, and inter-layer connectivity.
    """
    def __init__(self, **kwargs):
        super().__init__(**kwargs)

    def call(self, inputs, **kwargs):
        """TODO: docstring

        Args:
            inputs (TODO): TODO

        **kwargs:
            TODO

        Returns:
            TODO
        """

        input_tensor, ref_tensor = inputs

        return self.resize_like(input_tensor, ref_tensor)


    def resize_like(self, input_tensor, ref_tensor):
        """ Resize an image tensor to the same size/shape as a reference image tensor

        Args:
            input_tensor: (image tensor) Input image tensor that will be resized
            ref_tensor: (image tensor) Reference image tensor that we want to resize the input tensor to.

        Returns:
            reshaped tensor
        """
        reshaped_tensor = tf.image.resize(images=input_tensor,
                                          size=tf.shape(ref_tensor)[1:3],
                                          method=tf.image.ResizeMethod.NEAREST_NEIGHBOR,
                                          preserve_aspect_ratio=False,
                                          antialias=False)
        return reshaped_tensor
# ############################################################# #

# ############ DEFINE MODEL USING TF.KERAS FN API  ############ #

# INPUTS
model_input_1 = tf.keras.layers.Input(shape=(160,160,3))
model_input_2 = tf.keras.layers.Input(shape=(160,160,3))

# OUTPUTS
model_output_1 = tf.keras.layers.Conv2D(filters=64, 
                                        kernel_size=(1, 1), 
                                        use_bias=False,
                                        kernel_initializer='he_normal',
                                        name='conv_name_base')(model_input_1)
model_output_2 = ResizeLike(name="resize_layer")([model_input_2, model_output_1])

# MODEL
model = tf.keras.models.Model(inputs=[model_input_1, model_input_2],
                                      outputs=model_output_2,
                                      name="test_model")

# ############################################################# #

# ######### TRY TO UTILIZE PREDICT WITH DUMMY INPUT  ########## #

dummy_input = [np.ones((1,160,160,3)), np.zeros((1,160,160,3))]
model.predict(x=dummy_input) # >>>>ERROR OCCURS HERE<<<<

# ############################################################# #

想法??

提前致谢!!

如果您希望我提供其他任何东西,请告诉我。

最佳答案

您可以尝试以下步骤:

  • 更改 resize_like如下:
    def resize_like(inputs):
        input_tensor, ref_tensor = inputs
        reshaped_tensor = tf.image.resize(images=input_tensor,
                                  size=tf.shape(ref_tensor)[1:3],
    
        method=tf.image.ResizeMethod.NEAREST_NEIGHBOR,
                                  preserve_aspect_ratio=False,
                                  antialias=False,
                                  name=None)
        return reshaped_tensor
    
  • 然后,在 Lambda层:
    model_output_2 = tf.keras.layers.Lambda(function=resize_like)([model_input_2, model_output_1])
    
  • 关于Tensorflow 2.1.0 - 函数构建代码之外的操作正在传递一个 "Graph"张量,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/60551145/

    相关文章:

    Tensorflow:用 tf.Variable 替换/提供图形的占位符?

    machine-learning - 你能让 LSTM 手动忘记上下文吗?

    python - 在以下 CNN 模型中,全连接层之前的 reshape 是如何工作的?

    python - 如何在keras中的LSTM自动编码器中获取middel层的输出

    python - siamese-net 中的自定义组合铰链/kb-divergence 损失函数无法生成有意义的说话人嵌入

    java - Mahout SparseMatrix 的密度

    python - 值错误 : Output tensors to a Model must be the output of a TensorFlow `Layer`

    deep-learning - a.sub(or*a.grad) 实际上做了什么?

    python - keras 中不兼容的密集层错误

    tensorflow - 检查 Tensorflow(批处理版本)中张量列表中存在的值的好方法是什么?