tensorflow - 如何使用 tf.saved_model 加载模型并调用预测函数 [TENSORFLOW 2.0 API]

标签 tensorflow model save load predict

我对 tensorflow 非常陌生,尤其是 2.0,因为关于该 API 的示例还不够多,但它似乎比 1.x 方便得多
到目前为止,我设法使用 tf.estimator api 训练了一个线性模型,然后设法使用 tf.estimator.exporter 保存它。

之后我想使用 tf.saved_model api 加载这个模型,我想我成功了,但我对我的程序有一些疑问,所以这里是我的代码的快速浏览:

所以我有一个使用 tf.feature_column api 创建的功能数组,它看起来像这样:

feature_columns = 
[NumericColumn(key='geoaccuracy', shape=(1,), default_value=None, dtype=tf.float32, normalizer_fn=None),
 NumericColumn(key='longitude', shape=(1,), default_value=None, dtype=tf.float32, normalizer_fn=None),
 NumericColumn(key='latitude', shape=(1,), default_value=None, dtype=tf.float32, normalizer_fn=None),
 NumericColumn(key='bidfloor', shape=(1,), default_value=None, dtype=tf.float32, normalizer_fn=None),
 VocabularyListCategoricalColumn(key='adid', vocabulary_list=('115', '124', '139', '122', '121', '146', '113', '103', '123', '104', '147', '114', '149', '148'), dtype=tf.string, default_value=-1, num_oov_buckets=0),
 VocabularyListCategoricalColumn(key='campaignid', vocabulary_list=('36', '31', '33', '28'), dtype=tf.string, default_value=-1, num_oov_buckets=0),
 VocabularyListCategoricalColumn(key='exchangeid', vocabulary_list=('1241', '823', '1240', '1238'), dtype=tf.string, default_value=-1, num_oov_buckets=0),
...]

之后,我以这种方式使用我的特征列数组定义了一个估计器,并对其进行训练。到这里为止,没问题。
linear_est = tf.estimator.LinearClassifier(feature_columns=feature_columns)

训练我的模型后,我想保存它,所以这里开始怀疑,这是我的处理方式,但不确定这是正确的方法:
serving_input_parse = tf.feature_column.make_parse_example_spec(feature_columns=feature_columns)

""" view of the variable : serving_input_parse = 
 {'adid': VarLenFeature(dtype=tf.string),
 'at': VarLenFeature(dtype=tf.string),
 'basegenres': VarLenFeature(dtype=tf.string),
 'bestkw': VarLenFeature(dtype=tf.string),
 'besttopic': VarLenFeature(dtype=tf.string),
 'bidfloor': FixedLenFeature(shape=(1,), dtype=tf.float32, default_value=None),
 'browserid': VarLenFeature(dtype=tf.string),
 'browserlanguage': VarLenFeature(dtype=tf.string)
 ...} """

# exporting the model :
linear_est.export_saved_model(export_dir_base='./saved',
 serving_input_receiver_fn=tf.estimator.export.build_parsing_serving_input_receiver_fn(serving_input_receiver_fn),
 as_text=True)

现在我尝试加载它,但我不知道如何使用加载的模型使用来自 Pandas 数据帧的原始数据对其进行预测
loaded = tf.saved_model.load('saved/1573144361/')

还有一件事,我试图查看模型的签名,但我无法真正理解我的输入形状发生了什么
MetaGraphDef with tag-set: 'serve' contains the following SignatureDefs:

signature_def['classification']:
  The given SavedModel SignatureDef contains the following input(s):
    inputs['inputs'] tensor_info:
        dtype: DT_STRING
        shape: (-1)
        name: input_example_tensor:0
  The given SavedModel SignatureDef contains the following output(s):
    outputs['classes'] tensor_info:
        dtype: DT_STRING
        shape: (-1, 2)
        name: head/Tile:0
    outputs['scores'] tensor_info:
        dtype: DT_FLOAT
        shape: (-1, 2)
        name: head/predictions/probabilities:0
  Method name is: tensorflow/serving/classify

signature_def['predict']:
  The given SavedModel SignatureDef contains the following input(s):
    inputs['examples'] tensor_info:
        dtype: DT_STRING
        shape: (-1)
        name: input_example_tensor:0
  The given SavedModel SignatureDef contains the following output(s):
    outputs['all_class_ids'] tensor_info:
        dtype: DT_INT32
        shape: (-1, 2)
        name: head/predictions/Tile:0
    outputs['all_classes'] tensor_info:
        dtype: DT_STRING
        shape: (-1, 2)
        name: head/predictions/Tile_1:0
    outputs['class_ids'] tensor_info:
        dtype: DT_INT64
        shape: (-1, 1)
        name: head/predictions/ExpandDims:0
    outputs['classes'] tensor_info:
        dtype: DT_STRING
        shape: (-1, 1)
        name: head/predictions/str_classes:0
    outputs['logistic'] tensor_info:
        dtype: DT_FLOAT
        shape: (-1, 1)
        name: head/predictions/logistic:0
    outputs['logits'] tensor_info:
        dtype: DT_FLOAT
        shape: (-1, 1)
        name: linear/linear_model/linear/linear_model/linear/linear_model/weighted_sum:0
    outputs['probabilities'] tensor_info:
        dtype: DT_FLOAT
        shape: (-1, 2)
        name: head/predictions/probabilities:0
  Method name is: tensorflow/serving/predict

signature_def['regression']:
  The given SavedModel SignatureDef contains the following input(s):
    inputs['inputs'] tensor_info:
        dtype: DT_STRING
        shape: (-1)
        name: input_example_tensor:0
  The given SavedModel SignatureDef contains the following output(s):
    outputs['outputs'] tensor_info:
        dtype: DT_FLOAT
        shape: (-1, 1)
        name: head/predictions/logistic:0
  Method name is: tensorflow/serving/regress

signature_def['serving_default']:
  The given SavedModel SignatureDef contains the following input(s):
    inputs['inputs'] tensor_info:
        dtype: DT_STRING
        shape: (-1)
        name: input_example_tensor:0
  The given SavedModel SignatureDef contains the following output(s):
    outputs['classes'] tensor_info:
        dtype: DT_STRING
        shape: (-1, 2)
        name: head/Tile:0
    outputs['scores'] tensor_info:
        dtype: DT_FLOAT
        shape: (-1, 2)
        name: head/predictions/probabilities:0
  Method name is: tensorflow/serving/classify

最佳答案

saved_model.load(...) documentation演示了这样的基 native 制:

imported = tf.saved_model.load(path)
f = imported.signatures["serving_default"]
print(f(x=tf.constant([[1.]])))

我自己还是个新手,但 serving_default似乎是使用 saved_model.save(...) 时的默认签名.

(我的理解是saved_model.save(...)不保存模型,它保存图形。为了解释图形,您需要显式存储“签名”图形上的定义操作。如果您不显式地这样做那么“serve_default”将是您唯一的签名。)

我在下面提供了一个实现。有几个细节值得注意:
  • 输入 需求成为张量;所以我需要手动进行转换。
  • 输出是字典。该文档将其描述为“具有从签名 key 到功能的签名属性映射的可跟踪对象”。

  • 就我而言,字典的键是一个相对随意的“dense_83”。这似乎有点……具体。所以我概括了使用迭代器忽略键的解决方案:

    import tensorflow as tf
    input_data = tf.constant(input_data, dtype=tf.float32)
    prediction_tensors = signature_collection.signatures["serving_default"](input_data)
    for _, values in prediction_tensors.items():
        predictions = values.numpy()[0]
        return predictions
    raise Exception("Expected a response from predict(...).")
    

    关于tensorflow - 如何使用 tf.saved_model 加载模型并调用预测函数 [TENSORFLOW 2.0 API],我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/58755970/

    相关文章:

    javascript - 从tensorflow js模型获取原始输出

    jquery - MVC 动态部分 View 数据在 IE7 中消失

    jquery - 使用jQuery拖放后如何获取坐标位置?

    c++ - 将对象 vector 保存到文件 C++

    ios - 子上下文未保存在核心数据中

    python - Keras:无法在我的模型中使用自定义损失函数

    python - 尝试使用之前训练的 tf.keras 模型作为预训练,但得到“ValueError : Input 0 of layer dense_3 is incompatible with the laye

    python - 使用 BERT 对单词位置进行序列标记

    python - 如何在 Django 中定义跨年、月的日期范围?

    c# - Entity Framework 代码优先 : how to add more props on model and update database?