python - Tensorflow a2.0.0 : Converting CSV to a tfrecord, 创建一个 Keras 模型,该模型使用来自大型源的管道数据,将权重存储到 CSV 文件中?

标签 python python-3.x tensorflow tensorflow-datasets

我正在通过 Andrew NG 在 Coursera 上的讲座学习机器学习。该类(class)使用 Matlab,它非常适合理解机器学习模型并对其进行原型(prototype)设计,但速度相当慢。我目前正在研究 Tensorflow,因为它支持 GPU 利用率和数据管道,这应该会加快我的模型速度。

但是,我完全迷失了这一点。文档没有详细介绍,示例代码没有注释,最重要的是,Tensorflow 刚刚发布了 Alpha2.0,它显着改变了 API(许多旧的 StackOverflow 线程没有帮助)。

我的目标是:

  1. 将大型 (10GB+) CSV 文件转换为 tfrecords(在某处找到) 这有好处吗?)
  2. 创建一个ks.dataset,在多个线程中读取数据并 将其输送到模型
  3. 创建一个模型,使用我的 GPU 从所述数据集中学习
  4. 将学习到的参数导出到文件

目前,我只能构建 keras 模型

model = keras.Sequential([
    keras.layers.Conv2D(filters=3, activation='relu',
                        kernel_regularizer=keras.regularizers.l2(0.001),
                        kernel_size=28,
                        padding="same",
                        input_shape=(28, 28, 1)),
    keras.layers.Flatten(),
    keras.layers.Dropout(0.09),
    keras.layers.Dense(10, activation='softmax', kernel_regularizer=keras.regularizers.l2(lambd)),
    keras.layers.Dropout(0.09)])

model.compile(optimizer='adam',
              loss='sparse_categorical_crossentropy',
              metrics=['accuracy'])

model.fit(x_train, y_train, epochs=47, batch_size=256)
test_loss, test_acc = model.evaluate(x_test, y_test)
print('\nTest accuracy:', test_acc)

此时任何事情都会有帮助!我应该研究哪些功能对我的目标至关重要?

最佳答案

经过 24 小时不间断的研究,我终于把所有的碎片粘到了拼图上。 API 很棒,但缺少文档。

将 CSV 转换为 tfrecord:

import tensorflow as tf
import numpy as np
import pandas as pd # For reading .csv
from datetime import datetime # For knowing how long does each read/write take

def _bytes_feature(value):
    # Returns a bytes_list from a string / byte.
    if isinstance(value, type(tf.constant(0))):
        value = value.numpy()  # BytesList won't unpack a string from an EagerTensor.
    return tf.train.Feature(bytes_list=tf.train.BytesList(value=[value]))


def _float_feature(value):
    # Returns a float_list from a float / double.
    # If a list of values was passed, a float list feature with the entire list will be returned
    if isinstance(value, list):
        return tf.train.Feature(float_list=tf.train.FloatList(value=value))

    return tf.train.Feature(float_list=tf.train.FloatList(value=[value]))


def _int64_feature(value):
    # Returns an int64_list from a bool / enum / int / uint.
    return tf.train.Feature(int64_list=tf.train.Int64List(value=[value]))


def serialize_example(pandabase):
    # Serializes inputs from a pandas dataset (read in chunks)

    # Creates a mapping of the features from the header row of the file
    base_chunk = pandabase.get_chunk(0)
    num_features = len(base_chunk.columns)
    features_map = {}
    for i in range(num_features):
        features_map.update({'feature' + str(i): _float_feature(0)})

    # Set writing options with compression
    options = tf.io.TFRecordOptions(compression_type=tf.io.TFRecordCompressionType.ZLIB,
                                    compression_level=9)
    with tf.io.TFRecordWriter('test2.tfrecord.zip', options=options) as writer:
    # Convert the chunk to a numpy array, and write each row to the file in a double for loop
        for chunk in pandabase:
            nump = chunk.to_numpy()
            for row in nump:
                ii = 0
                for elem in row:
                    features_map['feature' + str(ii)] = _float_feature(float(elem))
                    ii += 1
                myProto = tf.train.Example(features=tf.train.Features(feature=features_map))
                writer.write(myProto.SerializeToString())


start = datetime.now()
bk1 = pd.read_csv("Book2.csv", chunksize=2048, engine='c', iterator=True)    
serialize_example(bk1)
end = datetime.now()
print("- consumed time: %ds" % (end-start).seconds)

对于从 tfrecords 和使用 GPU 进行机器学习: 关注 this正确设置指南 然后使用此代码:

# Recreate the feature mappings (Must be similar to the one used to write the tfrecords)
_NUMCOL = 5
feature_description = {}
for i in range(_NUMCOL):
    feature_description.update({'feature' + str(i): tf.io.FixedLenFeature([], tf.float32)})

# Parse the tfrecords into the form (x, y) or (x, y, weights) to be used with keras
def _parse_function(example_proto):
    dic = tf.io.parse_single_example(example_proto, feature_description)
    y = dic['feature0']
    x = tf.stack([dic['feature1'],
                   dic['feature2'],
                   dic['feature3'],
                   dic['feature4']], axis=0)
    return x, y

# Let tensorflow autotune the training speed
AUTOTUNE = tf.data.experimental.AUTOTUNE
# creat a tfdataset from the recorded file, set parallel reads to number of cores for best running speed
myData = tf.data.TFRecordDataset('test.tfrecord.zip', compression_type='ZLIB',
                                 num_parallel_reads=2)
# Map the data to a form useable by keras (using _parse_function), cache the data, shuffle, and read the data in batches  
myData = myData.map(_parse_function, num_parallel_calls=AUTOTUNE)
myData = myData.cache()
myData = myData.shuffle(buffer_size=8192)
batches = 16385
myData = myData.batch(batches).prefetch(buffer_size=AUTOTUNE)

model = keras.Sequential([
    keras.layers.Dense(100, activation='softmax', kernel_regularizer=keras.regularizers.l2(lambd)),
    keras.layers.Dense(10, activation='softmax', kernel_regularizer=keras.regularizers.l2(lambd)),
    keras.layers.Dense(1, activation='linear', kernel_regularizer=keras.regularizers.l2(lambd))])

model.compile(optimizer='adam',
              loss='mean_squared_error')
model.save('keras.HD5F')

关于python - Tensorflow a2.0.0 : Converting CSV to a tfrecord, 创建一个 Keras 模型,该模型使用来自大型源的管道数据,将权重存储到 CSV 文件中?,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/55436719/

相关文章:

python - 将 OctetString 数据解包到变量中以进行进一步处理

python - 使用python在excel中转​​换dicom标签时出错

python - 使用 Python 请求获取 HEAD 内容

python - 如何设置 header (用户代理)、检索网页、捕获重定向和接受 cookie?

python - 如何比较两个列表列表之间的对应位置

python迭代循环遍历数据框的列

python - 如何加快多处理队列的同时读写速度?

machine-learning - 如何在没有绑定(bind)权值的情况下实现卷积连接?

machine-learning - Tensorflow 批处理参数

python - 如何在高维中执行tensorflow segment_max