Skip to content

Latest commit

 

History

History
161 lines (97 loc) · 4.09 KB

File metadata and controls

161 lines (97 loc) · 4.09 KB

Note

After we have established the basic objects and methods in TensorFlow, we now want to establish the components that make up TensorFlow algorithms. We start by introducing computational graphs, and then move to loss functions and back propagation. We end with creating a simple classifier and then show an example of evaluating regression and classification algorithms.

下载本章 Jupyter Notebook <https://github.com/qmlcode/qml/blob/master/qml/wrappers.py>

计算图

/02_TensorFlow_Way/01_Operations_as_a_Computational_Graph/index

We show how to create an operation on a computational graph and how to visualize it using Tensorboard.

下载本章 Jupyter Notebook </02_TensorFlow_Way/01_Operations_as_a_Computational_Graph/01_operations_on_a_graph.ipynb>


分层嵌套操作

/02_TensorFlow_Way/02_Layering_Nested_Operations/index

We show how to create multiple operations on a computational graph and how to visualize them using Tensorboard.

image

下载本章 Jupyter Notebook </02_TensorFlow_Way/02_Layering_Nested_Operations/02_layering_nested_operations.ipynb>


多层操作

/02_TensorFlow_Way/03_Working_with_Multiple_Layers/index

Here we extend the usage of the computational graph to create multiple layers and show how they appear in Tensorboard.

image

下载本章 Jupyter Notebook </02_TensorFlow_Way/03_Working_with_Multiple_Layers/03_multiple_layers.ipynb>


载入损失函数

/02_TensorFlow_Way/04_Implementing_Loss_Functions/index

In order to train a model, we must be able to evaluate how well it is doing. This is given by loss functions. We plot various loss functions and talk about the benefits and limitations of some.

下载本章 Jupyter Notebook </02_TensorFlow_Way/04_Implementing_Loss_Functions/04_loss_functions.ipynb>


载入反向传播

/02_TensorFlow_Way/05_Implementing_Back_Propagation/index

Here we show how to use loss functions to iterate through data and back propagate errors for regression and classification.

下载本章 Jupyter Notebook </02_TensorFlow_Way/05_Implementing_Back_Propagation/05_back_propagation.ipynb>


随机和批量训练

/02_TensorFlow_Way/06_Working_with_Batch_and_Stochastic_Training/index

TensorFlow makes it easy to use both batch and stochastic training. We show how to implement both and talk about the benefits and limitations of each.

image

下载本章 Jupyter Notebook </02_TensorFlow_Way/06_Working_with_Batch_and_Stochastic_Training/06_batch_stochastic_training.ipynb>


结合训练

/02_TensorFlow_Way/07_Combining_Everything_Together/index

We now combine everything together that we have learned and create a simple classifier.

下载本章 Jupyter Notebook </02_TensorFlow_Way/07_Combining_Everything_Together/07_combining_everything_together.ipynb>


模型评估

/02_TensorFlow_Way/08_Evaluating_Models/index

Any model is only as good as it's evaluation. Here we show two examples of (1) evaluating a regression algorithm and (2) a classification algorithm.

下载本章 Jupyter Notebook </02_TensorFlow_Way/08_Evaluating_Models/08_evaluating_models.ipynb>

本章学习模块

tensorflow.zeros

tensorflow.zeros


tensorflow.ones

tensorflow.ones