目的:线性拟合 y=0.1x+0.3 , 每20步训练,输出w, b。
import tensorflow as tf
import numpy as np# create data
x_data = np.random.rand(100).astype(np.float32)
y_data = x_data * 0.1 + 0.3# create model
Weights = tf.Variable(tf.random_uniform([1], -1.0, 1.0))
biases = tf.s([1]))
y = Weights * x_data + biases# cal loss
loss = tf.reduce_mean(tf.square(y-y_data))
optimizer = tf.train.GradientDescentOptimizer(0.5)
tain = optimizer.minimize(loss)# use model
init = tf.global_variable_initializer() #初始化之前定义的Variable
sess = tf.Session() #创建会话,用session执行init初始化步骤
sess.run(init)# train
for step in range(200):sess.run(train)if step % 20 == 0:print(step, sess.run(Weights), sess.run(biases))
功能:加载两个tensorflow,建立两个matrix,输出两个matrix相乘的结果。
import tensorflow as tf# create two matrixes
matrix1 = tf.constant([[3,3]])
matrix2 = tf.constant([[2],[2]])
product = tf.matmul(matrix1, matrix2)# method of open Session
with tf.Session() as sess:result = sess.run(product)print(result) #[[12]]
在Tensorflow中,变量必须定义是用tf.Variable说明。
import numpy as np
import matplotlib.pyplot as plt
import tensorflowpat.v1 as tf
tf.disable_v2_behavior()v1 = tf.Variable(0,name='age') #定义变量,值为0,名字为age
c1 = tf.constant(1) #定义常量
v2 = tf.add(v1,c1)
update = tf.assign(v1,v2)# 若定义了Variable就一定要initialize
init = tf.global_variables_initializer()# 使用Session启动
with tf.Session() as sess:sess.run(init)for _ in range(3):sess.run(update)print(sess.run(v1))
placeholder
是Tensorflow中的占位符,暂时存储变量。
Tensorflow如果想从外部传入data,需要用tf.placeholder()
,然后以这种形式传输数据sess.run(**, feed_dict={input: **})
.
import numpy as np
import matplotlib.pyplot as plt
import tensorflowpat.v1 as tf
tf.disable_v2_behavior()# 定义两个碗
input1 = tf.placeholder(tf.float32)
input2 = tf.placeholder(tf.float32)output = tf.multiply(input1,input2)with tf.Session() as sess:print(sess.run(output,feed_dict={input1:[3.],input2:[8.]})) # [24.]
import tensorflow as tf
import numpy as np
import matplotlib.pyplot as plt# fake data
x = np.linspace(-5, 5, 200) # x data, shape=(100, 1)# following are popular activation functions
y_relu = lu(x)
y_sigmoid = tf.nn.sigmoid(x)
y_tanh = tf.nn.tanh(x)
y_softplus = tf.nn.softplus(x)
# y_softmax = tf.nn.softmax(x) softmax is a special kind of activation function, it is about probabilitysess = tf.Session()
y_relu, y_sigmoid, y_tanh, y_softplus = sess.run([y_relu, y_sigmoid, y_tanh, y_softplus])# plt to visualize these activation function
plt.figure(1, figsize=(8, 6))
plt.subplot(221)
plt.plot(x, y_relu, c='red', label='relu')
plt.ylim((-1, 5))
plt.legend(loc='best')plt.subplot(222)
plt.plot(x, y_sigmoid, c='red', label='sigmoid')
plt.ylim((-0.2, 1.2))
plt.legend(loc='best')plt.subplot(223)
plt.plot(x, y_tanh, c='red', label='tanh')
plt.ylim((-1.2, 1.2))
plt.legend(loc='best')plt.subplot(224)
plt.plot(x, y_softplus, c='red', label='softplus')
plt.ylim((-0.2, 6))
plt.legend(loc='best')plt.show()
本文发布于:2024-02-03 03:50:50,感谢您对本站的认可!
本文链接:https://www.4u4v.net/it/170690345148460.html
版权声明:本站内容均来自互联网,仅供演示用,请勿用于商业和其他非法用途。如果侵犯了您的权益请与我们联系,我们将在24小时内删除。
留言与评论(共有 0 条评论) |