您好,登錄后才能下訂單哦!
這期內容當中小編將會給大家帶來有關使用tensorflow怎么實現邏輯回歸模型,文章內容豐富且以專業的角度為大家分析和敘述,閱讀完這篇文章希望大家可以有所收獲。
邏輯回歸模型
邏輯回歸是應用非常廣泛的一個分類機器學習算法,它將數據擬合到一個logit函數(或者叫做logistic函數)中,從而能夠完成對事件發生的概率進行預測。
import numpy as np import tensorflow as tf import matplotlib.pyplot as plt from tensorflow.examples.tutorials.mnist import input_data #下載好的mnist數據集存在F:/mnist/data/中 mnist = input_data.read_data_sets('F:/mnist/data/',one_hot = True) print(mnist.train.num_examples) print(mnist.test.num_examples) trainimg = mnist.train.images trainlabel = mnist.train.labels testimg = mnist.test.images testlabel = mnist.test.labels print(type(trainimg)) print(trainimg.shape,) print(trainlabel.shape,) print(testimg.shape,) print(testlabel.shape,) nsample = 5 randidx = np.random.randint(trainimg.shape[0],size = nsample) for i in randidx: curr_img = np.reshape(trainimg[i,:],(28,28)) curr_label = np.argmax(trainlabel[i,:]) plt.matshow(curr_img,cmap=plt.get_cmap('gray')) plt.title(""+str(i)+"th Training Data"+"label is"+str(curr_label)) print(""+str(i)+"th Training Data"+"label is"+str(curr_label)) plt.show() x = tf.placeholder("float",[None,784]) y = tf.placeholder("float",[None,10]) W = tf.Variable(tf.zeros([784,10])) b = tf.Variable(tf.zeros([10])) # actv = tf.nn.softmax(tf.matmul(x,W)+b) #計算損失 cost = tf.reduce_mean(-tf.reduce_sum(y*tf.log(actv),reduction_indices=1)) #學習率 learning_rate = 0.01 #隨機梯度下降 optm = tf.train.GradientDescentOptimizer(learning_rate).minimize(cost) #求1位置索引值 對比預測值索引與label索引是否一樣,一樣返回True pred = tf.equal(tf.argmax(actv,1),tf.argmax(y,1)) #tf.cast把True和false轉換為float類型 0,1 #把所有預測結果加在一起求精度 accr = tf.reduce_mean(tf.cast(pred,"float")) init = tf.global_variables_initializer() """ #測試代碼 sess = tf.InteractiveSession() arr = np.array([[31,23,4,24,27,34],[18,3,25,4,5,6],[4,3,2,1,5,67]]) #返回數組的維數 2 print(tf.rank(arr).eval()) #返回數組的行列數 [3 6] print(tf.shape(arr).eval()) #返回數組中每一列中最大元素的索引[0 0 1 0 0 2] print(tf.argmax(arr,0).eval()) #返回數組中每一行中最大元素的索引[5 2 5] print(tf.argmax(arr,1).eval()) J""" #把所有樣本迭代50次 training_epochs = 50 #每次迭代選擇多少樣本 batch_size = 100 display_step = 5 sess = tf.Session() sess.run(init) #循環迭代 for epoch in range(training_epochs): avg_cost = 0 num_batch = int(mnist.train.num_examples/batch_size) for i in range(num_batch): batch_xs,batch_ys = mnist.train.next_batch(batch_size) sess.run(optm,feed_dict = {x:batch_xs,y:batch_ys}) feeds = {x:batch_xs,y:batch_ys} avg_cost += sess.run(cost,feed_dict = feeds)/num_batch if epoch % display_step ==0: feeds_train = {x:batch_xs,y:batch_ys} feeds_test = {x:mnist.test.images,y:mnist.test.labels} train_acc = sess.run(accr,feed_dict = feeds_train) test_acc = sess.run(accr,feed_dict = feeds_test) #每五個epoch打印一次信息 print("Epoch:%03d/%03d cost:%.9f train_acc:%.3f test_acc: %.3f" %(epoch,training_epochs,avg_cost,train_acc,test_acc)) print("Done")
程序訓練結果如下:
Epoch:000/050 cost:1.177228655 train_acc:0.800 test_acc: 0.855 Epoch:005/050 cost:0.440933891 train_acc:0.890 test_acc: 0.894 Epoch:010/050 cost:0.383387268 train_acc:0.930 test_acc: 0.905 Epoch:015/050 cost:0.357281335 train_acc:0.930 test_acc: 0.909 Epoch:020/050 cost:0.341473956 train_acc:0.890 test_acc: 0.913 Epoch:025/050 cost:0.330586549 train_acc:0.920 test_acc: 0.915 Epoch:030/050 cost:0.322370980 train_acc:0.870 test_acc: 0.916 Epoch:035/050 cost:0.315942993 train_acc:0.940 test_acc: 0.916 Epoch:040/050 cost:0.310728854 train_acc:0.890 test_acc: 0.917 Epoch:045/050 cost:0.306357428 train_acc:0.870 test_acc: 0.918 Done
上述就是小編為大家分享的使用tensorflow怎么實現邏輯回歸模型了,如果剛好有類似的疑惑,不妨參照上述分析進行理解。如果想知道更多相關知識,歡迎關注億速云行業資訊頻道。
免責聲明:本站發布的內容(圖片、視頻和文字)以原創、轉載和分享為主,文章觀點不代表本網站立場,如果涉及侵權請聯系站長郵箱:is@yisu.com進行舉報,并提供相關證據,一經查實,將立刻刪除涉嫌侵權內容。