1. 程式人生 > >tensorflow下已經初始化,但還是存在 Attempting to use uninitialized value的解決方案

tensorflow下已經初始化,但還是存在 Attempting to use uninitialized value的解決方案

執行tensorflow後,程式報錯:

FailedPreconditionError (see above for traceback): Attempting to use uninitialized value beta1_power
     [[Node: beta1_power/read = Identity[T=DT_FLOAT, _class=["loc:@Adam/Assign_1"], _device="/job:localhost/replica:0/task:0/device:GPU:0"](beta1_power)]

程式如下,已經修改正確:​​​​​​​

  1. #機器學習框架搭建
    import tensorflow as tf
    
    W = tf.Variable(tf.zeros([2, 1]), name="weights")
    b = tf.Variable(0., name="bias")
    
    
    def inference(X):
        return tf.matmul(X, W) + b
    
    
    def loss(X, Y):
        Y_predicted = inference(X)
        return tf.reduce_sum(tf.squared_difference(Y, Y_predicted))
    
    
    def inputs():
        weight_age = [[84, 46], [73, 20], [65, 52], [70, 30], [76, 57], [69, 25], [63, 28], [72, 36], [79, 57], [75, 44],
                      [27, 24], [89, 31], [65, 52], [57, 23], [59, 60], [69, 48], [60, 34], [79, 51], [75, 50], [82, 34],
                      [59, 46], [67, 23], [85, 37], [55, 40], [63, 30]]
        blood_fat_content = [354, 190, 405, 263, 451, 302, 288, 385, 402, 365, 209, 290, 346, 254, 395, 434, 220, 374, 308,
                             220, 311, 181, 274, 303, 244]
        return tf.to_float(weight_age), tf.to_float(blood_fat_content)
    
    
    def train(total_loss):
        learning_rate = 0.0000001
        #return tf.train.AdamOptimizer(0.0000001).minimize(total_loss)
        return tf.train.GradientDescentOptimizer(learning_rate).minimize(total_loss)
    
    
    def evaluate(sess, X, Y):
        print(sess.run(inference([[80., 25.]])))
        print(sess.run(inference([[65., 25.]])))
    
    
    with tf.Session() as sess:
        X, Y = inputs()
        
        #init = tf.global_variables_initializer()
        total_loss = loss(X, Y)
        train_op = train(total_loss)
    
        coord = tf.train.Coordinator()
        threads = tf.train.start_queue_runners(sess=sess, coord=coord)
        init = tf.global_variables_initializer()#初始化所在的位置至關重要,以本程式為例,使用adam優化器時,會主動建立變數。
        #因此,如果這時的初始化位置在建立adam優化器之前,則adam中包含的變數會未初始化,然後報錯。本行初始化時,可以看到Adam
        #已經宣告,古不會出錯
        sess.run(init)
        training_steps = 10000
        saver = tf.train.Saver()  #模型儲存和恢復,當把改行放入for迴圈中以後,會發現程式執行速度明顯變慢
        for step in range(training_steps):
            #saver = tf.train.Saver()
            sess.run(train_op)
            if step % 10 == 0:
                print("loss", sess.run(total_loss))
    
            if step % 1000 == 0:
                saver.save(sess, r"E:\tf_project\練習\model_save_dir\my-model", global_step=step)
    
        evaluate(sess, X, Y)
        saver.save(sess, r"E:\tf_project\練習\model_save_dir\my-model", global_step=training_steps)
    
        coord.request_stop()
        coord.join(threads)
        sess.close()
  2. 在寫程式初始化的時候一定要注意初始化的位置,並不是說自己的計算圖搭建完成就要初始化,因為其中的元件可能本身也要申請變數,比如adam,這時最好是在程式執行前面新增變數的初始化,因為可以幫助你把後來模組中宣告的變數同時進行初始化。可以參考上面程式,初始化位置。