LoginSignup
2
0

More than 5 years have passed since last update.

Gradient Clipping x LSTM ( Batch Normalization, Zoneout ) の実装に関するメモ

Last updated at Posted at 2018-05-30

ポイント

  • LSTMをベースに Gradient Clipping ( by norm ) を実装し、効果を検証。
  • 効果を確認できず。今後、追加検証。

レファレンス

1. Recurrent Batch Normalization
2. Zoneout: Regularizing RNNs by Randomly Preserving Hidden Activations
3. Zoneout に関するメモ

検証方法

  • Base (no regularization)、Base + Recurrent Batch Normalization、Base + Zoneout に適用し、効果を比較。          

データ

MNIST handwritten digits

from tensorflow.examples.tutorials.mnist import input_data
mnist = input_data.read_data_sets('***/mnist', \
                                     one_hot = True)

検証結果

数値計算例:

  • n_units = 100
  • learning_rate = 0.01
  • batch_size = 64
  • zoneout_prob = 0.2

Base ( no clipping )
image.png

Base ( clip_norm = 0.5 )
image.png

Base ( clip_norm = 1.0 )
image.png

Batch Normalization ( no clipping )
image.png

Batch Normalization ( clip_norm = 0.5 )
image.png

Batch Normalization ( clip_norm = 1.0 )
image.png

Zoneout ( no clipping )
image.png

Zoneout ( clip_norm = 0.5 )
image.png

Zoneout ( clip_norm = 1.0 )
image.png

サンプルコード

  def training(self, loss, learning_rate, clip_norm):
    optimizer = tf.train.AdamOptimizer(learning_rate = \
                  learning_rate)

    grads_and_vars = optimizer.compute_gradients(loss)
    clipped_grads_and_vars = [(tf.clip_by_norm(grad, \
               clip_norm = clip_norm), var) for grad, \
               var in grads_and_vars]
    train_step = \
      optimizer.apply_gradients(clipped_grads_and_vars)

    return train_step
2
0
0

Register as a new user and use Qiita more conveniently

  1. You get articles that match your needs
  2. You can efficiently read back useful information
  3. You can use dark theme
What you can do with signing up
2
0