Anneal LSTM Teacher Forcing steps

In LSTM-based sequence to sequence model Teacher Forcing is often used, i.e., using ground truth as the input to the decoder instead of using predicted value from last step during training. Is there an algorithm to anneal the teacher forcing steps? E.g. to reduce the number of teacher forcing steps gradually during training ?