Long Short-term memory is a kind of recurrent neural networks. It has additional gates and its' architecture is bellow:
This architecture can avoid the gradient vanishing and keep the long-term memory, so this is a very powerful neural network.
We can get these formulations: