I’ll talk about two independent works on classical and quantum neural networks connected by information theory. In the first part of the talk, I’ll treat sequence models as one-dimensional classical statistical mechanical systems and analyze the scaling behavior of mutual information. I'll provide a new perspective on why recurrent neural networks are not good at natural language processing. In the second part of the talk, I’ll study information scrambling dynamics when quantum neural networks are trained by classical gradient descent algorithm. For many problems, this hybrid quantum-classical training process consists of two stages where information scrambles very differently in the network.
- Condensed Matter
- Scientific Series