Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun of Microsoft Research introduced deep residual learning in their 2015 paper, enabling training of networks with over 150 layers. The architecture, known as ResNet, won first place in the ImageNet Large Scale Visual Recognition Challenge. Subsequent work by Ashish Vaswani, Noam Shazeer, and colleagues at Google Brain on the Transformer architecture fundamentally changed natural language processing. Their paper "Attention Is All You Need" has been cited over 100,000 times according to Google Scholar.