Transformers Provably Learn Sparse Token Selection While Fully-Connected Nets Cannot

التفاصيل البيبلوغرافية
العنوان: Transformers Provably Learn Sparse Token Selection While Fully-Connected Nets Cannot
المؤلفون: Wang, Zixuan, Wei, Stanley, Hsu, Daniel, Lee, Jason D.
سنة النشر: 2024
المجموعة: Computer Science
Mathematics
Statistics
مصطلحات موضوعية: Statistics - Machine Learning, Computer Science - Information Theory, Computer Science - Machine Learning
الوصف: The transformer architecture has prevailed in various deep learning settings due to its exceptional capabilities to select and compose structural information. Motivated by these capabilities, Sanford et al. proposed the sparse token selection task, in which transformers excel while fully-connected networks (FCNs) fail in the worst case. Building upon that, we strengthen the FCN lower bound to an average-case setting and establish an algorithmic separation of transformers over FCNs. Specifically, a one-layer transformer trained with gradient descent provably learns the sparse token selection task and, surprisingly, exhibits strong out-of-distribution length generalization. We provide empirical simulations to justify our theoretical findings.
نوع الوثيقة: Working Paper
URL الوصول: http://arxiv.org/abs/2406.06893
رقم الأكسشن: edsarx.2406.06893
قاعدة البيانات: arXiv