Transformers and Attention-Based Deep Networks


Temizel A., Taşkaya Temizel T.

  • Dersin Düzeyi: Yüksek Lisans
  • Tasarlanan Ders Kodu: DI725
  • Öğretim Türü: Örgün Öğretim (Normal Öğretim)
  • Dersin Kapsamı: Teorik ve Uygulama
  • Akademik Yıl: 2023 - 2024
  • Ders İçeriği:

    This course explores advanced concepts and applications of transformers and attention-based models in various domains, focusing particularly on natural language processing (NLP), time series and computer vision as well as unified vision and language understanding. It covers topics such as attention, vanilla transformer, large language models (LLM), LLM frameworks, NLP applications with LLM, Unified Vision-Language Understanding and Multimodal Transformers, Distillation and data-efficient transformers, explainability, flash attention, in-context learning, prompting, and ethical concerns. The course aims to give both theoretical and practical aspects of the topics and present real-world use cases.