
My lecture talk. In this lecture, I introduce soft-attention mechanism for language modeling, and visual question answering. I also show my most recent work on hard-attention, where a subset of inputs is selected rather than a weighted mask over the whole image. Finally, I show basics of self-attention mechanism that is commonly used, e.g., in Transformer.