There’s infinite data around us, so attention mechanisms are becoming ever more important. One of the foundational breakthroughs that led to the current LLM technology, the Transformer architecture is a practical implementation of an attention mechanism:
Irrespective of the title, as a society we need to become more conscious of the attention mechanisms. For example, the fixation on gender and race: what does it do?