Skip to main content

COMPEL Glossary / transformer-architecture

Transformer Architecture

The transformer is the neural network architecture that powers modern large language models and many other state-of-the-art AI systems.

What this means in practice

Introduced in a landmark 2017 paper ('Attention Is All You Need'), transformers use an 'attention mechanism' that processes relationships between all parts of the input simultaneously, enabling superior performance on sequence-based tasks like language understanding and generation. Transformers replaced earlier architectures (recurrent neural networks) because they train much faster on parallel hardware and scale more effectively. For transformation leaders, the transformer's significance is strategic: it is the technical foundation underlying the generative AI capabilities reshaping enterprise strategy, and understanding its capabilities and limitations helps evaluate vendor claims and assess technology risk.

Why it matters

The transformer architecture is the technical foundation underlying the generative AI capabilities reshaping enterprise strategy. Understanding its capabilities and limitations helps leaders evaluate vendor claims, assess technology risk, and make informed strategic decisions about AI investment. As transformer-based models become central to enterprise operations, leaders who understand this architecture make more effective governance and investment decisions.

How COMPEL uses it

Transformer architecture understanding is built into COMPEL's certification curriculum, with progressive technical literacy from Level 1 foundations through Level 3 architectural decision-making. During the Model stage, transformer capabilities inform use case feasibility assessment. The Technology pillar evaluates infrastructure requirements for transformer-based workloads, and the Evaluate stage assesses whether transformer-powered systems deliver expected performance.

Related Terms

Other glossary terms mentioned in this entry's definition and context.