正解:B,C,D
A transformer model is the foundational architecture behind many modern natural language processing systems such as GPT and BERT. It processes text data through multiple key stages. According to the Microsoft Azure AI Fundamentals (AI-900) curriculum and Microsoft Learn materials, the major stages of a transformer-based large language model are tokenization, embedding calculation, and next token prediction.
* Tokenization (C) - The first step converts raw text into smaller units called tokens (words, subwords, or characters). This process allows the model to handle text in a structured numerical form rather than as raw language.
* Embedding Calculation (B) - After tokenization, the tokens are mapped into high-dimensional numeric vectors, known as embeddings. These embeddings capture semantic relationships between words and phrases so that the model can understand context and meaning.
* Next Token Prediction (D) - This stage is the heart of transformer operation, where the model predicts the next likely token in a sequence based on prior tokens. Repeated next-token predictions enable text generation, summarization, or translation.
Options A (object detection) and E (anonymization) are incorrect because they relate to vision and privacy workflows, not language modeling.