LLM Architecture Evolution from 2023 to 2026: Tokenizers, Positional Encoding, Attention, MoE, Normalization, and Activation Functions