FlashAttention and GPGPU Acceleration in Transformers ⚡