To mitigate toxicity, developers employ alignment techniques, fine-tuning LLMs to discourage harmful outputs. While alignment ...