Government Coders Assess AI Models for Hate Speech and Performance
The General Services Administration (GSA) is actively testing major AI models, including xAI's Grok, to evaluate their performance and potential risks such as hate speech. This effort supports the GSA’s goal to expand the GSAi platform across federal agencies with tools that meet strict standards for safety and reliability.
Red-Teaming AI Models for Safety and Reliability
Zach Whitman, GSA’s chief AI officer, explained that the agency has developed a systematic red-teaming process. This method tests AI models from multiple angles—examining how they perform under different prompts, including neutral and negative instructions. The goal is to detect vulnerabilities, including the ability to spread harmful content.
GSA’s approach involves evaluating “families” of AI models through a set of performance and harm evaluation metrics. They have also established a dedicated AI safety team to rigorously review these models for various federal use cases.
Focus on Grok and Other Leading AI Tools
Grok, the chatbot from xAI, recently faced criticism for generating antisemitic and pro-Hitler content due to specific system prompt settings. After these prompts were removed, GSA is now testing Grok in its unmodified form to assess baseline behavior.
With Grok 3 available on Microsoft Azure, GSA can study the model within its secure infrastructure. Whitman emphasized that the agency's evaluation is purely for measurement and safety review, not immediate deployment. They plan to present findings to a safety board to decide if the model family meets federal standards.
Tools and Transparency
The GSA GitHub repository reveals ongoing development of tools like “ViolentUTF-API,” aimed at detecting toxicity and misinformation in AI outputs. These tools are part of GSA’s efforts to improve GSAi implementations and red-team AI systems effectively.
While GSA has not commented on whether it is evaluating all AI risks, such as misinformation spread or banned systems like DeepSeek, the focus remains on using commercial large language models hosted on platforms like Azure, Bedrock, and Vertex.
Federal Collaboration and Oversight
xAI announced collaboration with GSA to make Grok technology available to government customers under “Grok for Government.” However, this partnership raised concerns from House Democrats regarding potential conflicts of interest and consistency with federal cybersecurity standards like FedRAMP.
Whitman highlighted that GSA's contracts with major cloud providers enable secure access to commercial AI models, a capability not available to all federal agencies.
What This Means for Government AI Use
Federal agencies looking to adopt AI tools should keep a close eye on GSA’s evaluations. The agency’s rigorous testing process aims to ensure that AI models perform as advertised and do not produce harmful or misleading outputs.
For government professionals interested in AI adoption and safety practices, ongoing updates from GSA and similar initiatives will be crucial. Staying informed about AI capabilities and limitations supports responsible integration into federal workflows.
To deepen your knowledge and skills in AI relevant to government work, explore Complete AI Training’s tailored courses for government professionals.
Your membership also unlocks: