# OpenAI Announces New Research on Evaluating Code-Trained Language Models
OpenAI has published research focused on evaluating large language models that have been trained on code, signaling continued advancement in AI-powered programming tools.
The announcement, shared via OpenAI's official Twitter account, highlights the company's efforts to better understand and measure the capabilities of AI systems designed to work with programming languages. These models, which include systems like Codex that powers GitHub Copilot, have become increasingly important tools for software developers.
The research addresses a critical need in the AI development community: establishing reliable methods to assess how well these models actually perform coding tasks. As AI coding assistants become more prevalent in software development workflows, understanding their strengths, limitations, and potential errors becomes essential for both developers and organizations relying on these tools.
This evaluation work matters because it helps ensure AI coding tools are safe, reliable, and effective. Proper assessment methods