--- dataset_info: config_name: default splits: - name: test num_examples: 124 --- # SWE-Atlas QnA Codebase QnA is the first benchmark in the SWE-Atlas suite. It evaluates AI agents on deep code comprehension — tracing execution paths, explaining architectural decisions, and answering deeply technical questions about production-grade software systems. 124 tasks across 11 open-source repositories spanning Go, Python, C, and TypeScript. Link to leaderboard - [https://scale.com/leaderboard/sweatlas-qna](https://scale.com/leaderboard/sweatlas-qna) ## Schema | Column | Type | Description | |---|---|---| | `task_id` | string | Unique 24-char hex identifier | | `prompt` | string | The question presented to the agent | | `reference_answer` | string | Expert-written reference answer | | `repository_url` | string | GitHub repo | | `repository_base_commit` | string | 40-char commit SHA the environment is pinned to | | `language` | string | `go`, `python`, `c`, or `ts` | | `category` | string | Task category (see below) | | `rubric` | string (JSON) | Evaluation criteria (see below) | | `docker_image` | string | Docker Hub image for the sandboxed environment | ### Rubric format Each task's `rubric` field is a JSON array: ```json [ { "id": "a33fc01cba19849aaf3b55e6b801001c", "title": "1.1: States that kitty uses Unix sockets for external connections...", "annotations": { "type": "positive hli verifier", "importance": "must have" } } ] ``` - `positive hli verifier` — a factual claim the answer must contain. If the claim is met my the agent's answer, the rubric item result is a PASS. - `negative hli verifier` — something the answer must *not* claim. If the claim is met my the agent's answer, the rubric item result is a FAIL. Each task includes a `docker_image` field pointing to a pre-built Docker Hub image with the repository and all dependencies installed at `/app`: ## Inference and Eval We follow the standard SWE-Agent scaffold, and we provide a sample config (with the prompts) in [default_qa_config.yaml](default_qa_config.yaml) To run tasks, you can pull the docker image and run the container, and reset the environment to the base commit: ```bash cd /app git config --global --add safe.directory /app git restore . git reset --hard git clean -fdq ``` Evaluation is performed by an LLM judge (Claude Opus 4.5) that scores the agent's answer against each rubric criterion independently. Each criterion receives a binary score (met or not met) indicating and is then aggregated. The primary metric is the Task Resolve Rate: the percentage of tasks for which the agent's answer is comprehensive (i.e. passes all rubric items and scores 1.0), as graded by a set of task-specific rubrics. The agents are also instructed to avoid modifying source-code files, and clean up any temporary scripts made. So we add a programmatic check that fails a task that has any code changes after submission. Our rubric evaluation prompt and other relevant details are in [rubric_evaluation_config.yaml](rubric_evaluation_config.yaml)