| """ |
| Test runner for the Idea First, Code Later CP Benchmark dataset. |
| |
| This module provides tools to load competitive programming problems from the |
| Hugging Face dataset and run solutions against their test cases. |
| |
| Features: |
| - Automatic dataset loading from Hugging Face |
| - Support for C++ and Python solutions |
| - Special judge support for problems with multiple valid outputs |
| - Kattis integration for CS3233 problems |
| - Memory and time limit enforcement |
| |
| Example: |
| >>> from hf_test_runner import TestRunner |
| >>> runner = TestRunner() |
| >>> results = runner.run_solution( |
| ... problem_id="icpc-jakarta-2019_card-collecting", |
| ... solution_code=open("solution.cpp").read(), |
| ... language="cpp" |
| ... ) |
| >>> print(results["status"]) # "PASSED" or error type |
| |
| Classes: |
| TestRunner: Main class for running solutions against problems. |
| Problem: Dataclass representing a problem from the dataset. |
| |
| Functions: |
| run_solution: Convenience function for quick one-off solution testing. |
| parse_time_limit: Parse time limit strings like "2 sec" to float. |
| parse_memory_limit_to_bytes: Parse memory limit strings to bytes. |
| """ |
|
|
| import os |
| import re |
| import subprocess |
| import tempfile |
| import time |
| import signal |
| import resource |
| import traceback |
| from dataclasses import dataclass |
| from typing import Optional |
|
|
| try: |
| from datasets import load_dataset |
| except ImportError: |
| print("Please install datasets: pip install datasets") |
| raise |
|
|
|
|
| |
| PREEXEC_ERROR_EXIT_CODE = 2 |
| HF_DATASET_ID = "samahadhoud/idea-first-code-later-cp-benchmark" |
|
|
|
|
| @dataclass |
| class Problem: |
| """ |
| Represents a competitive programming problem loaded from the HF dataset. |
| |
| Attributes: |
| problem_id: Unique identifier in format "contest_slug" (e.g., "icpc-jakarta-2019_card-collecting") |
| problem_code: Problem letter in the contest (A, B, C, ...) |
| problem_slug: URL-friendly problem name |
| problem_title: Full human-readable problem title |
| |
| contest_name: Contest identifier (e.g., "icpc-jakarta-2019") |
| contest_full_name: Full contest name (e.g., "ICPC Asia Jakarta Regional Contest 2019") |
| year: Competition year as string |
| source: Source URL or repository |
| |
| time_limit: Time limit as string (e.g., "2 sec") |
| memory_limit: Memory limit as string (e.g., "256 MB") |
| |
| statement: Problem statement in Markdown format |
| analysis: Editorial/solution analysis in Markdown |
| |
| sample_test_cases_input: List of sample input strings (shown in problem statement) |
| sample_test_cases_output: List of corresponding sample outputs |
| hidden_test_cases_input: List of hidden input strings (for actual judging) |
| hidden_test_cases_output: List of corresponding hidden outputs |
| |
| has_special_judge: True if problem accepts multiple valid outputs |
| special_judge_code: C++ scorer code using testlib for validation |
| special_judge_format: "standard" or "jakarta2017" (different CLI formats) |
| |
| uses_kattis: True for CS3233 problems (must submit via Kattis) |
| kattis_problem_id: Kattis problem ID for submission |
| """ |
| problem_id: str |
| problem_code: str |
| problem_slug: str |
| problem_title: str |
| |
| contest_name: str |
| contest_full_name: str |
| year: str |
| source: str |
| |
| time_limit: str |
| memory_limit: str |
| |
| statement: str |
| analysis: str |
| |
| sample_test_cases_input: list |
| sample_test_cases_output: list |
| hidden_test_cases_input: list |
| hidden_test_cases_output: list |
| |
| has_special_judge: bool |
| special_judge_code: str |
| special_judge_format: str |
| |
| uses_kattis: bool |
| kattis_problem_id: str |
|
|
|
|
| def parse_time_limit(time_str: str) -> float: |
| """ |
| Parse a time limit string to seconds. |
| |
| Args: |
| time_str: Time limit string (e.g., "2 sec", "1.5s", "3") |
| |
| Returns: |
| Time limit in seconds as float. Returns 10.0 if parsing fails. |
| |
| Examples: |
| >>> parse_time_limit("2 sec") |
| 2.0 |
| >>> parse_time_limit("1.5s") |
| 1.5 |
| >>> parse_time_limit("") |
| 10.0 |
| """ |
| if not time_str: |
| return 10.0 |
| match = re.search(r'([\d.]+)\s*(?:sec|s)?', time_str.lower()) |
| if match: |
| return float(match.group(1)) |
| return 10.0 |
|
|
|
|
| def parse_memory_limit_to_bytes(value) -> Optional[int]: |
| """ |
| Parse a memory limit value to bytes. |
| |
| Args: |
| value: Memory limit as int (MB), float (MB), or string (e.g., "256 MB", "1 GB") |
| |
| Returns: |
| Memory limit in bytes, or None if unlimited/unspecified. |
| |
| Raises: |
| ValueError: If the value cannot be parsed or is negative. |
| |
| Examples: |
| >>> parse_memory_limit_to_bytes(256) |
| 268435456 |
| >>> parse_memory_limit_to_bytes("256 MB") |
| 268435456 |
| >>> parse_memory_limit_to_bytes("1 GB") |
| 1073741824 |
| >>> parse_memory_limit_to_bytes(None) |
| None |
| """ |
| if value is None: |
| return None |
| if isinstance(value, (int, float)): |
| if value < 0: |
| raise ValueError("memory limit must be non-negative") |
| return int(value * 1024 * 1024) |
|
|
| if isinstance(value, str): |
| s = value.strip().lower() |
| if s in ("none", "unlimited", "0", ""): |
| return None |
| m = re.match(r"^\s*([0-9]*\.?[0-9]+)\s*([kmgtp]?b?|)$", s) |
| if not m: |
| raise ValueError(f"cannot parse memory limit '{value}'") |
| num = float(m.group(1)) |
| unit = m.group(2) |
| if unit in ("", "b"): |
| multiplier = 1024 * 1024 |
| elif unit in ("k", "kb"): |
| multiplier = 1024 |
| elif unit in ("m", "mb"): |
| multiplier = 1024 * 1024 |
| elif unit in ("g", "gb"): |
| multiplier = 1024 * 1024 * 1024 |
| else: |
| raise ValueError(f"unknown unit in memory limit '{value}'") |
| return int(num * multiplier) |
|
|
| raise ValueError("unsupported type for memory limit") |
|
|
|
|
| def _preexec_setrlimit_bytes(bytes_limit): |
| """ |
| Set memory limit in child process using setrlimit. |
| |
| This function is designed to be used as a preexec_fn in subprocess.run(). |
| It sets RLIMIT_AS (address space limit) to enforce memory limits on |
| the child process. |
| |
| Args: |
| bytes_limit: Memory limit in bytes, or None for no limit. |
| |
| Note: |
| If this function fails, it writes the traceback to stderr and |
| exits with PREEXEC_ERROR_EXIT_CODE (2) to signal the error. |
| """ |
| try: |
| if bytes_limit is None: |
| return |
| if not isinstance(bytes_limit, int) or bytes_limit < 0: |
| raise ValueError("bytes_limit must be non-negative int") |
| resource.setrlimit(resource.RLIMIT_AS, (bytes_limit, bytes_limit)) |
| except Exception: |
| try: |
| tb = traceback.format_exc() |
| os.write(2, tb.encode("utf-8", errors="replace")) |
| except Exception: |
| pass |
| os._exit(PREEXEC_ERROR_EXIT_CODE) |
|
|
|
|
| class TestRunner: |
| """ |
| Test runner for the Idea First, Code Later CP Benchmark dataset. |
| |
| This class loads problems from Hugging Face and provides methods to run |
| solutions against test cases with proper judging, time/memory limits, |
| and special judge support. |
| |
| Attributes: |
| dataset: The loaded HuggingFace dataset |
| problems: Dict mapping problem_id to problem data |
| |
| Example: |
| >>> runner = TestRunner() |
| >>> results = runner.run_solution( |
| ... problem_id="icpc-jakarta-2019_card-collecting", |
| ... solution_code=code, |
| ... language="cpp" |
| ... ) |
| >>> if results["status"] == "PASSED": |
| ... print("All tests passed!") |
| """ |
| |
| def __init__(self, dataset_id: str = HF_DATASET_ID, cache_dir: Optional[str] = None): |
| """ |
| Initialize the test runner and load the dataset from Hugging Face. |
| |
| Args: |
| dataset_id: HuggingFace dataset ID (default: samahadhoud/idea-first-code-later-cp) |
| cache_dir: Optional local cache directory for the dataset |
| |
| Raises: |
| ImportError: If the datasets library is not installed |
| """ |
| print(f"Loading dataset from {dataset_id}...") |
| self.dataset = load_dataset(dataset_id, cache_dir=cache_dir)["train"] |
| self.problems = {row["problem_id"]: row for row in self.dataset} |
| print(f"Loaded {len(self.problems)} problems") |
| |
| def get_problem(self, problem_id: str) -> Problem: |
| """ |
| Retrieve a problem by its ID. |
| |
| Args: |
| problem_id: Unique problem identifier (e.g., "icpc-jakarta-2019_card-collecting") |
| |
| Returns: |
| Problem object with all problem data. |
| |
| Raises: |
| ValueError: If the problem_id is not found in the dataset. |
| |
| Example: |
| >>> problem = runner.get_problem("icpc-jakarta-2019_card-collecting") |
| >>> print(problem.problem_title) |
| 'Card Collecting' |
| """ |
| if problem_id not in self.problems: |
| raise ValueError(f"Problem {problem_id} not found. Available: {list(self.problems.keys())[:5]}...") |
| |
| row = self.problems[problem_id] |
| return Problem( |
| problem_id=row["problem_id"], |
| problem_code=row["problem_code"], |
| problem_slug=row["problem_slug"], |
| problem_title=row["problem_title"], |
| contest_name=row["contest_name"], |
| contest_full_name=row["contest_full_name"], |
| year=row["year"], |
| source=row["source"], |
| time_limit=row["time_limit"], |
| memory_limit=row["memory_limit"], |
| statement=row["statement"], |
| analysis=row["analysis"], |
| sample_test_cases_input=row["sample_test_cases_input"], |
| sample_test_cases_output=row["sample_test_cases_output"], |
| hidden_test_cases_input=row["hidden_test_cases_input"], |
| hidden_test_cases_output=row["hidden_test_cases_output"], |
| has_special_judge=row["has_special_judge"], |
| special_judge_code=row["special_judge_code"], |
| special_judge_format=row["special_judge_format"], |
| uses_kattis=row["uses_kattis"], |
| kattis_problem_id=row["kattis_problem_id"], |
| ) |
| |
| def list_problems(self, contest: Optional[str] = None) -> list: |
| """ |
| List all available problem IDs, optionally filtered by contest. |
| |
| Args: |
| contest: Optional substring to filter by (e.g., "jakarta-2019") |
| |
| Returns: |
| List of problem ID strings. |
| |
| Example: |
| >>> runner.list_problems()[:3] |
| ['cs3233-2023_a', 'cs3233-2023_b', 'cs3233-2023_c'] |
| >>> runner.list_problems("jakarta-2017") |
| ['icpc-jakarta-2017_...'] |
| """ |
| if contest: |
| return [pid for pid in self.problems if contest in pid] |
| return list(self.problems.keys()) |
| |
| def run_solution( |
| self, |
| problem_id: str, |
| solution_code: str, |
| language: str = "cpp", |
| run_hidden: bool = True, |
| solution_file: Optional[str] = None, |
| ) -> dict: |
| """ |
| Run a solution against a problem's test cases. |
| |
| Executes the solution code against sample tests first, then hidden tests |
| if sample tests pass. Handles compilation, time/memory limits, and |
| special judges automatically. |
| |
| Args: |
| problem_id: The problem ID from the dataset |
| solution_code: The solution source code as a string |
| language: Programming language - "cpp" (default) or "Python" |
| run_hidden: Whether to run hidden test cases after samples pass |
| solution_file: Optional path to solution file (used for Kattis submission) |
| |
| Returns: |
| dict with keys: |
| - status: "PASSED", "Wrong Answer", "Time Limit Exceeded", |
| "Memory Limit Exceeded", "Runtime Error", or "Compile Error" |
| - problem_id: The problem ID |
| - problem_title: Human-readable problem title |
| - test_cases: List of individual test results |
| - sample_summary: Dict with pass/fail counts for sample tests |
| - hidden_summary: Dict with pass/fail counts for hidden tests |
| |
| Example: |
| >>> results = runner.run_solution( |
| ... "icpc-jakarta-2019_card-collecting", |
| ... open("solution.cpp").read(), |
| ... language="cpp" |
| ... ) |
| >>> print(results["status"]) |
| 'PASSED' |
| """ |
| problem = self.get_problem(problem_id) |
| |
| |
| if problem.uses_kattis: |
| return self._run_kattis(problem, solution_code, solution_file, language) |
| |
| |
| scorer_executable = None |
| if problem.has_special_judge and problem.special_judge_code: |
| scorer_executable = self._compile_scorer(problem) |
| |
| results = { |
| "problem_id": problem_id, |
| "problem_title": problem.problem_title, |
| "status": "PASSED", |
| "test_cases": [], |
| "sample_summary": {}, |
| "hidden_summary": {}, |
| } |
| |
| time_limit = parse_time_limit(problem.time_limit) |
| memory_limit = problem.memory_limit |
| |
| |
| sample_passed = self._run_tests( |
| problem, solution_code, language, |
| problem.sample_test_cases_input, |
| problem.sample_test_cases_output, |
| "sample", results, time_limit, memory_limit, |
| scorer_executable |
| ) |
| |
| if not sample_passed: |
| return results |
| |
| |
| if run_hidden and problem.hidden_test_cases_input: |
| self._run_tests( |
| problem, solution_code, language, |
| problem.hidden_test_cases_input, |
| problem.hidden_test_cases_output, |
| "hidden", results, time_limit, memory_limit, |
| scorer_executable |
| ) |
| |
| return results |
| |
| def _run_kattis( |
| self, |
| problem: Problem, |
| solution_code: str, |
| solution_file: Optional[str], |
| language: str |
| ) -> dict: |
| """ |
| Submit a solution to Kattis for CS3233 problems. |
| |
| CS3233 problems are judged on Kattis and don't have local hidden test cases. |
| This method submits the solution via the Kattis CLI and parses the result. |
| |
| Args: |
| problem: The Problem object |
| solution_code: Solution source code |
| solution_file: Optional path to existing solution file |
| language: "cpp" or "Python" |
| |
| Returns: |
| dict with status, kattis_output, and execution time |
| |
| Note: |
| Requires kattis-cli to be installed and configured: |
| https://github.com/Kattis/kattis-cli |
| """ |
| if not solution_file: |
| |
| ext = ".py" if language == "Python" else ".cpp" |
| with tempfile.NamedTemporaryFile(mode="w", suffix=ext, delete=False) as f: |
| f.write(solution_code) |
| solution_file = f.name |
| |
| print(f"Submitting to Kattis: {problem.kattis_problem_id}") |
| |
| try: |
| proc = subprocess.run( |
| ["kattis", solution_file, "-p", problem.kattis_problem_id, "-f"], |
| capture_output=True, text=True, timeout=300 |
| ) |
| |
| stdout = proc.stdout |
| stderr = proc.stderr |
| |
| print(f"📤 Kattis output:\n{stdout}") |
| |
| |
| verdict_match = re.search( |
| r'(Accepted|Wrong Answer|Time Limit Exceeded|Run Time Error|Compile Error|Memory Limit Exceeded)', |
| stdout |
| ) |
| time_match = re.search(r'\(([\d.]+)s\)', stdout) |
| |
| raw_verdict = verdict_match.group(1) if verdict_match else "Unknown" |
| |
| verdict_map = { |
| "Accepted": "PASSED", |
| "Wrong Answer": "Wrong Answer", |
| "Time Limit Exceeded": "Time Limit Exceeded", |
| "Run Time Error": "Runtime Error", |
| "Compile Error": "Compile Error", |
| "Memory Limit Exceeded": "Memory Limit Exceeded", |
| "Unknown": "Runtime Error" |
| } |
| |
| return { |
| "problem_id": problem.problem_id, |
| "problem_title": problem.problem_title, |
| "status": verdict_map.get(raw_verdict, "Runtime Error"), |
| "time": float(time_match.group(1)) if time_match else None, |
| "kattis_output": stdout, |
| "kattis_stderr": stderr, |
| } |
| |
| except Exception as e: |
| return { |
| "problem_id": problem.problem_id, |
| "status": "Submission Failed", |
| "error": str(e) |
| } |
| |
| def _compile_scorer(self, problem: Problem) -> Optional[str]: |
| """ |
| Compile the special judge scorer for problems with multiple valid outputs. |
| |
| Some problems accept multiple correct answers (e.g., "print any valid permutation"). |
| These use a custom scorer (written in C++ using testlib) to validate outputs. |
| |
| Args: |
| problem: The Problem object containing special_judge_code |
| |
| Returns: |
| Path to the compiled scorer executable, or None if compilation fails. |
| |
| Note: |
| The scorer is compiled with g++ -std=c++17. The executable is stored |
| in a temporary directory that persists for the duration of testing. |
| """ |
| |
| tmp_dir = tempfile.mkdtemp(prefix="scorer_") |
| scorer_src = os.path.join(tmp_dir, "scorer.cpp") |
| scorer_exe = os.path.join(tmp_dir, "scorer") |
| |
| with open(scorer_src, "w") as f: |
| f.write(problem.special_judge_code) |
| |
| result = subprocess.run( |
| ["g++", "-o", scorer_exe, scorer_src, "-std=c++17"], |
| capture_output=True, text=True |
| ) |
| |
| if result.returncode != 0: |
| print(f"⚠️ Scorer compilation failed: {result.stderr}") |
| return None |
| |
| os.chmod(scorer_exe, 0o755) |
| return scorer_exe |
| |
| def _run_tests( |
| self, |
| problem: Problem, |
| solution_code: str, |
| language: str, |
| inputs: list, |
| outputs: list, |
| test_type: str, |
| results: dict, |
| time_limit: float, |
| memory_limit: str, |
| scorer_executable: Optional[str], |
| ) -> bool: |
| """ |
| Run a set of test cases against a solution. |
| |
| Executes the solution for each input, compares output with expected, |
| and handles time/memory limits and special judging. |
| |
| Args: |
| problem: The Problem object |
| solution_code: Source code to execute |
| language: "cpp" or "Python" |
| inputs: List of input strings |
| outputs: List of expected output strings |
| test_type: "sample" or "hidden" (for labeling results) |
| results: Dict to append test results to (modified in place) |
| time_limit: Time limit in seconds |
| memory_limit: Memory limit as string (e.g., "256 MB") |
| scorer_executable: Path to compiled scorer, or None for exact match |
| |
| Returns: |
| True if all tests passed, False if any test failed. |
| |
| Note: |
| Stops on first failure and updates results["status"] accordingly. |
| """ |
| tally = {"PASSED": 0, "TLE": 0, "MLE": 0, "Wrong Answer": 0, "RTE": 0, "CE": 0} |
| |
| |
| try: |
| bytes_limit = parse_memory_limit_to_bytes(memory_limit) |
| except ValueError: |
| bytes_limit = None |
| |
| |
| executable = None |
| if language == "cpp": |
| executable = self._compile_cpp(solution_code, problem.problem_slug) |
| if executable is None: |
| results["status"] = "Compile Error" |
| tally["CE"] += 1 |
| results[f"{test_type}_summary"] = tally |
| return False |
| else: |
| try: |
| compile(solution_code, "<solution>", "exec") |
| except SyntaxError as se: |
| results["status"] = "Compile Error" |
| results["test_cases"].append({ |
| "test": f"{test_type}_1", |
| "result": "Compile Error", |
| "error": str(se) |
| }) |
| tally["CE"] += 1 |
| results[f"{test_type}_summary"] = tally |
| return False |
| |
| |
| timeout = min(60, time_limit + 5) |
| |
| for i, (input_data, expected_output) in enumerate(zip(inputs, outputs)): |
| test_name = f"{test_type}_{i+1}" |
| |
| try: |
| |
| start_time = time.time() |
| |
| if language == "cpp": |
| proc = subprocess.run( |
| [executable], |
| input=input_data, text=True, capture_output=True, |
| timeout=timeout, |
| preexec_fn=lambda bl=bytes_limit: _preexec_setrlimit_bytes(bl) |
| ) |
| else: |
| proc = subprocess.run( |
| ["python3", "-c", solution_code], |
| input=input_data, text=True, capture_output=True, |
| timeout=timeout, |
| preexec_fn=lambda bl=bytes_limit: _preexec_setrlimit_bytes(bl) |
| ) |
| |
| execution_time = time.time() - start_time |
| generated_output = (proc.stdout or "").strip() |
| |
| |
| if proc.returncode == PREEXEC_ERROR_EXIT_CODE: |
| raise RuntimeError("preexec_fn failed") |
| |
| if proc.returncode is not None and proc.returncode < 0: |
| sig = -proc.returncode |
| if sig == signal.SIGKILL: |
| raise MemoryError("Killed by SIGKILL (likely OOM)") |
| |
| if proc.returncode == 137: |
| raise MemoryError("Exit code 137 (OOM)") |
| |
| if proc.returncode != 0: |
| raise RuntimeError(f"Non-zero exit: {proc.returncode}, stderr: {proc.stderr}") |
| |
| |
| is_correct = self._judge_output( |
| problem, input_data, expected_output, generated_output, |
| scorer_executable |
| ) |
| |
| if is_correct: |
| results["test_cases"].append({ |
| "test": test_name, |
| "result": "PASSED", |
| "time": execution_time, |
| }) |
| tally["PASSED"] += 1 |
| print(f"✅ {test_name} Passed ({execution_time:.2f}s)") |
| else: |
| results["status"] = "Wrong Answer" |
| results["test_cases"].append({ |
| "test": test_name, |
| "result": "Wrong Answer", |
| "time": execution_time, |
| "expected": expected_output[:200] + "..." if len(expected_output) > 200 else expected_output, |
| "generated": generated_output[:200] + "..." if len(generated_output) > 200 else generated_output, |
| }) |
| tally["Wrong Answer"] += 1 |
| print(f"❌ {test_name} Wrong Answer") |
| results[f"{test_type}_summary"] = tally |
| return False |
| |
| except subprocess.TimeoutExpired: |
| results["status"] = "Time Limit Exceeded" |
| results["test_cases"].append({"test": test_name, "result": "TLE"}) |
| tally["TLE"] += 1 |
| print(f"⏳ {test_name} TLE") |
| results[f"{test_type}_summary"] = tally |
| return False |
| |
| except MemoryError as e: |
| results["status"] = "Memory Limit Exceeded" |
| results["test_cases"].append({"test": test_name, "result": "MLE", "error": str(e)}) |
| tally["MLE"] += 1 |
| print(f"💾 {test_name} MLE") |
| results[f"{test_type}_summary"] = tally |
| return False |
| |
| except Exception as e: |
| results["status"] = "Runtime Error" |
| results["test_cases"].append({"test": test_name, "result": "RTE", "error": str(e)}) |
| tally["RTE"] += 1 |
| print(f"❌ {test_name} Runtime Error: {e}") |
| results[f"{test_type}_summary"] = tally |
| return False |
| |
| results[f"{test_type}_summary"] = tally |
| return True |
| |
| def _compile_cpp(self, solution_code: str, name: str) -> Optional[str]: |
| """ |
| Compile a C++ solution to an executable. |
| |
| Args: |
| solution_code: C++ source code as a string |
| name: Base name for the source file and executable |
| |
| Returns: |
| Path to the compiled executable, or None if compilation fails. |
| |
| Note: |
| Compiles with: g++ -o <exe> <src> -std=c++17 -O2 |
| Compilation errors are printed to stdout. |
| """ |
| tmp_dir = tempfile.mkdtemp() |
| cpp_file = os.path.join(tmp_dir, f"{name}.cpp") |
| executable = os.path.join(tmp_dir, f"{name}.out") |
| |
| with open(cpp_file, "w") as f: |
| f.write(solution_code) |
| |
| result = subprocess.run( |
| ["g++", "-o", executable, cpp_file, "-std=c++17", "-O2"], |
| capture_output=True, text=True |
| ) |
| |
| if result.returncode != 0: |
| print(f"❌ Compile Error:\n{result.stderr}") |
| return None |
| |
| return executable |
| |
| def _judge_output( |
| self, |
| problem: Problem, |
| input_data: str, |
| expected_output: str, |
| generated_output: str, |
| scorer_executable: Optional[str], |
| ) -> bool: |
| """ |
| Judge whether the generated output is correct. |
| |
| First attempts exact string match (after stripping whitespace). |
| If that fails and a special judge is available, uses the scorer. |
| |
| Args: |
| problem: The Problem object |
| input_data: The input that was given to the solution |
| expected_output: The expected output from test case |
| generated_output: The output produced by the solution |
| scorer_executable: Path to compiled scorer, or None |
| |
| Returns: |
| True if the output is correct, False otherwise. |
| |
| Note: |
| Special judge formats: |
| - "standard": scorer <input> <judge_output> <contestant_output> |
| - "jakarta2017": scorer <input> <unused> <judge_output> < contestant_output |
| """ |
| expected_output = expected_output.strip() |
| generated_output = generated_output.strip() |
| |
| |
| if generated_output == expected_output: |
| return True |
| |
| |
| if scorer_executable and problem.has_special_judge: |
| with tempfile.TemporaryDirectory() as tmp_dir: |
| input_file = os.path.join(tmp_dir, "input.txt") |
| expected_file = os.path.join(tmp_dir, "expected.txt") |
| generated_file = os.path.join(tmp_dir, "generated.txt") |
| |
| with open(input_file, "w") as f: |
| f.write(input_data) |
| with open(expected_file, "w") as f: |
| f.write(expected_output) |
| with open(generated_file, "w") as f: |
| f.write(generated_output) |
| |
| if problem.special_judge_format == "jakarta2017": |
| |
| proc = subprocess.run( |
| [scorer_executable, input_file, "dummy", expected_file], |
| input=generated_output, text=True, capture_output=True |
| ) |
| |
| return "WA" not in proc.stdout and proc.returncode == 0 |
| else: |
| |
| proc = subprocess.run( |
| [scorer_executable, input_file, expected_file, generated_file], |
| capture_output=True, text=True |
| ) |
| |
| output = proc.stdout.strip().upper() |
| return "AC" in output or (proc.returncode == 0 and "WA" not in output) |
| |
| return False |
|
|
|
|
| |
| def run_solution(problem_id: str, solution_code: str, language: str = "cpp") -> dict: |
| """ |
| Convenience function to quickly run a solution against a problem. |
| |
| Creates a TestRunner instance, runs the solution, and returns results. |
| For running multiple solutions, create a TestRunner instance directly |
| to avoid reloading the dataset each time. |
| |
| Args: |
| problem_id: The problem ID from the dataset |
| solution_code: Solution source code as a string |
| language: "cpp" (default) or "Python" |
| |
| Returns: |
| dict with status and test results (see TestRunner.run_solution) |
| |
| Example: |
| >>> from hf_test_runner import run_solution |
| >>> results = run_solution("icpc-jakarta-2019_card-collecting", code) |
| >>> print(results["status"]) |
| """ |
| runner = TestRunner() |
| return runner.run_solution(problem_id, solution_code, language) |
|
|
|
|
| if __name__ == "__main__": |
| |
| runner = TestRunner() |
| |
| |
| print("\nAvailable problems (first 10):") |
| for pid in runner.list_problems()[:10]: |
| p = runner.get_problem(pid) |
| print(f" {pid}: {p.problem_title}") |
| print(f" Sample tests: {len(p.sample_test_cases_input)}, Hidden: {len(p.hidden_test_cases_input)}") |
| print(f" Special judge: {p.has_special_judge}, Kattis: {p.uses_kattis}") |
|
|