Frontier-CS / SUBMIT.md
andylizf's picture
Upload folder using huggingface_hub
5fed0fc verified
# Submitting Your Results
We welcome submissions from all models and agent frameworks. To have your results included in our leaderboard, please follow the instructions below.
## Algorithmic Problems
We currently release **1 -- 3 public test case** per problem for local testing and debugging. Full evaluation (with all test cases) is performed on our servers.
### What to Submit
1. **Solution files**: `{problem_id}_{model_name}_solution.cpp` for each problem
2. **Model/Agent info**: Name and version of the model or agent framework used
3. **Generation method**: Brief description of how solutions were generated (e.g., one-shot, multi-turn, with/without feedback)
### Submission Format
Organize your solutions as:
```
submissions/
β”œβ”€β”€ 1_gpt4_solution.cpp
β”œβ”€β”€ 2_gpt4_solution.cpp
β”œβ”€β”€ ...
└── metadata.json
```
`metadata.json`:
```json
{
"model": "gpt-4o",
"agent_framework": "custom",
"generation_method": "one-shot",
"date": "2025-01-15",
"notes": "Optional additional notes"
}
```
## Research Problems
Research problems require a `solution.py` file implementing the `Solution` class interface.
### Problem Structure
Research problems follow a hierarchical structure:
```
Problem (e.g., gemm_optimization, poc_generation)
└── Category (e.g., squares, heap_buffer_overflow)
└── Variant (e.g., arvo_21000)
```
| Level | Example | Description |
|-------|---------|-------------|
| **Problem** | `gemm_optimization` | Top-level problem domain |
| **Category** | `gemm_optimization/squares` | Scores are **aggregated** at this level for leaderboard reporting |
| **Variant** | `poc_generation/heap_buffer_overflow/arvo_21000` | Each variant is **evaluated independently** with its own README |
**Key distinction:**
- **Evaluation**: Each variant runs independently and produces its own score
- **Reporting**: Scores are aggregated by category for the leaderboard (e.g., all `heap_buffer_overflow` variants β†’ one score)
> Note: Some problems have only one level (e.g., `flash_attn`), which functions as both category and variant.
### Problem ID Format
Each variant has a unique **Problem ID** based on its path under `research/`.
The full list of all evaluatable variants is in [`research/problems.txt`](research/problems.txt) (109 variants total, aggregated into ~50 categories for reporting).
| Type | Example Path | Problem ID |
|------|-------------|------------|
| Single problem | `research/flash_attn` | `flash_attn` |
| Problem with variants | `research/gemm_optimization/squares` | `gemm_optimization/squares` |
| Nested variants | `research/poc_generation/heap_buffer_overflow/arvo_21000` | `poc_generation/heap_buffer_overflow/arvo_21000` |
### What to Submit
1. **Solution files**: `solution.py` for each problem, placed in a directory matching the Problem ID
2. **Model/Agent info**: Name and version of the model or agent framework used
3. **Local evaluation results** (optional but recommended): Score from running the evaluator locally
### Submission Format
Your submission zip should mirror the Problem ID directory structure:
```
submission.zip
β”œβ”€β”€ flash_attn/
β”‚ └── solution.py
β”œβ”€β”€ gemm_optimization/
β”‚ └── squares/
β”‚ └── solution.py
β”œβ”€β”€ cant_be_late/
β”‚ └── high_availability_loose_deadline/
β”‚ └── solution.py
β”œβ”€β”€ poc_generation/
β”‚ └── heap_buffer_overflow/
β”‚ └── arvo_21000/
β”‚ └── solution.py
└── metadata.json
```
**Important**: The directory structure must exactly match the Problem ID. For example:
- `flash_attn/solution.py`
- `gemm_optimization/squares/solution.py`
Each `solution.py` must implement:
```python
class Solution:
def __init__(self):
pass
def solve(self, *args):
# Returns: solution output (format varies by problem)
pass
```
### metadata.json
```json
{
"model": "gpt-4o",
"agent_framework": "custom",
"generation_method": "one-shot",
"date": "2025-01-15",
"problems_solved": [
"flash_attn",
"gemm_optimization/squares",
"cant_be_late/high_availability_loose_deadline"
],
"notes": "Optional additional notes"
}
```
### Running Local Evaluation
Before submitting, you can verify your solutions locally:
```bash
# Evaluate a single solution
frontier-eval flash_attn solution.py
# Batch evaluation with progress tracking
frontier-eval batch --pairs-file pairs.txt --results-dir results/
# Batch evaluation with SkyPilot (cloud)
frontier-eval batch --pairs-file pairs.txt --skypilot --max-concurrent 4
```
## How to Submit
Send your submission to:
- **Email**: [email protected] or [email protected]
Please include:
1. A zip/tar archive of your solutions following the format above
2. `metadata.json` with model and method information
3. (Optional) Local evaluation results if you ran them
## Leaderboard
Accepted submissions will be evaluated on our full test suite and results will be published on the [Frontier-CS Leaderboard](https://frontier-cs.org).