|
|
|
|
|
""" |
|
|
Code Completion Benchmark for CodeReality-1T Dataset |
|
|
Evaluates code completion models using Pass@k metrics |
|
|
""" |
|
|
|
|
|
import json |
|
|
import os |
|
|
import re |
|
|
import random |
|
|
from typing import Dict, List, Tuple, Optional |
|
|
from collections import defaultdict |
|
|
|
|
|
def load_dataset_sample(data_dir: str, sample_size: int = 200) -> List[Dict]: |
|
|
"""Load sample of repositories with code files.""" |
|
|
print(f"🔍 Loading sample of {sample_size} repositories with code files...") |
|
|
|
|
|
repositories = [] |
|
|
files = [f for f in os.listdir(data_dir) if f.endswith('.jsonl')] |
|
|
random.shuffle(files) |
|
|
|
|
|
for filename in files[:15]: |
|
|
file_path = os.path.join(data_dir, filename) |
|
|
try: |
|
|
with open(file_path, 'r', encoding='utf-8', errors='ignore') as f: |
|
|
for line in f: |
|
|
if len(repositories) >= sample_size: |
|
|
break |
|
|
try: |
|
|
repo_data = json.loads(line) |
|
|
|
|
|
if has_code_files(repo_data): |
|
|
repositories.append(repo_data) |
|
|
except json.JSONDecodeError: |
|
|
continue |
|
|
except Exception as e: |
|
|
continue |
|
|
|
|
|
if len(repositories) >= sample_size: |
|
|
break |
|
|
|
|
|
print(f"✅ Loaded {len(repositories)} repositories with code files") |
|
|
return repositories |
|
|
|
|
|
def has_code_files(repo: Dict) -> bool: |
|
|
"""Check if repository contains code files.""" |
|
|
code_extensions = {'.py', '.js', '.java', '.cpp', '.c', '.go', '.rs', '.ts'} |
|
|
|
|
|
files = repo.get('files', []) |
|
|
for file_obj in files: |
|
|
if isinstance(file_obj, dict): |
|
|
file_path = file_obj.get('path', '') |
|
|
if any(file_path.endswith(ext) for ext in code_extensions): |
|
|
return True |
|
|
return False |
|
|
|
|
|
def extract_function_snippets(repo: Dict, language: str = 'python') -> List[Dict]: |
|
|
"""Extract function definitions for completion tasks.""" |
|
|
snippets = [] |
|
|
|
|
|
|
|
|
patterns = { |
|
|
'python': r'def\s+(\w+)\s*\([^)]*\):\s*', |
|
|
'javascript': r'function\s+(\w+)\s*\([^)]*\)\s*{', |
|
|
'java': r'(?:public|private|protected)?\s*(?:static)?\s*\w+\s+(\w+)\s*\([^)]*\)\s*{', |
|
|
'cpp': r'\w+\s+(\w+)\s*\([^)]*\)\s*{', |
|
|
} |
|
|
|
|
|
if language not in patterns: |
|
|
return snippets |
|
|
|
|
|
pattern = patterns[language] |
|
|
extension_map = { |
|
|
'python': '.py', |
|
|
'javascript': '.js', |
|
|
'java': '.java', |
|
|
'cpp': '.cpp' |
|
|
} |
|
|
|
|
|
target_ext = extension_map[language] |
|
|
|
|
|
files = repo.get('files', []) |
|
|
for file_obj in files: |
|
|
if isinstance(file_obj, dict): |
|
|
file_path = file_obj.get('path', '') |
|
|
content = file_obj.get('content', '') |
|
|
|
|
|
if file_path.endswith(target_ext) and content: |
|
|
matches = list(re.finditer(pattern, content, re.MULTILINE)) |
|
|
|
|
|
for match in matches: |
|
|
start_pos = match.start() |
|
|
function_name = match.group(1) |
|
|
|
|
|
|
|
|
lines_before = content[:start_pos].split('\n') |
|
|
context_lines = lines_before[-5:] if len(lines_before) >= 5 else lines_before |
|
|
context = '\n'.join(context_lines) |
|
|
|
|
|
|
|
|
remaining_content = content[start_pos:] |
|
|
lines = remaining_content.split('\n') |
|
|
|
|
|
function_lines = [] |
|
|
indent_level = None |
|
|
|
|
|
for i, line in enumerate(lines): |
|
|
if i == 0: |
|
|
function_lines.append(line) |
|
|
continue |
|
|
|
|
|
|
|
|
if indent_level is None and line.strip(): |
|
|
indent_level = len(line) - len(line.lstrip()) |
|
|
|
|
|
|
|
|
if line.strip() and indent_level is not None: |
|
|
current_indent = len(line) - len(line.lstrip()) |
|
|
if current_indent <= indent_level and not line.strip().startswith(('if', 'for', 'while', 'try', 'except', 'else', 'elif')): |
|
|
break |
|
|
|
|
|
function_lines.append(line) |
|
|
|
|
|
|
|
|
if len(function_lines) > 20: |
|
|
break |
|
|
|
|
|
function_body = '\n'.join(function_lines) |
|
|
|
|
|
|
|
|
if len(function_lines) > 3: |
|
|
snippets.append({ |
|
|
'function_name': function_name, |
|
|
'context': context, |
|
|
'prompt': function_lines[0], |
|
|
'completion': '\n'.join(function_lines[1:]), |
|
|
'file_path': file_path, |
|
|
'language': language |
|
|
}) |
|
|
|
|
|
return snippets |
|
|
|
|
|
def simple_code_completion_model(prompt: str, language: str) -> List[str]: |
|
|
"""Simple rule-based code completion for demonstration.""" |
|
|
completions = [] |
|
|
|
|
|
|
|
|
templates = { |
|
|
'python': [ |
|
|
" pass", |
|
|
" return None", |
|
|
" # TODO: implement this function\n pass", |
|
|
" result = None\n return result", |
|
|
" # Implementation needed\n raise NotImplementedError()" |
|
|
], |
|
|
'javascript': [ |
|
|
" return null;", |
|
|
" // TODO: implement\n return;", |
|
|
" throw new Error('Not implemented');", |
|
|
" var result = null;\n return result;", |
|
|
" console.log('Function called');\n return;" |
|
|
], |
|
|
'java': [ |
|
|
" return null;", |
|
|
" // TODO: implement this method\n return null;", |
|
|
" throw new UnsupportedOperationException();", |
|
|
" Object result = null;\n return result;", |
|
|
" System.out.println(\"Method called\");\n return null;" |
|
|
] |
|
|
} |
|
|
|
|
|
if language in templates: |
|
|
|
|
|
return templates[language] |
|
|
else: |
|
|
return ["// TODO: implement"] |
|
|
|
|
|
def evaluate_completion_quality(predicted: str, actual: str) -> float: |
|
|
"""Simple evaluation of completion quality.""" |
|
|
|
|
|
pred_lines = [line.strip() for line in predicted.split('\n') if line.strip()] |
|
|
actual_lines = [line.strip() for line in actual.split('\n') if line.strip()] |
|
|
|
|
|
if not actual_lines: |
|
|
return 0.0 |
|
|
|
|
|
|
|
|
score = 0.0 |
|
|
|
|
|
|
|
|
empty_indicators = ['pass', 'todo', 'not implemented', 'null', 'return;', 'return null'} |
|
|
pred_empty = any(indicator in predicted.lower() for indicator in empty_indicators) |
|
|
actual_empty = any(indicator in actual.lower() for indicator in empty_indicators) |
|
|
|
|
|
if pred_empty and actual_empty: |
|
|
score += 0.8 |
|
|
elif not pred_empty and not actual_empty: |
|
|
|
|
|
pred_keywords = set(re.findall(r'\b\w+\b', predicted.lower())) |
|
|
actual_keywords = set(re.findall(r'\b\w+\b', actual.lower())) |
|
|
|
|
|
if actual_keywords: |
|
|
keyword_overlap = len(pred_keywords & actual_keywords) / len(actual_keywords) |
|
|
score += keyword_overlap * 0.6 |
|
|
|
|
|
|
|
|
line_ratio = min(len(pred_lines), len(actual_lines)) / max(len(pred_lines), len(actual_lines)) |
|
|
score += line_ratio * 0.4 |
|
|
|
|
|
return min(score, 1.0) |
|
|
|
|
|
def calculate_pass_at_k(completion_results: List[Tuple[List[str], str]], k: int = 1) -> float: |
|
|
"""Calculate Pass@k metric.""" |
|
|
if k <= 0: |
|
|
return 0.0 |
|
|
|
|
|
total_passed = 0 |
|
|
|
|
|
for completions, ground_truth in completion_results: |
|
|
|
|
|
top_k_completions = completions[:k] |
|
|
|
|
|
|
|
|
passed = False |
|
|
for completion in top_k_completions: |
|
|
quality_score = evaluate_completion_quality(completion, ground_truth) |
|
|
if quality_score > 0.5: |
|
|
passed = True |
|
|
break |
|
|
|
|
|
if passed: |
|
|
total_passed += 1 |
|
|
|
|
|
return total_passed / len(completion_results) if completion_results else 0.0 |
|
|
|
|
|
def run_completion_benchmark(repositories: List[Dict]) -> Dict: |
|
|
"""Run code completion benchmark.""" |
|
|
print("🧮 Running code completion benchmark...") |
|
|
|
|
|
results = { |
|
|
'total_repositories': len(repositories), |
|
|
'completion_tasks': [], |
|
|
'language_stats': defaultdict(int), |
|
|
'pass_at_1': 0.0, |
|
|
'pass_at_3': 0.0, |
|
|
'pass_at_5': 0.0, |
|
|
'average_quality': 0.0 |
|
|
} |
|
|
|
|
|
completion_results = [] |
|
|
quality_scores = [] |
|
|
|
|
|
|
|
|
for repo in repositories: |
|
|
for language in ['python', 'javascript', 'java']: |
|
|
snippets = extract_function_snippets(repo, language) |
|
|
|
|
|
for snippet in snippets[:2]: |
|
|
results['language_stats'][language] += 1 |
|
|
|
|
|
|
|
|
completions = simple_code_completion_model(snippet['prompt'], language) |
|
|
ground_truth = snippet['completion'] |
|
|
|
|
|
completion_results.append((completions, ground_truth)) |
|
|
|
|
|
|
|
|
if completions: |
|
|
quality = evaluate_completion_quality(completions[0], ground_truth) |
|
|
quality_scores.append(quality) |
|
|
|
|
|
results['completion_tasks'].append({ |
|
|
'function_name': snippet['function_name'], |
|
|
'language': language, |
|
|
'prompt_length': len(snippet['prompt']), |
|
|
'completion_length': len(ground_truth) |
|
|
}) |
|
|
|
|
|
|
|
|
results['pass_at_1'] = calculate_pass_at_k(completion_results, 1) |
|
|
results['pass_at_3'] = calculate_pass_at_k(completion_results, 3) |
|
|
results['pass_at_5'] = calculate_pass_at_k(completion_results, 5) |
|
|
results['average_quality'] = sum(quality_scores) / len(quality_scores) if quality_scores else 0.0 |
|
|
|
|
|
return results |
|
|
|
|
|
def print_benchmark_results(results: Dict): |
|
|
"""Print formatted benchmark results.""" |
|
|
print("=" * 60) |
|
|
print("🎯 CODE COMPLETION BENCHMARK RESULTS") |
|
|
print("=" * 60) |
|
|
|
|
|
print(f"Total repositories: {results['total_repositories']}") |
|
|
print(f"Completion tasks: {len(results['completion_tasks'])}") |
|
|
|
|
|
print(f"\n📊 Pass@k Metrics:") |
|
|
print(f" Pass@1: {results['pass_at_1']:.3f}") |
|
|
print(f" Pass@3: {results['pass_at_3']:.3f}") |
|
|
print(f" Pass@5: {results['pass_at_5']:.3f}") |
|
|
print(f" Average Quality: {results['average_quality']:.3f}") |
|
|
|
|
|
print(f"\n🔤 Language Distribution:") |
|
|
for language, count in sorted(results['language_stats'].items(), key=lambda x: x[1], reverse=True): |
|
|
print(f" {language}: {count} functions") |
|
|
|
|
|
print(f"\n💡 Insights:") |
|
|
print("- This is a simplified demonstration benchmark") |
|
|
print("- Real evaluation requires more sophisticated code execution") |
|
|
print("- CodeReality-1T provides diverse, noisy code for robust testing") |
|
|
print("- Consider functional correctness testing for production models") |
|
|
|
|
|
def main(): |
|
|
"""Run code completion benchmark.""" |
|
|
print("🚀 CodeReality-1T Code Completion Benchmark") |
|
|
print("=" * 60) |
|
|
|
|
|
|
|
|
data_dir = "/mnt/z/CodeReality_Final/unified_dataset" |
|
|
sample_size = 100 |
|
|
|
|
|
if not os.path.exists(data_dir): |
|
|
print(f"❌ Dataset directory not found: {data_dir}") |
|
|
print("Please update the data_dir path to point to your CodeReality-1T dataset") |
|
|
return |
|
|
|
|
|
|
|
|
repositories = load_dataset_sample(data_dir, sample_size) |
|
|
|
|
|
if not repositories: |
|
|
print("❌ No repositories loaded. Check dataset path.") |
|
|
return |
|
|
|
|
|
|
|
|
results = run_completion_benchmark(repositories) |
|
|
|
|
|
|
|
|
print_benchmark_results(results) |
|
|
|
|
|
|
|
|
output_file = "code_completion_results.json" |
|
|
with open(output_file, 'w') as f: |
|
|
|
|
|
results_json = { |
|
|
'total_repositories': results['total_repositories'], |
|
|
'completion_tasks': results['completion_tasks'], |
|
|
'language_stats': dict(results['language_stats']), |
|
|
'pass_at_1': results['pass_at_1'], |
|
|
'pass_at_3': results['pass_at_3'], |
|
|
'pass_at_5': results['pass_at_5'], |
|
|
'average_quality': results['average_quality'] |
|
|
} |
|
|
json.dump(results_json, f, indent=2) |
|
|
|
|
|
print(f"\n💾 Results saved to: {output_file}") |
|
|
|
|
|
if __name__ == "__main__": |
|
|
main() |