Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Tenable updates Vulnerability Priority Rating scoring method to flag fewer vulnerabilities as critical

      July 24, 2025

      Google adds updated workspace templates in Firebase Studio that leverage new Agent mode

      July 24, 2025

      AI and its impact on the developer experience, or ‘where is the joy?’

      July 23, 2025

      Google launches OSS Rebuild tool to improve trust in open source packages

      July 23, 2025

      EcoFlow’s new portable battery stations are lighter and more powerful (DC plug included)

      July 24, 2025

      7 ways Linux can save you money

      July 24, 2025

      My favorite Kindle tablet just got a kids model, and it makes so much sense

      July 24, 2025

      You can turn your Google Photos into video clips now – here’s how

      July 24, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      Blade Service Injection: Direct Service Access in Laravel Templates

      July 24, 2025
      Recent

      Blade Service Injection: Direct Service Access in Laravel Templates

      July 24, 2025

      This Week in Laravel: NativePHP Mobile and AI Guidelines from Spatie

      July 24, 2025

      Retrieve the Currently Executing Closure in PHP 8.5

      July 24, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      FOSS Weekly #25.30: AUR Poisoned, Linux Rising, PPA Explained, New Open Source Grammar Checker and More

      July 24, 2025
      Recent

      FOSS Weekly #25.30: AUR Poisoned, Linux Rising, PPA Explained, New Open Source Grammar Checker and More

      July 24, 2025

      How to Open Control Panel in Windows 11

      July 24, 2025

      How to Shut Down Windows 11

      July 24, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Machine Learning»TikTok Researchers Introduce SWE-Perf: The First Benchmark for Repository-Level Code Performance Optimization

    TikTok Researchers Introduce SWE-Perf: The First Benchmark for Repository-Level Code Performance Optimization

    July 22, 2025

    Introduction

    As large language models (LLMs) advance in software engineering tasks—ranging from code generation to bug fixing—performance optimization remains an elusive frontier, especially at the repository level. To bridge this gap, researchers from TikTok and collaborating institutions have introduced SWE-Perf—the first benchmark specifically designed to evaluate the ability of LLMs to optimize code performance in real-world repositories.

    Unlike prior benchmarks focused on correctness or function-level efficiency (e.g., SWE-Bench, Mercury, EFFIBench), SWE-Perf captures the complexity and contextual depth of repository-scale performance tuning. It provides a reproducible, quantitative foundation to study and improve the performance optimization capabilities of modern LLMs.

    Image source: https://arxiv.org/abs/2507.12415

    Why SWE-Perf Is Needed

    Real-world codebases are often large, modular, and intricately interdependent. Optimizing them for performance requires understanding of cross-file interactions, execution paths, and computational bottlenecks—challenges beyond the scope of isolated function-level datasets.

    LLMs today are largely evaluated on tasks like syntax correction or small function transformations. But in production environments, performance tuning across repositories can yield more substantial system-wide benefits. SWE-Perf is explicitly built to measure LLM capabilities in such settings.

    Image source: https://arxiv.org/abs/2507.12415

    Dataset Construction

    SWE-Perf is constructed from over 100,000 pull requests across high-profile GitHub repositories. The final dataset covered 9 repositories including:

    • 140 curated instances demonstrating measurable and stable performance improvements.
    • Complete codebases pre- and post-optimization.
    • Target functions categorized as oracle (file-level) or realistic (repo-level).
    • Unit tests and Docker environments for reproducible execution and performance measurement.
    • Expert-authored patches used as gold standards.

    To ensure validity, each unit test must:

    1. Pass before and after the patch.
    2. Show statistically significant runtime gains over 20 repetitions (Mann-Whitney U test, p < 0.1).

    Performance is measured using minimum performance gain (δ), isolating statistical improvements attributable to the patch while filtering noise.

    Benchmark Settings: Oracle vs. Realistic

    • Oracle Setting: The model receives only the target functions and corresponding files. This setting tests localized optimization skills.
    • Realistic Setting: The model is given an entire repository and must identify and optimize performance-critical paths autonomously. This is a closer analog to how human engineers work.

    Evaluation Metrics

    SWE-Perf defines a three-tier evaluation framework, reporting each metric independently:

    1. Apply: Can the model-generated patch be applied cleanly?
    2. Correctness: Does the patch preserve functional integrity (all unit tests pass)?
    3. Performance: Does the patch yield measurable runtime improvement?

    The metrics are not aggregated into a single score, allowing more nuanced evaluation of tradeoffs between syntactic correctness and performance gains.

    Experimental Results

    The benchmark evaluates several top-tier LLMs under both oracle and realistic settings:

    Model Setting Performance (%)
    Claude-4-opus Oracle 1.28
    GPT-4o Oracle 0.60
    Gemini-2.5-Pro Oracle 1.48
    Claude-3.7 (Agentless) Realistic 0.41
    Claude-3.7 (OpenHands) Realistic 2.26
    Expert (Human Patch) – 10.85

    Notably, even the best-performing LLM configurations fall significantly short of human-level performance. The agent-based method OpenHands, built on Claude-3.7-sonnet, outperforms other configurations in the realistic setting but still lags behind expert-crafted optimizations.

    Key Observations

    • Agent-based frameworks like OpenHands are better suited for complex, multi-step optimization, outperforming direct model prompts and pipeline-based approaches like Agentless.
    • Performance degrades as the number of target functions increases—LLMs struggle with broader optimization scopes.
    • LLMs exhibit limited scalability in long-runtime scenarios, where expert systems continue to show performance gains.
    • Patch analysis shows LLMs focus more on low-level code structures (e.g., imports, environment setup), while experts target high-level semantic abstractions for performance tuning.

    Conclusion

    SWE-Perf represents a pivotal step toward measuring and improving the performance optimization capabilities of LLMs in realistic software engineering workflows. It uncovers a significant capability gap between existing models and human experts, offering a strong foundation for future research in repository-scale performance tuning. As LLMs evolve, SWE-Perf can serve as a north star guiding them toward practical, production-ready software enhancement at scale.


    Check out the Paper, GitHub Page and Project. All credit for this research goes to the researchers of this project.

    Sponsorship Opportunity: Reach the most influential AI developers in US and Europe. 1M+ monthly readers, 500K+ community builders, infinite possibilities. [Explore Sponsorship]

    The post TikTok Researchers Introduce SWE-Perf: The First Benchmark for Repository-Level Code Performance Optimization appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleThis AI Paper from Alibaba Introduces Lumos-1: A Unified Autoregressive Video Generator Leveraging MM-RoPE and AR-DF for Efficient Spatiotemporal Modeling
    Next Article Allen Institute for AI-Ai2 Unveils AutoDS: A Bayesian Surprise-Driven Engine for Open-Ended Scientific Discovery

    Related Posts

    Machine Learning

    How to Evaluate Jailbreak Methods: A Case Study with the StrongREJECT Benchmark

    July 24, 2025
    Machine Learning

    AI Guardrails and Trustworthy LLM Evaluation: Building Responsible AI Systems

    July 24, 2025
    Leave A Reply Cancel Reply

    For security, use of Google's reCAPTCHA service is required which is subject to the Google Privacy Policy and Terms of Use.

    Continue Reading

    CVE-2025-43575 – Adobe Acrobat Reader Out-of-Bounds Write Arbitrary Code Execution Vulnerability

    Common Vulnerabilities and Exposures (CVEs)

    Apple Appeals App Store Ruling in Epic Games Case

    Security

    Anker issues recall for popular power bank due to fire risk – stop using it now

    News & Updates

    CVE-2025-5365 – Campcodes Online Hospital Management System SQL Injection Vulnerability

    Common Vulnerabilities and Exposures (CVEs)

    Highlights

    CVE-2024-24915 – Check Point SmartConsole Unsecured Credentials

    June 29, 2025

    CVE ID : CVE-2024-24915

    Published : June 29, 2025, 12:15 p.m. | 3 hours, 2 minutes ago

    Description : Credentials are not cleared from memory after being used. A user with Administrator permissions can execute memory dump for SmartConsole process and fetch them.

    Severity: 6.1 | MEDIUM

    Visit the link for more details, such as CVSS details, affected products, timeline, and more…

    RL^V: Unifying Reasoning and Verification in Language Models through Value-Free Reinforcement Learning

    May 13, 2025

    CVE-2025-4720 – SourceCodester Student Result Management System Remote Path Traversal Vulnerability

    May 15, 2025

    CVE-2025-38156 – Marvell MT7996 Null Pointer Dereference Vulnerability (WiFi)

    July 3, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.