About the projects: We are building LLM evaluation and training datasets to train LLM to work on realistic software engineering problems. One of our approaches, in this project, is to build verifiable SWE tasks based on public repository histories in a synthetic approach with human-in-the-loop while expanding the dataset coverage to different types of tasks in terms of programming language, difficulty level,etc.
We are looking for experienced software engineers (tech lead level) who are familiar with high-quality public GitHub repositories and can contribute to this project. This role involves hands-on software engineering work, including development environment automation, issue triaging, and evaluating test coverage and quality
Turing is one of the world’s fastest-growing AI companies accelerating the advancement and deployment of powerful AI systems. You’ll be at the forefront of evaluating how LLMs interact with real code, influencing the future of AI-assisted software development. This is a unique opportunity to blend practical software engineering with AI research.
Evaluating unit test coverage and quality.
Modify and run codebases locally to assess LLM performance in bug-fixing scenarios.
Opportunities to lead a team of junior engineers to collaborate on projects.
Strong experience with Python or Javascript.
Working knowledge of other major languages (e.g., C++, Java, GO etc.) Proficiency with Git, Docker, and basic software pipeline setup.
Comfortable running, modifying, and testing real-world projects locally.
Experience contributing to or evaluating open-source projects is a plus.
Previous participation in LLM research or evaluation projects.
Experience building or testing developer tools or automation agents.
Perks of Freelancing With Turing:
Work in a fully remote environment.
Opportunity to work on cutting-edge AI projects with leading LLM companies.
Potential for contract extension based on performance and project needs.
Contractor position (no medical/paid leave)
Duration of contract :