CodeCaptain: AI-Powered Code Analysis & Performance Evaluation Tool
Explore CodeCaptain, an AI-driven tool for code health assessment, PR risk analysis, and developer performance tracking, built by Team BNM at Geekathon 2024.
Author

Date

Book a call
Editor’s Note: This blog introduces CodeCaptain, an AI-powered code analysis and performance evaluation tool developed by Team BNM during Geekathon 2024. It provides comprehensive code health assessment, pull request risk analysis, user performance tracking, and commit insights, enabling developers to improve efficiency and maintain high-quality code. While the language has been refined for clarity, the content remains true to the original vision and enthusiasm of the team.
Team Composition: Vishal Singh, Sachin Singh, Shivam Pundir, Roshan Ojha, Devanshi Garg
Hello, everyone. I am Shivam from Team BNM.
At Geekathon 2024, our team developed CodeCaptain, a powerful code analysis and performance evaluation tool designed to help developers track code health, analyze pull requests, assess user contributions, and gain deep commit insights. By automating performance tracking and risk assessment, CodeCaptain simplifies codebase management and enhances collaboration for engineering teams.

How It Works
CodeCaptain integrates seamlessly with GlueStack UI, providing a dashboard that displays key metrics such as total branches, commits, and contributors. The activity overview visualizes commit trends over different periods, allowing users to track development consistency.
The metrics section evaluates code health scores and pulls request analysis. Users can review PR risk levels, categorized as low, medium, or high risk, ensuring developers can prioritize code improvements efficiently. (Screenshot of PR analysis screen to be added)
The user performance module ranks contributors based on their commit history and impact. By selecting a developer, users receive performance scores, strength analysis, areas for improvement, and a summary of contributions, offering valuable insights for individual and team growth.
Key Features
CodeCaptain offers comprehensive commit analysis, assessing code complexity, test coverage, coding standards, and review quality. To optimize AI efficiency, it processes commit comments in a streaming fashion, ensuring smooth and scalable analysis. The tool also tracks historical contributions and development patterns, helping teams monitor progress and maintain high coding standards.
Live Demonstration

Users begin by logging into CodeCaptain's dashboard, where they can explore code health insights, contributor performance, and commit analytics. Developers can view risk assessments for pull requests, check contributor rankings, and analyze commit complexity in real-time.
For instance, selecting Viraj, the top contributor at GlueStack UI, generates a performance score and personalized feedback. Similarly, assessing Sanket Sahu, GeekyAnts’ CTO, provides insights into his contributions, strengths, and improvement areas.
Commit analysis is performed dynamically, considering multiple factors such as complexity, standards, and test coverage. The system ensures structured feedback by streaming commits in a controlled manner, preventing overload on the AI model.

Post-Hackathon Enhancements
Following Geekathon 2024, we enhanced CodeCaptain by transitioning from cloud-based AI to a local large language model (LLM), QUEN 2.5. This upgrade improved data privacy, control, and efficiency, making code analysis 2.5 to 3 times faster. By optimizing fine-tuning processes, the tool now analyzes larger codebases with greater accuracy, further strengthening its capabilities for development teams.


Why CodeCaptain Matters
Maintaining code quality and tracking developer performance is critical for any engineering team. CodeCaptain automates these processes, reducing manual effort and ensuring code consistency, better collaboration, and informed decision-making. Its AI-driven approach to risk assessment, user analysis, and commit tracking makes it a valuable tool for teams striving to improve software quality.
Final Thoughts
Building CodeCaptain during Geekathon 2024 was an incredible experience. Our goal was to create a developer-friendly tool that enhances code review, team performance tracking, and project insights. We are excited to continue refining it and expanding its features to further optimize the developer workflow.
A huge thank you to GeekyAnts for organizing this hackathon and providing us the platform to innovate.
Project GItHib link: https://git.geekyants.com/tanmayj/REST-Assured
Related Articles.
More from the engineering frontline.
Dive deep into our research and insights on design, development, and the impact of various trends to businesses.

Apr 14, 2026
The Keyboard Bounce of Death: Handling Inputs on Complex React Native Screens
Fix the React Native ‘Keyboard Bounce of Death.’ Learn why inputs jump and how to build smooth, production-ready forms with modern architecture.

Apr 9, 2026
From RFPs to Revenue: How We Built an AI Agent Team That Writes Technical Proposals in 60 Seconds
GeekyAnts built DealRoom.ai — four AI agents that turn RFPs into accurate technical proposals in 60 seconds, with real-time cost breakdowns and scope maps.

Apr 6, 2026
How We Built an AI System That Automates Senior Solution Architect Workflows
Discover how we built a 4-agent AI co-pilot that converts complex RFPs into draft technical proposals in 15 minutes — with built-in conflict detection, assumption surfacing, and confidence scoring.

Apr 6, 2026
AI Code Healer for Fixing Broken CI/CD Builds Fast
A deep dive into how GeekyAnts built an AI-powered Code Healer that analyzes CI/CD failures, summarizes logs, and generates code-level fixes to keep development moving.

Apr 2, 2026
A Real-Time AI Fraud Decision Engine Under 50ms
A deep dive into how GeekyAnts built a real-time AI fraud detection system that evaluates transactions in milliseconds using a hybrid multi-agent approach.

Apr 1, 2026
Building an Autonomous Multi-Agent Fraud Detection System in Under 200ms
GeekyAnts built a 5-agent fraud detection pipeline that makes decisions in under 200ms — 15x cheaper than single-model systems, with full explainability built in.