Performance FAQs
Git AI's performance characteristics and impact on git operations. Learn about overhead, scaling, and benchmarks for enterprise deployments.
Git AI is designed to have minimal impact on the speed of git operations when running in proxy mode. The system is built in Rust for low overhead, fast execution, and minimal memory usage.
Testing Performance Yourself
You can measure Git AI's overhead on any command using the profiling flag:
GIT_AI_PROFILE=1 git-ai <command>
This will show you the exact overhead Git AI adds to each operation.
Proxy Mode Overhead
When running in proxy mode, Git AI adds approximately 10-20ms of overhead per git command. This overhead primarily comes from starting the git process and piping arguments to the underlying binary.
For git commands that Git AI hooks into (commit, rebase, cherry-pick, reset, merge, fetch, and push), Git AI adds anywhere from 20-250ms on large commits in big repos.
Scalability
How does Git AI scale with repository size?
Git AI operations scale with the size of changes (diffs), not the size of the repository. This means performance remains consistent whether you're working in a small project or a massive monorepo.
Checkpoint operations (tracking AI vs human changes):
- Scale with the number of files changed
- Independent of total repository size
- Typically complete in low hundreds of milliseconds
Commit operations (generating authorship logs):
- Scale with the size of the commit diff
- The slowest operation is generating the authorship log on post-commit
- Typical range: 20-300ms depending on commit size
- Most commits complete in under 100ms
Example Timing Breakdown
For a typical commit with moderate changes:
- Proxy overhead: 10-20ms
- Pre-commit checkpoint: 5-15ms
- Git commit execution: (varies by repo)
- Post-commit authorship log generation: 20-100ms
- Total Git AI overhead: ~35-135ms
Git Notes Performance
Git AI stores authorship information using git notes, which sync alongside your commits. Notes have excellent performance characteristics even in large repositories.
Sync Performance
Git note merges are fast and scale linearly with the number of new notes, not total notes in the repository:
Total Notes | New Notes | Sync Time |
---|---|---|
1,000 | 10 | 0.007s |
1,000 | 1,000 | 0.010s |
10,000 | 10 | 0.007s |
10,000 | 1,000 | 0.010s |
50,000 | 10 | 0.008s |
50,000 | 1,000 | 0.010s |
100,000 | 10 | 0.009s |
100,000 | 1,000 | 0.011s |
As shown in the benchmarks above, sync time remains under 15ms even with 100,000 existing notes.
Storage Considerations
Repository size impact:
- Notes are stored in git's object database
- Compressed and deduplicated (packed) like other git objects
Ref limits:
- Notes are stored in a commit tree structure, not as individual refs
- No practical limit on the number of notes
- Won't hit git's ref limits even in very large repositories
Performance Best Practices
For Individual Developers
- Use the wrapper method - The git wrapper approach has better performance than hook-based methods
- Keep agents updated - Newer versions of coding agents with the latest Git AI hooks will be most optimized. Run
git-ai install-hooks
to update your hooks.
For Enterprise Deployments
- Monitor initial rollout - Measure performance during pilot deployments
- Configure strategically - Use repository allowlists to control where Git AI runs (see Enterprise Configuration)
- Sync notes efficiently - Notes sync automatically after push/fetch with minimal overhead
Common Performance Questions
Does Git AI slow down git operations noticeably?
For most operations, no. The 10-20ms proxy overhead is imperceptible. The largest impact is during commits (20-300ms for authorship log generation), which is still fast enough to feel instant for typical workflows.
How much disk space do authorship logs use?
Authorship logs are very compact. For a typical repository with thousands of commits, authorship notes add less than 1% to the total repository size. Git's compression and deduplication keeps the overhead minimal.
Can Git AI handle large monorepos?
Yes. Because Git AI scales with diff size rather than repository size, it works well even in massive monorepos. Operations remain fast regardless of how many files exist in the repository.
Getting Help
If you experience performance issues with Git AI:
- Use
GIT_AI_PROFILE=1
to measure actual overhead - Open an issue on the GitHub repository
- Schedule a call with the maintainers for enterprise support