The tool is quite powerful, I did some testing and it refactored my code beautifully. A question though, is there a limit on how large my codebase can be since there is a limit of tokens in GPT I/O

Hey! We did some texting on https://github.com/twitter/the-algorithm without issues(*), that was around 150K lines of code.

And it works great, I wish I knew more domain details about twitter to be able to really leverage it, but it quickly finds parts of your system and how to build features like "removing echo chambers from the-algorithm by recommending posts to a broader audience that is likely to have opposing views and interact with others."

I think one of our users tried indexing their internal code on about 200k LOC.

The outputs are however limited to 4k or 8k depending on what model you use.

*(initial indexing might be slow and re-indexing might need to be disabled)