Also, there is enough open source code available that I would hope Anthropic doesn’t feel the need to train their models on potentially litigious code base.
The problem with this statement is twofold. Firstly, it is unrealistic to assume that leading AI companies are staying entirely above board in terms of code licensing. With how widespread AI is, this makes it all the harder for developers to enforce their licenses when many developers inevitably violate their terms without knowing. Even if that code is open source, licensing terms typically require attribution that an AI is unlikely to provide for every segment of code cobbled together.
When the developers that had their code taken and reused are unable to know who reused it, it is disingenuous to work under a 'take first, ask later' mentality.