Also pictured here: Anthropic stating out loud their models will just give out all the "secret" and "secured" internal data to anyone who asks.
Of course, that's by design. LLMs can't have any barrier between data and instructions, so they can never be secure.
