Yes, that would be a misuse of the tech. A big problem with AI is that the general public just can’t get their heads around what’s going on. They insist on trying to use it as a truth engine, and then when it fails at that they declare that AI is useless.
Take the current conversation. There are a lot of misconceptions. For example, there seems to be a big confusion between ‘training’ an AI and ‘fine tuning it’ or using its context to analyze material that you give it.
In the first case (which ended in Sept or Oct. '21 for GPT 3.5 and 4.0), material that is fed into it winds up modifying an unspecified number of parameters. Anywhere from none to billions. The training material is not retained, but having read it ChatGPT can use what it learned, much in the way a human does. It’s not memorization and verbatim repeat. It’s learning. For example, it’s possible for a chatbot to read something, make a zillion parameter changes, then read something else that invalidates what it read, overwriting many of the changes it made betore. Or it could read something small that pushes it over a tipping point and causes a massive reorganization of its parameter structure. We don’t know. But there’s no verbatim copying of anything really going on.
In the case of ‘fine tuning’, new layers are added to the model, and these layers can be modified by what is read. But the original layers of the trained model remain intact.
When using ChatGPT as a user, you have access to the ‘context’, which is a chunk of memory used for ChatGPT’s prompt. You can use this for fine-tuning of a sort (it also uses additional layers but doesn’t modify the main model), for instance by feeding it a transcript and asking it to format it. When the session is done, those layers are destroyed, and the original model was never modified, so nothing is changed at all afterwards, no IP is violated.
Like humans, it can repeat small passages verbatim. Just like I can repeat the entire poem ‘Jabberwocky’ on command. If someone asked me to write a nonsense poem, no doubt my attempt would lean heavily on my memory of ‘Jabberwocky’. But it’s not plagairism to have read it, or even to have memorized it word for word, or even using the knowledge of it to inform new work. It would only be plagairism if I posted it essentially intact and claimed it was my own work, or tried to sell it for profit without permission. I don’t see why an AI should be held to a different standard.
Now, it’s possible (and likely) that a record of your session is recorded. But that is true of ANYTHING on the Internet. Do you keep proprietary info in Google Docs? Guess what? When you delete something in Google Docs, it is not deleted. It’s just marked deleted so you don’t see it. Google has the ability to look through everything you upload to them. Ever send a private message on Twitter? Twitter employees can and do read them. As does the government.