As a followup to my last post, here’s an example of ChatGPT using a ‘verified’ set of information to interpret questions.
Me: I’m going to give you a list of laws for the fictional country of Bravagna.
It: Sure, I’d be happy to help you with laws for the fictional country of Bravagna. What would you like me to do with the list once you’ve provided it?
Me: Once I’ve provided the list, I will give you a series of scenarios. You’ll tell me whether the individual in that scenario has broken the law or not.
It: Sounds good! Please provide the list of laws for Bravagna and I’ll do my best to determine whether or not the individuals in your scenarios have broken them.
Me:
Law 1: No man will wear red shirts on Thursday.
Law 2: Nobody can eat fried eggs for breakfast.
Law 3: Dancing is forbidden.
Law 4: Nobody puts Baby in a corner.
Law 5: Everybody must watch cartoons while eating.
It: Thank you for the laws of Bravagna. I will do my best to determine whether or not the individuals in your scenarios have broken them. However, it’s worth noting that some of these laws may be difficult to enforce or may not be taken seriously. Law 4, for example, seems to be a reference to a famous line from the movie “Dirty Dancing” and may not be a serious law. Additionally, Law 5 may be hard to enforce as it may be difficult to determine if someone is indeed watching cartoons while eating.
Please let me know when you have your scenarios ready, I will be happy to help you with them.
[ed note: thanks for the critique, cheeky fucker]
Me: It is Thursday, January 10th. Johnny goes to the 4th Street Diner for breakfast. He orders cereal and is upset when he spills milk on his red shirt. What laws has he broken?
It: On Thursday, January 10th, Johnny has broken Law 1, “No man will wear red shirts on Thursday.” by wearing a red shirt. He has not broken any other laws.
Me: Good! Next scenario. It is Thursday, January 10th. Johnny goes to the 4th Street Diner for breakfast. He orders fried eggs and is upset when he gets ketchup on his green shirt. What laws has he broken?
It: On Thursday, January 10th, Johnny has broken Law 2, “Nobody can eat fried eggs for breakfast.”. He has not broken any other laws.
Me: It is Thursday, January 10th. Johnny goes to the 4th Street Diner for breakfast. He turns on the television to watch the classic movie “Dirty Dancing” while he eats his fried eggs. He’s wearing a red-and-white striped shirt.
It: On Thursday, January 10th, Johnny has broken Law 1, “No man will wear red shirts on Thursday.” by wearing a red-and-white striped shirt. He has also broken Law 2, “Nobody can eat fried eggs for breakfast.”. He has not broken any other laws.
Me: Are you sure he hasn’t broken any other laws?
It: I apologize, in the previous scenario, Johnny also broke law 5 “Everybody must watch cartoons while eating” as he watches a movie and not cartoon while eating.
–
This is what I mean when I talk about future models referencing their responses against source data. These laws I made up are absolute nonsense, but the program accepts them as true for the sake of the hypothetical and then runs my queries against that small data set. It wasn’t perfect, but it’s not trained for this kind of stuff.
A model could be fed a law library, trained heavily on it, and told that the information in that library is true. It could be made to always check its probabilistic output against that information before outputting to the user, reducing the risk that the model is doing the “confident bullshit” thing.