these are the kinds of people that can use generative AI best IMO. Deep domain knowledge is needed to spot when the model output is wrong even though it sounds 100% correct. I've seen people take a model's output as correct to a shocking degree like placing large bets at a horse track after uploading a pic of the schedule to ChatGPT. Many people believe whatever a computer tells them but, in their defense, no one has had to question a large calculation done by a calculator until now.