This is just an FYI to be careful when using AI for research. I wanted to look up resources that I could verify about cow mutilations. I had this “conversation” with MS ”CoPilot” (aka “Bing”):
_**Me**_: What percentage of mutilated cattle were ensanguined where no blood was found?”
_**Copilot **_: According to Howard Burgess, nearly 90 percent of mutilated cattle are found under supposedly unusual, usually bloodless circumstances …
_**Me**_: Who is Howard Burgess?
_**Copilot**_: There are several people named Howard Burgess. If you’re looking for a specific Howard Burgess, could you provide more details?
_**Me**_: The Howard Burgess you mentioned in your first answer
_**Copilot**_: I apologize for the confusion, but I don’t have specific information about a Howard Burgess related to the topic of cattle mutilation. The information I provided earlier was based on general observations and reports about cattle mutilation, not attributed to a specific individual named Howard Burgess.
_**Me**_: How did you arrive at the name Howard Burgess?
_**Copilot**_: I apologize for the confusion. It appears there was a mistake in my previous message. The name “Howard Burgess” was incorrectly included. I don’t have any information about a person named Howard Burgess related to the topic of cattle mutilation.
_**Me**_: In the first answer, how did you arrive at the 90% figure?
_**Copilot**_: I apologize for the confusion. The figure of 90% was an error in my response. I don’t have specific data on the percentage of mutilated cattle found in bloodless conditions.
_**Me**_: Did you just make up the name Howard Burgess and the 90% figure in your first answer?
_**Copilot**_: I apologize for the confusion. The name “Howard Burgess” and the figure of 90% were errors in my response. I strive to provide accurate and reliable information, and I regret any misunderstanding caused by this mistake.
I just wanted to show that any information received from a chatbot cannot be assumed to be correct …*or even true*! This is an example what is known in AI circles as a “hallucination”. The AI did not have an answer so it simply made things up. In some sense it lied. I bring this up as a cautionary warning about chatbots,