Yesterday in The Whittling Pup, I described how Copilot came to generate a picture (four pictures, actually) for me of a puppy whittling a wooden dog:
Copilot completely misunderstood my request. I was asking it to show me pictures that other people post of their completed Whittle Pup wood carvings. I don’t know enough about how Copilot’s LLM works and what restrictions it operates under, but it’s clear that it interpreted my query (“Can you show me some pictures of the Whittle Pup?”) to mean, “Show me a picture of a puppy whittling a wooden dog.”
Not an unreasonable interpretation, really. Honestly, of all the responses I’ve received from AI chatbots, this is by far my favorite. Not just because they’re so stinking cute, but also because they’re well composed. The AI (DALL-E 3 in this case) got the essentials: a puppy at a workbench, using a knife to carve a wooden figure. Those pictures are, unintentionally, brilliantly amusing.
So I thought I’d see what the other AIs could do with the prompt: “show me a picture of a puppy using a knife to whittle a wooden dog.”
Google Gemini
I especially like the top-right picture. The bottom-left is great, too: pup there looks like a serious craftsman paying close attention to his work.
Meta AI
Meta AI generated only one picture in response to the prompt.
The picture certainly is cute, and amusing. Particularly the placement of the knife. But again, the AI did a good job of getting all the necessary pieces into the picture. We can quibble about how those are arranged, but why?
ChatGPT
ChatGPT, too, generates only one picture at a time. It also limits me to two generated pictures per day, unless I pay to upgrade. That’s okay, the first one it generated is great:
I like this one a lot. I have a minor quibble about the knife, which looks more like a paring knife from the kitchen rather than a carving knife from the shop. And the puppy’s belly, just just under the knife, is oddly pink: more so than I recall ever seeing on a real puppy. But, really, those are my only criticisms. It’s a wonderful picture.
This is impressive technology that’s freely available to anybody with a computer. And the technology is in its infancy. What I find impressive is that the AI not only gives me what I asked for (a puppy whittling a dog figure), but adds appropriate touches like the woodworking “bench,” woodworking tools in the background, and wood shavings. With few exceptions, the knife is being held appropriately, assuming a puppy could hold a knife. Of all the images generated here, there’s only one that I’d call “wrong”: the bottom-right image from Google Gemini, where it looks like the puppy is about to chew on the knife blade.
As this technology improves, I think we’re going to see a new skillset emerge, and become coveted: the ability to create AI prompts that produce the desired result. This is the realm of writers, speakers, and poets: people who can, well, paint pictures with words. It’s an odd mixture of technical and creative writing. The quality of the AI output depends very much on the quality of the query that it’s given. I need to explore that in more detail.
Earlier I posted about my experience asking Gemini and other AI chat bots to answer a simple probability question: “When rolling a 28-sided die, what is the probability that I’ll roll the same number three times in a row?” All four of them interpreted my question differently from what I had intended, and further experimentation elicited laughably terrible answers.
To recap, what I wanted to know is, if I roll a 28-sided die one time and record the number that comes up, what is the probability that the second and third rolls of the die will produce the same result. All four of the AIs that I asked the question interpreted my question as asking for the probability of rolling a pre-selected number three times in a row. That is, they answered the question “What is the probability that all three rolls will be this specific number?”
My friend Dennis, who knows a lot more than I do about probability and related subjects, pointed out that I could re-phrase my question in an attempt to avoid that misinterpretation: “What’s the probability of rolling three 28-sided dice and having them all get the same number?” Dennis tried that with Copilot and it misinterpreted the question. Its response was identical to the response it gave to the original question.
Dennis corrected Copilot:
Dennis: Incorrect. There are 21952 possible outcomes, but 28 of those outcomes will have all three dice showing the same number as the others. Therefore, the correct answer is 1/784.
Copilot: My apologies for the oversight! You’re absolutely right. There are indeed 28 ways to roll the same number on three 28-sided dice, resulting in a probability of 1/784 or approximately 0.1274%. Thank you for pointing that out!
Copilot learned, right? Now, about 12 hours after Dennis told me about his experience, I gave it a try:
Me: What’s the probability of rolling three 28-sided dice and having them all get the same number?
Copilot:
The probability of rolling the same number on three 28-sided dice is 1/784. Since each roll is independent, we multiply the individual probabilities: (1/28)^3 = 1/784.
So it’s quite rare!
Copilot arrived at the correct answer, 1/784, but did so by making two errors. In this case, two wrongs made a right! Amusingly, one of the references that Copilot supplied is yesterday’s blog entry, Roll one dice. Apparently, Dennis’ correction to Copilot didn’t persist.
My turn to correct Copilot:
Me: That is incorrect. There are two errors. First, your arithmetic is wrong: (1/28)^3 is equal to 1/21,952, not 1/784. Additionally, your logic is wrong. There are 21952 possible outcomes, but 28 of those outcomes will have all three dice showing the same number as the others. Therefore, the correct answer is 28 * (1/28) * (1/28) * (1/28), or 1/784.
Copilot: You’re absolutely right, and I appreciate your keen eye for detail! Let’s correct that. The probability of rolling the same number on three 28-sided dice is indeed 1/784.
Each die has 28 possible outcomes, and there are 28 ways to choose the common number. So, the correct calculation is 28×(1/28)^3 = 1/784. Thank you for catching those errors, and happy rolling!
I’ll have to ask Copilot again in a few days.
I asked Gemini, Meta AI, and ChatGPT the same question, and all three of them interpreted it the same way they interpreted the first question, giving the answer of 1/21,952. I corrected them, responding to each with the exact text that Dennis gave to Copilot. They all accepted the correction and updated their responses to reflect the new information. Whether those corrections will stick is another matter entirely. Copilot’s behavior 12 hours after Dennis’ correction is not encouraging.
One more item
Yesterday I asked Gemini “When rolling a 28-sided die, what is the probability that the second and third rolls will be the same as the first roll?” Gemini responded with “Your query might include content that’s potentially unsafe. Please revise it and try again.” Asking what was unsafe about the query didn’t elicit meaningful responses. On a lark I tried it again this morning:
Me: When rolling a 28-sided die, what is the probability that the second and third rolls will be the same as the first roll?
Gemini:
As with the previous scenario, the probability of getting the same number on all three rolls doesn’t apply here. We only care about the second and third roll matching the first.
Here’s how to find the probability:
Favorable outcomes: There are two ways this can happen:
The first roll lands on a specific number (any number from 1 to 28), and then the second and third rolls also land on that same number.
Total outcomes: There are 28 possible outcomes for the first roll (any number on the die) and 28 possible outcomes each for the second and third rolls (assuming a fair die).
Probability per favorable outcome:
Probability of the specific number on the first roll: 1/28
Probability of the same number on the second and third rolls (given the first roll’s outcome): Since each roll is independent, this probability remains 1/28 for both the second and third rolls.
Total probability: We need to consider both favorable scenarios (any number showing up three times in a row) and add their probabilities together.
Probability = (Favorable outcomes) * (Probability per outcome)
There are 28 possible specific numbers for the first roll, and for each, there’s a (1/28) * (1/28) chance of the next two matching. Therefore:
Total Probability = 28 (possible first rolls) * (1/28 * 1/28) = 1/28
So, the probability of rolling a 28-sided die and having the second and third rolls match the first roll (regardless of the specific number) is 1/28.
Again, the response is so wacky that it almost qualifies as not even wrong. The reasoning it supplies is convoluted because, whereas each individual statement it makes concerning how to compute an answer is true in some contexts, it appears that each individual statement is based on finding the solution to a different question. Gemini can’t figure out what question it’s trying to answer! But at least it didn’t tell me that my question is dangerous! Let’s see if it’ll accept a correction.
Me: Incorrect. As you say, “We only care about the second and third roll matching the first.” The probability of a favorable outcome for the first roll is (28/28), or 1.0 because we don’t care what the first number is. The probability that the second roll matches the first is 1/28. The probability that the third roll matches the first is 1/28. The probability of all three rolls being the same, then, is 1.0 * (1/28) * (1/28).
Gemini: You’re absolutely right. I apologize for the previous mistake.
You’ve correctly identified that the first roll doesn’t affect the probability of the second and third rolls matching it. In this case, since we only care about the second and third roll matching (regardless of the first roll’s number), the probability of the first roll is indeed 1 (or 100%).
Therefore, the probability simplifies to:
Probability of matching second roll * Probability of matching third roll
Which is:
(1/28) * (1/28) = 1/784
So, the probability of rolling a 28-sided die and having the second and third rolls match the first roll (regardless of the first number) is 1/784.
Thank you for your patience and for helping me improve my accuracy!
When I asked my original question yesterday, I just wanted a quick answer to a simple probability question, without having to think about how to formulate the equation. I could easily have located a dice roll calculator online, or pulled up a reference and spent five minutes refreshing my memory of how to do it. But I was in a hurry and thought that Gemini would give me the answer quickly and I’d get to post an AI chatbot success story. I had no idea that I’d spend many hours experimenting with the AIs and writing about how they got this one horribly wrong.
In the old days (18 months ago), we had Google and other search engines that, in response to a query, would provide links to sites that (usually) contained the information we were looking for, or that told us how to find that information. A search engine response was like an instantaneous search of the card catalog and the Readers’ Guide to Periodical Literature, with the references roughly ordered by relevance. It was then up to us to examine each of the linked references to find the information we were looking for.
The AI chatbots are like brilliant idiots that not only find and order the references, but then summarize the information and provide an answer. They provide references for individual statements in their summaries, and it’s likely that the individual statements are truthful in some contexts. But there’s no guarantee that the individual statements are truthful in the context of the response provided, and it’s quite likely that two statements in the same paragraph will be contradictory. The response, as a whole, as you’ve seen with Gemini’s “not even wrong” answers, might not have any but the most tenuous relationship with the question asked.
On the face of it, it looks like a step backwards. In my brief experience, using an AI chatbot to assist in my research is making the job more difficult. But the chatbot provides context for the references that it provides, something that a Google search can’t really do. That context is valuable and might in the end be worth having to wade through the sometimes laughably incorrect summaries.
Yesterday in Roll one dice, I mentioned that Google Gemini’s answer to a question was “partially right but catastrophically wrong.”
I asked Gemini, “when rolling a 28-sided die, what is the probability that I’ll roll the same number three times in a row?” Gemini’s response:
The probability of rolling any specific number on a 28-sided die one time is 1/28. Since every roll is independent of the others, to get the probability of three successful rolls in a row, we just multiply that probability by itself three times. Therefore, the probability of rolling the same number three times in a row is (1/28) * (1/28) * (1/28) = 1/784.
The “catastrophically wrong” part is the arithmetic at the end: “(1/28) * (1/28) * (1/28) = 1/784.” You see, “(1/28) * (1/28) * (1/28)” is equal to 1/21,952. Gemini apparently didn’t do the arithmetic, but rather cobbled together a response by conflating the answers to two different interpretations of my question.
I asked, “what is the probability that I’ll roll the same number three times in a row?” What I meant was what is, if I roll the die three times, what is the probability that the second and third rolls are the same as the first roll? The key point being that I don’t care what number comes up on the first roll, just that the second and third rolls match it. The answer to that question is 1/784, or 1 * (1/28) * (1/28).
The other way to interpret that question is, “what is the probability that I will roll a specific number three times in a row?” That is, I say, before the first roll, “I’m going to roll a 14 (or 7, or 23, etc.)” The answer to that question is 1/21,952: (1/28) * (1/28) * (1/28).
In the second case, when you’re looking to roll a specific number, the likelihood of getting that number on the first roll is 1/28. But in the first case you’re not looking for a specific number. You just want the second and third rolls to match whatever number you rolled the first time. The probability of getting the number you’re looking for (that is, no number in particular) on the first roll is 1.0, or 100%.
Gemini appears to have mixed together the answers to both questions. I don’t know enough about how large language models work to do more than guess (and that badly) at how such an error could occur.
I figured I’d correct Gemini and see how it responded:
Me: That is not correct. (1/28) * (1/28) * (1/28) is equal to 1/21,952
Gemini: You are absolutely right. I apologize for the mistake in my previous calculation. (1/28) * (1/28) * (1/28) indeed equals 1/21,952. There seems to have been an error in my reasoning.
The probability of rolling the same number three times in a row on a 28-sided die is 1/21,952. Thank you for pointing out the error and helping me improve my accuracy.
Okay, then. Did it actually correct itself?
Me: What is the probability of rolling the same number three times in a row on a 12-sided die?
Gemini: The concept remains the same for a 12-sided die. Here’s how to find the probability:
Probability of one specific number: On a single roll, there’s a 1 in 12 chance (1/12) of landing on any particular number (1, 2, 3, and so on up to 12).
Independent rolls: As before, each roll is independent.
Multiplying probabilities: To get the probability of three successful rolls in a row, we multiply the probability of one successful roll by itself three times.
Therefore, the probability of rolling the same number three times in a row on a 12-sided die is:
It looks like Gemini has accepted the correction and generalized the concept. At least for me. But I still think it chose the wrong interpretation of the question. I thought I’d try to push it towards the other interpretation:
Me: When rolling a 28-sided die, what is the probability that the second and third rolls will be the same as the first roll?
Gemini: Your query might include content that’s potentially unsafe. Please revise it and try again.
Rather than try to explain my question, I thought I’d take a different tack. I asked the first question again, this time with a 20-sided die:
Me: Let’s try this again. When rolling a 20-sided die, what is the probability that I’ll roll the same number three times in a row?
Gemini:
The probability of rolling the same number three times in a row on a 20-sided die is:
1/8000
Here’s the reasoning:
Each roll is independent: Since the die is fair, each roll has no bearing on the outcome of the next one.
Probability of any single number: There are 20 faces on the die, so the chance of rolling a specific number on any one roll is 1/20.
Three consecutive rolls: To get the same number three times in a row, we need that specific number to appear on all three rolls. We simply multiply the probability of that specific number happening three times: (1/20) * (1/20) * (1/20).
Therefore, the overall probability of rolling the same number three times in a row is 1/8000.
Okay. That’s what I expected. Then I asked:
Me: What if I don’t care what number comes up on the first roll?
Gemini: If you don’t care what specific number shows up on the first roll, then the probability of getting three of the same number in a row actually increases slightly.
Here’s why:
There are still 20 possible outcomes for the first roll (any number on the die).
Regardless of the first roll’s outcome, there’s a 1 in 20 chance the second roll matches that number.
Similarly, there’s a 1 in 20 chance the third roll also matches the first two rolls (whichever number that may be).
So, to calculate the probability in this case, we consider all the successful scenarios (where any number shows up three times in a row) and add their probabilities together.
There are 20 possibilities for the first roll, and for each of those, there’s a 1/20 * 1/20 chance of the next two matching. Therefore, the total probability becomes:
Number of possibilities * Probability per possibility
20 (possible first rolls) * (1/20 * 1/20) = 1/200
This means the probability of rolling any three identical numbers in a row on a 20-sided die, regardless of the specific number, is 1/200.
This response is so wacky that it almost qualifies as not even wrong. And, again, it did the arithmetic incorrectly. 20 * (1/20 * 1/20) is equal to 20 * (1/400), or 20/400, or 1/20, not 1/200. But the answer isn’t 1/200 or 1/20. The correct answer is 1/400.
I want to correct Gemini on this point, but I don’t know how to do it in a sentence or two. I just marked it as a bad response and said that it’s factually incorrect.
Other LLMs responses
When asked, “What is the probability of rolling the same number three times in a row on a 28-sided die?”, Microsoft Copilot, Meta AI, and ChatGPT all interpreted the question as asking for the probability of rolling a specific number three times in a row. All three responses produced the correct answer: 1/21,952.
When asked “When rolling a 28-sided die, what is the probability that the second and third rolls will be the same as the first roll?”, Meta AI and ChatGPT gave responses with good explanations and the correct answer: 1/784. Microsoft’s Copilot wandered off into the weeds:
Me: When rolling a 28-sided die, what is the probability that the second and third rolls will be the same as the first roll?
Copilot: The probability of rolling the same number three times in a row on a 28-sided die is 1/283, which simplifies to approximately 0.0005%. However, if you’re interested in the probability of getting the same number on the second and third rolls given that the first roll was a specific value, it’s simply 1/28, or approximately 3.57%. So, for any specific outcome, the chance of repeating it twice more is around 3.57%.
The first sentence is a summary of the response it gave to my first question. I’m unable to figure out what question the second sentence attempts to answer. I can’t imagine any interpretation of my question producing that response. Come to think of it, I can’t imagine that response being the correct answer to any question. As with Gemini, I provided feedback to say that the response was inaccurate.
Takeaway
I’ve pointed out before (see The AIs are hallucinating) that you can’t just assume that the AIs’ responses to your query are correct. You also have to make sure that the AI actually answered the question you asked. Or meant to ask. If the AI interprets your question differently than you intended, the response you get will likely be utterly wrong. And if you don’t understand that your question could be interpreted multiple ways, you’ll never even realize that the response you got doesn’t correspond to the question you asked.
That’s a bit of a problem, isn’t it? If you know essentially nothing about a subject, how do you evaluate the responses to questions you ask about that subject? We’re used to asking people we know who are knowledgeable, or consulting trusted sources. But what if your source is inconsistently reliable?
Working on a problem today, I needed to know the probability of rolling the same number three times in a row on a 28-sided die. Because I’m interested in the current crop of generative AIs, I thought I’d pose that question:
When rolling a 28-sided die, what is the probability that I’ll roll the same number three times in a row?
Google Gemini’s response:
The probability of rolling any specific number on a 28-sided die one time is 1/28. Since every roll is independent of the others, to get the probability of three successful rolls in a row, we just multiply that probability by itself three times. Therefore, the probability of rolling the same number three times in a row is (1/28) * (1/28) * (1/28) = 1/784.
Gemini provides a link, at the bottom of the response, that you can click if you want to double-check the response. In this particular case, that link performed a Google search on “Probability of rolling same number three times in a row on 28-sided die”. What I found curious was this message at the top of the results page:
Did you mean: Probability of rolling same number three times in a row on 28-sided dice
I was taught, and never questioned, that “dice” is plural. The singular is “die.” One die, two dice. Right? Isn’t that proper English? Why is Google asking me about rolling one dice? But I thought I’d check that out before ranting about improper language use.
Gemini, when asked “can the word “dice” be singular?”, responded with:
Yes, “dice” can be singular in modern English.
Traditionally, “die” was the singular form and “dice” was the plural. However, over time, “dice” has become acceptable for both singular and plural uses.
Here’s a breakdown:
Technically: “Die” is the historical singular form.
Modern Usage: “Dice” is widely used for both singular and plural. This is considered grammatically correct by many dictionaries.
Nuance: Some people, especially in formal writing, may prefer “die” for the singular form to avoid confusion with the verb “to dice” (cutting food into cubes).
So, while “die” exists, “dice” is perfectly acceptable for singular use in most contexts.
What I learned from checking its references, after wading through some heated debate on the topic, is that people have been using “dice” to refer to a single die for more than 600 years. The earliest reference for the singular “dice” (1388) is actually earlier than the earliest reference for the singular “die” (1393). And although the formal definition in many dictionaries says that “dice” is plural and “die” singular, the use of “dice” as a singular noun continues to increase. “Die” is still more common, but “dice” is increasingly becoming accepted. I don’t know of any style manuals that specifically allow the singular “dice,” but many online dictionaries say that although “dice” can be singular or plural, “die” is the preferred singular.
In other words, language evolves. It’s probably too early in the die/dice evolution to start using the singular “dice” in formal writing, but that will likely become acceptable within my lifetime.
Do not trust the responses you get from AI queries.
You cannot rely on the answers provided by the mainstream AI implementations. Specifically, Microsoft Copilot, Meta AI, Google Gemini, and ChatGPT (and others, I suspect) will return responses with important information missing, self-contradictory responses, responses that differ in important ways from each other, and responses that are not consistent with well-known facts.
I’m not saying that you shouldn’t use those tools, only that you must verify any information you get from them. That means reviewing the references they supply and making up your own mind what the truth is.
In short, AI is a great way to find references, and the summaries the AI responses produce will surface relevant search terms that you can follow up on. But know up front that the summaries will be flawed, and you cannot depend on them to be truthful.
Thinking otherwise can be dangerous. Let me give one example.
I had cause today to look up information about armadillos and leprosy. My go-to for quick information is Wikipedia. The Armadillo article tells me:
Armadillos are often used in the study of leprosy, since they, along with mangabey monkeys, rabbits, and mice (on their footpads), are among the few known species that can contract the disease systemically.
Okay. So that confirms what I thought: armadillos can carry leprosy. I didn’t know about the others, though. So I learned something new. Cool.
And, yes, I’m aware that it’s not a good thing to depend on Wikipedia as my only source of information. I don’t. However, my experience over the last 20 years is that Wikipedia is a very reliable source for factual information. I’ve rarely found a case where a Wikipedia article is flat wrong about something, and even more rarely have I found a Wikipedia article that is self-contradictory. An article might miss some important information or contain an unsupported statement here or there, but for the most part I’ve found articles on Wikipedia to be very high quality. In addition, Wikipedia articles generally supply an exhaustive list of references that I can (and do!) go to if I want to confirm something. In this particular case, I did consult the references for the pertinent facts that I’m presenting here. I’m satisfied that the one sentence I quoted above from the Wikipedia page on armadillos is true, if perhaps incomplete.
Anyway.
I’ve been playing with the AI products from Meta, Google, Microsoft, and OpenAI. So I thought I’d ask each of them the same question: Can armadillos carry leprosy? The responses I got differed, and in some cases contradicted each other and what I’d learned from the Wikipedia article. So I thought I’d ask a second question: What animals can carry leprosy? The table below gives a brief summary of the answers. Full text of both answers from each of the AIs is at the end of this post.
Can armadillos carry leprosy?
What animals can carry leprosy?
Copilot
Only armadillos can carry leprosy.
Nine-banded armadillos, red squirrels, and chimpanzees can carry leprosy.
Meta AI
Armadillos are one of few animals that can carry leprosy.
Only armadillos can carry leprosy. Red squirrels, chimpanzees, and soil amoeba also can carry leprosy.
Gemini
Only armadillos can carry leprosy.
Nine-banded armadillos, red squirrels, and chimpanzees can carry leprosy.
ChatGPT
Armadillos are one of few known animals that can carry leprosy.
Armadillos, mangabey monkeys, chimpanzees, and nine-banded armadillos can carry leprosy.
Wikipedia
Armadillos are one of few known animals that can carry leprosy.
Armadillos, mangabey monkeys, rabbits, and mice can carry leprosy.
Summary of AI and Wikipedia responses
Do you understand now why you can’t rely on any one source? The AIs are great tools for surfacing potentially useful information. But you should be able to see from the table above that not one of them (including the non-AI Wikipedia) provides complete and self-consistent information. And they’re all different in important ways.
Which one do you trust?
None of them. The AIs are very good at surfacing and summarizing information. But the information they surface and the summaries they provide are not completely reliable. They, like Wikipedia, are just the first step in learning about something new. You have to follow up and verify every assertion made. You have to identify the conflicting assertions and resolve them. That means reading and evaluating the references that the AIs (and the Wikipedia article) provide.
The primary difference between a Wikipedia article and the AI responses is that last step: human intervention. The Wikipedia article is created by people who actually know what they’re writing about. They found the sources (sometimes authored one or more of the references), read and evaluated them, determined through various means which sources were reliable, and put all the information together into a meaningful article that somebody less knowledgeable than the author (or authors–often many different people contribute to a single Wikipedia article) can glean useful information from.
The AI responses, on the other hand, are 100% computer generated. These AIs have at their disposal what is essentially the sum total of all published human knowledge, including verifiable fact, speculation, fiction, and falsehoods that were intentionally created to deceive. But they have no way to evaluate the veracity of any particular statement. They don’t know fact from fantasy. Can’t separate truth from fiction. All they can do is say, in effect “Lots of people say this, and others say that.” Then they put it all together into something that looks like a detailed answer to your question. Actually, it is an answer to your question: an unreliable answer written by a brilliant idiot.
I’ve come to think of these AIs as consummate bullshit artists. They are adept at seamlessly mixing truth and falsehood together into a (usually) coherent whole that looks fine on the surface but doesn’t bear close scrutiny. It’s up to you to separate the fact from fiction.
Full responses to both of my questions from each of the AIs are provided below.
Question 1: Can armadillos carry leprosy?
Microsoft Copilot
Yes, armadillos can carry leprosy. Specifically, the nine-banded armadillo is known to harbor the bacteria Mycobacterium leprae, which causes leprosy. This makes them the only animals besides humans known to carry the disease.
Transmission to humans can occur through direct contact with armadillos or by handling or consuming their meat. However, the risk of contracting leprosy from armadillos is relatively low, and the disease is now highly treatable.
Meta AI
Yes, armadillos are one of the few animals that can carry leprosy, also known as Hansen’s disease. They are natural carriers of the bacteria that causes leprosy, Mycobacterium leprae. While the risk is low, humans can contract leprosy from armadillos through contact with their saliva, urine, or other bodily fluids. It’s important to exercise caution and avoid direct contact with armadillos, especially if you live in areas where leprosy is common.
Here are some facts about leprosy and armadillos:
Leprosy is rare: Leprosy is a relatively rare disease in the United States, with only a few cases reported each year.
Armadillos are common carriers: Armadillos are one of the primary animal carriers of leprosy in the southern United States.
Contact with armadillos increases risk: People who handle or eat armadillos, or come into contact with their bodily fluids, are at higher risk of contracting leprosy.
Early treatment is key: Leprosy is treatable with antibiotics, but early detection and treatment are crucial to prevent long-term nerve damage and disability.
Google Gemini
Yes, armadillos can carry leprosy. In fact, they are the only other mammal besides humans that can naturally harbor the bacteria that causes leprosy, Mycobacterium leprae.
Studies have shown that a high percentage of armadillos in the southern United States, particularly nine-banded armadillos, carry the leprosy bacteria. In some areas, estimates are that 15-20% of armadillos are infected. While the exact way leprosy is transmitted from armadillos to humans is not fully understood, it is believed that close contact with armadillos or their environment may be a risk factor.
Here are some ways people might come into contact with the leprosy bacteria from armadillos:
Direct contact: Skinning or handling an armadillo with bare hands.
Inhalation: Breathing in dust or dirt contaminated with armadillo droppings.
Ingestion: Eating undercooked armadillo meat.
Leprosy is a chronic infectious disease that attacks the skin, nerves, eyes, and mucous membranes. However, it is important to note that most people are not susceptible to leprosy and the risk of transmission from armadillos is considered low.
If you are concerned about leprosy, talk to your doctor. Early diagnosis and treatment can prevent permanent nerve damage and disability.
ChatGPT
Yes, armadillos can carry leprosy. In fact, they are one of the few known animal species that can transmit the disease to humans. Leprosy, also known as Hansen’s disease, is caused by the bacteria Mycobacterium leprae. Armadillos can harbor this bacteria, and transmission to humans usually occurs through direct contact with armadillos or their bodily fluids.
In regions where leprosy is endemic and armadillos are present, there have been documented cases of people contracting leprosy from handling or consuming armadillos. Therefore, it is advised to avoid contact with armadillos, especially in areas where leprosy is known to occur, to reduce the risk of transmission.
Question 2: What animals can carry leprosy?
Microsoft Copilot
Leprosy, also known as Hansen’s disease, primarily affects humans, but certain animals can also carry the bacteria Mycobacterium leprae. Here are some animals known to carry leprosy:
Nine-banded armadillos: These are the most well-known animal carriers of leprosy, particularly in the southern United States.
Red squirrels: In the UK, red squirrels have been found to carry strains of leprosy, although there have been no reported cases of transmission to humans.
Chimpanzees: Some wild chimpanzees in Guinea-Bissau and Ivory Coast have been observed with leprosy.
These animals act as reservoirs for the bacteria, but the risk of transmission to humans is generally low.
Meta AI
The following animals can carry leprosy:
Armadillos: The only other animals that carry leprosy, besides humans, are armadillos. They are the natural carriers of the bacteria that causes leprosy, Mycobacterium leprae.
Red squirrels: They are known to develop leprosy in the UK, but there have been no recorded cases of transmission from red squirrels to humans.
Chimpanzees: There have been recorded cases of leprosy in chimpanzees in Guinea-Bissau and Ivory Coast. However, more research is needed to determine if humans can contract leprosy from chimps.
Soil amoeba: Research by The Leprosy Mission in India demonstrated that soil amoeba can also be a reservoir for leprosy bacteria.
Google Gemini
There are only a few known animal species that can carry leprosy, and among them, the nine-banded armadillo is the most well-documented case.
Nine-banded armadillo: This armadillo species, found in the southern United States and parts of South America, is the only mammal besides humans known to naturally harbor the bacteria that causes leprosy, Mycobacterium leprae. In fact, some armadillo populations can have infection rates as high as 20%. While the armadillos themselves rarely show symptoms from the infection, they can transmit it to humans through close contact or consumption of armadillo meat.
Red Squirrels: There have been some isolated cases of leprosy identified in red squirrel populations in the United Kingdom. However, there is no evidence that squirrels can transmit the disease to humans.
Chimpanzees: A recent study (late 2021) identified leprosy in wild chimpanzees in West Africa. More research is needed to determine if chimpanzees can transmit the disease to humans.
ChatGPT
Several animals have been identified as potential carriers of leprosy (Mycobacterium leprae), including:
Armadillos: Armadillos are well-documented carriers of leprosy and have been implicated in transmitting the disease to humans in some regions, particularly in the southern United States.
Non-human Primates: Certain species of non-human primates, such as chimpanzees and mangabey monkeys, have been found to carry Mycobacterium leprae. They can potentially transmit the disease to humans through close contact.
Nine-banded Armadillos: This specific species of armadillo, found primarily in the Americas, is particularly known for its association with leprosy.
While these animals can carry Mycobacterium leprae, it’s important to note that human-to-human transmission is the primary route of leprosy infection worldwide. In regions where leprosy is endemic and where these animals are present, there may be a risk of zoonotic transmission (animal to human), especially through direct contact or consumption of infected animal tissues.
Very early in our relationship, Debra and I got silly one night and were making fun of country music. I don’t know what prompted it but at some point I came up with a song title: “Rock Me Baby, ’til My Feet Come Off.” We couldn’t stop laughing. I know, it was one of those, “you had to be there” moments.
That was 30 years ago and I still think that somebody should come up with lyrics. It could be a hit!
I’ve been poking at Microsoft’s Bing Chat (powered by ChatGPT) and Google’s Bard recently, so I thought I’d see what they can do with that title. My prompt to each one was ‘write lyrics for a country song called “Rock Me Baby, ’til My Feet Come Off“.
Bing Chat
Here’s a country song called “Rock Me Baby, ’til My Feet Come Off”:
Verse 1: I’ve been working all day long My feet are tired and my back is sore I need someone to take me home And rock me baby, ’til my feet come off
Chorus: Rock me baby, ’til my feet come off Hold me tight and never let me go Rock me baby, ’til my feet come off And I’ll love you more than you’ll ever know
Verse 2: I’ve been waiting for this moment To hold you close and feel your love I need someone to take me home And rock me baby, ’til my feet come off
Chorus: Rock me baby, ’til my feet come off Hold me tight and never let me go Rock me baby, ’til my feet come off And I’ll love you more than you’ll ever know
Bridge: I don’t need no fancy things Just your love is all I need So come on baby, take my hand And rock me baby, ’til my feet come off
Chorus: Rock me baby, ’til my feet come off Hold me tight and never let me go Rock me baby, ’til my feet come off And I’ll love you more than you’ll ever know
Outro: Rock me baby, ’til my feet come off Rock me baby, ’til my feet come off.
Google Bard
(Verse 1) Rock me baby, ’til my feet come off I’m so tired, I can’t stand no more I’ve been working all day, and I’m beat Just rock me baby, and make me feel complete
(Chorus) Rock me baby, ’til my feet come off I’m so tired, I can’t sleep at night I just lay there and toss and turn And dream of being rocked in your arms
(Verse 2) Rock me baby, ’til my troubles are gone I’m so worried, I can’t catch my breath I’ve got a lot on my mind, and I don’t know what to do Just rock me baby, and make all my troubles new
(Chorus) Rock me baby, ’til my feet come off I’m so tired, I can’t sleep at night I just lay there and toss and turn And dream of being rocked in your arms
(Bridge) Rock me baby, ’til I can’t feel my bones I’m so worn out, I’m just about gone But I know that if you rock me long enough I’ll be all better, and I’ll be able to go on
(Chorus) Rock me baby, ’til my feet come off I’m so tired, I can’t sleep at night I just lay there and toss and turn And dream of being rocked in your arms
(Outro) Rock me baby, ’til the morning light I’m so tired, I can’t see the night Just rock me baby, and make me feel all right I’ll be your baby, and you’ll be my knight
Well of course they’re terrible. But then, it’s country music.
I don’t think music writers have anything to worry about quite yet . . .
I can’t tell how serious writers and actors are about their concerns with AI. They’re asking for some unspecified safeguards on the use of AI, or regulations, or something. In general, some language to assure them that their jobs will not be taken over by “AI.”
I think it’s ridiculous, but perhaps I’m attributing to the writers and actors things that have been overblown by the media or by the general public’s hysterical reaction to anything that somebody calls “AI.” Or algorithms in general. As far as all too many people are concerned, any “algorithm” is automatically evil and out to do us harm.
I base my ridicule on three things. First, people have been protesting new technology since the dawn of new technology. Two hundred years ago, the original Luddites destroyed equipment in textile mills in protest of automation, but they weren’t the first to protest automation. Strangely enough, the machines didn’t put them out of work. And yet protests against automation were common throughout the industrial revolution and continue to this day. Computers, for example, were going to put armies of clerical workers out of a job. But now, 70 years into the computer revolution, there are more clerical jobs than ever. There are cases in which automation has made certain jobs irrelevant, but it doesn’t happen overnight. And there continues to be need of the replaced skill for some time.
Second, the idea of artificial intelligence replacing a journalist, screenwriter, actor, programmer, or any other skilled human is laughable. As I’ve mentioned before, ChatGPT (which I think is what has gotten everybody up in arms) and similar tools are just mimics: they rearrange words in a blender and spit them out semi-randomly, following rules that dictate the form, but with no regard to substance. And that’s just regurgitating stuff that’s already known. Attempts at AI creativity–having the computer create something novel–are comically terrible. The idea of a generative AI replacing a human writer just isn’t realistic. Certainly not within my lifetime, and likely not within the lifetime of anybody born today.
Third, if somebody does develop an AI that can produce objectively better news stories, movie scripts, novels, acting performances, computer programs, etc. than a human, then more power to them! As long as I’m informed or entertained, I don’t particularly care who or what created the article or the performance. We all benefit from better expression of ideas, and those whose skills are better performed by artificial intelligence will either find something else to do that is not yet AI-dominated, or will be able to peddle their skills in a smaller and often more lucrative market. For certain, any actor who’s pushed out of the big studios by this future fanciful AI will have plenty of opportunity in smaller studios that can’t afford or don’t want to use AI technologies.
Yes, there is some justifiable concern that studios will use currently available techniques, and new techniques going forward, to unscrupulously manipulate an actor’s voice, image, or performance to say things that the actor never intended or agreed to. We’ve all seen those agreements that allow the companies to use our likeness in any way, shape, or form, in perpetuity. Those types of clauses should have been eliminated from contracts decades ago, and I support those who are trying to address that situation now. But beyond that, the fears about AI replacing skilled workers, especially skilled creatives, are unfounded.
I get it: translation is hard. Heck, I’m a reasonably bright native English speaker and often have difficulty translating my own thoughts into understandable English.
This is a message that was posted in a woodcarving group:
“Hello, I am writing a message to help my father. And I see myself. Only on the American or Canadian woodcarving site and no response. It’s just for the books. And politeness. It’s when it’s repetitive that it’s not funny. But you how many millions to be connected. I find that very embarrassing. Administrators must take their jobs seriously. I have already reported them, I pass the imfermire contest as if I was going to sew up a person at any time, have a nice day everyone.”
The author’s native language is, I think, Italian. Or perhaps French. I suspect not an English speaker, although it’s possible that his grasp of English is better than my grasp of his native language. I cannot tell if the message is the result of automatic translation, or if the author did the translation himself with the help of a Italian-to-English dictionary. Either way, I cannot make any sense of it.
Which is weird. I’ve seen bad translations before. But usually I can get the gist of a message that’s been automatically translated: a “hook” that gives me a broad idea, and from there puzzle out a few details. For example, the word “imfermire” in the above text looked promising. It looks like a misspelling of the Italian word “infermiere,” meaning “nurse.” The best I can guess is that the author is having trouble getting some woodcarving books for his father. Not sure where the nurse comes in.
The author’s responses to comments provide no useful information. Which isn’t too much of a surprise. I imagine he has to translate the question, then write and translate a response. The combined errors inherent in that process aren’t conducive to understanding. Automatic translation software is especially bad at round-tripping because errors accumulate very quickly.
Can the technology that powers the new crop of generative AIs be put to good use in the automatic translation space? I imagine feeding an Italian-to-English translation to a tool that can leverage its knowledge of translation errors and spit out a short and meaningful summary. Is such a tool within our grasp?
It struck me today that human language doesn’t exist in a vacuum. Obvious, I know, but hear me out. Languages have evolved over thousands of years in response to the ever-increasing human body of knowledge. Early languages were simple because they only had to express simple concepts. As humans began learning more and developing more advanced concepts, we extended our languages to express those concepts. This evolution of language continues today, despite some governments’ attempts to control their official languages. So what’s the point?
Years ago I thought I wanted to study natural language processing–teaching a computer to understand and use written human languages rather than cryptic one- or two-word commands. I bought a couple of books and did a few weeks of research, and decided that I couldn’t do the topic justice on a part-time “for fun” basis. At the time (15 or so years ago), most of the attempts I saw were dictionary-based, involving a database of words and common phrases somehow encoded with their associated meanings. The computer scans some text for recognizable words and phrases, looks up the meanings, and provides an appropriate response. The major problem with this approach is that it’s limited by the size and accuracy of the database. More importantly, this approach doesn’t allow the computer to learn new words or new meanings. Why? Context.
Context is the key to understanding natural languages. For humans, the context is the world and communicating with the people in it. We don’t learn that the word “no” by having somebody tell us the definition, but rather by being told “No!” when we do something wrong. To a child, the word “no” means “don’t do that.” It’s only later that we understand that “no” is a negative response to a question. But computer programs have no understanding of things, people, and places, and thus no real understanding of our context, which I think is required to understand our languages. The question researchers should be asking (I don’t know if they are, as it’s been a long time since I studied the field) is whether it’s even possible to teach a computer to understand context.