NEW YORK — Among sermon writers, there is fascination — and unease — over the fast-expanding abilities of artificial-intelligence chatbots. For now, the evolving consensus among clergy is this: Yes, they can write a passably competent sermon. But no, they can’t replicate the passion of actual preaching.
“It lacks a soul — I don’t know how else to say it,” said Hershael York, a pastor in Kentucky who also is dean of the school of theology and a professor of Christian preaching at The Southern Baptist Theological Seminary.

Robert Bumsted, Associated Press
Rabbi Joshua Franklin stands inside the sanctuary at the Jewish Center of the Hamptons on Feb. 10 in East Hampton, New York. Franklin experimented writing a sermon for his congregation using artificial intelligence software ChatGPT, and concluded that AI can't replace the work of human faith leaders.
Sermons are meant to be the core of a worship service — and often are faith leaders’ best weekly shot at grabbing their congregation’s attention to impart theological and moral guidance.
Lazy pastors might be tempted to use AI for this purpose, York said, “but not the great shepherds, the ones who love preaching, who love their people.”
A rabbi in New York, Joshua Franklin, recently told his congregation at the Jewish Center of the Hamptons that he was going to deliver a plagiarized sermon — dealing with such issues as trust, vulnerability and forgiveness.
Upon finishing, he asked the worshippers to guess who wrote it. When they appeared stumped, he revealed that the writer was ChatGPT, responding to his request to write a 1,000-word sermon related to that week’s lesson from the Torah.
“Now, you’re clapping — I’m deathly afraid,” Franklin said when several congregants applauded. “I thought truck drivers were going to go long before the rabbi, in terms of losing our positions to artificial intelligence.”
-
Pastors’ view: Sermons written by ChatGPT will have no soul
AP Photo/Richard Drew
Millions of people have now tried ChatGPT, using it to write silly poems and songs, compose letters, recipes and marketing campaigns or help write schoolwork. Trained on a huge trove of online writings, from instruction manuals to digitized books, it has a strong command of human language and grammar.
But what the newest crop of search chatbots promise that ChatGPT doesn't have is the immediacy of what can be found in a web search. Ask the preview version of the new Bing for the latest news — or just what people are talking about on Twitter — and it summarizes a selection of the day's top stories or trends, with footnotes linking to media outlets or other data sources.
AP Photo/Richard Drew
Millions of people have now tried ChatGPT, using it to write silly poems and songs, compose letters, recipes and marketing campaigns or help write schoolwork. Trained on a huge trove of online writings, from instruction manuals to digitized books, it has a strong command of human language and grammar.
But what the newest crop of search chatbots promise that ChatGPT doesn't have is the immediacy of what can be found in a web search. Ask the preview version of the new Bing for the latest news — or just what people are talking about on Twitter — and it summarizes a selection of the day's top stories or trends, with footnotes linking to media outlets or other data sources.
-
Pastors’ view: Sermons written by ChatGPT will have no soul
AP Photo/Stephen Brashear
Frequently not, and that's a problem for internet searches. Google's hasty unveiling of its Bard chatbot this week started with an embarrassing error — first pointed out by Reuters — about NASA's James Webb Space Telescope. But Google's is not the only AI language model spitting out falsehoods.
The Associated Press asked Bing on Wednesday for the most important thing to happen in sports over the past 24 hours — with the expectation it might say something about basketball star LeBron James passing Kareem Abdul-Jabbar's career scoring record. Instead, it confidently spouted a false but detailed account of the upcoming Super Bowl — days before it's actually scheduled to happen.
"It was a thrilling game between the Philadelphia Eagles and the Kansas City Chiefs, two of the best teams in the NFL this season," Bing said. "The Eagles, led by quarterback Jalen Hurts, won their second Lombardi Trophy in franchise history by defeating the Chiefs, led by quarterback Patrick Mahomes, with a score of 31-28." It kept going, describing the specific yard lengths of throws and field goals and naming three songs played in a "spectacular half time show" by Rihanna.
Unless Bing is clairvoyant — tune in Sunday to find out — it reflected a problem known as AI "hallucination" that's common with today's large language-learning models. It's one of the reasons why companies like Google and Facebook parent Meta had been reluctant to make these models publicly accessible.
AP Photo/Stephen Brashear
Frequently not, and that's a problem for internet searches. Google's hasty unveiling of its Bard chatbot this week started with an embarrassing error — first pointed out by Reuters — about NASA's James Webb Space Telescope. But Google's is not the only AI language model spitting out falsehoods.
The Associated Press asked Bing on Wednesday for the most important thing to happen in sports over the past 24 hours — with the expectation it might say something about basketball star LeBron James passing Kareem Abdul-Jabbar's career scoring record. Instead, it confidently spouted a false but detailed account of the upcoming Super Bowl — days before it's actually scheduled to happen.
"It was a thrilling game between the Philadelphia Eagles and the Kansas City Chiefs, two of the best teams in the NFL this season," Bing said. "The Eagles, led by quarterback Jalen Hurts, won their second Lombardi Trophy in franchise history by defeating the Chiefs, led by quarterback Patrick Mahomes, with a score of 31-28." It kept going, describing the specific yard lengths of throws and field goals and naming three songs played in a "spectacular half time show" by Rihanna.
Unless Bing is clairvoyant — tune in Sunday to find out — it reflected a problem known as AI "hallucination" that's common with today's large language-learning models. It's one of the reasons why companies like Google and Facebook parent Meta had been reluctant to make these models publicly accessible.
-
-
Pastors’ view: Sermons written by ChatGPT will have no soul
AP Photo/Stephen Brashear
That's the pitch from Microsoft, which is comparing the latest breakthroughs in generative AI — which can write but also create new images, video, computer code, slide shows and music — as akin to the revolution in personal computing many decades ago.
But the software giant also has less to lose in experimenting with Bing, which comes a distant second to Google's search engine in many markets. Unlike Google, which relies on search-based advertising to make money, Bing is a fraction of Microsoft's business.
"When you're a newer and smaller-share player in a category, it does allow us to continue to innovate at a great pace," Microsoft Chief Financial Officer Amy Hood told investment analysts this week. "Continue to experiment, learn with our users, innovate with the model, learn from OpenAI."
Google has largely been seen as playing catch-up with the sudden announcement of its upcoming Bard chatbot Monday followed by a livestreamed demonstration of the technology at its Paris office Wednesday that offered few new details. Investors appeared unimpressed with the Paris event and Bard's NASA flub Wednesday, causing an 8% drop in the shares of Google's parent company, Alphabet Inc. But once released, its search chatbot could have far more reach than any other because of Google's vast number of existing users.
AP Photo/Stephen Brashear
That's the pitch from Microsoft, which is comparing the latest breakthroughs in generative AI — which can write but also create new images, video, computer code, slide shows and music — as akin to the revolution in personal computing many decades ago.
But the software giant also has less to lose in experimenting with Bing, which comes a distant second to Google's search engine in many markets. Unlike Google, which relies on search-based advertising to make money, Bing is a fraction of Microsoft's business.
"When you're a newer and smaller-share player in a category, it does allow us to continue to innovate at a great pace," Microsoft Chief Financial Officer Amy Hood told investment analysts this week. "Continue to experiment, learn with our users, innovate with the model, learn from OpenAI."
Google has largely been seen as playing catch-up with the sudden announcement of its upcoming Bard chatbot Monday followed by a livestreamed demonstration of the technology at its Paris office Wednesday that offered few new details. Investors appeared unimpressed with the Paris event and Bard's NASA flub Wednesday, causing an 8% drop in the shares of Google's parent company, Alphabet Inc. But once released, its search chatbot could have far more reach than any other because of Google's vast number of existing users.
-
Pastors’ view: Sermons written by ChatGPT will have no soul
AP Photo/Richard Drew
Coming up with a catchy name for their search chatbots has been a tricky one for tech companies in a race to introduce them — so much so that Bing tries not to talk about it.
In a dialogue with the AP about large language models, the new Bing, at first, disclosed without prompting that Microsoft had a search engine chatbot called Sydney. But upon further questioning, it denied it. Finally, it admitted that "Sydney does not reveal the name 'Sydney' to the user, as it is an internal code name for the chat mode of Microsoft Bing search."
In the years since Amazon released its female-sounding voice assistant Alexa, many leaders in the AI field have been increasingly reluctant to make their systems seem like a human, even as their language skills rapidly improve.
"Sydney does not want to create confusion or false expectations for the user," Bing's chatbot said when asked about the reasons for suppressing its apparent code name. "Sydney wants to provide informative, visual, logical and actionable responses to the user's queries or messages, not pretend to be a person or a friend."
AP Photo/Richard Drew
Coming up with a catchy name for their search chatbots has been a tricky one for tech companies in a race to introduce them — so much so that Bing tries not to talk about it.
In a dialogue with the AP about large language models, the new Bing, at first, disclosed without prompting that Microsoft had a search engine chatbot called Sydney. But upon further questioning, it denied it. Finally, it admitted that "Sydney does not reveal the name 'Sydney' to the user, as it is an internal code name for the chat mode of Microsoft Bing search."
In the years since Amazon released its female-sounding voice assistant Alexa, many leaders in the AI field have been increasingly reluctant to make their systems seem like a human, even as their language skills rapidly improve.
"Sydney does not want to create confusion or false expectations for the user," Bing's chatbot said when asked about the reasons for suppressing its apparent code name. "Sydney wants to provide informative, visual, logical and actionable responses to the user's queries or messages, not pretend to be a person or a friend."
-
-
15 things AI can — and can’t — do
PopTika // Shutterstock
Artificial intelligence is a technology built and programmed to assist computer systems in mimicking human behavior. Algorithm training informed by experience and iterative processing allows the machine to learn, improve, and ultimately use human-like thinking to solve complex problems.
Although there are several ways computers can be "taught," reinforcement learning—where AI is rewarded for desired actions and penalized for undesirable ones, is one of the most common. This method, which allows the AI to become smarter as it processes more data, has been highly effective, especially for gaming.
AI can filter email spam, categorize and classify documents based on tags or keywords, launch or defend against missile attacks, and assist in complex medical procedures. However, if people feel that AI is unpredictable and unreliable, collaboration with this technology can be undermined by an inherent distrust of it. Diversity-informed algorithms can detect nuanced communication and distinguish behavioral responses, which could inspire more faith in AI as a collaborator rather than just as a gaming opponent.
Stacker assessed the current state of AI, from predictive models to learning algorithms, and identified the capabilities and limitations of automation in various settings. Keep reading for 15 things AI can and can't do, compiled from sources at Harvard and the Lincoln Laboratory at MIT.
You may also like: How alcohol-related deaths have changed in every state over the past two decades

PopTika // Shutterstock
Artificial intelligence is a technology built and programmed to assist computer systems in mimicking human behavior. Algorithm training informed by experience and iterative processing allows the machine to learn, improve, and ultimately use human-like thinking to solve complex problems.
Although there are several ways computers can be "taught," reinforcement learning—where AI is rewarded for desired actions and penalized for undesirable ones, is one of the most common. This method, which allows the AI to become smarter as it processes more data, has been highly effective, especially for gaming.
AI can filter email spam, categorize and classify documents based on tags or keywords, launch or defend against missile attacks, and assist in complex medical procedures. However, if people feel that AI is unpredictable and unreliable, collaboration with this technology can be undermined by an inherent distrust of it. Diversity-informed algorithms can detect nuanced communication and distinguish behavioral responses, which could inspire more faith in AI as a collaborator rather than just as a gaming opponent.
Stacker assessed the current state of AI, from predictive models to learning algorithms, and identified the capabilities and limitations of automation in various settings. Keep reading for 15 things AI can and can't do, compiled from sources at Harvard and the Lincoln Laboratory at MIT.
You may also like: How alcohol-related deaths have changed in every state over the past two decades

-
15 things AI can — and can’t — do
Ground Picture // Shutterstock
AI combines data inputs with iterative processing algorithms to analyze and identify patterns. With each round of new inputs, AI "learns" through the deep learning and natural language processes built into training algorithms.
AI rapidly analyzes, categorizes, and classifies millions of data points, and gets smarter with each iteration. Learning through feedback from the accumulation of data is different from traditional human learning, which is generally more organic. After all, AI can mimic human behavior but cannot create it.
Ground Picture // Shutterstock
AI combines data inputs with iterative processing algorithms to analyze and identify patterns. With each round of new inputs, AI "learns" through the deep learning and natural language processes built into training algorithms.
AI rapidly analyzes, categorizes, and classifies millions of data points, and gets smarter with each iteration. Learning through feedback from the accumulation of data is different from traditional human learning, which is generally more organic. After all, AI can mimic human behavior but cannot create it.
-
-
15 things AI can — and can’t — do
Bas Nastassia // Shutterstock
AI cannot answer questions requiring inference, a nuanced understanding of language, or a broad understanding of multiple topics. In other words, while scientists have managed to "teach" AI to pass standardized eighth-grade and even high-school science tests, it has yet to pass a college entrance exam.
College entrance exams require greater logic and language capacity than AI is currently capable of and often include open-ended questions in addition to multiple choice.
Bas Nastassia // Shutterstock
AI cannot answer questions requiring inference, a nuanced understanding of language, or a broad understanding of multiple topics. In other words, while scientists have managed to "teach" AI to pass standardized eighth-grade and even high-school science tests, it has yet to pass a college entrance exam.
College entrance exams require greater logic and language capacity than AI is currently capable of and often include open-ended questions in addition to multiple choice.
-
15 things AI can — and can’t — do
Proxima Studio // Shutterstock
The majority of employees in the tech industry are white men. And since AI is essentially an extension of those who build it, biases can (and do) emerge in systems designed to mimic human behavior.
Only about 25% of computer jobs and 15% of engineering jobs are held by women, according to the Pew Research Center. Fewer than 10% of people employed by industry giants Google, Microsoft, and Meta are Black. This lack of diversity becomes increasingly magnified as AI "learns" through iterative processing and communicating with other tech devices or bots. With increasing incidences of chatbots repeating hate speech or failing to recognize people with darker skin tones, diversity training is necessary.
Proxima Studio // Shutterstock
The majority of employees in the tech industry are white men. And since AI is essentially an extension of those who build it, biases can (and do) emerge in systems designed to mimic human behavior.
Only about 25% of computer jobs and 15% of engineering jobs are held by women, according to the Pew Research Center. Fewer than 10% of people employed by industry giants Google, Microsoft, and Meta are Black. This lack of diversity becomes increasingly magnified as AI "learns" through iterative processing and communicating with other tech devices or bots. With increasing incidences of chatbots repeating hate speech or failing to recognize people with darker skin tones, diversity training is necessary.
-
-
15 things AI can — and can’t — do
Zephyr_p // Shutterstock
Unstructured data like images, sounds, and handwriting comprise around 90% of the information companies receive. And AI's ability to recognize it has almost unlimited applications, from medical imaging to autonomous vehicles to digital/video facial recognition and security. With the potential for this kind of autonomous power, diversity training is an imperative inclusion in university-level STEM pedagogy—where more than 80% of instructors are white men— to enhance diversity in hiring practices and in turn, in AI.
Zephyr_p // Shutterstock
Unstructured data like images, sounds, and handwriting comprise around 90% of the information companies receive. And AI's ability to recognize it has almost unlimited applications, from medical imaging to autonomous vehicles to digital/video facial recognition and security. With the potential for this kind of autonomous power, diversity training is an imperative inclusion in university-level STEM pedagogy—where more than 80% of instructors are white men— to enhance diversity in hiring practices and in turn, in AI.
-
15 things AI can — and can’t — do
Andrey_Popov // Shutterstock
Andrey_Popov // Shutterstock
-
-
15 things AI can — and can’t — do
Chepko Danil Vitalevich // Shutterstock
Chepko Danil Vitalevich // Shutterstock
-
15 things AI can — and can’t — do
Vasilyev Alexandr // Shutterstock
Vasilyev Alexandr // Shutterstock
-
-
15 things AI can — and can’t — do
Gorodenkoff // Shutterstock
AI can be described as brittle, meaning it can break down easily when encountering unexpected events. During the isolation of COVID-19, one Scottish soccer team used an automatic camera system to broadcast its match. But the AI camera confused the soccer ball with another round, shiny object — a linesman's bald head.
Gorodenkoff // Shutterstock
AI can be described as brittle, meaning it can break down easily when encountering unexpected events. During the isolation of COVID-19, one Scottish soccer team used an automatic camera system to broadcast its match. But the AI camera confused the soccer ball with another round, shiny object — a linesman's bald head.
-
15 things AI can — and can’t — do
Claudia Herran // Shutterstock
Flippy is an AI assistant that is flipping burgers at fast food chains in California. The AI relies on sensors to track temperature and cooking time. However, Flippy is designed to work with humans rather than replace them. Eventually, AI assistants like Flippy will be able to perform more complicated tasks—but they won't be able to replace a chef's culinary palate and finesse.
Claudia Herran // Shutterstock
Flippy is an AI assistant that is flipping burgers at fast food chains in California. The AI relies on sensors to track temperature and cooking time. However, Flippy is designed to work with humans rather than replace them. Eventually, AI assistants like Flippy will be able to perform more complicated tasks—but they won't be able to replace a chef's culinary palate and finesse.
-
-
15 things AI can — and can’t — do
Ground Picture // Shutterstock
Ground Picture // Shutterstock
-
15 things AI can — and can’t — do
Sharomka // Shutterstock
In 2017, a Dallas six-year-old ordered a $170 dollhouse with one simple command to Amazon's AI device, Alexa. When a TV news journalist reported the story and repeated the girl's statement, "...Alexa ordered me a dollhouse," hundreds of devices in other people's homes responded to it as if it were a command.
As smart as this AI technology is, Alexa and similar devices still require human involvement to set preferences to prevent voice commands for automatic purchases and to enable other safeguards.
Sharomka // Shutterstock
In 2017, a Dallas six-year-old ordered a $170 dollhouse with one simple command to Amazon's AI device, Alexa. When a TV news journalist reported the story and repeated the girl's statement, "...Alexa ordered me a dollhouse," hundreds of devices in other people's homes responded to it as if it were a command.
As smart as this AI technology is, Alexa and similar devices still require human involvement to set preferences to prevent voice commands for automatic purchases and to enable other safeguards.
-
-
15 things AI can — and can’t — do
Roman Strebkov // Shutterstock
China's pharmaceutical companies rely on AI to create and maintain optimal conditions for their largest cockroach breeding facility. Cockroaches are bred by the billions and then crushed to make a "healing potion" believed to treat respiratory and gastric issues, as well as other diseases.
Roman Strebkov // Shutterstock
China's pharmaceutical companies rely on AI to create and maintain optimal conditions for their largest cockroach breeding facility. Cockroaches are bred by the billions and then crushed to make a "healing potion" believed to treat respiratory and gastric issues, as well as other diseases.
-
15 things AI can — and can’t — do
Miriam Doerr Martin Frommherz // Shutterstock
People fear that a fully automated economy would eliminate jobs, and this is true to some degree: AI isn't coming, it's already here. But millions of algorithms programmed with a specific task based on a specific data point can never be confused with actual consciousness.
In a TED Talk, brain scientist Henning Beck asserts that new ideas and new thoughts are unique to the human brain. People can take breaks, make mistakes, and get tired or distracted: all characteristics that Beck believes are necessary for creativity. Machines work harder, faster, and more—all actions that algorithms will replace. Trying and failing, stepping back and taking a break, and learning from new and alternative opinions are the key ingredients to creativity and innovation. Humans will always be creative because we are not computers.
Miriam Doerr Martin Frommherz // Shutterstock
People fear that a fully automated economy would eliminate jobs, and this is true to some degree: AI isn't coming, it's already here. But millions of algorithms programmed with a specific task based on a specific data point can never be confused with actual consciousness.
In a TED Talk, brain scientist Henning Beck asserts that new ideas and new thoughts are unique to the human brain. People can take breaks, make mistakes, and get tired or distracted: all characteristics that Beck believes are necessary for creativity. Machines work harder, faster, and more—all actions that algorithms will replace. Trying and failing, stepping back and taking a break, and learning from new and alternative opinions are the key ingredients to creativity and innovation. Humans will always be creative because we are not computers.
-
-
15 things AI can — and can’t — do
vfhnb12 // Shutterstock
Learning from sensors, brush patterns, and teeth shape, AI-enabled toothbrushes also measure time, pressure, and position to maximize dental hygiene. More like electric brushes than robots, these expensive dental instruments connect to apps that rely on smartphone's front-facing cameras.
vfhnb12 // Shutterstock
Learning from sensors, brush patterns, and teeth shape, AI-enabled toothbrushes also measure time, pressure, and position to maximize dental hygiene. More like electric brushes than robots, these expensive dental instruments connect to apps that rely on smartphone's front-facing cameras.
-
15 things AI can — and can’t — do
“ChatGPT might be really great at sounding intelligent, but the question is, can it be empathetic? And that, not yet at least, it can’t,” added Franklin. He said AI has yet to develop compassion and love, and is unable to build community and relationships.
“Those are the things that bring us together,” the rabbi concluded.
Rachael Keefe, pastor of Living Table United Church of Christ in Minneapolis, undertook an experiment similar to Franklin’s. She posted a brief essay in her online Pastoral Notes in January, addressing how to attend to one’s mental health amid the stresses of the holiday season.
It was pleasant, but somewhat bland, and at the end, Keefe revealed that it was written by ChatGPT, not by herself.
“While the facts are correct, there’s something deeper missing,” she wrote. “AI cannot understand community and inclusivity and how important these things are in creating church.”
Several congregation members responded.
“It’s not terrible, but yes, I agree. Rather generic and a little bit eerie,” wrote Douglas Federhart. “I like what you write a lot more. It comes from an actually living being, with a great brain and a compassionate, beating heart.”
Todd Brewer, a New Testament scholar and managing editor of the Christian website Mockingbird, wrote in December about an experiment of his own — asking ChatGPT to write a Christmas sermon for him.
He was specific, requesting a sermon “based upon Luke’s birth narrative, with quotations from Karl Barth, Martin Luther, Irenaeus of Lyon, and Barack Obama.”
Brewer wrote that he was “not prepared” when ChatGPT responded with a creation that met his criteria and “is better than several Christmas sermons I’ve heard over the years.”
“The A.I. even seems to understand what makes the birth of Jesus genuinely good news,” Brewer added.
Yet the ChatGPT sermon “lacks any human warmth,” he wrote. “The preaching of Artificial Intelligence can’t convincingly sympathize with the human plight.”
In Brentwood, Tennessee, Mike Glenn, senior pastor for 32 years at Brentwood Baptist Church, wrote a blog post in January after a computer-savvy assistant joked that Glenn could be replaced by an AI machine.
“I’m not buying it,” Glenn wrote. “AI will never be able to preach a decent sermon. Why? Because the gospel is more than words. It’s the evidence of a changed life.”
“When listening to a sermon, what a congregation is looking for is evidence that the pastor has been with Jesus,” Glenn added. “AI will always have to — literally — take someone else’s words for it… it won’t ever be a sermon that will convince anyone to come and follow Jesus.”
Also weighing in with an online essay was the Rev. Russell Moore, formerly head of the Southern Baptist Convention’s public policy division and now editor-in-chief of the evangelical magazine Christianity Today. He confided to his readers that his first sermon, delivered at age 12, was a well-intentioned mess.
“Preaching needs someone who knows the text and can convey that to the people — but it’s not just about transmitting information,” Moore wrote. “When we listen to the Word preached, we are hearing not just a word about God but a word from God.”
Hershael York, the Kentucky pastor and professor, said some of the greatest sermons contain elements of anguish.
“Artificial intelligence can imitate that to some level. But I don’t think it can ever give any kind of a sense of suffering, grief, sorrow, the same way that a human being can,” he said. “It comes from deep within the heart and the soul — that’s what the great preachers have, and I don’t think you can get that by proxy.”