CalFire firefighters cut a fire line during an uncontrolled fire using hand tools and chain saws at the Hughes Fire in Castaic, on Jan. 22, 2025. Photo by David McNew, Getty Images
In summary
The bot fails at some basic questions about fires. Cal Fire says it is working on fixes.
California government agencies are going all-in on generative artificial intelligence tools after Gov. Gavin Newsom’s 2023 executive order to improve government efficiency with AI. One deployment recently touted by the governor is a chatbot from the California Department of Forestry and Fire Protection, the primary agency tasked with coordinating the state’s wildfire response.
The chatbot, which Cal Fire says is independent of Newsom’s order, is meant to give Californians better access to “critical fire prevention resources and near-real-time emergency information,” according to a May release from Newsom’s office. But CalMatters found that it fails to accurately describe the containment of a given wildfire, doesn’t reliably provide information such as a list for evacuation supplies and can’t tell users about evacuation orders.
Newsom has announced AI applications for traffic, housing and customer service to be implemented in the coming months and years. But Cal Fire’s chatbot issues raise questions about whether agencies are following best practices.
“Evaluation is not an afterthought,” said Daniel Ho, law professor at Stanford University whose research focuses on government use of AI. “It should be part of the standard expectation when we pilot and roll out a system like this.”
The chatbot uses the Cal Fire website and the agency’s ReadyForWildfire.org to generate answers. It can tell users about topics such as active wildfires, the agency, fire preparedness tips and Cal Fire’s programs. It was built by Citibot, a South Carolina-based company that sells AI-powered chatbots for local government agencies across the country. Cal Fire plans to host the tool until at least 2027, according to procurement records.
“It really was started with the intent and the goal of having a better-informed public about Cal Fire,” said Issac Sanchez, deputy chief of communications for the agency.
When CalMatters asked Cal Fire’s bot questions about what fires were currently active and basic information about the agency, it returned accurate answers. But for other information, CalMatters found that the chatbot can give different answers when the wording of the query changes slightly, even if the meaning of the question remains the same.
For example, an important way Californians can prepare for fire season is assembling a bag of emergency supplies should they need to evacuate. Only “What should I have in my evacuation kit?” returned a specific list of items from Cal Fire’s chatbot. Variations of the question that included “go bag,” “wildfire ready kit” and “fire preparedness kit” instead returned either a prompt to visit Cal Fire’s “Ready for Wildfire” site, which has that information, or a message saying “I’m not sure about the specific items you should have” and the wildfire site link. Two of those terms are present on the site the chatbot referenced.
And while the chatbot didn’t generate incorrect answers in any of the queries CalMatters made, it doesn’t always pull the most up-to-date information.
When asked if the Ranch Fire, a 4,293-acre fire in San Bernardino County, was contained, the chatbot said that the “latest” update as of June 10 showed the fire was 50% contained. At the time CalMatters queried the chatbot, the information was six days out of date – the fire was 85% contained by then.
Similarly, when asked about current job openings at the agency, the chatbot said there weren’t any. A search on the state’s job site showed two positions at Cal Fire accepting applications at the time.
Mila Gascó-Hernandez is research director for the University at Albany’s Center for Technology in government and has studied how public agencies use AI-powered chatbots. Two key factors she uses to evaluate such chatbots are the accuracy of information they provide and how consistently they answer the same questions even if the question is asked in different ways.
“If a fire is coming and you need to know how to react to it, you do need both accuracy and consistency in the answer,” she said. “You’re not going to think about ‘what’s the nice way to ask the chatbot?’”
Currently, the chatbot is unable to provide information about evacuation orders associated with fires. When asked who issues evacuation orders, it sometimes correctly said law enforcement, while other times said it didn’t know. Cal Fire’s Sanchez said it’s reasonable to expect the chatbot to be able to answer questions about evacuations.
If there are no evacuation orders for a particular fire, he said, “the answer should be ‘there doesn’t appear to be any evacuations associated with this incident.’”
Sanchez said he and his team of about four people tested the chatbot before it went out by submitting questions they expected the public to ask. Cal Fire is currently making improvements to the bot’s answers by combing through the queries people make and ensuring that the chatbot correctly surfaces the needed answer.
When CalMatters asked the bot “What can you help me with?” in early May, it responded, “Sorry I don’t have the answer to that question right now” and asked if CalMatters had questions about information on the Cal Fire site. By mid-June, that answer was updated to being able to “provide answers to questions related to information located on this page such as details about current fires, CAL FIRE job classifications, examination requirements and CAL FIRE’s various programs.”
“The big message we want to get across,” Sanchez said, “is be patient.”
But experts said the process of kicking the tires on a chatbot should happen long before procurement begins.
The preferred process, Stanford’s Ho said, is to establish criteria for how the chatbot should perform before a vendor is selected so there are clear benchmarks to evaluate the tool. Ideally, those benchmarks are created by an independent third party. There should also be an evaluation of the benefits and risks before the chatbot is released.
And in a best-case scenario, the public would be involved before launch, Albany’s Gascó-Hernandez said. Agencies interested in using chatbots should identify the questions the public is likely to ask the AI tool ahead of time, ensure those are representative of the expected population the agency serves and refine the chatbot by having members of the public pilot the system to ensure it provides the kind of information they seek.
“These user engagement and user experiences are very important so the citizen ends up using the chabot,” she said.
Malena Carollo investigates broken systems and wrongdoing.
Her most recent investigation found that a sweeping change to an algorithm deciding who gets a life-saving liver transplant hurt patients in... More by Malena Carollo
Republish
California launched an AI chatbot for wildfires. It can’t answer one crucial question
We love that you want to share our stories with your readers. Hundreds of publications republish our work on a regular basis.
All of the articles at CalMatters are available to republish for free, under the following conditions:
Give prominent credit to our journalists: Credit our authors at the top of the article and any other byline areas of your publication. In the byline, we prefer “By Author Name, CalMatters.” If you’re republishing guest commentary (example) from CalMatters, in the byline, use “By Author Name, Special for CalMatters.”
Credit CalMatters at the top of the story: At the top of the story’s text, include this copy: “This story was originally published by CalMatters. Sign up for their newsletters.” If you are republishing commentary, include this copy instead: “This commentary was originally published by CalMatters. Sign up for their newsletters.” If you’re republishing in print, omit the second sentence on newsletter signups.
Do not edit the article, including the headline,except to reflect relative changes in time, location and editorial style. For example, “yesterday” can be changed to “last week,” and “Alameda County” to “Alameda County, California” or “here.”
If you add reporting that would help localize the article, include this copy in your story: “Additional reporting by [Your Publication]” and let us know at republish@calmatters.org.
If you wish to translate the article, please contact us for approval at republish@calmatters.org.
Photos and illustrations by CalMatters staff or shown as “for CalMatters” may only be republished alongside the stories in which they originally appeared. For any other uses, please contact us for approval at visuals@calmatters.org.
Photos and illustrations from wire services like the Associated Press, Reuters, iStock are not free to republish.
Do not sell our stories, and do not sell ads specifically against our stories. Feel free, however, to publish it on a page surrounded by ads you’ve already sold.
Sharing a CalMatters story on social media? Please mention @CalMatters. We’re on X, Facebook, Instagram, TikTok and BlueSky.
If you’d like to regularly republish our stories, we have some other options available. Contact us at republish@calmatters.org if you’re interested.
Have other questions or special requests? Or do you have a great story to share about the impact of one of our stories on your audience? We’d love to hear from you. Contact us at republish@calmatters.org.
Gift this article
Cal Fire made an AI chatbot. Don't ask it about evacuations
The bot fails at some basic questions. Cal Fire says it is working on fixes. Experts wonder if it launched too soon.
CalMatters
California, explained
Malena Carollo
Malena Carollo investigates broken systems and wrongdoing. Her most recent investigation found that a sweeping change to an algorithm deciding who gets a life-saving liver transplant hurt patients in several poorer, more rural states in the South and Midwest. Other states, however, benefited significantly—including several whose transplant leaders orchestrated the change behind the scenes. The investigation was copublished with The Washington Post. Her previous work prompted an investigation by the Florida Attorney General’s office into an auto lender’s predatory practices, highlighted a state-sanctioned loan program saddling homeowners with significant debt, and showed that Florida’s largest power company was among the deadliest for customers. Her work was the basis for an episode of Last Week Tonight with John Oliver. Malena mentors other journalists through the Investigative Reporters & Editors program with Journalism Mentors. She earned her master’s degree at the Columbia University Graduate School of Journalism, where she studied investigative and data reporting and practical digital security. When she isn’t working, she enjoys hiking. She is based in North Carolina.