CalFire firefighters cut a fire line during an uncontrolled fire using hand tools and chain saws at the Hughes Fire in Castaic, on Jan. 22, 2025. Photo by David McNew, Getty Images
En resumen
The bot fails at some basic questions about fires. Cal Fire says it is working on fixes.
California government agencies are going all-in on generative artificial intelligence tools after Gov. Gavin Newsom's 2023 executive order to improve government efficiency with AI. One deployment recently touted by the governor is a chatbot from the California Department of Forestry and Fire Protection, the primary agency tasked with coordinating the state’s wildfire response.
The chatbot, which Cal Fire says is independent of Newsom’s order, is meant to give Californians better access to “critical fire prevention resources and near-real-time emergency information,” according to a May release from Newsom’s office. But CalMatters found that it fails to accurately describe the containment of a given wildfire, doesn’t reliably provide information such as a list for evacuation supplies and can’t tell users about evacuation orders.
Newsom has announced AI applications for traffic, alojamiento y customer service to be implemented in the coming months and years. But Cal Fire’s chatbot issues raise questions about whether agencies are following best practices.
“Evaluation is not an afterthought,” said Daniel Ho, law professor at Stanford University whose research focuses on government use of AI. “It should be part of the standard expectation when we pilot and roll out a system like this.”
The chatbot uses the Cal Fire website and the agency’s ReadyForWildfire.org to generate answers. It can tell users about topics such as active wildfires, the agency, fire preparedness tips and Cal Fire’s programs. It was built by Citibot, a South Carolina-based company that sells AI-powered chatbots for local government agencies across the country. Cal Fire plans to host the tool until at least 2027, according to procurement records.
“It really was started with the intent and the goal of having a better-informed public about Cal Fire,” said Issac Sanchez, deputy chief of communications for the agency.
When CalMatters asked Cal Fire’s bot questions about what fires were currently active and basic information about the agency, it returned accurate answers. But for other information, CalMatters found that the chatbot can give different answers when the wording of the query changes slightly, even if the meaning of the question remains the same.
For example, an important way Californians can prepare for fire season is assembling a bag of emergency supplies should they need to evacuate. Only “What should I have in my evacuation kit?” returned a specific list of items from Cal Fire’s chatbot. Variations of the question that included “go bag,” “wildfire ready kit” and “fire preparedness kit” instead returned either a prompt to visit Cal Fire’s “Ready for Wildfire” site, which has that information, or a message saying “I’m not sure about the specific items you should have” and the wildfire site link. Two of those terms are present on the site the chatbot referenced.
And while the chatbot didn’t generate incorrect answers in any of the queries CalMatters made, it doesn’t always pull the most up-to-date information.
When asked if the Ranch Fire, a 4,293-acre fire in San Bernardino County, was contained, the chatbot said that the “latest” update as of June 10 showed the fire was 50% contained. At the time CalMatters queried the chatbot, the information was six days out of date – the fire was 85% contained by then.
Similarly, when asked about current job openings at the agency, the chatbot said there weren’t any. A search on the state’s job site showed two positions at Cal Fire accepting applications at the time.
Mila Gascó-Hernandez is research director for the University at Albany’s Center for Technology in government and has studied how public agencies use AI-powered chatbots. Two key factors she uses to evaluate such chatbots are the accuracy of information they provide and how consistently they answer the same questions even if the question is asked in different ways.
“If a fire is coming and you need to know how to react to it, you do need both accuracy and consistency in the answer,” she said. “You’re not going to think about ‘what’s the nice way to ask the chatbot?’”
Currently, the chatbot is unable to provide information about evacuation orders associated with fires. When asked who issues evacuation orders, it sometimes correctly said law enforcement, while other times said it didn’t know. Cal Fire’s Sanchez said it’s reasonable to expect the chatbot to be able to answer questions about evacuations.
If there are no evacuation orders for a particular fire, he said, “the answer should be ‘there doesn’t appear to be any evacuations associated with this incident.’”
Sanchez said he and his team of about four people tested the chatbot before it went out by submitting questions they expected the public to ask. Cal Fire is currently making improvements to the bot’s answers by combing through the queries people make and ensuring that the chatbot correctly surfaces the needed answer.
When CalMatters asked the bot “What can you help me with?” in early May, it responded, “Sorry I don’t have the answer to that question right now” and asked if CalMatters had questions about information on the Cal Fire site. By mid-June, that answer was updated to being able to “provide answers to questions related to information located on this page such as details about current fires, CAL FIRE job classifications, examination requirements and CAL FIRE’s various programs.”
“The big message we want to get across,” Sanchez said, “is be patient.”
But experts said the process of kicking the tires on a chatbot should happen long before procurement begins.
The preferred process, Stanford’s Ho said, is to establish criteria for how the chatbot should perform before a vendor is selected so there are clear benchmarks to evaluate the tool. Ideally, those benchmarks are created by an independent third party. There should also be an evaluation of the benefits and risks before the chatbot is released.
And in a best-case scenario, the public would be involved before launch, Albany’s Gascó-Hernandez said. Agencies interested in using chatbots should identify the questions the public is likely to ask the AI tool ahead of time, ensure those are representative of the expected population the agency serves and refine the chatbot by having members of the public pilot the system to ensure it provides the kind of information they seek.
“These user engagement and user experiences are very important so the citizen ends up using the chabot,” she said.
Malena Carollo investiga sistemas defectuosos y malas prácticas. Su investigación más reciente descubrió que un cambio radical en un algoritmo que decide quién recibe un trasplante de hígado que podría salvarle la vida... More by Malena Carollo
Republish
California launched an AI chatbot for wildfires. It can’t answer one crucial question
We love that you want to share our stories with your readers. Hundreds of publications republish our work on a regular basis.
All of the articles at CalMatters are available to republish for free, under the following conditions:
Give prominent credit to our journalists: Credit our authors at the top of the article and any other byline areas of your publication. In the byline, we prefer “By Author Name, CalMatters.” If you’re republishing guest commentary (example) from CalMatters, in the byline, use “By Author Name, Special for CalMatters.”
Credit CalMatters at the top of the story: At the top of the story’s text, include this copy: “This story was originally published by CalMatters. Inscribirse for their newsletters.” If you are republishing comentario, include this copy instead: “This commentary was originally published by CalMatters. Inscribirse for their newsletters.” If you’re republishing in print, omit the second sentence on newsletter signups.
Do not edit the article, including the headline,except to reflect relative changes in time, location and editorial style. For example, “yesterday” can be changed to “last week,” and “Alameda County” to “Alameda County, California” or “here.”
If you add reporting that would help localize the article, include this copy in your story: “Additional reporting by [Your Publication]” and let us know at republish@calmatters.org.
If you wish to translate the article, please contact us for approval at republish@calmatters.org.
Photos and illustrations by CalMatters staff or shown as “for CalMatters” may only be republished alongside the stories in which they originally appeared. For any other uses, please contact us for approval at visuals@calmatters.org.
Photos and illustrations from wire services like the Associated Press, Reuters, iStock are not free to republish.
Do not sell our stories, and do not sell ads specifically against our stories. Feel free, however, to publish it on a page surrounded by ads you’ve already sold.
Sharing a CalMatters story on social media? Please mention @CalMatters. We’re on X, Facebook, Instagram, TikTok and BlueSky.
If you’d like to regularly republish our stories, we have some other options available. Contact us at republish@calmatters.org if you’re interested.
Have other questions or special requests? Or do you have a great story to share about the impact of one of our stories on your audience? We’d love to hear from you. Contact us at republish@calmatters.org.
Gift this article
Cal Fire made an AI chatbot. Don't ask it about evacuations
The bot fails at some basic questions. Cal Fire says it is working on fixes. Experts wonder if it launched too soon.
CalMatters
California, explicó
Malena Carollo
Malena Carollo investiga sistemas defectuosos y malas prácticas. Su investigación más reciente descubrió que un cambio radical en un algoritmo que decide quién recibe un trasplante de hígado que salva vidas perjudicó a pacientes en varios estados más pobres y rurales del sur y el medio oeste. Sin embargo, otros estados se beneficiaron significativamente, incluidos varios cuyos líderes de trasplantes orquestaron el cambio entre bastidores. La investigación fue publicada conjuntamente con The Washington Post. Su trabajo anterior motivó una investigación de la oficina del Fiscal General de Florida sobre las prácticas predatorias de un prestamista de automóviles, destacó un programa de préstamos aprobado por el estado que cargaba a los propietarios de viviendas con una deuda significativa y mostró que la compañía eléctrica más grande de Florida era una de las más mortales para los clientes. Su trabajo fue la base de un episodio de Last Week Tonight con John Oliver. Malena asesora a otros periodistas a través del programa Investigative Reporters & Editors con Journalism Mentors. Obtuvo su maestría en la Escuela de Posgrado de Periodismo de la Universidad de Columbia, donde estudió informes de investigación y datos y seguridad digital práctica. Cuando no está trabajando, disfruta del senderismo. Vive en Carolina del Norte.