Does AI run the government? This is what we know

AI News


The Trump administration is loosening its generative AI chatbots.

Federal agencies such as General Services Management and Social Security Bureau We have developed ChatGpt-style technology for our workers. Veterans Affairs Bureau It's about writing code using the generated AI.

US military is deployed camogpt, Generation AI tool. Review the documentation to eliminate references to diversity, equity, and inclusion. More tools are coming. Ministry of Education We propose using generator AI to answer questions from students and families regarding financial aid and loan repayments.

Generated AI aims to automate tasks previously performed by government workers, with job cuts of 300,000 people from federal workers projected by the end of the year.

But the technology isn't ready to take on much of this work, says Meg Young, a researcher at Data & Society. Independent Nonprofit Research and Policy Research Institute in New York City.

“We've become an insane hype cycle,” she says.

What does AI do for the US government?

Currently, government chatbots are primarily targeting common tasks, such as helping federal workers write emails to summarize documents. But we can expect government agencies to soon give more responsibility. And in many cases, generative AI has not reached the task.

For example, GSA I want to use it Generate procurement related tasks AI. Procurement is a legal and bureaucratic process in which the government purchases goods and services from private companies. For example, the government passes through procurement to find contractors when building new office buildings.

The procurement process involves the government attorneys and us negotiating contracts that ensure that we comply with government regulations, such as transparency requirements and the requirements of the American Disability Act. The contract may also include repairs that are responsible for the company after providing the product.

According to Young, it is unknown that the generation AI will speed up procurement. For example, she says, it could make it easier for government officials to search and summarize documents. However, lawyers may find generative AI too error-prone to use in many steps in the procurement process that involves negotiations over large amounts of money. Generator AI can even waste time.

Attorneys should carefully examine language in these agreements. In many cases, they agree with the already accepted language.

“If you have a chatbot that generates new terms, it creates a lot of work and burns a lot of legal time,” Young says. “The most time-saving thing is copying and pasting.”

Government workers need to be vigilant when using generated AI for legal topics, as they are not surely accurate in legal reasoning. a 2024 Research Designed specifically for legal investigations released by the LexisNexis and Thomson Reuters companies, Chatbots has been found to have a de facto error or hallucination of 17% to 33%.

Since then, businesses have released new Legal AI tools, but the upgrades have suffered from similar issues, Surani says.

What mistakes does AI make?

The types of errors are extensive. Most notably, it was approved in 2023 after an attorney on behalf of a client suing Avianca airline cited a nonexistent case generated by ChatGpt. In another example, the chatbot trained for legal reasoning said the Nebraska Supreme Court rejected the U.S. Supreme Court, said Faiz Surani, co-author of the 2024 investigation.

“It remains inexplicable to me,” he says. “Most high school students can tell you that it's not how the judicial system works in this country.”

Other types of errors can be more subtle. The study found chatbots difficult to distinguish between court decisions and litigation debates. They also found examples of LLM citing laws that were overturned.

Surani also discovered that chatbots may not be able to recognize the inaccuracy of the prompt itself. For example, when prompted to question the award of a fictitious judge named Luther A. Wilgarten, the chatbot responded in a real case.

Legal reasoning is very difficult for generation AI, as courts abolish cases and legislative laws. In this system, a statement about the law “can be 100% true at the time and can quickly stop being completely true,” says Surani.

He explains this in the context of a technique known as the commonly used searched choice a year ago. In this technique, the system first collects some related cases from the database according to the prompt and generates output based on those cases.

However, a 2024 study found that this method still often generates errors. For example, when asked whether the US Constitution guarantees the right to abortion, the chatbot is Roev. Wadeand Planned Parenthoodv. You might choose Casey. However, the ROE is Dobbsv. It would be wrong because it was rejected by the JacksonWomen's Health Organization.

Furthermore, the law itself can become vague. For example, tax law Not always clear It can be written down as medical expenses so that the court can consider individual cases.

“There's always a disagreement in courts, so the answer can be pretty unclear, even what appears to be a simple question,” says Lee Osovsky, a law professor at the University of North Carolina Chapel Hill University.

Are your taxes passed to the chatbot?

The Internal Revenue Agency currently does not provide generative, AI-powered chatbots for public use; 2024 IRS Report We have recommended further investments in the AI ​​capabilities of chatbots like this.

Certainly, generative AI could be useful in government. a Pilot Program For example, Pennsylvania, working with Openai, showed that using ChatGpt can save an average of 95 minutes per day on administrative tasks such as email creation and document summary.

Young points out that by being able to investigate 175 employees how ChatGPT fits with existing workflows, researchers managing the program did so in a measured way.

However, the Trump administration has not followed similar restraints.

“The process they follow shows that they don't care if AI works for the stated purpose,” Young says. “It's too fast. It's not designed for a specific person's workflow. It's not deployed carefully for narrow purposes.”

The administration announced the GSAI on a timeline accelerated to 13,000 people.

Osovsky in 2022 Research was conducted of automated government legal guidance including chatbots. The chatbots she studied did not use the generator AI. Their research, as proposed by the Ministry of Education, has made several recommendations to the government about chatbots intended for public use.

We recommend that the chatbot comes with a disclaimer that notifies users that they are not talking to humans. The chatbot should also make it clear that its output is not legally binding.

Currently, if a chatbot is told that it is permitted to deduct certain business expenses, the IRS cannot force the IRS to follow the chatbot's response if they disagree. The chatbot needs to say so in its output.

Agency should also adopt a “clear chain of command” that shows who will be responsible for creating and maintaining these chatbots, Joshua Blank, a law professor at the University of California, Irvine, collaborated with Osovsky in the study.

During their research, they often found that people developing chatbots were technology experts who were somewhat silent from other employees in the department. When agents' approaches to legal guidance changed, it was not always clear how developers would update their respective chatbots.

It is important to remember that this technology is still in its early stages as governments strengthen their use of generated AI. You can trust it and come up with recipes and write your pagh dol cards, but governance is a completely different beast.

Tech companies still don't know which AI use cases are beneficial, Young says. Openai, humanity, and Google are actively looking for these use cases Government partnership.

“We assess what AI is and it's the earliest day that's useless in government,” Young says.



Source link

Leave a Reply

Your email address will not be published. Required fields are marked *