India's largest platform and marketplace for GCCs & AI

Sign in

India's largest platform and marketplace for GCCs & AI

3AI Digital Library

What if AI scans legislation and allocates funds to agencies?

3AI February 15, 2021

New Treasury Department software points the way. But research suggests that it’s impossible to show that an artificial ‘superintelligence’ can be contained

.

If, like me, you’re worried about how members of Congress are supposed to vote on a stimulus bill so lengthy and complex that nobody can possibly know all the details, fear not — the Treasury Department will soon be riding to the rescue.

But that scares me a little too.

Let me explain. For the past few months, the department’s Bureau of the Fiscal Service has been testing software designed to scan legislation and correctly allocate funds to various agencies and programs in accordance with congressional intent — a process known as issuing Treasury warrants. Right now, human beings must read each bill line by line to work out where the money goes. If the program can be made to work, the savings will be significant.

Alas, there’s a big challenge. Plenty of tools exist for extracting data from HTML files (and, of course, XML files), but Congress initially publishes legislation only in PDF form; XML or HTML versions often arrive only weeks later. As many a business knows, scraping data from PDFs generally requires human intervention, leading to the possibility of copy errors. The trouble is that PDFs have no standard data format. Even “simple” methods for extraction generally are designed to work only if the data in question is already presented within the PDF in tabular form.

Treasury’s ambitious hope, however, is that its software, when fully operational, will be able to scan new legislation in its natural language form, figure out where the money is supposed to go and issue the appropriate warrants far more swiftly than humans could. The faster the warrants are issued, the sooner the agency that’s supposed to receive the money can start spending.

Pretty cool stuff.

Yet this snapshot of the future inspires a wicked train of thought. Suppose that the Treasury Department software — which you are free to describe as artificial intelligence or not, depending on your taste — is later replaced by a better program, then by a better one and finally by one that can mimic the working general intelligence of the human mind.

What’s to stop this future AI from deciding on its own that Congress was wrong to give another billion to Agency A when, in the judgment of the program, Agency B needs it more? The program makes a tiny adjustment in a gigantic spending bill, and given that nobody’s actually read it, nobody’s the wiser.

Sounds improbable, right? HAL 9000 meets “Person of Interest” meets Skynet?

Not so fast.

For technophiles like me, recent achievements in AI are exciting, even breathtaking. AI is credited with reorganizing supply chains to help overcome disruptions caused by the pandemic. Deep learning systems may be able to discover coronary plaques more accurately than clinicians.

So why worry? After all, most of those in the field, including my professors when I studied artificial intelligence as an undergraduate, are confident that tight programming will keep even the most advanced artificial intelligence from escaping the bounds set by its creators. (Think Isaac Asimov’s Laws of Robotics.)

But there have long been dissenters, even among the experts. The prospect of an out-of-control AI has haunted researchers in the field for almost as long as it’s haunted science fiction writers. One thinks of Joseph Weizenbaum’s “Computer Power and Human Reason,” published back in 1976, or even Norbert Wiener’s classic “God and Golem, Inc.,” based on lectures the author delivered in 1962.

All of which brings us to an unnerving paper published last month by six AI researchers who argue that it is impossible to show that an artificial “superintelligence” can be contained. The authors are an international group, representing universities in Germany, Spain, and Chile, as well as the U.S. According to their analysis, no matter how tightly an AI may be programmed, if it indeed possesses generalized reasoning skills “far surpassing” those of the most gifted humans, what they call “total containment” turns out to be incapable of formal proof.

Using what is known as computability theory, they hypothesize a superintelligent AI that incorporates a fundamental command never to harm humans. (Asimov again.) The programming will then require a function that decides whether a particular action will harm humans or not. They proceed to show that even if it’s possible “to articulate in a precise programming language” a perfect set of “control strategies” to implement this function, there’s no way to know for sure whether the strategies will in fact constrain the AI. (The proof, although technical, is rather elegant, and fun to read.)

Don’t get me wrong: I’m not arguing that the Treasury Department should abandon its quest for a system that extracts data from PDFs, any more than I’m suggesting that any of the countless researchers working on various aspects of AI should halt. I continue to find the prospect of true artificial intelligence as exciting as ever.

What concerns me, however, is the way that public critiques of AI tend to pick around the edges rather than go to the heart of the matter. We often charge nascent AI systems with enhancing bias — for example, by exacerbating rather than correcting disparities in the distribution of health care. Such issues are of undeniable public importance. But as the authors of the paper on computability remind us, you don’t have to be either a technophobe or a fan of apocalyptic steampunk sci-fi to see that the time for public conversation about the containability of AI is now, not later.

 

 

Picture from freepik.com

    3AI Trending Articles

  • The Impact of AI on Global Economy: Trends and Initiatives to Shape the Future

    Featured Article: Author: Princy Gupta, Data Analytics Leader, Honeywell Artificial Intelligence (AI) is revolutionizing the global economy, providing unprecedented opportunities for businesses and governments to thrive in the age of data. According to McKinsey’s latest report, AI is projected to contribute $13 trillion to the global economy by 2030, a figure equivalent to the output […]

  • AI readiness – CEO’s most strategic agenda

    For the larger mass of professionals, the words “artificial intelligence,” or AI, often conjure up images of robots, the sorts of robots that might someday take their jobs. But at the enterprise level, AI means something different. It has enormous power and potential: it can disrupt, innovate, enhance, and in many cases totally transform a business. […]

  • Is your Enterprise AI Ready: Strategic considerations for the CXOs

    Author: Sameer Dhanrajani, CEO at AIQRATE advisory & consulting | AI advisor, business builder | Author | Speaker |Columnist| President at 3AI | Ex-CSO: Fractal, Global Business Leader: Cognizant, Country Head: Fidelity National Financial At the enterprise level, AI assumes enormous power and potential , it can disrupt, innovate, enhance, and in many cases totally […]

  • BigBasket sues a bootstrapped startup Daily Basket for using the word basket in its name

    Founder of Daily Basket Ramesh Vel has launched a website – “bbisabully”, calling out the billion dollar startup. Vel wrote on the website that BigBasket is fighting with a mere two person startup and said that it doesn’t have the ‘mighty power & money’ BigBasket has. BigBasket sent Daily Basket a cease and desist letter, […]