I think this treats the capabilities of LLMs far too credulously. In reality they are not nearly this good, one cannot simply dumb the annotated US Code and CFR into them and hope to get anything useful out.
Ironically this is precisely what LLMs would be good at - coordinating issues and highlighting differences across vast reams of structured legal text. Especially the newer models released the last few weeks which are far better at quantitative reasoning. We should welcome the attempt rather than preemptively try to shut it down.
So, I think it'll do mostly okay with the USC and CFR (though I bet you run into a bunch of weird problems as the sheer number of different definitions of the same words may drive it a bit bananas). Where I guess it will totally crap out is the caselaw on those. First, just accessing it all is going to be tricky, as I really doubt westlaw/lexis/casetext are going to sell their databases. Second, it's just a giant confusing mess. My guess is almost anything is going to come back 'well, it could be illegal, or not, depending on execution and a bunch of facts, including most notably, which judge you happen to pull and in which circuit you're doing this!'
I'd like it to work, even if it might put me out of a job, but I doubt it.
I think people see chatgpt and think AI solutions are just feeding data directly as a big PDF to the same LLM theyre using. I'd advise folks to look up agent layers that break these processes into discrete parts handled by discrete agents which drive actual industry-specific solutions. There's arguably nothing more suited to modern agentic AI than the legal field, where the gaps and how one part of the law impacts decisions or ramifications in the other are actually pretty clear when the pieces are all laid out. The idea that the law is a 'giant confusing mess' that will make LLMs return 'illegal' results is....ill informed.
It's probably a perception left over from when some lawyers got into trouble when the AI would make up cases (AI "hallucinations") in citations when they had the AI writing briefs that were presented to a court. It made the news a few times when the lawyers who did that got sanctioned. There are, as you say, specific AIs now for the legal field, and many hours of continuing legal education have been devoted to teaching lawyers how AI can be used safely.
Trying to reconcile Supreme Court jurisprudence would be challenging enough, much less taking on the rest of the federal courts... But the first task, I agree, is reconciling the Constitution, U.S. Code, and Code of Federal Regulations (CFR) to flag areas of conflict for review by human attorneys. Teaching an A.I. to abstract "the rule" from case law is probably not even necessary for DOGE at this point.
For example, DOGE co-lead Ramaswamy has said that part of their review is to determine what aspects of the administrative state's current operations conflict with the recent SCOTUS U.S. v. EPA "major questions" decision. Are there things that agencies are doing that implicate "major questions" and which lack a clear statement from Congress in the agency's enabling legislation to support them? In other words, does what the agency is doing conflict with the law which created the agency? Or in a constitutional sense, has the Executive Branch encroached on Congress' lawmaking power, seizing more regulatory discretion than was granted them?
As the DOGE project looks for cost savings and deregulatory targets, with the objective of shrinking the size and cost of the administrative state, things which agencies are doing that appear to implicate "major questions" but which lack legislative support are obvious choices to eliminate.
As another example, are there agencies which have accrued over the decades through legislation passed by successive Congresses whose areas of authority overlap or conflict, or which are redundant and could be consolidated? So that's another type of conflict. The DOGE project leads have talked about reducing the hundreds of agencies which currently exist to less than a hundred.
"The DOGE project leads have talked about reducing the hundreds of agencies which currently exist to less than a hundred."
These people don't know what these agencies are, don't know how they overlap, or how they came to exist in the first place. Shades of Rick Perry not understanding that DOE was responsible for so much of American nuclear policy/safety. We don't have to credit these tech bro libertarian fever dreams, we really don't.
Does a DOGE AI model really even need to consider what any other courts besides SCOTUS have said about what constitutional law is? Besides the text of the SCOTUS cases themselves, there's an abundance of commentary on Supreme Court cases to potentially draw on (including dissents), without the need for the AI to do case analysis itself. As far as comparing statutes or regulations to the Constitution, it could be useful to do it the way the Supreme Court would examine a statute (or regulation, in the absence of Chevron deference). The first pass might just be textual analysis.
I mean, the Supreme Court takes so few cases that the concern I hear is essentially never 'what will the Supreme Court think about this,' it's 'what will the District Court and 9th/5th (choose your bugaboo) Circuit think of this?'
Now, maybe if you're just trying to argue agency/agency lawyers into agreeing with you, that's the case you want to make, but as a practical matter I don't think 'the AI says you'll probably win in ~4-10 (Lopez Bright/Sackett) years if the Supreme Court takes it up,' convinces anyone.
This ignores existing rulings from lower courts which can control what agencies can do. It also sort of assumes that the Set is consistent across cases, which it isn't.
The primary challenge is simply building agents that handle discrete parts of the law and others that tell them what takes precedence and when and why. I imagine the company that builds those basic agents will be a lawyer-known name by this time next year.
strong disagree. also, all the argument and assumptions are based on one or two interaction and current state. We miss the point that as the time is passing, those are becoming more efficient and improving at large scale. so we can only play that card for limited time and some handful of instnaces.
This is completely wrong—as an LLM researcher I'm convinced the latest LLMs are extremely good at legal and policy analysis, and given that we see better models being released every 3-6 months on average we should expect significant improvement over DOGE's term.
Count me among the doubters. Musk hasn’t been able to effectively productize AI for Twitter. And he is known for announcing his ideas/intentions far too soon, so I wouldn’t expect him to be tight lipped about this if that’s what they’re doing. And finally, having worked legislatively, I can only imagine that designing an LLM to actually cut through the red tape is going to be far more complex than could be spun up in a short timeframe. To really cut through the mess, you’d have to consider case law, other precedents, etc. A lot of human knowledge would have to go into shaping the AI itself, which takes time.
What about being able to trace summary information back to its legal/regulatory sources? My impression is that AI algorithms, or LLMs, at any rate have to discard their original material and source references (?)
What kind of summary information? We may be talking about different use cases. FWIW, agencies usually cite to authorizing statute when making regulation. Even in sub-regulatory guidance.
An anecdote for the naysayers here. A couple weeks ago I was at a carahsoft event highlighting the impacts of the current admins PMA goal around CX. They focused on program level stuff - lower paper processing times at USCIS, more digital tools for HUD, call wait times down at SSA, etc.
During Q&A I asked the question of where the persistent CX work is happening. In industry companies have built huge systems connecting touchpoints of users across online and offline interactions,and use AI to help personalize experiences. This is well trodden ground of over a decade of available tools, systems, and best practices. Where was that happening in these agenies mandated to fix their overall CX issues?
The answer was blank states. The HUD guy talked about how personalization was bad due to equity gaps it raises. The DHS guy said 'huh that would be interesting to look at'
These problems are endemic, and I don't think that fed agencies truly understand just how far they are away from boilerplate capability in every other sector. Turning tail on new initiatives like DOGE pre emptively because you don't like the people or systems involved just continues building a moat around agency systems and processes from the 90s. The arguments for keeping our agencies working like they are now gets worse every day that AI advances.
These were the people leading CX at their agencies under the EO mandate that has been in place for years. OMB was there are well. If it's not then it's not happening, and its clearly not happening.
These sorts of things are actually policy choices we should be extremely critical about. The public sector is at the bottom of CX capabilities across every industry polled but instead of using tried and true methods they decided to take the money and put towards programs that they wanted to fix anyway, with the hand-wave of "it impacts constituents therefore it's CX." It's ridiculous and borderline negligent.
I’m not doubting that you could develop an algorithm to do this kind of work; I’m doubting that we have enough good data to feed that algorithm. DoD (for example) runs on PDFs and PPTs - not good structured data.
Show me you can find enough machine readable data to trace a JROC Requirement from validation through the acquisition process and out to TRL-9 deployment to an operational unit and I’ll be the first one to buy you a beer.
I understand the intellectual value of taking this initiative seriously.
But I've seen Elon Musk's raw, unvarnished thoughts for years now and I know what he thinks about everything under the sun. To put a fine point on it, he's an idiot, massively out of his depth, and seemingly hopped up on a serious cocktail of stimulants. He cares less about government efficiency than he seems to care about the size of breasts of women in video games.
I think if given the opportunity, what he wants to do to the federal government is what he did to Twitter/X, namely fire a ton of people whose jobs he doesn't understand. But if the experience of Twitter/X is prologue, then expect the experience of working with the federal government to get so much worse. Twitter/X is basically unusable right now; the feed is mostly low-grade spam, search doesn't function, notifications recommends nothing you're interested in (including recently sending you posts from Elon himself whether or not you follow him), bots are more prevalent than ever before he took over, and generally the real users are being overrun by vile out-and-proud Nazi accounts. Let's not forget that he's lost billions on this endeavor and his only path to recouping his investment is through some kind of corrupt exercise of his newfound influence with the government.
Oh, but he might tack some kind of slipshod AI product onto the government? Just what we don't need. Until AI can figure out its hallucination problem, these technologies can't reliably be used for government. We've seen this play out repeatedly with court cases, for instance, as lawyers and experts use AI to write filings or depositions and they turn out to cite completely fake sources.
So while I appreciate Jennifer's willingness to give this not-actually-a-government agency some benefit of the doubt, I think it's remarkably charitable to do so.
You’re going to need to at least have one person who has actually read it all to do some QA on AI outputs. (Raises hand)
I didn’t find it too bothersome to keep it all in my head (the 5000+ pages necessary to launch any digital system in govt) but it became very clear that I was in the minority in that regard. Eventually I would figure out many COs couldn’t even be bothered to memorize their default procurement clauses.
I can't imagine anything that will generate more anti-AI sentiment than DOGE suddenly unleashing opaque and likely untrustworthy Musk AI on government rules and regs. His mode now is break things and move fast; the first problematic recommendation will among other things drive states to each self-regulate AI inconsistently.
I hope you're right...but I do not believe that is likely. I think you can produce a medium useful tool on the CFR/USC that provides some places to start, though the history of lawyers using AI to try to determine what the law is is a minefield of blown up careers.
But frankly, for most of the questions people actually want to ask, the answer is either--yes, the annoying thing compliance wants you to do is actually a requirement. Unless you can call it an emergency, you have to do 106 compliance on any undertaking, sorry and unless you can get to no potential to effect, you have to consult on it.
Or, 'it's grey, do you want to take the risk of being sued and potentially losing?' And there you run into the problem that leadership tends to, even more than the lawyers, hate being sued and losing. To be fair, they also hate being sued and winning. Or suing. I hate to sound like a broken record, but I think if you can unjam the judiciary and make things actually move efficiently there, things get a lot less painful on most fronts. Like, an APA case is never going to be instantaneous, but for most big actions, an agency knows if a lawsuit is coming. If you can rely on a court to move it along fast, I bet you see a lot more agencies building the admin record as they make the decision and filing it ahead of time, or even (and if it was me, I might do this, publicly posting it along with the ROD to avoid arguments about how the plaintiffs need lots of time to review the admin record). But when you're already into either 'this litigation will never end' or 'the entire thing is going to be fought out in a motion for preliminary injunction/temporary restraining order context' litigation just eats so much time and money, even for an agency represented by DOJ that it's often not worth it (especially since the only people who hate losing more than leadership are DOJ attorneys).
This is somewhat true, and frankly pretty strange given the level of job security people generally have. I've never seen anyone disciplined in any way for deliberately choosing to accept risk, though I don't claim to be all knowing, or have any statistical backing for that.
Speaking only for myself, I've actually gotten somewhat less risk averse as I've gotten more experienced as I realized that law school paranoia aside, no one rushes to litigation and most of the time, even if people think you're wrong, the worst they'll do is yell at you and send an angry letter. It's not fun and if you're trying to advance, maybe there's some hidden penalties, but I'm as high as I'm likely to get (or want to get) so...who cares?
But, to be fair to other people, we dealt with a case where we had an active cattle encroachment on federal property, and I don't think we ever got through a meeting discussing it without referencing Bundy et. assholes...which was fair enough, but it all turned out fine. Bundy's a big deal because it's unusual!
I work in environmental review, so my perception is biased accordingly and might not reflect other contexts. But most of my jobs have been in or adjacent to government (including local and non-US) and there does seem to be some self-selection along these lines in the agencies I’m familiar with. It’s great that you, and maybe your organization, haven’t been sucked in to this kind of culture.
Wow, this is quite the double-edged sword! We have a real or maybe not real regulatory issue barring a significant service upgrade here (that departments can’t share information like address changes with other departments, leading to endless confusion, lost benefits etc). Using AI to untangle that web of bureaucratic barriers could give that process a real boost. Thanks for posting this.
Getting all the case law and then parsing through the often conflicting opinions is more than a little challenging for an LLM. The risk of illusions is going to be pretty huge.
How would an LLM deal with text, history and tradition, an entirely subjective standard invented by 6 people?
1. Re: “you never want to have to rely on analysis done by an adversary when you’re negotiating” — This is a huge a problem with budgeting in many states and strong-mayor cities, where there’s often exceedingly little legislative capacity for budgetary matters (or, in many cases, anything at all). I was reading just yesterday that in Chicago, “Though City Council members can vote yes or no on a budget, they have no authority to shape legislation — as aldermen pointed out in the latest budget process. … In the mayor’s corner are dozens of budget office employees who can shape the budget according to the mayor’s policy agenda. The City Council, meanwhile, has three people in the City Council Office of Financial Analysis.” And I’ve heard many similar stories over the years about other jurisdictions.
2. Re: Impoundment — In DC there was a big kerfuffle a few years ago because there was a project to build a protected bike lane on a particular street, which was included in the budget that was passed by council and signed by the mayor, then the mayor simply refused to let the project go forward. Eventually the council grew quite frustrated and debated passing emergency legislation demanding that the Department of Transportation move forward with it, but that ended up falling through in early 2020. Then, over a year later, the mayor announced an abrupt U-turn and said that the project would happen after all. I have to imagine this probably happens in other cities and states as well, with zero consequences, and of course far less media attention than what we see at the federal level.
I don’t expect DOGE to have fancy AI tools, but Elon Musk certainly does have access to a lot of people who are willing to jump on random urgent initiatives he has. And he’s proven that he’s willing to redirect groups from one of his companies to work on impulsive new projects at the spur of the moment. And he’s willing to put a lot of legal resources behind idiosyncratic interpretations of the law, and sometimes succeeds. So I wouldn’t be surprised if he manages to do some intensive analysis of some legal issues here.
"I hear rumors that Elon has about 40 or so engineers already squirreled away in a building near Lafayette Square. I have no special knowledge, but what I imagine them doing is exactly what darulharb suggested: spinning up AI systems that are going to change the playing field in ways few in Washington understand."
All it would take is for an enterprising individual to look at electricity usage in nearby office buildings. Anyone doing AI work would be using far more kilowatts than those doing normal office tasks (even spreadsheet financial analysis). In the end, there is going to be a huge amount of litigation regarding DOGE and any recommendations that come out. As one who filed lots of comments on notice and comment rulemaking during my working career, I am not terribly concerned about this whole effort. Yes, there needs to be a streamlining of regulations but it needs to be carefully thought out.
Putting aside your framing (which, using “lawmaking,” begs the question), I think delegation is good and proper and constitutional, and absolutely essential in a modern economy. Congress retains sufficient checks on agency action—the CRA, amending statutes, hearings, appropriations, advice and consent on nominees—to ensure the agencies don’t stray too far from statutory guidance. It’s not a faultless system, but the alternative is much much worse.
Could some agencies be consolidated? Yes, absolutely, and they should be. But the DOGE spitballing to cut the number of agencies by 80% isn’t close to reasonable.
How many new government agencies have been created that MCs want to take credit for? Warren and the CFPB, who else? I don’t think the incentives to create agencies are that strong.
What possible basis do you have for extending the benefit of the doubt to Musk / DOGE? You see the carelessness and malice with which he spreads misinformation on Twitter and you think "hm, I bet this guy's really focused on the public interest"?
I think this treats the capabilities of LLMs far too credulously. In reality they are not nearly this good, one cannot simply dumb the annotated US Code and CFR into them and hope to get anything useful out.
Ironically this is precisely what LLMs would be good at - coordinating issues and highlighting differences across vast reams of structured legal text. Especially the newer models released the last few weeks which are far better at quantitative reasoning. We should welcome the attempt rather than preemptively try to shut it down.
So, I think it'll do mostly okay with the USC and CFR (though I bet you run into a bunch of weird problems as the sheer number of different definitions of the same words may drive it a bit bananas). Where I guess it will totally crap out is the caselaw on those. First, just accessing it all is going to be tricky, as I really doubt westlaw/lexis/casetext are going to sell their databases. Second, it's just a giant confusing mess. My guess is almost anything is going to come back 'well, it could be illegal, or not, depending on execution and a bunch of facts, including most notably, which judge you happen to pull and in which circuit you're doing this!'
I'd like it to work, even if it might put me out of a job, but I doubt it.
I think people see chatgpt and think AI solutions are just feeding data directly as a big PDF to the same LLM theyre using. I'd advise folks to look up agent layers that break these processes into discrete parts handled by discrete agents which drive actual industry-specific solutions. There's arguably nothing more suited to modern agentic AI than the legal field, where the gaps and how one part of the law impacts decisions or ramifications in the other are actually pretty clear when the pieces are all laid out. The idea that the law is a 'giant confusing mess' that will make LLMs return 'illegal' results is....ill informed.
It's probably a perception left over from when some lawyers got into trouble when the AI would make up cases (AI "hallucinations") in citations when they had the AI writing briefs that were presented to a court. It made the news a few times when the lawyers who did that got sanctioned. There are, as you say, specific AIs now for the legal field, and many hours of continuing legal education have been devoted to teaching lawyers how AI can be used safely.
Trying to reconcile Supreme Court jurisprudence would be challenging enough, much less taking on the rest of the federal courts... But the first task, I agree, is reconciling the Constitution, U.S. Code, and Code of Federal Regulations (CFR) to flag areas of conflict for review by human attorneys. Teaching an A.I. to abstract "the rule" from case law is probably not even necessary for DOGE at this point.
Why do you think "areas of conflict" is the primary problem?
For example, DOGE co-lead Ramaswamy has said that part of their review is to determine what aspects of the administrative state's current operations conflict with the recent SCOTUS U.S. v. EPA "major questions" decision. Are there things that agencies are doing that implicate "major questions" and which lack a clear statement from Congress in the agency's enabling legislation to support them? In other words, does what the agency is doing conflict with the law which created the agency? Or in a constitutional sense, has the Executive Branch encroached on Congress' lawmaking power, seizing more regulatory discretion than was granted them?
As the DOGE project looks for cost savings and deregulatory targets, with the objective of shrinking the size and cost of the administrative state, things which agencies are doing that appear to implicate "major questions" but which lack legislative support are obvious choices to eliminate.
As another example, are there agencies which have accrued over the decades through legislation passed by successive Congresses whose areas of authority overlap or conflict, or which are redundant and could be consolidated? So that's another type of conflict. The DOGE project leads have talked about reducing the hundreds of agencies which currently exist to less than a hundred.
"The DOGE project leads have talked about reducing the hundreds of agencies which currently exist to less than a hundred."
These people don't know what these agencies are, don't know how they overlap, or how they came to exist in the first place. Shades of Rick Perry not understanding that DOE was responsible for so much of American nuclear policy/safety. We don't have to credit these tech bro libertarian fever dreams, we really don't.
I don't know how you do that for the constitution...at all, without doing case law...
Does a DOGE AI model really even need to consider what any other courts besides SCOTUS have said about what constitutional law is? Besides the text of the SCOTUS cases themselves, there's an abundance of commentary on Supreme Court cases to potentially draw on (including dissents), without the need for the AI to do case analysis itself. As far as comparing statutes or regulations to the Constitution, it could be useful to do it the way the Supreme Court would examine a statute (or regulation, in the absence of Chevron deference). The first pass might just be textual analysis.
I mean, the Supreme Court takes so few cases that the concern I hear is essentially never 'what will the Supreme Court think about this,' it's 'what will the District Court and 9th/5th (choose your bugaboo) Circuit think of this?'
Now, maybe if you're just trying to argue agency/agency lawyers into agreeing with you, that's the case you want to make, but as a practical matter I don't think 'the AI says you'll probably win in ~4-10 (Lopez Bright/Sackett) years if the Supreme Court takes it up,' convinces anyone.
This ignores existing rulings from lower courts which can control what agencies can do. It also sort of assumes that the Set is consistent across cases, which it isn't.
I don't want to shut anything down, by all means try. But its not a deus ex machina that will just solve these problems.
Why do you think "coordinating issues and highlighting differences across vast reams of structured legal text" is the primary challenge?
The primary challenge is simply building agents that handle discrete parts of the law and others that tell them what takes precedence and when and why. I imagine the company that builds those basic agents will be a lawyer-known name by this time next year.
I’ve been using LLMs for similar stuff—it’s actually great for pulling out key details from long documents
“Hey Claude, give me a rundown of the call/put ratios in this market report.”
“Hey Claude, what does this case (pasted) say about that law (attached)”
You need a sharp analyst BEHIND the prompts, but it’s a genuinely useful tool.
strong disagree. also, all the argument and assumptions are based on one or two interaction and current state. We miss the point that as the time is passing, those are becoming more efficient and improving at large scale. so we can only play that card for limited time and some handful of instnaces.
This is completely wrong—as an LLM researcher I'm convinced the latest LLMs are extremely good at legal and policy analysis, and given that we see better models being released every 3-6 months on average we should expect significant improvement over DOGE's term.
Are you also a legal scholar?
Count me among the doubters. Musk hasn’t been able to effectively productize AI for Twitter. And he is known for announcing his ideas/intentions far too soon, so I wouldn’t expect him to be tight lipped about this if that’s what they’re doing. And finally, having worked legislatively, I can only imagine that designing an LLM to actually cut through the red tape is going to be far more complex than could be spun up in a short timeframe. To really cut through the mess, you’d have to consider case law, other precedents, etc. A lot of human knowledge would have to go into shaping the AI itself, which takes time.
What about being able to trace summary information back to its legal/regulatory sources? My impression is that AI algorithms, or LLMs, at any rate have to discard their original material and source references (?)
What kind of summary information? We may be talking about different use cases. FWIW, agencies usually cite to authorizing statute when making regulation. Even in sub-regulatory guidance.
I was just wondering if the summary information generated by an AI would, well, have footnotes / citations
An anecdote for the naysayers here. A couple weeks ago I was at a carahsoft event highlighting the impacts of the current admins PMA goal around CX. They focused on program level stuff - lower paper processing times at USCIS, more digital tools for HUD, call wait times down at SSA, etc.
During Q&A I asked the question of where the persistent CX work is happening. In industry companies have built huge systems connecting touchpoints of users across online and offline interactions,and use AI to help personalize experiences. This is well trodden ground of over a decade of available tools, systems, and best practices. Where was that happening in these agenies mandated to fix their overall CX issues?
The answer was blank states. The HUD guy talked about how personalization was bad due to equity gaps it raises. The DHS guy said 'huh that would be interesting to look at'
These problems are endemic, and I don't think that fed agencies truly understand just how far they are away from boilerplate capability in every other sector. Turning tail on new initiatives like DOGE pre emptively because you don't like the people or systems involved just continues building a moat around agency systems and processes from the 90s. The arguments for keeping our agencies working like they are now gets worse every day that AI advances.
Probably not the right audience for that question.
These were the people leading CX at their agencies under the EO mandate that has been in place for years. OMB was there are well. If it's not then it's not happening, and its clearly not happening.
These sorts of things are actually policy choices we should be extremely critical about. The public sector is at the bottom of CX capabilities across every industry polled but instead of using tried and true methods they decided to take the money and put towards programs that they wanted to fix anyway, with the hand-wave of "it impacts constituents therefore it's CX." It's ridiculous and borderline negligent.
I’m not doubting that you could develop an algorithm to do this kind of work; I’m doubting that we have enough good data to feed that algorithm. DoD (for example) runs on PDFs and PPTs - not good structured data.
Good point
Show me you can find enough machine readable data to trace a JROC Requirement from validation through the acquisition process and out to TRL-9 deployment to an operational unit and I’ll be the first one to buy you a beer.
I understand the intellectual value of taking this initiative seriously.
But I've seen Elon Musk's raw, unvarnished thoughts for years now and I know what he thinks about everything under the sun. To put a fine point on it, he's an idiot, massively out of his depth, and seemingly hopped up on a serious cocktail of stimulants. He cares less about government efficiency than he seems to care about the size of breasts of women in video games.
I think if given the opportunity, what he wants to do to the federal government is what he did to Twitter/X, namely fire a ton of people whose jobs he doesn't understand. But if the experience of Twitter/X is prologue, then expect the experience of working with the federal government to get so much worse. Twitter/X is basically unusable right now; the feed is mostly low-grade spam, search doesn't function, notifications recommends nothing you're interested in (including recently sending you posts from Elon himself whether or not you follow him), bots are more prevalent than ever before he took over, and generally the real users are being overrun by vile out-and-proud Nazi accounts. Let's not forget that he's lost billions on this endeavor and his only path to recouping his investment is through some kind of corrupt exercise of his newfound influence with the government.
Oh, but he might tack some kind of slipshod AI product onto the government? Just what we don't need. Until AI can figure out its hallucination problem, these technologies can't reliably be used for government. We've seen this play out repeatedly with court cases, for instance, as lawyers and experts use AI to write filings or depositions and they turn out to cite completely fake sources.
So while I appreciate Jennifer's willingness to give this not-actually-a-government agency some benefit of the doubt, I think it's remarkably charitable to do so.
Can I like this comment more than once?
I’m not a linguist, but I believe the proper pronunciation of “DOGE” is “doosh”.
LOL
You’re going to need to at least have one person who has actually read it all to do some QA on AI outputs. (Raises hand)
I didn’t find it too bothersome to keep it all in my head (the 5000+ pages necessary to launch any digital system in govt) but it became very clear that I was in the minority in that regard. Eventually I would figure out many COs couldn’t even be bothered to memorize their default procurement clauses.
Can we get pro bono credit for volunteering for this? 🤔
I can't imagine anything that will generate more anti-AI sentiment than DOGE suddenly unleashing opaque and likely untrustworthy Musk AI on government rules and regs. His mode now is break things and move fast; the first problematic recommendation will among other things drive states to each self-regulate AI inconsistently.
"Move fast and break things" will go over well, when the thing being broken is people's social security payments.
I hope you're right...but I do not believe that is likely. I think you can produce a medium useful tool on the CFR/USC that provides some places to start, though the history of lawyers using AI to try to determine what the law is is a minefield of blown up careers.
But frankly, for most of the questions people actually want to ask, the answer is either--yes, the annoying thing compliance wants you to do is actually a requirement. Unless you can call it an emergency, you have to do 106 compliance on any undertaking, sorry and unless you can get to no potential to effect, you have to consult on it.
Or, 'it's grey, do you want to take the risk of being sued and potentially losing?' And there you run into the problem that leadership tends to, even more than the lawyers, hate being sued and losing. To be fair, they also hate being sued and winning. Or suing. I hate to sound like a broken record, but I think if you can unjam the judiciary and make things actually move efficiently there, things get a lot less painful on most fronts. Like, an APA case is never going to be instantaneous, but for most big actions, an agency knows if a lawsuit is coming. If you can rely on a court to move it along fast, I bet you see a lot more agencies building the admin record as they make the decision and filing it ahead of time, or even (and if it was me, I might do this, publicly posting it along with the ROD to avoid arguments about how the plaintiffs need lots of time to review the admin record). But when you're already into either 'this litigation will never end' or 'the entire thing is going to be fought out in a motion for preliminary injunction/temporary restraining order context' litigation just eats so much time and money, even for an agency represented by DOJ that it's often not worth it (especially since the only people who hate losing more than leadership are DOJ attorneys).
The risk aversion of government at all levels, and its relationship to the exasperating slowness of administrative procedures, cannot be overstated.
This is somewhat true, and frankly pretty strange given the level of job security people generally have. I've never seen anyone disciplined in any way for deliberately choosing to accept risk, though I don't claim to be all knowing, or have any statistical backing for that.
Speaking only for myself, I've actually gotten somewhat less risk averse as I've gotten more experienced as I realized that law school paranoia aside, no one rushes to litigation and most of the time, even if people think you're wrong, the worst they'll do is yell at you and send an angry letter. It's not fun and if you're trying to advance, maybe there's some hidden penalties, but I'm as high as I'm likely to get (or want to get) so...who cares?
But, to be fair to other people, we dealt with a case where we had an active cattle encroachment on federal property, and I don't think we ever got through a meeting discussing it without referencing Bundy et. assholes...which was fair enough, but it all turned out fine. Bundy's a big deal because it's unusual!
I work in environmental review, so my perception is biased accordingly and might not reflect other contexts. But most of my jobs have been in or adjacent to government (including local and non-US) and there does seem to be some self-selection along these lines in the agencies I’m familiar with. It’s great that you, and maybe your organization, haven’t been sucked in to this kind of culture.
Wow, this is quite the double-edged sword! We have a real or maybe not real regulatory issue barring a significant service upgrade here (that departments can’t share information like address changes with other departments, leading to endless confusion, lost benefits etc). Using AI to untangle that web of bureaucratic barriers could give that process a real boost. Thanks for posting this.
Getting all the case law and then parsing through the often conflicting opinions is more than a little challenging for an LLM. The risk of illusions is going to be pretty huge.
How would an LLM deal with text, history and tradition, an entirely subjective standard invented by 6 people?
A couple local/state angles here:
1. Re: “you never want to have to rely on analysis done by an adversary when you’re negotiating” — This is a huge a problem with budgeting in many states and strong-mayor cities, where there’s often exceedingly little legislative capacity for budgetary matters (or, in many cases, anything at all). I was reading just yesterday that in Chicago, “Though City Council members can vote yes or no on a budget, they have no authority to shape legislation — as aldermen pointed out in the latest budget process. … In the mayor’s corner are dozens of budget office employees who can shape the budget according to the mayor’s policy agenda. The City Council, meanwhile, has three people in the City Council Office of Financial Analysis.” And I’ve heard many similar stories over the years about other jurisdictions.
2. Re: Impoundment — In DC there was a big kerfuffle a few years ago because there was a project to build a protected bike lane on a particular street, which was included in the budget that was passed by council and signed by the mayor, then the mayor simply refused to let the project go forward. Eventually the council grew quite frustrated and debated passing emergency legislation demanding that the Department of Transportation move forward with it, but that ended up falling through in early 2020. Then, over a year later, the mayor announced an abrupt U-turn and said that the project would happen after all. I have to imagine this probably happens in other cities and states as well, with zero consequences, and of course far less media attention than what we see at the federal level.
I don’t expect DOGE to have fancy AI tools, but Elon Musk certainly does have access to a lot of people who are willing to jump on random urgent initiatives he has. And he’s proven that he’s willing to redirect groups from one of his companies to work on impulsive new projects at the spur of the moment. And he’s willing to put a lot of legal resources behind idiosyncratic interpretations of the law, and sometimes succeeds. So I wouldn’t be surprised if he manages to do some intensive analysis of some legal issues here.
Maybe. How's he doing with the Boring Company?
Ummm... so Musk isn't elected so he's offering his "expertise" gratis for the good of the country ?
"I hear rumors that Elon has about 40 or so engineers already squirreled away in a building near Lafayette Square. I have no special knowledge, but what I imagine them doing is exactly what darulharb suggested: spinning up AI systems that are going to change the playing field in ways few in Washington understand."
All it would take is for an enterprising individual to look at electricity usage in nearby office buildings. Anyone doing AI work would be using far more kilowatts than those doing normal office tasks (even spreadsheet financial analysis). In the end, there is going to be a huge amount of litigation regarding DOGE and any recommendations that come out. As one who filed lots of comments on notice and comment rulemaking during my working career, I am not terribly concerned about this whole effort. Yes, there needs to be a streamlining of regulations but it needs to be carefully thought out.
The AI processing will probably take place off-premises in data centers.
Yeah the energy processing isn't happening at the office buildings
Putting aside your framing (which, using “lawmaking,” begs the question), I think delegation is good and proper and constitutional, and absolutely essential in a modern economy. Congress retains sufficient checks on agency action—the CRA, amending statutes, hearings, appropriations, advice and consent on nominees—to ensure the agencies don’t stray too far from statutory guidance. It’s not a faultless system, but the alternative is much much worse.
Could some agencies be consolidated? Yes, absolutely, and they should be. But the DOGE spitballing to cut the number of agencies by 80% isn’t close to reasonable.
How many new government agencies have been created that MCs want to take credit for? Warren and the CFPB, who else? I don’t think the incentives to create agencies are that strong.
What possible basis do you have for extending the benefit of the doubt to Musk / DOGE? You see the carelessness and malice with which he spreads misinformation on Twitter and you think "hm, I bet this guy's really focused on the public interest"?