Given that that's a task you want to do, it's at least the right kind of task (language processing) for an LLM. The proposals from the comment starting this thread aren't.
If grant classification is trying to drive a car non-stop (including not stopping for gas) from NY to LA, stuffing LLMs into weapons is more like trying to drive that same car from NY to London. They're just not the proper kind of tool for that, and it's not the same class of error.
If people on Hacker News are uncertain about what is and isn't a suitable task for these models then the non technical people making these decisions surely are as well.
You're saying that weapons are designed by incompetents, and that enthusiasts have a reasoned understanding of the capabilities and limitations of the latest thing they're going "ooh shiny" about.
That's fundamentally not a language processing task. That's a decision making task with a huge impact on individual scientists and the scientific community. Not something that should be delegated to a machine, no matter how sophisticated.
If grant classification is trying to drive a car non-stop (including not stopping for gas) from NY to LA, stuffing LLMs into weapons is more like trying to drive that same car from NY to London. They're just not the proper kind of tool for that, and it's not the same class of error.