Environment & Energy
In reply to the discussion: Projected 10-30% Increase In Natural Gas Power Generation Because Muh AI Datacenters Must Be Fed!!! Oh, And Coal, Too [View all]NNadir
(36,196 posts)...often to support my scientific work and given the richness of the literature, its vast scope, I certainly wish I had something like CCU-Lama, which I described in the science forum: CCU-Llama.
Who's going to monitor the "correct use" of servers? The Trump administration?
It's funny, because just the other day I was having a conversation with another scientist about whether we should always trust the sophisticated software we use, both on line and in house, to interpret mass spec data. I'm so old, of course, that I remember sitting with a pencil and paper and calculating the mass of potential fragments and then looking at the data to see if such a mass was there visually. I could spend a week or more with a complex compound in that way. Now, in less than a few minutes, I can see all the PTMs and sequences from a very large protein, no trouble at all. I almost never find a result that seems to be invalidated by experiment, unless it involves an isobaric species, and now their are ways around that as well. The public servers, like Uniprot, do, must do something very much like AI, although honestly I don't know how it works, just that it's as fast as hell and I have direct experience with it being perfectly correct on multiple occasions, for example, finding the exact correct species of associated with a highly conserved protein found across many living things when I was blinded. And trust me, the protein in question is highly conserved across a wide range of species, from single cell organisms to human beings.
However, we can and do, set false discovery rates in the use of the software, and that is designed to establish the error parameters. The fact that there is a "false discovery rate," means that we have to be careful with the data, it is not determinative so much as (highly) suggestive.
Your link, by the way, refers to an article referring to a paper, this one: Terwilliger, T.C., Liebschner, D., Croll, T.I. et al. AlphaFold predictions are valuable hypotheses and accelerate but do not replace experimental structure determination. Nat Methods 21, 110116 (2024). It certainly doesn't discount the value of Alphafold, remarking that it often produces results that are remarkably similar to crystallized proteins, which, as the authors note does not necessarily correspond to protein structure in vivo.
To wit:
From the conclusion of the paper:
To me, this doesn't read like a dismissal of Alphafold, but rather a wise cautionary suggestion as to how it should be used.
Of course, in the history of science, there have been many calculations that proved to not hold up to experimental data. Experiment always prevails over theory, or should anyway. One should always check results against theory, and in fact, that is what automated machine learning does, compares data with theory to determine whether theory holds, and adjusts the theory accordingly, but yes, the output of this process needs human review.
None of this means that there is something corrupt or illegitimate with data centers. My remark about my son's work was intended not to be "right" or "wrong," but rather to suggest that we ought to be careful with how we judge technologies. Sure there are kids who produce term papers on ChatGPT. That doesn't mean that ChatGPT is evil. I have an assistant, not a scientist, who brings me text from it regularly, with my knowledge. It often fails the Turing test, but recognizing that it fails the Turing test often, it can help unblock writer's block. We never use it directly in our reports, but it suggests, not defines, a path.
Edit history
Recommendations
0 members have recommended this reply (displayed in chronological order):