I’ve recently been experimenting with AI ‘agents’, i.e., Claude Code or other LLMs via Opencode. It’s, imo, very good at automating boilerplate code, especially data wrangling, plotting, cleaning up code for packages etc. In general, far superior to copy and pasting from chat interfaces. Ofc, there are various downsides, but I assume that most researchers will adopt this in one form or another.
A very useful feature are ‘skills’. These are local instructions (markdown files), LLM access depending on the context of the current chat. These ‘skills’ can be very sophisticated and cover quite complex workflows. One could, for example, instruct the LLM to prefer certain analysis / modeling strategies over other, with very concrete examples, code, validations strategies, etc. However, the current statistics ‘skills’ that one can find online (here, here), are very generic, or even read like AI generated slop, mostly or only use Python libraries which are no way close to R libraries for most non-ML applications. This will of course not stop people from using these and increasing the output of subpar research.
I think it would be very worthwhile to create such ‘skills’, but which actually support the implementation of best practices, e.g., suggest and implement strategies found in RMS from @f2harrell. It’s probably also a good idea to provide such a ‘skill’ when releasing a package for R, to make correct and widespread application more likely. Of course there are downsides: people will not read the original sources as often, leading to worse understanding, less traffic to websites hosting these books etc. However, the alternative is that people use LLMs to produce worse analyses, and still don’t read up on things.
These better ‘skills’ would of course not stop the increasing flood of articles that should not be published, but it might be able to increase the quality somewhat. Curious to hear your thoughts.