Submitted by adventurousprogram4 t3_zyiib1 in MachineLearning
Has the research community embraced any of the frameworks or findings published by Anthropic at all? Google Scholar seems to indicate no, but I'm curious. I work on the applied side and not on the research side, so I don't have a good sense for how influential their work on interpretability is.
The motivation for my question is that they have a huge amount of funding (although how long that will last after SBF's downfall remains to be seen) and a lot of press attention and fans in the rationalist/EA communities, but my feeling is that their work is largely not being adopted or cited in AI research. If I am correct in this, I'm curious if this is because it is seen as unoriginal, incorrect, or misguided? Or is there something else going on?
AGI_aint_happening t1_j26847d wrote
As a former interpretability researcher who has skimmed their work but not read it closely, I just don't find it terribly interesting or novel. Also, frankly, I find the writing style for the papers pretty hard to parse (as they don't follow standard paper formats) and a tad grandiose, as they tend to avoid standard things like comparing against other methods or citing other work. Relatedly, I think their choice to avoid peer review has impacted how people perceive their work, and limited its distribution.