I'm organizing a workshop on "Emerging models in scientific communication and discussion" in CNS Antwerp, 19-20 July 2017. This is somewhat related to my new year resolution.
This is the first issue of my journal of theoretical neuroscience.
This month, I have selected 4 papers on spike initiation (1-4), 1 classical paper on the theory of brain energetics (5), and 1 paper on bibliometrics (6). Three of the papers on spike initiation (1-3) have in common that they are about the relation between geometry (morphology of the neuron and spatial distribution of channels) and excitability. Spikes are initiated in a small region called the axon initial segment (AIS), and this region is very close to the soma. Thus there is a discontinuity in both the geometry (big soma, thin axon) and the spatial distribution of channels (lots in the AIS). It has great impact on excitability, but this has not been very deeply explored theoretically. In fact, as I have discussed in a recent review (Brette, 2015), most theory on excitability (dynamical systems theory) has been developed on isopotential models, and so is largely obsolete. So there is much to do on spike initiation theory that takes into account the soma-AIS system.
1. Evans MD, Tufo C, Dumitrescu AS and MS Grubb. (2017). Myosin II activity is required for structural plasticity at the axon initial segment. (Comment on PubPeer)
A number of studies have shown that the AIS can move over hours or days, with various manipulations such as depolarizing the neuron (as in this study) or stimulating it optogenetically. Two open questions: what are the molecular mechanisms involved in this displacement? Is it actually a displacement or is it just that stuff is removed at one end and inserted at the other end? The same lab previously addressed the first question, showing the involvement of somatic L-type calcium channels and calmodulin. This study shows that myosin (the stuff of muscle, except not the type expressed in muscles) is involved, which strongly suggests that it is an actual displacement; this is in line with previous studies showing that dendrites and axons are contractile structures (e.g. Roland et al. (2014)). This and previous studies start to provide building blocks for a model of activity-dependent structural plasticity of the AIS (working on it!).
2. Hamada M, Goethals S, de Vries S, Brette R, Kole M (2016). Covariation of axon initial segment location and dendritic tree normalizes the somatic action potential. (Comment on PubPeer)
Full disclosure: I am an author of this paper. In the lab, we are currently interested in the relation between neural geometry and excitability. In particular, what is the electrical impact of the location of the axon initial segment (AIS)? Experimentally, this is a difficult question because manipulations of AIS geometry (distance, length) also induce changes in Nav channel and other channel properties, in particular phosphorylation (Evans et al., 2015). So this is typically a good question for theorists. I have previously shown that moving the AIS away from the soma should make the neuron more excitable (lower spike threshold), everything else being equal (Brette, 2013). Here we look at what happens after axonal spike initiation, when the current enters the soma (I try to avoid the term “backpropagate”, see Telenczuk et al., 2016). The basic insight is simple: when the axonal spike is fully developed, the voltage gradient between soma and start of the AIS should be roughly 100 mV, and so the axonal current into the soma should be roughly 100 mV divided by resistance between soma and AIS, which is proportional to AIS distance. Next, to charge a big somatodendritic compartment, you need a bigger current. So we predict that big neurons should have a more proximal AIS. This is what the data obtained by Kole’s lab show in this paper (along with many other things, as our theoretical work is a small part of the paper – as often, most of the theory ends up in the supplementary).
3. Michalikova M, Remme MWH and R Kempter. (2017). Spikelets in Pyramidal Neurons: Action Potentials Initiated in the Axon Initial Segment That Do Not Activate the Soma. (Comment on PubPeer)
Using simulations of detailed models, the authors propose to explain the observation of spikelets in vivo (small all-or-none events) by the failed propagation of axonal spikes to the soma. Under certain circumstances, they show that a spike generated at the distal axonal initiation site may fail to reach the somatic threshold for AP generation, so that only the smaller axonal spike is observed at the soma. This paper provides a nice overview of the topic and I found the study convincing. There is in fact a direct relation to our paper discussed above (Hamada et al., 2016): this study shows how the axonal spike can fail to trigger the somatic spike, which explains why the AIS needs to be placed at the right position to prevent this. One can argue (speculatively) that if AIS position is indeed tuned to produce the right amount of somatic depolarization, then sometimes this should fail and result in a spikelet (algorithm: if no spikelet, move AIS distally; if spikelet, move AIS proximally).
4. Mensi S, Hagens P, Gerstner W and C Pozzorini (2016). Enhanced Sensitivity to Rapid Input Fluctuations by Nonlinear Threshold Dynamics in Neocortical Pyramidal Neurons. (Comment on PubPeer)
I had to love this paper, because the authors basically experimentally confirm every theoretical prediction we had made in our paper on spike threshold adaptation (Platkiewicz and Brette, 2011). Essentially, what we had done is derive the dynamics of spike threshold from the dynamics of Nav channel inactivation. There were a number of non-trivial predictions, such as the shortening of the effective integration time constant, sensitivity to input variance, the specific way in which spike threshold depends on membrane potential, and the interaction between spike-triggered and subthreshold adaptation (that we touched upon in the discussion). This study uses a non-parametric model-fitting approach in cortical slices to empirically derive the dynamics of spike threshold (indirectly, based on responses to fluctuating currents), and the results are completely in line with our theoretical predictions.
5. Attwell D and SB Laughlin (2001). An energy budget for signaling in the grey matter of the brain (Comment on PubPeer and Pubmed Commons).
This is an old but important paper on energetics of the brain, in particular: how much does it cost to maintain the resting potential? How much does it cost to propagate a spike? The paper explains some theoretical ideas to do these estimations, and is also a good source for relevant empirical numbers. It is important though to look at follow-up studies, which have addressed some issues, for example action potential efficiency is underestimated in this study. One problem in this study is the estimation of the cost of the resting potential, which I think is just wrong (see my detailed comment on Pubmed Commons; I believe the authors will soon respond). Unfortunately, I think it is really hard to estimate this cost by theoretical means; it would require knowing the permeability at rest to various ions, most importantly in the axon.
6. Brembs B, Button K and M Munafò (2013). Deep Impact: Unintended consequences of journal rank. (Comment on PubPeer)
The authors look at the relation between journal rank (derived from impact factor) and various indicators, for example effect sizes reported, statistical power, etc. In summary, they found that the only thing journal rank strongly correlates with is the proportion of retractions and frauds. Another interesting finding is about the predictive power of journal rank on future citations. There is obviously a positive correlation since impact factor measures the number of citations. But it is really quite small (see my post on this). What is most interesting is that the predictive power started increasing in the 1960s, when the impact factor was introduced. This strongly suggests that, rather than being a quality indicator, the impact factor biases the citations of papers (increases the visibility of otherwise equally good papers). This paper also shows evidence of manipulation of impact factors by journals (including Current Biology, whose impact factor went from 7 to 12 after its acquisition by Elsevier), and is generally a good source of references on the subject.
I recently reported that one of my papers has been rejected by PLoS Computational Biology after 10 months and 4 revisions, on the ground of general interest. This has generated a little buzz. A colleague mentioned it on his blog. As a result, the editor of my paper contacted him directly to tell his version of the story, which my colleague has now reported on his blog.
Unfortunately, the editor’s story is “misleading”, to be polite. It is a shame that the review process is confidential, as it allows the journal to hide what actually happens behind their closed doors. Nevertheless, I have asked the journal for the authorization to publish the content of my appeal and their response, where I explain what happened in more detail (and more accurately). They have accepted. I have removed names of the persons involved. This illustrates one of the flaws of the current peer-review system (see this post for how it could work better).
(Just one note: the editor has apparently told my colleague that the third reviewer was a collaborator, so they could not take into account his review. Well, that’s a lie. I know because he chose to sign his review. The "collaboration" was the scientist sending me published data.)
So here it is.
Re: Manuscript PCOMPBIOL-D-16-00007R4
So after 10 months and 4 revisions, our paper has been rejected, following the recommendation of one reviewer, because it is not considered of broad enough interest. I quote from the final editorial message: “We regret that the specific hypothesis that your manuscript is geared to dispute does not reach that level of general interest.”.
These facts being recalled, it should be obvious enough that the editorial process has gone very wrong. There were no more technical criticisms already after revision 2, on July 8th, 4 months ago, and the paper should have been accepted then. I have repeatedly asked the editors to explain why we were required to justify novelty and significance after having been required to do so much work on technical aspects. But the editors have refused to answer this simple query. Frankly, I was expecting a bit more respect for the authors that make this journal, and I do not think that explaining the journal’s policy and the decisions is so much to ask. All I know is Michael Eisen’s view, founding editor of this journal, who has cared to comment “I agree - a paper going out for review should mean it is of interest”.
This editorial process has gone beyond anything I have ever witnessed in my career in terms of absurdity and waste. Why scientists (“peers”) would voluntarily make each other’s life so unnecessarily hard instead of cooperating and debating is beyond my understanding. In the end it appears that the ego of one (important?) reviewer matters more than science, and that is very sad. This being said, I have been notified that appeals are only considered when “a) a reviewer or editor is thought to have made a significant factual error” or “b) his/her objectivity is compromised by a documented competing interest”, and since bureaucracy apparently beats reason and ethics, I will now explain how this applies.
I have already explained at length the factual errors of the first reviewer, who is apparently the only one that is trusted by the editors. This editorial message repeats some of them (no, we are not criticizing simulation results of a particular model, but the biophysical interpretation (what goes on physically), and we did so in several state-of-the-art biophysical models, not one). I will therefore focus on case (b), and attach my previous letter to the editors for reference; please also read the responses to reviewers as regards case (a), in particular to reviewer-editor Dr. YYY who has unfortunately not cared to reply to our point-by-point response that he had required from us. The editorial decisions that have led to rejecting the paper on the basis of general interest after 10 months are so bizarre that I am compelled to question senior editor Dr. YYY’s objectivity – I presume that Dr. XXX, who sent the paper for review in the first place, does consider the paper of interest. The sequence of facts speaks for itself:
- On June 6th (revision #2), the editorial message reads “We understand that Reviewer 2 was very enthusiastic, and Reviewer 3 had relatively minor comments, but we both stress that addressing Reviewer 1's reservations are essential. Indeed, it is only fair to say that it seems to us that it will be challenging to address these comments in the context of the presented results.”. The exclusive reliance on one reviewer and the presumption that we could not address the comments is rather surprising. Nonetheless, the editorial message that followed was exclusively about the match with experimental data, not about interest (“the reviewer's point about (apparently) unrealistic voltage dependencies of the currents […]”). We did successfully address these comments, pointing out that the reviewer had made factual errors (such as misreading the figure he was commenting, and discussing the results of an experimental paper he had not opened).
- On July 8th (revision #3), the editorial message was now asking to explain the novelty compared to what we had done in the past (and published in the same journal), blindly following the 3-sentence report of reviewer #1, and making no mention whatsoever to the fact that we had just answered the major (and flawed) criticisms on experimental observations, which constituted the previous editorial message. At this point we complained that we were asked to justify the novelty of our study 7 months after submission, especially when it was explicit in the introduction; nonetheless, we complied and explained again.
- On August 25th (revision #4), we were appalled to read that, instead of finally accepting the paper, senior editor Dr. YYY decided to nominate himself as a reviewer, admitting that “the latest revision is first one he has had the chance to read”. The report was not an assessment of the novelty of the paper, as would have been logical since this was what the previous editorial message was about. Instead, it was a 6 pages long report full of technical queries, making negative criticisms that, for most of them, had already been addressed in previous reports, and asking for substantial modifications of the paper.
- At this point, I replied to the editorial message and obtained no response; as the message stated “If you would like to discuss anything, please don't hesitate to contact either of us directly”, I emailed Dr. YYY, and he started his response as follows: “To answer your email, allow me to be brief, because this sort of exchange should really be going through the journal, and indeed that will be the case from now on.”. Nonetheless, we exchanged a few emails, in which he offered no explanation; in the end we agreed that I would write a point-by-point response to his six-page review, but not modify the paper. I submitted it, together with a response to the first reviewer, and a letter to the editors, on September 22nd.
- Three weeks later, on October 10th, I received a message where I was asked to edit the letter so that it could be passed on to the reviewers. Apparently the editors had not noticed the response to reviewers. It still took them three weeks to read a letter, which, considering the history of this paper, does not strike me as very respectful. I complained to Dr. YYY, who replied “We believe that you have been adequately notified by the PLoS administrative team concerning the status of your revision.”. I had to exchange several emails with Dr. XXX who realized the error. I received no apology from Dr. YYY.
- On November 11th, I received the reject decision, together with the response of reviewer #1 and, oddly enough, of reviewer #3 to which I had not replied (since there was no remaining comment). He also was surprised, since he wrote “I don’t have the expertise, authority or, honestly, the time to judge whether the new comments from Reviewers 1 & 4 are fair, or whether the authors’ responses have fully addressed them – this is clearly a job for the Editors (although hopefully not for the Editor who just became a Reviewer)”. But, editor-reviewer #4 Dr. YYY did not bother replying to my point-by-point response, which he had explicitly required.
- The final decision comes with excuses that are frankly hard to swallow. One is that the editors had failed to see the word “models” in the title. In 10 months and 4 revisions! Who can seriously believe that? And yes, the paper is about models – it is a computational biology journal (note that we have also successfully related the models to experimental observations, on request of the editors). The other excuse is that an anonymous reviewer (reviewer #2) had a conflict of interest and his reviews had to be dismissed. I am of course fine with that decision (let me simply state for the record that none of reviewers I have suggested are in such a position). But this happened in April, more than 6 months ago. Quite appropriately, the editor Dr. XXX asked for another reviewer who identified himself (Dr. ZZZ). Dr. ZZZ wrote a positive review, and in addition he read our responses to the other reviewers and wrote “the revisions of the manuscript in response to the other reviewers' comments seem entirely appropriate.”. At this point, given that no objection had been raised by any reviewer or editor on methods, results or clarity, the paper should have been accepted. Instead, the editors decided to follow a nonsensical comment from reviewer #1 alone: “unlikely to be of broad interest to the computational biology field, but could be of interest to computational neuroscientists”, which was not even consistent with his/her own first positive assessment (“this is an interesting paper”). Given that Dr. XXX sent the paper for review in the first place, this decision must originate from Dr. YYY (who at this point had not read the paper, by his own admission). I am compelled to conclude that Dr. YYY has not been objective, and in fact has been actively blocking our paper. Unfortunately, this is not the first time I witness the questionable attitude of Dr. YYY, as he has recently been a reviewer for an essay I wrote. The review process was extremely long, went over multiple rounds with massive lists of requests, where Dr. YYY basically wanted to rewrite the text to follow his own views and style. During the review process, Dr. YYY contacted me directly by email to discuss the paper, going so far as asking for co-authorship (“Indeed, the level of suggestions are approaching collaboration on this paper- something I would be happy with but I assume is not what you have in mind.”). In the same email, and while the review process was not over, he asked me for an experimental collaboration – which of course I have not followed up. I had to ask the editor to intervene to stop the madness – which he did: “Indeed your paper has been unduly delayed and I have asked the reviewer to answer me within 24 hours.”. I apologize for disclosing these email excerpts, but I have no other choice since I am asked to provide documentation. It is clear that, had I imagined that Dr. YYY could be chosen as a reviewer (which seemed unlikely given his recent track record), I would have opposed him. But I did not anticipate that he would nominate himself, or overthrow the editor’s decision without even reading my paper (by his own admission).
Therefore, I am asking that Dr. YYY is replaced by a new senior editor with a more reasonable attitude.
As far as I can see: 1) the three reviewers were initially positive on the interest of the paper; 2) the editor Dr. XXX, who as far as I can tell is the only scientist involved in this process who is a member of the computational biology community, supported our paper since he sent it for review; 3) one reviewer, who seems to be an experimental electrophysiologist (unfortunately he or she has decided to remain anonymous), reverted his subjective opinion on the paper’s interest after we have pointed out the errors in his/her report, and even then, still judged the paper interesting for the computational neuroscience community. I have failed to see to how the decision is “not trivial to reach”.
Attached: Letter to the editors from September 16th
Letter to the editors, September 16th
In the previous revision, I raised serious objections regarding the abusive attitude of reviewer #1. These objections have apparently been completely dismissed, but what I have been most disappointed about is the total lack of response to these objections. I am writing this letter in the hope that this time it will be given some consideration.
This manuscript has been submitted 8 months ago. This is the fourth major revision that we have been asked to make. The responses are now totaling more than 25 pages, much longer than the article itself. We have now entered a phase where a large part of the responses consist in citing previous revisions where the issues have already been addressed. This revision reaches a new level, where a fourth reviewer is added and repeats mostly questions that we have already answered in previous revisions. Why a fourth reviewer is considered necessary after 8 months of revision is not clear, when none of the three reviewers has raised any serious concern.
I have officially asked a detailed explanation for this peculiar decision. The only response I have obtained so far is that there was “a tie” between “conflicting reviews”. So apparently the editorial decision has been based on a vote between reviewers. This is yet what I read on the journal’s website:
If reviewers appear to disagree fundamentally, the editors may choose to share all the reviews with each of the reviewers and by this means elicit additional comment that may help the editors to make a decision. That said, decisions are not necessarily made according to majority rule. Instead, the editors evaluate the recommendations and comments of the reviewers alongside comments by the authors and material that may not have been made available to those reviewers.
If one followed this process, then one would realize that:
- None of the three reviewers has any remaining objection about results, methods, or clarity of the text.
- Reviewer #2 and #3 have an overall very positive assessment of the paper and in particular of its interest. Rev #2: “This is a great revision. The authors have clarified and addressed all my previous concerns. […] I strongly believe the study is publishable as it stands”; Rev# 3: “This is a very clear and logically presented manuscript dealing with a key question in fundamental cellular neuroscience”.
- On his/her first report, reviewer #1 also made a positive assessment of the paper and of its interest: “This is a clearly written manuscript that addresses an interesting question regarding the nature of spike initiation. Specifically, the authors propose a plausible explanation […] This is an interesting paper.”.
- After two rounds of technical revisions, in which we pointed out the reviewer’s errors and to which no objection has been made, reviewer #1 changed his mood and now concludes, without any argument: “unlikely to be of broad interest to the computational biology field, but could be of interest to computational neuroscientists” (sic).
- Reviewer #3 has read our responses to the two other reviewers and concluded: “the revisions of the manuscript in response to the other reviewers' comments seem entirely appropriate.” From these facts, it appears clearly that there are in fact 3 convergent reviews. All 3 reviewers have concluded that results and methods are rigorous and the text is well written. All 3 reviewers have found the paper interesting. It might be that reviewer #1 has “voted” negatively; however I would expect the editorial decision to be based on the content of reviews and responses, which in this case is convergent, and not on the mood of one reviewer, which in this case is inconsistent between the reports. It is my understanding that an editorial decision should be based on arguments and facts, not on the reviewer’s emotions.
Nonetheless, we have replied in detail, again, to all criticisms. We have pointed out in particular the factual errors of reviewer #1. To help the editors, we have underlined the important points. We would appreciate if the editors checked for themselves whether reviewer #1 is right or not. We have also replied to reviewer/senior editor Dr YYY, although I deeply regret that this fourth version is “the first one he has had the chance to read”.
Finally, I would like to call your attention on the conclusion of reviewer #1, on which his/her recommendation is based, which requires in my opinion a clarification from the journal: “Finally, now in their third revision, the authors acknowledge that this work strongly builds on the previous resistive-coupling hypothesis, and tests whether this hypothesis is compatible with sharp spike onset (a view they have already proposed), vs the alternative proposed by Yu, of back propagation. This very specific theoretical result I feel is unlikely to be of broad interest to the computational biology field, but could be of interest to computational neuroscientists” (Please see also our response, pointing out that the said acknowledgement was clear already in the very first version.)
This recommendation makes some important presumptions about this journal’s editorial views. Therefore I would very much like to know if this journal:
- also considers that proposing a hypothesis is more important than testing one, and that only the former should be published;
- considers that interesting computational neuroscience studies do not belong to this journal. I would also very much like to know if this journal considers that it is ok for a reviewer to ask for substantial technical revisions when he/she has already decided that the paper should not be published anyway. This has been indeed a lot of work for a decision ultimately based on the mood of one reviewer.
As I have argued in this letter, it is very clear that, given the content of the reports of the 3 reviewers and of our responses, this manuscript should have been accepted already. After 8 months and 4 revisions, and no serious objection on the manuscript, I can only hope very much that this journal does not confuse rigorous peer review with author harassment.
Again, I am hoping that this letter will be seriously taken into consideration, and even perhaps responded to.
Response of the editors-in-chief
Dear Dr. Brette,
Thank you for your response to the recent decision on your paper “The Basis of Sharp Spike Onset in Standard Biophysical Models”. The manuscript and your appeal letter have been carefully evaluated by Dr. XXX and the journal’s Editors-in-Chief.
We understand your frustration regarding the length and complexity of the review process, and we would like to apologize for the time taken to reach a final decision.
We would like to provide some further clarification on how the editorial decision was reached. The manuscript addresses the issue - how do cortical neuronal action potentials rise so sharply? – and after an initial evaluation, Dr. XXX found it interesting enough to merit sending out for review, so that the reviewers could assess the technical solidity of the work and the conceptual advance proposed. The paper received mixed reviews, and hence merited a revision. After several rounds of revision, Reviewer 1 remained unconvinced. In order to aid the review process, Dr. YYY volunteered to evaluate the paper in depth, and his opinion concurred with that of Reviewer 1. Dr. XXX also re-read the paper and came to the conclusion that this manuscript is critically close conceptually to the previous PLOS publications - in fact the idea was laid out clearly and beautifully in the 2013 and 2015 PLOS papers. The present manuscript is an implementation of this idea, showing that other biophysically realistic models used to examine the spike sharpness issue show the mechanism that was suggested in the 2013 and 2015 PLOS papers.
We regret that this did not become fully clear before the third revision, and we understand your disappointment with the final outcome.
However, we agree that the findings of the paper are not significant enough for PLOS Computational Biology, and we will not be reconsidering the paper. We are sorry not to be more encouraging, but we hope that you can understand the reasons for this decision.
I wish I could make the world a better place. I would like to prevent climate change and wars; but that’s not so easy on a personal level. What I can try to do more modestly as a scientist, is to make the scientific world a better place. We have all heard the phrase “publish or perish”. We all complain that careers are made by publishing in “high-impact” journals who favor story-telling and grand claims, and generally select papers arbitrarily (let alone that they do not even predict the impact of papers they publish); a trend that has been increasingly strong and has very negative impact on how research is done, including serious ethical problems. But what do we do concretely about it? For most of us including myself, not much. We keep on submitting to those journals, and we say we have no choice because that is how we are evaluated (for positions or grants). But who evaluates us? Surely there are some political aspects to it (mostly for funding), but the truth is, we are evaluated by peers. In brief, we are inflicting this madness on ourselves.
So: let us stop complaining and try to change the situation. I have previously exposed a vision of how the academic publishing system could look like without journals (by the way, this is not an original thought, fortunately). How to make it happen?
Where we are now
We should be hopeful, because many good things are happening:
- Preprint servers are getting more and more attention. In biology, a number of journals are now accepting direct submissions from biorxiv, including all the PLoS journals, PNAS, eLife, the Journal of Neuroscience. This ought to boost submissions of preprints.
- A few journals have started publishing the reviews along with the accepted paper, for example eLife, eNeuro and Nature Communications.
- More generally, open access to both paper and data is getting more and more common and enforced.
- A new journal, F1000 Research, now practices post-publication review. The paper is indexed in pubmed once two reviewers have approved it.
- Very significantly, the Wellcome Trust has opened a journal for its grantees, Wellcome Open Research, based on post-publication review (in partnership with F1000), with this statement “The expectation is that this, and other similar funder platforms that are expected to emerge, will ultimately combine into one central platform that ensures that assessment can only be done at the article level”.
- Finally: Pubpeer, started just a few years ago. A simple idea: to let anyone comment on any paper, anonymously or not, and let the authors know and respond. You should install their browser plugin. This is an individual initiative but it has already made a big impact, in particular by showing that the “big journals” are not better than the other ones in preventing flaws or frauds. It also addresses the concern that open reviews would be too nice: anyone who finds serious flaws can spot them anonymously and the authors will have to consider them. Pubmed commons is similar, but with signed comments.
What we can do now on a personal level
- Put every single paper you write on a “preprint” server before you submit to a journal.
- Put all your data online, see eg OpenAIRE.
- Remove journal names from the publications in your website. People who care about them will find out anyway.
- Start a literature search routine that does not involve looking at tables of contents; a few ideas in this Twitter thread; you could also have an author alert on Google Scholar.
- Write comments on Pubpeer; including on “pre-prints”.
- Send your papers to a journal with open post-publication review. I know this one is difficult, because the community still cares about impact factors. But at least you can favor those with public reviews (eg eLife, Nature Communications; I would prefer the former as it is non-profit). Instead of sending your papers to Frontiers, send them to F1000 Research; or at least eNeuro.
At the local community level, we can advocate for post-publication review. For example, the INCF has opened a channel on F1000 Research. Maybe we could have a computational neuroscience channel there, sponsored by the OCNS. It is too bad that F1000 Research is for-profit rather than institutional, but currently I do not know of other options.
What we can do on a more global scale
Open post-publication review potentially addresses the issue of recognition, but it does not address the issue of visibility. One concern I have by submitting in F1000 Research (for example), is that my intended readership will not know about it. There are so many papers published each year, one does not even have the time to read the title of all of them. This is one role journals have fulfilled: to select papers worthy of interest for a given community. But since we do not need journals anymore to publish anything, editorial selection and publication need not be coupled anymore. So here is my proposition. We make an independent website which lets any scientist, or possibly any group of scientists, be their own journal. That is, make a selection of papers they find interesting (including preprints). We provide a number of tools to make this as simple as possible: linking to pubpeer and pubmed commons, searching/browsing, alerting authors whose work is selected, email alerts and RSS feeds, etc. Papers are preferentially linked to the preprint if it exists, so as to completely bypass the journals. We could also let authors suggest their own paper for editorial selection. Basically, we provide all the services a journal typically has. This will be made increasingly easier as public open reviews become more widespread. These new “journals” could be run by an individual scientist, or a lab (eg linked to a journal club), or possibly a scientific society or group of scientists. Let us call any of these an “editor”. I would be happy for example to follow the selections of a few authors I respect, and that would be probably more valuable to me that the selection made by any journal, of which very few typically catch my attention in a given table of contents.
I am hoping that it goes as follows:
- People start using these individual journals, because it provides relevant information to them.
- As a result, papers in less well-known journals and preprints start getting more attention, and more citations.
- People take the habit of putting their papers on preprint servers because they get immediate attention.
- Editors progressively stop selecting papers published in journals because they have already selected them when they were preprints.
- As editors are also committee members, journal names start to matter less in evaluating research.
- Traditional journals disappear; instead, we have direct publication (formerly known as preprints) + open public reviews, both anonymous and signed.
How to get it started?
One simple idea to get it started is to make automatic channels for the actual conventional journals. For each journal, we list the table of contents, linked to preprint versions instead of the journal website, and to pubpeer, etc. If it’s convenient, people might start to use it, especially if it allows free access to the papers (legally, since we would use preprints). Then to get people to use the non-conventional channels, we provide suggestions based on content similarity (ie “you may also like...”).
How about this resolution?
In 2016, the academic publication system worked mostly as follows :
- Scientists do some research and write a paper about it to share their findings with the world.
- The paper is submitted to a journal, chosen for its prestige.
- The editor, possibly a professional editor (ie not a professional scientist), judges that the paper is not worth of review: go back to (2).
- The editor sends the paper to 2-3 reviewers, asking them to review the paper quickly (the more prestigious, the quicker). Reviewers are chosen based on the editor’s expertise and network.
- Reviewers have too many reviews to write already: go back to (4).
- The reviewers (or one reviewer) find the paper uninteresting, or bad, or they “don’t believe it”: go back to (2).
- The reviews are mixed: a third reviewer is asked; go back to (4).
- The reviewers find the paper interesting enough but have requests for the authors; they should entirely rewrite their paper; they should do new experiments; they should cite the reviewers; they should remove an uninteresting figure; they should add some discussion points.
- After some work, the authors submit a revised paper.
- Actually, reviewers didn’t think the paper was that interesting anyway: go back to (2).
- The reviewers have more requests: go back to (9).
- The paper is accepted: congratulations!
- The format of references does not follow the journal’s style: please rewrite the reference list.
- The author’s paper now belongs to Elsevier/Springer Nature/etc. It will be sold by the publisher to academic institutions, with a margin of 30-40%.
- (recent alternative option) The authors pay publication fees of about $2000. The paper is now freely accessible.
- A knowledgeable reader disagrees with the main conclusion of the paper: she writes a letter to the editor.
- After careful consideration, the letter is rejected.
- Another reader, not an expert of the field, reads the paper in a prestigious journal and believes it is highly regarded in the community, until she talks with colleagues and finds out that the paper is very criticized.
- The prestigious paper is commented in mainstream journals. Experts rant about the journalists.
- The first author applies for a position. The committee examines CVs of all candidates and selects those with publications in prestigious journals. The senior author sends an excellent letter of recommendation. The committee values it if the senior author is famous.
- The senior author applies for a grant. The committee selects candidates with a good “track record”, ie prestigious publications.
This publication model is wasteful, full of perverse incentives, opaque, and misleading. What I want to do here is not one more critique of this system, but rather give a vision of what the publication system could be and, in my opinion, will be, perhaps within a decade. This is a vision of a publication system centered on individual articles, rather than on journals. It is made possible today by information technology (aka “the internet”). What stands in its way is, in a word: conservatism. It works as follows:
- Scientists do some research and write a paper about it to share their findings with the world.
- The paper is uploaded to an archive site, together with the data. All is freely accessible to anyone.
- The authors invite an unlimited number of professional scientists (say, authors of at least one paper) to write a formal, signed, review. The reviews are public.
- Any other scientist can contribute a signed or anonymous review or comment, including comments on the public reviews (aka “scientific debate”).
- Authors respond to the reviews and may submit a revised version.
- Authors decide when to stop the revision process; reviewers express their satisfaction, or not.
- Any scientist can recommend the paper if she finds it worth reading. Tools are made available to search for papers, possibly using the number of recommendations, possibly weighted by scientist reputation (all data are freely accessible and third parties provide searching tools).
- A knowledgeable reader disagrees with the main conclusion of the paper: she writes a public critique, signed or anonymous, after reexamining the paper’s raw data. The authors respond publicly.
- Another reader, not an expert of the field, reads the paper. She notices that many knowledgeable scientists disagree with the conclusions and can read the detailed arguments and the authors’ response.
- The first author applies for a position. For every publication, the committee reads the reaction of the community (formal reviews, comments, recommendations) and uses it to evaluate the value of the work along all the dimensions addressed in the reactions (importance, originality, technicality, etc). The committee also asks the candidate what her main contributions are in her view and why. The candidate can point to published reactions to her work that support her claims.
- The senior author applies for a grant or a promotion. The committee reads the reaction of the community to the candidate’s previous work. They might also look at review invitations and reviewing activity.
In the post-journal world, a paper is submitted a single time (but of course it can be revised), and in contrast with the journal model, the reaction of the community is known. As a consequence, quality stamps (journal names) are obsolete. No review is thrown away. As a consequence, scientists save time on both submitting and reviewing. Scientists are not incentivized to make grand claims. In addition, the system saves several billions of dollars each year (eg US libraries spend $2 billion each year on subscriptions), not counting the time saved on wasted reviews and submissions.
All of this is clearly possible with today’s technology. The next question I will try to address is: how to move quickly to the post-journal world?