What is the point?

Introduction

Why do I even bother to do research and submit to “1st Tier Conferences” when it is going to get rejected anyway? What is the point? My advisor says that the review process is very very noisy and we should not get disheartened; we should make improvements and resubmit again. He is probably right about the extreme noise because of the number of people submitting and the occasional experimentation in the review process. But I find the idea of “resubmitting again and again till your paper gets outdated and then someone pity-accepts the paper” inelegant and disheartening. I would have been fine with improving the work if there were actual pitfalls that needed to be corrected. But sigh! That is never the case.

I am starting to find the process of review and rebuttal very emotionally draining as well as pointless. I’m not saying it does not work but it has not worked for me even once properly. By now I have gone through 4 cycles of reviews and rebuttals and every time it is just heart-breaking. There are only so many pieces that your heart can break into before it becomes irreparable. Besides, I realised that the immense amount of energy, effort, and time that goes into creating a work does not matter to a reviewer. If a reviewer has made up their mind in the initial review, it is highly improbable that they change their mind in the final review. Further, it does not even seem that it matters to the Area Chair/ Program Chair if the paper has any relevance.

I am going to share the experience of the review of one of my papers. I think they are important to understand where I am coming from but feel free to skip them to the conclusion if you want.

Example

In this example, I talk about how a work of mine that got preposterous reviews. In this, we had an insight that (text-free) diffusion model features have emergent recognition powers, which makes diffusion process a self-supervised pretraining method. We first analysed the diversity in these features to be able to come up with methods to harness this knowledge in a meaningful manner. Then we also showed competitive results on multiple tasks to show its effectiveness.

Some of the reviews were completely unrelated, eg. asking why we were text-free when the paper is in a self-supervised representation learning category, or why we didn’t compare/mention methods that use text; while others are unfair eg. disregarding a paper for not being SOTA, or questioning a hypothesis when we show empirical results; while some other are just plain wrong eg. misquoting papers, wrongly citing papers, saying that our method is the same as some completely different paper which models generation hence incremental, or disregarding the main heavy training stage of the SOTA competitor as an advantage we have. In regards to these, they go on to give comments which are very demoralizing as well as demeaning. And the PC final review was the final nail in the coffin being an utter concoction of balderdash.

version 1

I had a paper D1 which was majorly based on one primary insight (which I think was quite an exciting finding). The reviewers questioned (1) the motivation of the paper, (2) the fact that we don’t use “text”, (3) the performance gap, and (4) there being concurrent papers that explore one of all the aspects that we look at (obviously they do it differently). One of them asked to do 640 GPU hours of ablation (which we didn’t have the resources or time to do) and further goes on to say “I am not sure all reviewers will sufficiently raise their score for this paper to get through, but maybe this discussion is still useful to make a better paper.” (which is quite disheartening to get before the final review). Finally, the program chair rejected it quoting the “lack of thoroughness of this experimental study”, and the results being “mildly interesting” and “not [..] strong, impactful”. There was some constructive criticism as well which asked us to show results on more tasks and use the insight to build a method.

version 2

We took their advice for the next iteration of the paper, D2. In D2, D1 was merely a subpart of the analysis section. (1,2) We rewrote the introduction to handle the motivation and the “text-free” nature besides adding “text-free” in the title itself. (3) We introduced 2 new methods and showed their effectiveness on 5 tasks where were are now within +-2% of every method we compare against. (4) We also added a section in the related works talking about concurrent works that handle only one task using text-based models. So, we addressed all the flaws we had in the previous D1 and hence the paper should get in this time, right? Right?!

You guessed it. The answer is NO!

Initial reviews -

The reviewers praise our writing, experiments, and analysis.

Rebuttal -

We pointed out to the AC that rejection based on SOTA is against the guidelines and that we were being questioned for “text-free” in “Self-Supervised Representation Learning” category. Also pointed out that R2 makes incorrect statements.

Final reviews -

Discussion and Conclusion

I don’t think I have a conclusion because I am both confused and disappointed. How do you have faith in such a system where scientific evidence is not enough to convince the community to accept a work? Here the paper will get accepted only if your views align with those of the reviewers (who apparently can represent the whole community). Here the reviewers already make up their minds based on their beliefs and will not change it even with evidence bolstering your point. Here new insights are useless if you do not beat the SOTA. And if you do beat the SOTA your work is not novel enough. I can no longer even have faith in the AC/PC system as well because that also seems noisy. The chairs seem to not properly read the paper or the rebuttals to base their final judgment. I appreciate the reviewers, ACs, PCs to take out their time to undergo the review process but I am losing all my faith in the outcome and effectiveness of this system. Anyway, people are using LLMs to write their reviews, might as well remove the middle-“MAN”. Scaling may solve video generation but it seems to fail miserably in peer-review.

When a very similar study was done on a very small scale with results on CiFAR, MNIST and consequently got performance comparable to standard methods, even though not SOTA, it ended up getting an “Oral paper”. But us showing that this insight does not hold in larger datasets like ImageNet out of the box was “mildly interesting” and proposing methods to make it comparable to standard methods was “unconvincing [..] performance improvements”, “unjustified design choices”, and “incremental contributions”.

I don’t even understand why they gave this paper borderline in the first place if they didn’t want to accept it. Were they expecting that we would make it SOTA in the rebuttal period or did I say things in the rebuttal that they didn’t agree with? The PC came up with new unfair reasons to reject the paper without proper justification.

One conclusion could be that the system is extremely flawed. But the flip side would be that I’m not good enough for this field. My ideas are boring, incremental, and unacceptable for this fast-moving community. All ideas anyway get outdated and are declared useless because of concurrent works in a similar domain (even though they can never be the exact same). Anyway, there is no breathing space, the reviews/rebuttals further choke all the enthusiasm out of you. To catch up, currently, I put all my hours in the day toward research. And still, I lag in the amount of work that I produce and out of that even less gets accepted.

In the past three years of my PhD I have dedicated most of my time to work, leading to the loss of my hobbies. I knew that PhD would be difficult but it is also draining me from inside. Yesterday, when I asked what was the point of any of this it was going to get rejected anyway. My advisor said that you should do it because it makes you happy. And doing research makes me immensely happy but at the same time, these unjustified rejections make me extremely sad and I’m starting to question if the latter is more. So,back to the question – what is the point?


  1. These references are from the original paper.