PreSumm icon indicating copy to clipboard operation
PreSumm copied to clipboard

Problem with generating longer summaries

Open v-zmiycharov opened this issue 5 years ago • 3 comments

Hi there! Great job on this fantastic implementation! I am trying to use PreSumm algorithm and fine tune it on dataset which has average full text length of 10000 tokens and summary of 800 tokens. I have 1 question:

  • Do I need to set max_pos parameter to >10000 in order to read the entire content of full documents And 1 issue:
  • Even though I set max_length parameter to 1200 the algorithm always generates summaries with about 120 tokens which reduces recall. How do I change this?

v-zmiycharov avatar Jan 09 '21 21:01 v-zmiycharov

I have the same Questions.

@nlpyang Can you kindly help us in this regard? If we want to summarize a document with 3000 token, we just need to set max-pos = 3000 in order to consider the whole document?

AyeshaSarwar avatar Jun 08 '21 19:06 AyeshaSarwar

I have the same Questions.

@nlpyang Can you kindly help us in this regard? If we want to summarize a document with 3000 token, we just need to set max-pos = 3000 in order to consider the whole document?

Since the max_pos of BERT-base is 512, so if you have longer input, I think you have to modify the model yourself, like using a sliding window.

Ga091an avatar Jun 25 '21 02:06 Ga091an

No, but the documentation says that if you have a longer input such as 800 so we can use max-pos = 800. And thus my Q is if I pass 4000 then it will consider first 4000 words?

AyeshaSarwar avatar Jun 29 '21 07:06 AyeshaSarwar