import nltk from sumy.parsers.plaintext import PlaintextParser from sumy.summarizers.luhn import LuhnSummarizer from sumy.nlp.tokenizers import Tokenizer nltk.download('punkt') def summarize(file, SENTENCES_COUNT): sumarizer = LuhnSummarizer() with open(file.name) as f: doc = f.read() sentences_ = [] doc_ = PlaintextParser(doc, Tokenizer("en")).document for sentence in sumarizer(doc_, SENTENCES_COUNT): sentences_.append(str(sentence)) summm_ = " ".join(sentences_) return summm_