Update README.md
Browse files
README.md
CHANGED
@@ -64,7 +64,7 @@ The following hyperparameters were used during training:
|
|
64 |
```python
|
65 |
from transformers import pipeline
|
66 |
|
67 |
-
summarizer = pipeline("summarization", model="bart-large-cnn-finetuned")
|
68 |
|
69 |
text = """
|
70 |
The paper "Attention is All You Need" revolutionized the field of natural language processing (NLP) by introducing the Transformer architecture, which relies solely on attention mechanisms to model long-range dependencies in sequential data. Prior to this, models like recurrent neural networks (RNNs) and convolutional neural networks (CNNs) were the primary tools for sequence modeling, but they suffered from limitations such as difficulty in parallelization and the vanishing gradient problem. The Transformer, however, breaks free from these constraints by using a self-attention mechanism, which allows it to attend to different parts of a sequence simultaneously, leading to more efficient training and better performance on tasks such as machine translation, text summarization, and language modeling.
|
|
|
64 |
```python
|
65 |
from transformers import pipeline
|
66 |
|
67 |
+
summarizer = pipeline("summarization", model="luluw/bart-large-cnn-finetuned")
|
68 |
|
69 |
text = """
|
70 |
The paper "Attention is All You Need" revolutionized the field of natural language processing (NLP) by introducing the Transformer architecture, which relies solely on attention mechanisms to model long-range dependencies in sequential data. Prior to this, models like recurrent neural networks (RNNs) and convolutional neural networks (CNNs) were the primary tools for sequence modeling, but they suffered from limitations such as difficulty in parallelization and the vanishing gradient problem. The Transformer, however, breaks free from these constraints by using a self-attention mechanism, which allows it to attend to different parts of a sequence simultaneously, leading to more efficient training and better performance on tasks such as machine translation, text summarization, and language modeling.
|