- Jobayer Ahmmed
- Jahid Hasan
- Open a Linux Terminal
- Clone the repo:
git clone https://github.com/SigmaWe/SueNes_RE.git - Go to SueNes_RE directory:
cd SueNes_RE - Give execution permission to run.sh file:
chmod +x run.sh - Finally, run the script:
source run.sh
We trained two different models from the same checkpoint. One is using Tensorflow and other one is using PyTorch. The run.sh scipt runs all the python files for training the two models and testing them with sample data. For testing, we call our trained model with three pairs of document and summary. The original scores and the predicted scores are shown in the terminal.
The rest of the part is step-by-step instructions.
The transformer directory contains code for training transformer-based models with different datasets.
The datasets were generated using sentence delete or word delete techniques
mentioned in the SueNes paper.
You can create virtual environment using Python or Conda.
git clone https://github.com/JobayerAhmmed/SueNes.gitcd SueNespython3 -m venv .venvsource .venv/bin/activatepip install -r requirements.txtpython -m spacy download en_core_web_smpip install transformers datasets scikit-learn evaluate pyyaml h5py- Issue: replace
from keras.saving.hdf5_formatbyfrom tensorflow.python.keras.saving.hdf5_formatat line 39 of.venv/lib/python3.10/site-packages/transformers/modeling_tf_utils.py
- Create venv following this documentation
pip install tensorflow tensorflow-datasets tensorflow_hub- Install PyTorch following this documentation
pip install joblib numpy nltk matplotlib bs4 spacy stanzapython -m spacy download en_core_web_smpip install transformers datasets scikit-learn evaluate pyyaml h5py
mkdir exp exp/data exp/resultcd prepython3 sentence_scramble.py
Code for the model is in bert_tiny_cnndm_tf.py file.
This model is trained from checkpoint found in
prajjwal1/bert-tiny.
Data is generated from CNN Daily Mail dataset using
SueNes.
Only sentence delete technique, defined in
SueNes paper,
is used for data generation.
Only 10% data is considered from CNN Daily Mail dataset's train split
for generating train split for our experiment.
cd transformerpython3 bert_tiny_cnndm_tf.py
python3 bert_tiny_cnndm_tf_wrap.py
Code for the model is in bert_tiny_cnndm_pt.py file.
This model is trained from checkpoint found in
prajjwal1/bert-tiny.
Data is generated from CNN Daily Mail dataset using
SueNes.
Only sentence delete technique, defined in
SueNes paper,
is used for data generation.
Only 10% data is considered from CNN Daily Mail dataset's train split
for generating train split for our experiment.
cd transformerpython3 bert_tiny_cnndm_pt.py
python3 bert_tiny_cnndm_pt_wrap.py