Fine-tung gpt-2 run_clm.py stops early

Hi, I’m running run_clm.py to fine-tune gpt-2 form the huggingface library, following the language_modeling example:

!python run_clm.py \
    --model_name_or_path gpt2 \
    --train_file train.txt \
    --validation_file test.txt \
    --do_train \
    --do_eval \
    --output_dir /tmp/test-clm

This is the output, the process seemed to be started but there was the ^C appeared to stop the process:

The following columns in the training set don't have a corresponding argument in `GPT2LMHeadModel.forward` and have been ignored: .
The following columns in the evaluation set don't have a corresponding argument in `GPT2LMHeadModel.forward` and have been ignored: .
***** Running training *****
  Num examples = 2318
  Num Epochs = 3
  Instantaneous batch size per device = 8
  Total train batch size (w. parallel, distributed & accumulation) = 8
  Gradient Accumulation steps = 1
  Total optimization steps = 870
  0% 0/870 [00:00<?, ?it/s]^C

Here’s my environment info:

  • transformers version: 3.4.0
  • Platform: Linux-4.19.112±x86_64-with-Ubuntu-18.04-bionic
  • Python version: 3.6.9
  • Tensorflow version: 1.14
  • Using GPU in script?: yes

What would be the possible triggers of the early stopping?

Hi, if you are running Tensorflow, don’t you need TFGPT2LMHeadModel rather than GPT2LMHeadModel ?

The ^C means you hit Control+C, which cancels any process.

@rgwatwormhill Thank you for your response, I will try importing the different model.
@sgugger Thank you for your response, the ^C actually appeared by itself without hitting control + C

Hi @kelichiu,

another thought: you might need Tensorflow version 2 instead of version 1.14. See this page