Adjust training parameters
This commit is contained in:
parent
47407b9fb6
commit
d47afd47a3
2
bot.py
2
bot.py
|
@ -19,7 +19,7 @@ args = parser.parse_args()
|
|||
|
||||
|
||||
tokenizer = AutoTokenizer.from_pretrained('gpt2-large')
|
||||
model = AutoModelForCausalLM.from_pretrained(args.model, torch_dtype=float16).to('cuda')
|
||||
model = AutoModelForCausalLM.from_pretrained(args.model).to('cuda')
|
||||
|
||||
|
||||
if args.input is None:
|
||||
|
|
4
train.py
4
train.py
|
@ -47,7 +47,7 @@ lm_dataset = tokenized_dataset.map(group_texts, batched=True)
|
|||
# Create and train the model
|
||||
model = AutoModelForCausalLM.from_pretrained('gpt2-large',
|
||||
torch_dtype=float16, low_cpu_mem_usage=True).to('cuda')
|
||||
trainer = Trainer(model, TrainingArguments(output_dir=args.output, per_device_train_batch_size=1,
|
||||
gradient_accumulation_steps=8), default_data_collator, lm_dataset['train'])
|
||||
trainer = Trainer(model, TrainingArguments(output_dir=args.output, per_device_train_batch_size=1),
|
||||
default_data_collator, lm_dataset['train'])
|
||||
trainer.train()
|
||||
trainer.save_model()
|
||||
|
|
Loading…
Reference in a new issue