from transformers import GPT2LMHeadModel, GPT2Tokenizer model_name = "gpt2" # You can also use "gpt2-medium", "gpt2-large", "gpt2-xl" # Download the tokenizer and model tokenizer = GPT2Tokenizer.from_pretrained(model_name) model = GPT2LMHeadModel.from_pretrained(model_name) # Optional test input_ids = tokenizer.encode("The world is", return_tensors="pt") output = model.generate(input_ids, max_length=20) print(tokenizer.decode(output[0]))