File size: 1,171 Bytes
5fa1a76 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 |
thon from transformers import PLBartForConditionalGeneration, PLBartTokenizer tokenizer = PLBartTokenizer.from_pretrained("uclanlp/plbart-python-en_XX", src_lang="python", tgt_lang="en_XX") example_python_phrase = "def maximum(a,b,c):NEW_LINE_INDENTreturn max([a,b,c])" inputs = tokenizer(example_python_phrase, return_tensors="pt") model = PLBartForConditionalGeneration.from_pretrained("uclanlp/plbart-python-en_XX") translated_tokens = model.generate(**inputs, decoder_start_token_id=tokenizer.lang_code_to_id["en_XX"]) tokenizer.batch_decode(translated_tokens, skip_special_tokens=True)[0] "Returns the maximum value of a b c." Resources Text classification task guide Causal language modeling task guide Translation task guide Summarization task guide PLBartConfig [[autodoc]] PLBartConfig PLBartTokenizer [[autodoc]] PLBartTokenizer - build_inputs_with_special_tokens PLBartModel [[autodoc]] PLBartModel - forward PLBartForConditionalGeneration [[autodoc]] PLBartForConditionalGeneration - forward PLBartForSequenceClassification [[autodoc]] PLBartForSequenceClassification - forward PLBartForCausalLM [[autodoc]] PLBartForCausalLM - forward |