yonsei-sslab / EmailBERT

RoBERTa finetuned on Enron email dataset & aeslc email dataset

Home Page:https://huggingface.co/snoop2head/EmailBERT-base

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

EmailBERT

Pretrained RoBERTa finetuned on Enron email dataset & aeslc email dataset with Masked Language Modeling train objective.

How to use

from transformers import AutoTokenizer, pipeline

model_checkpoint = "snoop2head/EmailBERT-base"
tokenizer = AutoTokenizer.from_pretrained(model_checkpoint)
mask_filler = pipeline(
    "fill-mask", model=model_checkpoint
)
text = f"nPlease {tokenizer.mask_token} the following distribution list with updates:\nPhillip Allen (pallen@enron.com)\nMike Grigsby (mike.grigsby@enron.com)\nKeith Holst (kholst@enron.com)\nMonique Sanchez\nThank you for your help\nPhillip Allen'"

preds = mask_filler(text)

for pred in preds:
    print(f"Mask Filled: {pred['sequence']}")

References

About

RoBERTa finetuned on Enron email dataset & aeslc email dataset

https://huggingface.co/snoop2head/EmailBERT-base


Languages

Language:Python 100.0%