Edit model card

This is a version of flan-t5-xl fine-tuned on the KELM Corpus to take in sentences and output triplets of the form subject-relation-object to be used for knowledge graph generation.

The model uses custom tokens to delimit triplets:

special_tokens = ['<triplet>', '</triplet>', '<relation>', '<object>']
tokenizer.add_tokens(special_tokens)

You can use it like this:

model = model.to(device)
model.eval()

new_input = "Hugging Face, Inc. is an American company that develops tools for building applications using machine learning.",
inputs = tokenizer(new_input, return_tensors="pt")

with torch.no_grad():
    outputs = model.generate(input_ids=inputs["input_ids"].to("cuda"))
    print(tokenizer.batch_decode(outputs.detach().cpu().numpy(), skip_special_tokens=False)[0])

Output: <pad><triplet> Hugging Face <relation> instance of <object> Business </triplet></s>

This model still isn't perfect, and may make mistakes! I'm working on fine-tuning it for longer and on a more diverse set of data.

Downloads last month
10
Inference Examples
Inference API (serverless) does not yet support peft models for this pipeline type.

Model tree for bew/t5_sentence_to_triplet_xl

Base model

google/flan-t5-xl
Adapter
(37)
this model

Dataset used to train bew/t5_sentence_to_triplet_xl