--- library_name: transformers tags: - text summarization license: apache-2.0 language: - en metrics: - rouge pipeline_tag: text2text-generation --- # Model Card for Post-Disaster Digital Help Desk Summarization Model This model is designed to summarize digital help desk conversations in post-disaster scenarios, specifically tailored for non-profit organizations providing aid. It is based on the BART model, fine-tuned using parameter-efficient methods like LoRa adapters. ## Model Details ### Model Description This is a parameter efficient finetuned model based on the fine-tuning of the BART model. the methodology used is the LoRa adapter. this model focuses on automated text summarization of digital helpdesk conversations in post-disaster assistance scenarios in order to improve the efficiency and quality of the information gathered to provide timely and effective support to the affected people. ## Uses The model is designed to summarize digital help desk conversations for nonprofit organizations in post-disaster assistance scenarios, helping digital help desk staff to quickly extract key information and reduce the time it takes to manually write high-quality summaries. ## Bias, Risks, and Limitations Generated summaries may contain certain errors, such as the inclusion of sensitive information, and require manual secondary correction to ensure accuracy and privacy protection. ## How to Get Started with the Model Use the code below to get started with the model. ```python # install package !pip install transformers[torch] -U !pip install -q -U peft import torch from transformers import AutoModelForSeq2SeqLM, AutoTokenizer from huggingface_hub import notebook_login # login to hugging_face notebook_login() # use model on GPU device = "cuda" if torch.cuda.is_available() else "cpu" # load base model model_name = "knkarthick/MEETING_SUMMARY" model = AutoModelForSeq2SeqLM.from_pretrained(model_name).to(device) tokenizer = AutoTokenizer.from_pretrained(model_name) # load trained adapter adapter_id = "Joaaaane/510_ABW_LoRaAdapter_PostDisasterConv" model.load_adapter(adapter_id) # set the model to evaluation mode model.eval() input_text = """ PA: Hello, I need urgent housing help as a refugee from Ukraine. Can you assist? agent: Hello, thank you for reaching out to the Red Cross. We’re here to help with housing. agent: Have you registered with the local authorities yet? PA: Yes, but they mentioned delays, and we need something soon. It's urgent. agent: We have temporary shelters available. How many are with you, and are there any special needs? PA: It's just me and my elderly mother; we need accessible housing. agent: We can arrange for accessible temporary shelter. I’ll expedite your request and aim to place you within a few days. agent: I'll also connect you with a Ukrainian-speaking volunteer to help with your paperwork and make your mother more comfortable. PA: Thank you so much. This help means a lot to us right now. agent: You're welcome! Expect a call from our volunteer by tomorrow. We’ll make sure you both are settled quickly. PA: Thanks again. Looking forward to resolving this soon. """ # tokenized inputs inputs = tokenizer(input_text, return_tensors="pt", max_length=1024, truncation=True).to(device) # generate summary tokens outputs = model.generate(inputs['input_ids'], max_length=62, num_beams=5, early_stopping=True) # decode tokens summary = tokenizer.decode(outputs[0], skip_special_tokens=True) print("Generated Summary:", summary) ``` ## Training Details ### Training Data Data provided by 510, an initiative of the Netherlands Red Cross (all confidential data has been masked). ### Testing Data Data provided by 510, an initiative of the Netherlands Red Cross (all confidential data has been masked). ### Metrics ROUGE Score ### Results | Metric | Before LoRA | After LoRA | |--------------------|-------------|------------| | **ROUGE 1** | 22.50 | 28.30 | | **ROUGE 2** | 4.96 | 8.64 | | **ROUGE L** | 17.24 | 22.50 | ## Citation Base model: https://huggingface.co/knkarthick/MEETING_SUMMARY