Soaring-8b / README.md
minnmamin's picture
Upload LlamaForCausalLM
e48b23f verified
|
raw
history blame
2.85 kB
metadata
library_name: Llama3-8b-FlyingManual-Tutor
tags:
  - llama3
  - flying-manual
  - ai-tutoring
  - llama-factory

Model Card for Llama3-8b-FlyingManual-Tutor

This model is a fine-tuned version of the Llama3-8b model, specifically trained on the FlyingManual dataset to serve as an AI tutor for aviation-related subjects. It is designed to provide guidance and nudge users when they answer questions incorrectly.

Model Details

Model Description

  • Developed by: Canarie Teams
  • Model type: Large Language Model (LLM) for AI Tutoring
  • Language(s) (NLP): English (primary), potentially others depending on the FlyingManual dataset
  • Finetuned from model: Llama3-8b by Meta AI

How to Get Started with the Model

from transformers import AutoTokenizer, AutoModelForCausalLM

model_name = "path/to/your/Llama3-8b-FlyingManual-Tutor"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

# Example usage for tutoring
def tutor_interaction(question, user_answer):
    prompt = f"Question: {question}\nUser Answer: {user_answer}\nTutor Response:"
    inputs = tokenizer(prompt, return_tensors="pt")
    outputs = model.generate(**inputs, max_length=200)
    response = tokenizer.decode(outputs[0], skip_special_tokens=True)
    return response.split("Tutor Response:")[-1].strip()

# Example
question = "What are the primary flight controls of an aircraft?"
user_answer = "Steering wheel and gas pedal"
tutor_feedback = tutor_interaction(question, user_answer)
print(tutor_feedback)

Training Details

Training Data

The model was fine-tuned on the FlyingManual dataset, augmented with:

  • Sample Q&A pairs related to aviation topics
  • Examples of constructive feedback and explanations
  • Scenarios demonstrating correct and incorrect responses to aviation-related questions

Training Procedure

Preprocessing

  • Conversion of training data into a dialogue format suitable for tutoring interactions
  • Augmentation of data with tutoring-specific tokens or markers

Training Hyperparameters


Evaluation

Testing Data, Factors & Metrics

Testing Data

A held-out portion of the FlyingManual dataset, supplemented with:

  • A set of typical student questions and answers
  • Scenarios designed to test the model's ability to provide constructive feedback

Metrics

  • Human evaluation of tutoring quality (clarity, accuracy, helpfulness)
  • Task-specific metrics (e.g., ability to correctly identify and address user mistakes)

Results

[Provide the evaluation results here]

Environmental Impact

  • Hardware Type: 8 x NVIDIA A100 40GB GPUs

Model Card Authors

Canarie Teams

Model Card Contact

[Your contact information or a link to where people can reach out with questions]