Model Trained Using AutoTrain
This model was trained using AutoTrain. For more information, please visit AutoTrain.
Usage
from transformers import pipeline
pipe = pipeline("text-generation", "MadKFC/CPsyCode",device_map="auto")
messages = [
{"role": "user", "content": "I'm stupid!"},
{'role': "assistant", "content": "I'm so sorry to hear you're feeling that way! It's completely normal to have moments where we doubt ourselves or feel uncertain. But here's the thing: being 'stupid' is not a fact, it's an opinion. And opinions are subjective and can be influenced by many factors, including our own self-talk and the opinions of other\n\ns.You are not stupid. You are capable, unique, and deserving of kindness and compassion - from yourself and others. It's okay to make mistakes, and it's okay to not know something. That's how we learn and grow!\n\nWould you like to talk about what's going on"},
{"role": "user", "content": "I make a big mistake in the lab"},
]
answer = pipe(messages)
# Print the assistant's response
print(answer[0]['generated_text'][-1])
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Model tree for MadKFC/CPsyCode
Base model
meta-llama/Meta-Llama-3-8B-Instruct