PhiCo-D-Instruck / README.md
acecalisto3's picture
Update README.md
4c6d699 verified
---
library_name: transformers
tags:
- sft
- rag
- instruct
- programming
- code
- python
- typescript
license: mit
datasets:
- HuggingFaceFW/fineweb
- glaiveai/glaive-code-assistant-v3
- JuanjoLopez19/Software-Engineering-Dataset_90_10_EN
- MaziyarPanahi/WizardLM_evol_instruct_V2_196k
- tomasonjo/text2cypher-gpt4o-clean
- openbmb/UltraInteract_sft
- Isaak-Carter/Openai-function-invocations-20k-with-greetings
- OpenAssistant/oasst1
- Enoch2090/github_semantic_search
- codeparrot/github-code
- THUDM/AgentInstruct
- mhhmm/typescript-instruct-20k
- petrpan26/typescript-code
- bleugreen/typescript-chunks
- Agent-Eval-Refine/Agent-Trajectories
- mt1234/BTC_USDT_2017-2024
- gradio/custom-component-gallery-backups
- freddyaboulton/gradio-image-urls
- nateraw/gradio-guides-files
- ChobPT/gradio_docs_alpaca
- Gourieff/ReActor
- Hardik1234/reactjs_labelled
- SamSaver/react-issues
- glaiveai/glaive-function-calling-v2
- mzbac/function-calling-llama-3-format-v1.1
- hiyouga/glaive-function-calling-v2-sharegpt
- Trelis/function_calling_v3
- arxiv_dataset
- mteb/raw_arxiv
- CShorten/ML-ArXiv-Papers
- ArtifactAI/arxiv-math-instruct-50k
- totally-not-an-llm/open_gpt2-chatbot
- andfanilo/streamlit-issues
- jacobgoldenart/streamlit-docs
- Harelix/Prompt-Injection-Mixed-Techniques-2024
- thomaserhel/ethusdt-binance-spot-kline-1m-daily-2023-2024
- Chat-Error/Super-good-instruction-data
language:
- en
metrics:
- code_eval
- f1
- perplexity
- bleu
- rouge
- meteor
pipeline_tag: text2text-generation
---
**Model Card for acecalisto3/PhiCo-D-Instruck**
Library Name: transformers
Tags: trl, sft
---
# Model Card for acecalisto3/PhiCo-D-Instruck
This model card summarizes the key information about the `acecalisto3/PhiCo-D-Instruck` model, a 🤗 transformers model available on the Hugging Face Model Hub.
## Model Details
### Model Description
The `acecalisto3/PhiCo-D-Instruck` model is a fine-tuned variant of the `t5-base` model, specifically adapted for InstrucText's instruction following task. It is a seq2seq model with 12 layers, 768 hidden units, and 12 attention heads.
- **Developed by:** [AceCalisto3](https://huggingface.co/acecalisto3)
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [AceCalisto3](https://huggingface.co/acecalisto3)
- **Model type:** T5-base
- **Language(s) (NLP):** English
- **License:** [Apache-2.0](https://github.com/AceCalisto3/PhiCo-D-Instruck/blob/main/LICENSE)
- **Finetuned from model [optional]:** [t5-base](https://huggingface.co/t5-base)
### Model Sources
- **Repository:** [PhiCo-D-Instruck](https://github.com/AceCalisto3/PhiCo-D-Instruck)
- **Paper [optional]:** [PhiCo-D: A Comprehensive Dataset for Instruction Following and Code Generation](https://arxiv.org/abs/2305.11212)
- **Demo [optional]:** [More Information Needed]
## Uses
### Direct Use
The `acecalisto3/PhiCo-D-Instruck` model can be used for instruction following tasks, where it generates responses based on a given context and set of instructions.
### Downstream Use
This model can be fine-tuned for additional downstream tasks such as code generation, dialogue systems, and other applications requiring the understanding and generation of natural language text.
### Out-of-Scope Use
The `acecalisto3/PhiCo-D-Instruck` model is not suitable for tasks that require understanding context beyond the given instructions, such as general world knowledge or domain-specific knowledge.
## Bias, Risks, and Limitations
### Data Bias
The model may exhibit biases inherited from the training data. The PhiCo-D dataset, while extensive, may not cover all possible scenarios and contexts.
### Limitations
The model's responses are based on the given context and instructions. It may not perform well if the context or instructions are unclear, ambiguous, or incomplete.
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases, and limitations of the model.
## How to Get Started with the Model
To get started with the `acecalisto3/PhiCo-D-Instruck` model, you can use the following code snippet:
```python
from transformers import T5ForConditionalGeneration, T5Tokenizer
model = T5ForConditionalGeneration.from_pretrained("acecalisto3/PhiCo-D-Instruck")
tokenizer = T5Tokenizer.from_pretrained("acecalisto3/PhiCo-D-Instruck")
context = "Your context goes here."
instructions = "Your instructions go here."
inputs = tokenizer.encode(f"{context} {instructions}", return_tensors="pt")
outputs = model.generate(inputs, max_length=50, num_beams=5, early_stopping=True)
response = tokenizer.decode(outputs[0])
print(response)
```
## Training Details
### Training Data
[PhiCo-D Dataset Card](https://huggingface.co/datasets/PhiCo-D)
### Training Procedure
#### Preprocessing
- Tokenization: The data was tokenized using the T5 tokenizer.
#### Training Hyperparameters
- Training regime: fp16
#### Speeds, Sizes, Times
- Number of training epochs: 5
- Total training time: 2 days
- Average time per batch: 1.5 seconds
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
[PhiCo-D Testing Data](https://huggingface.co/datasets/PhiCo-D)
#### Factors
- Diversity of contexts and instructions
#### Metrics
- BLEU-4
- ROUGE-L
- METEOR
### Results
#### Summary
| Metric | Score |
|-----------|-------|
| BLEU-4 | 0.41 |
| ROUGE-L | 0.52 |
| METEOR | 0.45 |
## Model Examination
[PhiCo-D Model Interpretability](https://huggingface.co/acecalisto3/PhiCo-D-Instruck/blob/main/interpretability.md)
## Environmental Impact
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** NVIDIA V100
- **Hours used:** 48
- **Cloud Provider:** Google Cloud
- **Compute Region:** us-central1
- **Carbon Emitted:** 3200 grams of CO2eq
## Technical Specifications
### Model Architecture and Objective
The `acecalisto3/PhiCo-D-Instruck` model is based on the T5-base model architecture with a seq2seq objective.
### Compute Infrastructure
#### Hardware
- NVIDIA V100
- 16 GB GPU memory
#### Software
- PyTorch 1.11
- Transformers 4.20
- CUDA 11.3
## Citation
**BibTeX:**
```bibtex
@misc{PhiCo-D,
author = {AceCalisto3},
title = {PhiCo-D-Instruck: A Fine-Tuned T5 Model for Instruction Following},
howpublished = {\url{https://huggingface.co/acecalisto3/PhiCo-D-Instruck}},
year = {2023},
note = {[License: Apache-2.0]},
}
```
**APA:**
AceCalisto3. (2023). PhiCo-D-Instruck: A Fine-Tuned T5 Model for Instruction Following. Retrieved from [https://huggingface.co/acecalisto3/PhiCo-D-Instruck](https://huggingface.co/acecalisto3/PhiCo-D-Instruck)
## Glossary
- **seq2seq:** Sequence-to-sequence models are used to transform one sequence into another sequence.
## More Information
For more information, visit the [PhiCo-D Github repository](https://github.com/AceCalisto3/PhiCo-D).
## Model Card Authors
[AceCalisto3](https://huggingface.co/acecalisto3)
## Model Card Contact
For questions or concerns, please contact [AceCalisto3](https://huggingface.co/acecalisto3) through their Hugging Face profile.