--- license: apache-2.0 datasets: - yahma/alpaca-cleaned language: - en base_model: - tiiuae/Falcon3-10B-Instruct tags: - falcon3 - alpaca pipeline_tag: text-generation library_name: transformers --- Barcenas 10b Based on the tiiuae/Falcon3-10B-Instruct and trained with the yahma/alpaca-cleaned dataset. The objective of this new model is to explore finetuning on the new falcon 3 models. Made with ❤️ in Guadalupe, Nuevo Leon, Mexico 🇲🇽