---
inference: false
pipeline_tag: image-text-to-text
datasets:
- teamcraft/TeamCraft-Data-Dec
---
# TeamCraft-VLA-7B-Dec Model Card
TeamCraft-VLA-7B-Dec is a multi-modal vision-language action model designed for decentralized multi-agent collaborations. The model encodes multi-modal prompts specifying the task, one agent's visual observation and inventory at each timestep to generate actionable output for single agents under multi-agent settings.
## Usage
We provide a full environment with detailed running instruction on [GitHub](https://github.com/teamcraft-bench/teamcraft).
## Model details
The TeamCraft-VLA (Vision-Language-Action) architecture integrates a CLIP ViT-L/14 visual encoder with a linear projector for modality alignment and Vicuna-v1.5-7B (Llama 2.0) as the LLM backbone, combining visual and text embeddings to generate actions for multi-agent tasks.
**Model Type:**
- Vision-Language Action Model
**Model version:**
- v1.0
**Model date:**
- TeamCraft-VLA-7B-Dec is trained on September 2024
**Training dataset:**
- [Teamcraft decentralized full dataset](https://huggingface.co/datasets/teamcraft/TeamCraft-Data-Dec)
## Uses
### Direct use
- **Primary intended uses:**
The primary use of the TeamCraft-VLA-7B-Dec is research on multi-agents under multi-modal settings.
- **Primary intended users:**
The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, multi-agent system, and artificial intelligence.
### Out-of-Scope Use:
- The model is not designed for real-world decision-making or deployment in safety-critical systems.
- The model not be used for tasks requiring ethical reasoning, moral judgments, or any applications where improper actions could lead to harm or violation of regulations.
## License
Llama 2 is licensed under the LLAMA 2 Community License,
Copyright (c) Meta Platforms, Inc. All Rights Reserved.