File size: 1,050 Bytes
1247a15 02162cc 78c0f10 9ca90e4 02162cc 78c0f10 02162cc 1247a15 139296d 1247a15 402ed7d 1247a15 d159d67 402ed7d d159d67 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 |
---
license: apache-2.0
datasets:
- HuggingFaceFW/fineweb
- PleIAs/YouTube-Commons
- allenai/WildChat-1M
- Salesforce/xlam-function-calling-60k
- ShareGPT4Video/ShareGPT4Video
- OpenGVLab/ShareGPT-4o
- TempoFunk/webvid-10M
- MBZUAI/VideoInstruct-100K
- MaziyarPanahi/WizardLM_evol_instruct_V2_196k
- Isaak-Carter/J.O.S.I.E.v3.5
- NousResearch/dolma-v1_7-c4
- NousResearch/dolma-v1_7-cc_en_head
language:
- de
- en
library_name: mlx
tags:
- moe
- multimodal
- vision
- audio
- endtoend
- j.o.s.i.e.
---
# STILL IN BETA!!!
# This will be the repo for J.O.S.I.E.v4o
Like **OpenAIs GPT-4o**, it's natively Multimodal, based on the **NExT-GPT** combined with **ROPE**, **RMS Normalisation**, and **MoE**, parred with the **GPT-4o Tokenizer** from OpenAI.
This is a *future project* and will take it's time.
Further more, I will probably make a **UI application** with that model too.
Further updates comming soon!!!
Source code and more info will be available on my <a href="https://github.com/Goekdeniz-Guelmez/J.O.S.I.E.v4-o.git">GitHub Repo</a> |