File size: 1,050 Bytes
1247a15
 
02162cc
 
 
 
78c0f10
 
 
 
 
 
 
9ca90e4
 
02162cc
 
 
 
 
 
 
78c0f10
 
 
02162cc
1247a15
 
139296d
 
1247a15
 
 
 
 
402ed7d
1247a15
d159d67
 
 
402ed7d
d159d67
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
---
license: apache-2.0
datasets:
- HuggingFaceFW/fineweb
- PleIAs/YouTube-Commons
- allenai/WildChat-1M
- Salesforce/xlam-function-calling-60k
- ShareGPT4Video/ShareGPT4Video
- OpenGVLab/ShareGPT-4o
- TempoFunk/webvid-10M
- MBZUAI/VideoInstruct-100K
- MaziyarPanahi/WizardLM_evol_instruct_V2_196k
- Isaak-Carter/J.O.S.I.E.v3.5
- NousResearch/dolma-v1_7-c4
- NousResearch/dolma-v1_7-cc_en_head
language:
- de
- en
library_name: mlx
tags:
- moe
- multimodal
- vision
- audio
- endtoend
- j.o.s.i.e.
---

# STILL IN BETA!!!

# This will be the repo for J.O.S.I.E.v4o

Like **OpenAIs GPT-4o**, it's natively Multimodal, based on the **NExT-GPT** combined with **ROPE**, **RMS Normalisation**, and **MoE**, parred with the **GPT-4o Tokenizer** from OpenAI.
This is a *future project* and will take it's time.

Further more, I will probably make a **UI application** with that model too.

Further updates comming soon!!!



Source code and more info will be available on my <a href="https://github.com/Goekdeniz-Guelmez/J.O.S.I.E.v4-o.git">GitHub Repo</a>