Update modeling_navit_siglip.py
#26
by
toilaluan
- opened
- modeling_minicpmv.py +5 -2
- modeling_navit_siglip.py +0 -5
modeling_minicpmv.py
CHANGED
@@ -181,6 +181,7 @@ class MiniCPMV(MiniCPMVPreTrainedModel):
|
|
181 |
)
|
182 |
|
183 |
def _decode(self, inputs_embeds, tokenizer, attention_mask, decode_text=False, **kwargs):
|
|
|
184 |
terminators = [tokenizer.convert_tokens_to_ids(i) for i in self.terminators]
|
185 |
output = self.llm.generate(
|
186 |
inputs_embeds=inputs_embeds,
|
@@ -258,7 +259,7 @@ class MiniCPMV(MiniCPMVPreTrainedModel):
|
|
258 |
if stream:
|
259 |
result = self._decode_stream(model_inputs["inputs_embeds"], tokenizer, **kwargs)
|
260 |
else:
|
261 |
-
result = self._decode(model_inputs["inputs_embeds"], tokenizer, attention_mask, decode_text=
|
262 |
|
263 |
if return_vision_hidden_states:
|
264 |
return result, vision_hidden_states
|
@@ -360,12 +361,14 @@ class MiniCPMV(MiniCPMVPreTrainedModel):
|
|
360 |
"top_k": 100,
|
361 |
"temperature": 0.7,
|
362 |
"do_sample": True,
|
363 |
-
"repetition_penalty": 1.05
|
364 |
}
|
365 |
else:
|
366 |
generation_config = {
|
367 |
"num_beams": 3,
|
368 |
"repetition_penalty": 1.2,
|
|
|
|
|
369 |
}
|
370 |
|
371 |
if min_new_tokens > 0:
|
|
|
181 |
)
|
182 |
|
183 |
def _decode(self, inputs_embeds, tokenizer, attention_mask, decode_text=False, **kwargs):
|
184 |
+
from transformers import GenerationConfig
|
185 |
terminators = [tokenizer.convert_tokens_to_ids(i) for i in self.terminators]
|
186 |
output = self.llm.generate(
|
187 |
inputs_embeds=inputs_embeds,
|
|
|
259 |
if stream:
|
260 |
result = self._decode_stream(model_inputs["inputs_embeds"], tokenizer, **kwargs)
|
261 |
else:
|
262 |
+
result = self._decode(model_inputs["inputs_embeds"], tokenizer, attention_mask, decode_text=False, **kwargs)
|
263 |
|
264 |
if return_vision_hidden_states:
|
265 |
return result, vision_hidden_states
|
|
|
361 |
"top_k": 100,
|
362 |
"temperature": 0.7,
|
363 |
"do_sample": True,
|
364 |
+
"repetition_penalty": 1.05,
|
365 |
}
|
366 |
else:
|
367 |
generation_config = {
|
368 |
"num_beams": 3,
|
369 |
"repetition_penalty": 1.2,
|
370 |
+
"output_logits": True,
|
371 |
+
"output_scores": True,
|
372 |
}
|
373 |
|
374 |
if min_new_tokens > 0:
|
modeling_navit_siglip.py
CHANGED
@@ -142,11 +142,6 @@ SIGLIP_PRETRAINED_MODEL_ARCHIVE_LIST = [
|
|
142 |
# See all SigLIP models at https://huggingface.co/models?filter=siglip
|
143 |
]
|
144 |
|
145 |
-
if is_flash_attn_2_available():
|
146 |
-
from flash_attn import flash_attn_func, flash_attn_varlen_func
|
147 |
-
from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
|
148 |
-
|
149 |
-
|
150 |
# Copied from transformers.models.llama.modeling_llama._get_unpad_data
|
151 |
def _get_unpad_data(attention_mask):
|
152 |
seqlens_in_batch = attention_mask.sum(dim=-1, dtype=torch.int32)
|
|
|
142 |
# See all SigLIP models at https://huggingface.co/models?filter=siglip
|
143 |
]
|
144 |
|
|
|
|
|
|
|
|
|
|
|
145 |
# Copied from transformers.models.llama.modeling_llama._get_unpad_data
|
146 |
def _get_unpad_data(attention_mask):
|
147 |
seqlens_in_batch = attention_mask.sum(dim=-1, dtype=torch.int32)
|