Update modeling_navit_siglip.py

#26
by toilaluan - opened
Files changed (2) hide show
  1. modeling_minicpmv.py +5 -2
  2. modeling_navit_siglip.py +0 -5
modeling_minicpmv.py CHANGED
@@ -181,6 +181,7 @@ class MiniCPMV(MiniCPMVPreTrainedModel):
181
  )
182
 
183
  def _decode(self, inputs_embeds, tokenizer, attention_mask, decode_text=False, **kwargs):
 
184
  terminators = [tokenizer.convert_tokens_to_ids(i) for i in self.terminators]
185
  output = self.llm.generate(
186
  inputs_embeds=inputs_embeds,
@@ -258,7 +259,7 @@ class MiniCPMV(MiniCPMVPreTrainedModel):
258
  if stream:
259
  result = self._decode_stream(model_inputs["inputs_embeds"], tokenizer, **kwargs)
260
  else:
261
- result = self._decode(model_inputs["inputs_embeds"], tokenizer, attention_mask, decode_text=decode_text, **kwargs)
262
 
263
  if return_vision_hidden_states:
264
  return result, vision_hidden_states
@@ -360,12 +361,14 @@ class MiniCPMV(MiniCPMVPreTrainedModel):
360
  "top_k": 100,
361
  "temperature": 0.7,
362
  "do_sample": True,
363
- "repetition_penalty": 1.05
364
  }
365
  else:
366
  generation_config = {
367
  "num_beams": 3,
368
  "repetition_penalty": 1.2,
 
 
369
  }
370
 
371
  if min_new_tokens > 0:
 
181
  )
182
 
183
  def _decode(self, inputs_embeds, tokenizer, attention_mask, decode_text=False, **kwargs):
184
+ from transformers import GenerationConfig
185
  terminators = [tokenizer.convert_tokens_to_ids(i) for i in self.terminators]
186
  output = self.llm.generate(
187
  inputs_embeds=inputs_embeds,
 
259
  if stream:
260
  result = self._decode_stream(model_inputs["inputs_embeds"], tokenizer, **kwargs)
261
  else:
262
+ result = self._decode(model_inputs["inputs_embeds"], tokenizer, attention_mask, decode_text=False, **kwargs)
263
 
264
  if return_vision_hidden_states:
265
  return result, vision_hidden_states
 
361
  "top_k": 100,
362
  "temperature": 0.7,
363
  "do_sample": True,
364
+ "repetition_penalty": 1.05,
365
  }
366
  else:
367
  generation_config = {
368
  "num_beams": 3,
369
  "repetition_penalty": 1.2,
370
+ "output_logits": True,
371
+ "output_scores": True,
372
  }
373
 
374
  if min_new_tokens > 0:
modeling_navit_siglip.py CHANGED
@@ -142,11 +142,6 @@ SIGLIP_PRETRAINED_MODEL_ARCHIVE_LIST = [
142
  # See all SigLIP models at https://huggingface.co/models?filter=siglip
143
  ]
144
 
145
- if is_flash_attn_2_available():
146
- from flash_attn import flash_attn_func, flash_attn_varlen_func
147
- from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
148
-
149
-
150
  # Copied from transformers.models.llama.modeling_llama._get_unpad_data
151
  def _get_unpad_data(attention_mask):
152
  seqlens_in_batch = attention_mask.sum(dim=-1, dtype=torch.int32)
 
142
  # See all SigLIP models at https://huggingface.co/models?filter=siglip
143
  ]
144
 
 
 
 
 
 
145
  # Copied from transformers.models.llama.modeling_llama._get_unpad_data
146
  def _get_unpad_data(attention_mask):
147
  seqlens_in_batch = attention_mask.sum(dim=-1, dtype=torch.int32)