Nemil commited on
Commit
46a9783
β€’
1 Parent(s): 9a049b8

Upload 2 files

Browse files
Files changed (2) hide show
  1. app.py +7 -4
  2. requirements.txt +2 -1
app.py CHANGED
@@ -4,6 +4,9 @@ from PIL import Image
4
  import requests
5
  import traceback
6
 
 
 
 
7
  class Image2Text:
8
  def __init__(self):
9
  # Load the GIT coco model
@@ -188,7 +191,7 @@ class Social_Media_Captioner:
188
  self.model_loaded = False
189
 
190
 
191
- def inference(self, input_text: str, use_cached=True, cache_generation=True):
192
  if not self.model_loaded:
193
  raise Exception("Model not loaded")
194
 
@@ -230,7 +233,7 @@ class Social_Media_Captioner:
230
  raise Exception("Enter a valid input text to generate a valid prompt")
231
 
232
  return f"""
233
- Convert the given image description to social media worthy caption
234
  Description: {input_text}
235
  Caption:
236
  """.strip()
@@ -279,7 +282,7 @@ class Captions:
279
  image_description = self.image_to_text._generate_description(image, max_length=max_length_GIT)
280
  captions = self.LLM.inference(image_description, use_cached=use_cached_LLM, cache_generation=cache_generation_LLM)
281
  return captions
282
-
283
  caption_generator = Captions()
284
 
285
  import gradio as gr
@@ -293,4 +296,4 @@ iface = gr.Interface(
293
  outputs=gr.outputs.Textbox(label="Caption")
294
  )
295
 
296
- iface.launch()
 
4
  import requests
5
  import traceback
6
 
7
+ from huggingface_hub import login
8
+ login(token=HF_TOKEN)
9
+
10
  class Image2Text:
11
  def __init__(self):
12
  # Load the GIT coco model
 
191
  self.model_loaded = False
192
 
193
 
194
+ def inference(self, input_text: str, use_cached=True, cache_generation=True) -> str | None:
195
  if not self.model_loaded:
196
  raise Exception("Model not loaded")
197
 
 
233
  raise Exception("Enter a valid input text to generate a valid prompt")
234
 
235
  return f"""
236
+ Convert the given image description to a appropriate metaphoric caption
237
  Description: {input_text}
238
  Caption:
239
  """.strip()
 
282
  image_description = self.image_to_text._generate_description(image, max_length=max_length_GIT)
283
  captions = self.LLM.inference(image_description, use_cached=use_cached_LLM, cache_generation=cache_generation_LLM)
284
  return captions
285
+
286
  caption_generator = Captions()
287
 
288
  import gradio as gr
 
296
  outputs=gr.outputs.Textbox(label="Caption")
297
  )
298
 
299
+ iface.launch(share=True)
requirements.txt CHANGED
@@ -1,7 +1,7 @@
1
  evaluate
2
  jiwer
3
  huggingface_hub @ git+https://github.com/huggingface/huggingface_hub.git
4
- gradio==3.43.1
5
  bitsandbytes
6
  transformers @ git+https://github.com/huggingface/transformers.git
7
  peft @ git+https://github.com/huggingface/peft.git
@@ -11,3 +11,4 @@ safetensors
11
  torch
12
  xformers
13
  datasets
 
 
1
  evaluate
2
  jiwer
3
  huggingface_hub @ git+https://github.com/huggingface/huggingface_hub.git
4
+ gradio==3.36.0
5
  bitsandbytes
6
  transformers @ git+https://github.com/huggingface/transformers.git
7
  peft @ git+https://github.com/huggingface/peft.git
 
11
  torch
12
  xformers
13
  datasets
14
+ stable-diffusion