Spaces:
Paused
Paused
Update utils.py
Browse filesremoving lru_cache
utils.py
CHANGED
@@ -2,7 +2,6 @@ import subprocess
|
|
2 |
import os
|
3 |
import torch
|
4 |
from transformers import BitsAndBytesConfig, AutoConfig, AutoModelForCausalLM, LlavaNextForConditionalGeneration, LlavaForConditionalGeneration, PaliGemmaForConditionalGeneration, Idefics2ForConditionalGeneration
|
5 |
-
from functools import lru_cache
|
6 |
import spaces
|
7 |
|
8 |
os.environ["HF_HUB_ENABLE_HF_TRANSFER"] = "1"
|
@@ -26,7 +25,6 @@ ARCHITECTURE_MAP = {
|
|
26 |
|
27 |
# Function to get the model summary with caching and GPU support
|
28 |
@spaces.GPU
|
29 |
-
@lru_cache(maxsize=10)
|
30 |
def get_model_summary(model_name):
|
31 |
"""
|
32 |
Retrieve the model summary for the given model name.
|
|
|
2 |
import os
|
3 |
import torch
|
4 |
from transformers import BitsAndBytesConfig, AutoConfig, AutoModelForCausalLM, LlavaNextForConditionalGeneration, LlavaForConditionalGeneration, PaliGemmaForConditionalGeneration, Idefics2ForConditionalGeneration
|
|
|
5 |
import spaces
|
6 |
|
7 |
os.environ["HF_HUB_ENABLE_HF_TRANSFER"] = "1"
|
|
|
25 |
|
26 |
# Function to get the model summary with caching and GPU support
|
27 |
@spaces.GPU
|
|
|
28 |
def get_model_summary(model_name):
|
29 |
"""
|
30 |
Retrieve the model summary for the given model name.
|