Spaces:
Runtime error
Runtime error
Update
Browse files- app.py +4 -4
- constants.py +1 -1
- demo_list.py +143 -128
- list.yaml +88 -88
- tools/add_creation_timestamps.py +4 -3
app.py
CHANGED
@@ -6,11 +6,11 @@ import gradio as gr
|
|
6 |
|
7 |
from constants import (HARDWARE_CHOICES, OWNER_CHOICES, SDK_CHOICES,
|
8 |
SLEEP_TIME_CHOICES, STATUS_CHOICES, VISIBILITY_CHOICES)
|
9 |
-
from demo_list import DemoList
|
10 |
from restart_scheduler import RestartScheduler
|
11 |
from settings import HF_TOKEN, INTERVAL_MINUTES, SPACE_ID
|
12 |
|
13 |
-
demo_list = DemoList()
|
14 |
|
15 |
if SPACE_ID is not None and INTERVAL_MINUTES > 0:
|
16 |
scheduler = RestartScheduler(space_id=SPACE_ID,
|
@@ -85,7 +85,7 @@ with gr.Blocks(css='style.css') as demo:
|
|
85 |
value=OWNER_CHOICES,
|
86 |
type='value')
|
87 |
apply_button = gr.Button('Apply')
|
88 |
-
df = gr.Dataframe(value=demo_list.
|
89 |
datatype=demo_list.column_datatype,
|
90 |
type='pandas')
|
91 |
|
@@ -113,7 +113,7 @@ with gr.Blocks(css='style.css') as demo:
|
|
113 |
queue=False,
|
114 |
show_progress=False,
|
115 |
api_name=False)
|
116 |
-
apply_button.click(fn=demo_list.
|
117 |
inputs=[
|
118 |
status,
|
119 |
hardware,
|
|
|
6 |
|
7 |
from constants import (HARDWARE_CHOICES, OWNER_CHOICES, SDK_CHOICES,
|
8 |
SLEEP_TIME_CHOICES, STATUS_CHOICES, VISIBILITY_CHOICES)
|
9 |
+
from demo_list import DemoList, get_df_from_yaml
|
10 |
from restart_scheduler import RestartScheduler
|
11 |
from settings import HF_TOKEN, INTERVAL_MINUTES, SPACE_ID
|
12 |
|
13 |
+
demo_list = DemoList(get_df_from_yaml('list.yaml'))
|
14 |
|
15 |
if SPACE_ID is not None and INTERVAL_MINUTES > 0:
|
16 |
scheduler = RestartScheduler(space_id=SPACE_ID,
|
|
|
85 |
value=OWNER_CHOICES,
|
86 |
type='value')
|
87 |
apply_button = gr.Button('Apply')
|
88 |
+
df = gr.Dataframe(value=demo_list.df_prettified,
|
89 |
datatype=demo_list.column_datatype,
|
90 |
type='pandas')
|
91 |
|
|
|
113 |
queue=False,
|
114 |
show_progress=False,
|
115 |
api_name=False)
|
116 |
+
apply_button.click(fn=demo_list.filter,
|
117 |
inputs=[
|
118 |
status,
|
119 |
hardware,
|
constants.py
CHANGED
@@ -26,7 +26,7 @@ SDK_CHOICES = [
|
|
26 |
'docker',
|
27 |
]
|
28 |
SLEEP_TIME_INT_TO_STR = {
|
29 |
-
|
30 |
300: '5 minutes',
|
31 |
900: '15 minutes',
|
32 |
1800: '30 minutes',
|
|
|
26 |
'docker',
|
27 |
]
|
28 |
SLEEP_TIME_INT_TO_STR = {
|
29 |
+
0: 'null',
|
30 |
300: '5 minutes',
|
31 |
900: '15 minutes',
|
32 |
1800: '30 minutes',
|
demo_list.py
CHANGED
@@ -1,3 +1,4 @@
|
|
|
|
1 |
import datetime
|
2 |
import operator
|
3 |
import pathlib
|
@@ -10,123 +11,112 @@ from huggingface_hub import HfApi
|
|
10 |
from constants import (OWNER_CHOICES, SLEEP_TIME_INT_TO_STR,
|
11 |
SLEEP_TIME_STR_TO_INT, WHOAMI)
|
12 |
|
13 |
-
repo_dir = pathlib.Path(__file__).parent
|
14 |
|
15 |
-
|
16 |
-
class
|
17 |
-
|
18 |
-
|
19 |
-
|
20 |
-
|
21 |
-
|
22 |
-
|
23 |
-
|
24 |
-
|
25 |
-
|
26 |
-
|
27 |
-
|
28 |
-
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
|
33 |
-
]
|
34 |
-
|
35 |
-
|
36 |
-
|
37 |
-
self
|
38 |
-
|
39 |
-
self
|
40 |
-
|
41 |
-
@property
|
42 |
-
def column_names(self):
|
43 |
-
return list(map(operator.itemgetter(0), self.COLUMN_INFO))
|
44 |
-
|
45 |
-
@property
|
46 |
-
def column_datatype(self):
|
47 |
-
return list(map(operator.itemgetter(1), self.COLUMN_INFO))
|
48 |
|
49 |
@staticmethod
|
50 |
-
def
|
51 |
-
|
52 |
-
|
53 |
-
|
54 |
-
|
55 |
-
|
56 |
-
|
57 |
-
|
58 |
-
|
59 |
-
|
60 |
-
|
61 |
-
|
62 |
-
|
63 |
-
|
64 |
-
|
65 |
-
|
66 |
-
|
67 |
-
|
68 |
-
|
69 |
-
|
70 |
-
|
71 |
-
|
72 |
-
|
73 |
-
|
74 |
-
|
75 |
-
|
76 |
-
|
77 |
-
|
78 |
-
|
79 |
-
|
80 |
-
|
81 |
-
|
82 |
-
|
83 |
-
|
84 |
-
|
85 |
-
|
86 |
-
|
87 |
-
|
88 |
-
|
89 |
-
|
90 |
-
|
91 |
-
|
92 |
-
|
93 |
-
|
94 |
-
|
95 |
-
|
96 |
-
|
97 |
-
|
98 |
-
|
99 |
-
|
100 |
-
|
101 |
-
|
102 |
-
links = [self.create_link(link.split('/')[-1], link) for link in links]
|
103 |
return '\n'.join(links)
|
104 |
|
105 |
-
|
106 |
-
|
|
|
107 |
return '\n'.join(links)
|
108 |
|
109 |
-
|
|
|
110 |
return ', '.join(tags)
|
111 |
|
112 |
@staticmethod
|
113 |
def create_link(text: str, url: str) -> str:
|
114 |
return f'<a href={url} target="_blank">{text}</a>'
|
115 |
|
116 |
-
|
|
|
117 |
if text is None:
|
118 |
text = ''
|
119 |
class_name = f'{category_name}-{text.lower()}'
|
120 |
return f'<div class="{class_name}">{text}</div>'
|
121 |
|
122 |
-
@staticmethod
|
123 |
-
def format_timestamp(timestamp: str) -> str:
|
124 |
-
s = datetime.datetime.strptime(timestamp, '%Y-%m-%dT%H:%M:%S.000Z')
|
125 |
-
return s.strftime('%Y/%m/%d %H:%M:%S')
|
126 |
-
|
127 |
@staticmethod
|
128 |
def add_div_tag_to_replicas(replicas: int) -> str:
|
129 |
-
if replicas ==
|
130 |
return ''
|
131 |
if replicas == 1:
|
132 |
return '1'
|
@@ -139,10 +129,10 @@ class DemoList:
|
|
139 |
s = sleep_time_s.replace(' ', '-')
|
140 |
return f'<div class="sleep-time-{s}">{sleep_time_s}</div>'
|
141 |
|
142 |
-
def
|
143 |
new_rows = []
|
144 |
-
for _, row in
|
145 |
-
new_row = {
|
146 |
'status':
|
147 |
self.to_div(row.status, 'status'),
|
148 |
'hardware':
|
@@ -154,33 +144,58 @@ class DemoList:
|
|
154 |
'owner':
|
155 |
self.create_link(row.owner,
|
156 |
f'https://huggingface.co/{row.owner}'),
|
157 |
-
'arxiv':
|
158 |
-
self.get_arxiv_link(row.arxiv),
|
159 |
-
'github':
|
160 |
-
self.get_github_link(row.github),
|
161 |
-
'likes':
|
162 |
-
row.likes,
|
163 |
-
'tags':
|
164 |
-
self.get_tag_list(row.tags),
|
165 |
-
'last_modified':
|
166 |
-
self.format_timestamp(row.last_modified),
|
167 |
-
'created':
|
168 |
-
self.format_timestamp(row.created),
|
169 |
'sdk':
|
170 |
self.to_div(row.sdk, 'sdk'),
|
171 |
-
'sdk_version':
|
172 |
-
row.sdk_version,
|
173 |
'sleep_time':
|
174 |
self.add_div_tag_to_sleep_time(
|
175 |
SLEEP_TIME_INT_TO_STR[row.sleep_time], row.hardware),
|
176 |
'replicas':
|
177 |
self.add_div_tag_to_replicas(row.replicas),
|
|
|
|
|
|
|
|
|
|
|
|
|
178 |
}
|
179 |
new_rows.append(new_row)
|
180 |
-
|
181 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
182 |
|
183 |
-
def
|
184 |
self,
|
185 |
status: list[str],
|
186 |
hardware: list[str],
|
@@ -190,28 +205,28 @@ class DemoList:
|
|
190 |
visibility: list[str],
|
191 |
owner: list[str],
|
192 |
) -> pd.DataFrame:
|
193 |
-
|
194 |
-
df = self.df
|
195 |
|
196 |
if multiple_replicas:
|
197 |
-
df = df[df_raw.replicas > 1]
|
198 |
|
199 |
if visibility == ['public']:
|
200 |
-
df = df[~df_raw.private]
|
201 |
elif visibility == ['private']:
|
202 |
-
df = df[df_raw.private]
|
203 |
|
204 |
-
df = df[(df_raw.status.isin(status))
|
205 |
-
& (df_raw.
|
|
|
206 |
|
207 |
sleep_time_int = [SLEEP_TIME_STR_TO_INT[s] for s in sleep_time]
|
208 |
-
df = df[df_raw.sleep_time.isin(sleep_time_int)]
|
209 |
|
210 |
if set(owner) == set(OWNER_CHOICES):
|
211 |
pass
|
212 |
elif WHOAMI in owner:
|
213 |
-
df = df[df_raw.owner == WHOAMI]
|
214 |
else:
|
215 |
-
df = df[df_raw.owner != WHOAMI]
|
216 |
|
217 |
-
return df
|
|
|
1 |
+
import dataclasses
|
2 |
import datetime
|
3 |
import operator
|
4 |
import pathlib
|
|
|
11 |
from constants import (OWNER_CHOICES, SLEEP_TIME_INT_TO_STR,
|
12 |
SLEEP_TIME_STR_TO_INT, WHOAMI)
|
13 |
|
|
|
14 |
|
15 |
+
@dataclasses.dataclass(frozen=True)
|
16 |
+
class DemoInfo:
|
17 |
+
space_id: str
|
18 |
+
url: str
|
19 |
+
title: str
|
20 |
+
owner: str
|
21 |
+
sdk: str
|
22 |
+
sdk_version: str
|
23 |
+
likes: int
|
24 |
+
status: str
|
25 |
+
last_modified: str
|
26 |
+
sleep_time: int
|
27 |
+
replicas: int
|
28 |
+
private: bool
|
29 |
+
hardware: str
|
30 |
+
suggested_hardware: str
|
31 |
+
created: str = ''
|
32 |
+
arxiv: list[str] = dataclasses.field(default_factory=list)
|
33 |
+
github: list[str] = dataclasses.field(default_factory=list)
|
34 |
+
tags: list[str] = dataclasses.field(default_factory=list)
|
35 |
+
|
36 |
+
def __post_init__(self):
|
37 |
+
object.__setattr__(self, 'last_modified',
|
38 |
+
DemoInfo.convert_timestamp(self.last_modified))
|
39 |
+
object.__setattr__(self, 'created',
|
40 |
+
DemoInfo.convert_timestamp(self.created))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
41 |
|
42 |
@staticmethod
|
43 |
+
def convert_timestamp(timestamp: str) -> str:
|
44 |
+
try:
|
45 |
+
return datetime.datetime.strptime(
|
46 |
+
timestamp,
|
47 |
+
'%Y-%m-%dT%H:%M:%S.%fZ').strftime('%Y/%m/%d %H:%M:%S')
|
48 |
+
except ValueError:
|
49 |
+
return timestamp
|
50 |
+
|
51 |
+
@classmethod
|
52 |
+
def from_space_id(cls, space_id: str) -> 'DemoInfo':
|
53 |
+
api = HfApi()
|
54 |
+
space_info = api.space_info(repo_id=space_id)
|
55 |
+
card = space_info.cardData
|
56 |
+
runtime = space_info.runtime
|
57 |
+
resources = runtime['resources']
|
58 |
+
|
59 |
+
return cls(
|
60 |
+
space_id=space_id,
|
61 |
+
url=f'https://huggingface.co/spaces/{space_id}',
|
62 |
+
title=card['title'] if 'title' in card else '',
|
63 |
+
owner=space_id.split('/')[0],
|
64 |
+
sdk=card['sdk'],
|
65 |
+
sdk_version=card.get('sdk_version', ''),
|
66 |
+
likes=space_info.likes,
|
67 |
+
status=runtime['stage'],
|
68 |
+
last_modified=space_info.lastModified,
|
69 |
+
sleep_time=runtime['gcTimeout'] or 0,
|
70 |
+
replicas=resources['replicas'] if resources is not None else 0,
|
71 |
+
private=space_info.private,
|
72 |
+
hardware=runtime['hardware']['current']
|
73 |
+
or runtime['hardware']['requested'],
|
74 |
+
suggested_hardware=card.get('suggested_hardware', ''),
|
75 |
+
)
|
76 |
+
|
77 |
+
|
78 |
+
def get_df_from_yaml(path: pathlib.Path | str) -> pd.DataFrame:
|
79 |
+
with pathlib.Path(path).open() as f:
|
80 |
+
data = yaml.safe_load(f)
|
81 |
+
demo_info = []
|
82 |
+
for space_id in tqdm.auto.tqdm(list(data)):
|
83 |
+
base_info = DemoInfo.from_space_id(space_id)
|
84 |
+
info = DemoInfo(**(dataclasses.asdict(base_info) | data[space_id]))
|
85 |
+
demo_info.append(info)
|
86 |
+
return pd.DataFrame([dataclasses.asdict(info) for info in demo_info])
|
87 |
+
|
88 |
+
|
89 |
+
class Prettifier:
|
90 |
+
@staticmethod
|
91 |
+
def get_arxiv_link(links: list[str]) -> str:
|
92 |
+
links = [
|
93 |
+
Prettifier.create_link(link.split('/')[-1], link) for link in links
|
94 |
+
]
|
|
|
95 |
return '\n'.join(links)
|
96 |
|
97 |
+
@staticmethod
|
98 |
+
def get_github_link(links: list[str]) -> str:
|
99 |
+
links = [Prettifier.create_link('github', link) for link in links]
|
100 |
return '\n'.join(links)
|
101 |
|
102 |
+
@staticmethod
|
103 |
+
def get_tag_list(tags: list[str]) -> str:
|
104 |
return ', '.join(tags)
|
105 |
|
106 |
@staticmethod
|
107 |
def create_link(text: str, url: str) -> str:
|
108 |
return f'<a href={url} target="_blank">{text}</a>'
|
109 |
|
110 |
+
@staticmethod
|
111 |
+
def to_div(text: str | None, category_name: str) -> str:
|
112 |
if text is None:
|
113 |
text = ''
|
114 |
class_name = f'{category_name}-{text.lower()}'
|
115 |
return f'<div class="{class_name}">{text}</div>'
|
116 |
|
|
|
|
|
|
|
|
|
|
|
117 |
@staticmethod
|
118 |
def add_div_tag_to_replicas(replicas: int) -> str:
|
119 |
+
if replicas == 0:
|
120 |
return ''
|
121 |
if replicas == 1:
|
122 |
return '1'
|
|
|
129 |
s = sleep_time_s.replace(' ', '-')
|
130 |
return f'<div class="sleep-time-{s}">{sleep_time_s}</div>'
|
131 |
|
132 |
+
def __call__(self, df: pd.DataFrame) -> pd.DataFrame:
|
133 |
new_rows = []
|
134 |
+
for _, row in df.iterrows():
|
135 |
+
new_row = dict(row) | {
|
136 |
'status':
|
137 |
self.to_div(row.status, 'status'),
|
138 |
'hardware':
|
|
|
144 |
'owner':
|
145 |
self.create_link(row.owner,
|
146 |
f'https://huggingface.co/{row.owner}'),
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
147 |
'sdk':
|
148 |
self.to_div(row.sdk, 'sdk'),
|
|
|
|
|
149 |
'sleep_time':
|
150 |
self.add_div_tag_to_sleep_time(
|
151 |
SLEEP_TIME_INT_TO_STR[row.sleep_time], row.hardware),
|
152 |
'replicas':
|
153 |
self.add_div_tag_to_replicas(row.replicas),
|
154 |
+
'arxiv':
|
155 |
+
self.get_arxiv_link(row.arxiv),
|
156 |
+
'github':
|
157 |
+
self.get_github_link(row.github),
|
158 |
+
'tags':
|
159 |
+
self.get_tag_list(row.tags),
|
160 |
}
|
161 |
new_rows.append(new_row)
|
162 |
+
return pd.DataFrame(new_rows, columns=df.columns)
|
163 |
+
|
164 |
+
|
165 |
+
class DemoList:
|
166 |
+
COLUMN_INFO = [
|
167 |
+
['status', 'markdown'],
|
168 |
+
['hardware', 'markdown'],
|
169 |
+
['title', 'markdown'],
|
170 |
+
['owner', 'markdown'],
|
171 |
+
['arxiv', 'markdown'],
|
172 |
+
['github', 'markdown'],
|
173 |
+
['likes', 'number'],
|
174 |
+
['tags', 'str'],
|
175 |
+
['last_modified', 'str'],
|
176 |
+
['created', 'str'],
|
177 |
+
['sdk', 'markdown'],
|
178 |
+
['sdk_version', 'str'],
|
179 |
+
['suggested_hardware', 'markdown'],
|
180 |
+
['sleep_time', 'markdown'],
|
181 |
+
['replicas', 'markdown'],
|
182 |
+
['private', 'bool'],
|
183 |
+
]
|
184 |
+
|
185 |
+
def __init__(self, df: pd.DataFrame):
|
186 |
+
self.df_raw = df
|
187 |
+
self._prettifier = Prettifier()
|
188 |
+
self.df_prettified = self._prettifier(df).loc[:, self.column_names]
|
189 |
+
|
190 |
+
@property
|
191 |
+
def column_names(self):
|
192 |
+
return list(map(operator.itemgetter(0), self.COLUMN_INFO))
|
193 |
+
|
194 |
+
@property
|
195 |
+
def column_datatype(self):
|
196 |
+
return list(map(operator.itemgetter(1), self.COLUMN_INFO))
|
197 |
|
198 |
+
def filter(
|
199 |
self,
|
200 |
status: list[str],
|
201 |
hardware: list[str],
|
|
|
205 |
visibility: list[str],
|
206 |
owner: list[str],
|
207 |
) -> pd.DataFrame:
|
208 |
+
df = self.df_raw.copy()
|
|
|
209 |
|
210 |
if multiple_replicas:
|
211 |
+
df = df[self.df_raw.replicas > 1]
|
212 |
|
213 |
if visibility == ['public']:
|
214 |
+
df = df[~self.df_raw.private]
|
215 |
elif visibility == ['private']:
|
216 |
+
df = df[self.df_raw.private]
|
217 |
|
218 |
+
df = df[(self.df_raw.status.isin(status))
|
219 |
+
& (self.df_raw.hardware.isin(hardware))
|
220 |
+
& (self.df_raw.sdk.isin(sdk))]
|
221 |
|
222 |
sleep_time_int = [SLEEP_TIME_STR_TO_INT[s] for s in sleep_time]
|
223 |
+
df = df[self.df_raw.sleep_time.isin(sleep_time_int)]
|
224 |
|
225 |
if set(owner) == set(OWNER_CHOICES):
|
226 |
pass
|
227 |
elif WHOAMI in owner:
|
228 |
+
df = df[self.df_raw.owner == WHOAMI]
|
229 |
else:
|
230 |
+
df = df[self.df_raw.owner != WHOAMI]
|
231 |
|
232 |
+
return self._prettifier(df).loc[:, self.column_names]
|
list.yaml
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
-
|
2 |
arxiv:
|
3 |
- https://arxiv.org/abs/2307.01952
|
4 |
created: "2023-07-27T05:10:31.000Z"
|
@@ -7,7 +7,7 @@ https://huggingface.co/spaces/hysts/SD-XL:
|
|
7 |
tags:
|
8 |
- diffusion model
|
9 |
- text to image
|
10 |
-
|
11 |
arxiv:
|
12 |
- https://arxiv.org/abs/2307.09288
|
13 |
created: "2023-07-19T03:33:27.000Z"
|
@@ -17,26 +17,26 @@ https://huggingface.co/spaces/huggingface-projects/llama-2-7b-chat:
|
|
17 |
- chatbot
|
18 |
- LLM
|
19 |
- Llama 2
|
20 |
-
|
21 |
created: "2023-07-14T12:34:56.000Z"
|
22 |
tags:
|
23 |
- diffusion model
|
24 |
- text to image
|
25 |
-
|
26 |
created: "2023-07-14T09:48:26.000Z"
|
27 |
tags:
|
28 |
- diffusion model
|
29 |
- text to image
|
30 |
-
|
31 |
created: "2023-06-27T02:57:53.000Z"
|
32 |
tags:
|
33 |
- diffusion model
|
34 |
- text to video
|
35 |
-
|
36 |
created: "2023-06-15T17:16:16.000Z"
|
37 |
tags:
|
38 |
- list of papers
|
39 |
-
|
40 |
arxiv:
|
41 |
- https://arxiv.org/abs/2305.02463
|
42 |
created: "2023-05-05T07:40:18.000Z"
|
@@ -46,7 +46,7 @@ https://huggingface.co/spaces/hysts/Shap-E:
|
|
46 |
- diffusion model
|
47 |
- text to 3D
|
48 |
- image to 3D
|
49 |
-
|
50 |
arxiv:
|
51 |
- https://arxiv.org/abs/2302.05543
|
52 |
created: "2023-04-15T01:26:54.000Z"
|
@@ -57,14 +57,14 @@ https://huggingface.co/spaces/hysts/ControlNet-v1-1:
|
|
57 |
- diffusion model
|
58 |
- text2image
|
59 |
- image2image
|
60 |
-
|
61 |
created: "2023-03-30T01:36:56.000Z"
|
62 |
github:
|
63 |
- https://github.com/deep-floyd/IF
|
64 |
tags:
|
65 |
- diffusion model
|
66 |
- text2image
|
67 |
-
|
68 |
arxiv:
|
69 |
- https://arxiv.org/abs/2303.08320
|
70 |
created: "2023-03-19T04:30:43.000Z"
|
@@ -73,7 +73,7 @@ https://huggingface.co/spaces/damo-vilab/modelscope-text-to-video-synthesis:
|
|
73 |
tags:
|
74 |
- diffusion model
|
75 |
- text2video
|
76 |
-
|
77 |
arxiv:
|
78 |
- https://arxiv.org/abs/2303.06555
|
79 |
created: "2023-03-14T05:06:57.000Z"
|
@@ -84,7 +84,7 @@ https://huggingface.co/spaces/thu-ml/unidiffuser:
|
|
84 |
- text2image
|
85 |
- image2image
|
86 |
- image captioning
|
87 |
-
|
88 |
arxiv:
|
89 |
- https://arxiv.org/abs/2302.13848
|
90 |
created: "2023-03-09T03:28:49.000Z"
|
@@ -93,12 +93,12 @@ https://huggingface.co/spaces/ELITE-library/ELITE:
|
|
93 |
tags:
|
94 |
- diffusion model
|
95 |
- text2image
|
96 |
-
|
97 |
created: "2023-03-08T05:44:53.000Z"
|
98 |
tags:
|
99 |
- diffusion model
|
100 |
- text2image
|
101 |
-
|
102 |
arxiv:
|
103 |
- https://arxiv.org/abs/2303.00354
|
104 |
- https://arxiv.org/abs/2212.00490
|
@@ -108,7 +108,7 @@ https://huggingface.co/spaces/hysts/DDNM-HQ:
|
|
108 |
tags:
|
109 |
- diffusion model
|
110 |
- super resolution
|
111 |
-
|
112 |
arxiv:
|
113 |
- https://arxiv.org/abs/2302.05543
|
114 |
created: "2023-02-16T05:05:27.000Z"
|
@@ -118,7 +118,7 @@ https://huggingface.co/spaces/hysts/ControlNet-with-Anything-v4:
|
|
118 |
- diffusion model
|
119 |
- text2image
|
120 |
- image2image
|
121 |
-
|
122 |
arxiv:
|
123 |
- https://arxiv.org/abs/2302.05543
|
124 |
created: "2023-02-14T01:03:00.000Z"
|
@@ -128,7 +128,7 @@ https://huggingface.co/spaces/hysts/ControlNet:
|
|
128 |
- diffusion model
|
129 |
- text2image
|
130 |
- image2image
|
131 |
-
|
132 |
arxiv:
|
133 |
- https://arxiv.org/abs/2211.12572
|
134 |
created: "2023-02-13T06:20:32.000Z"
|
@@ -137,7 +137,7 @@ https://huggingface.co/spaces/hysts/PnP-diffusion-features:
|
|
137 |
tags:
|
138 |
- diffusion model
|
139 |
- image2image
|
140 |
-
|
141 |
arxiv:
|
142 |
- https://arxiv.org/abs/2301.12597
|
143 |
created: "2023-02-10T09:45:07.000Z"
|
@@ -146,7 +146,7 @@ https://huggingface.co/spaces/hysts/BLIP2-with-transformers:
|
|
146 |
tags:
|
147 |
- image captioning
|
148 |
- VQA
|
149 |
-
|
150 |
arxiv:
|
151 |
- https://arxiv.org/abs/2212.11565
|
152 |
created: "2023-02-09T02:39:44.000Z"
|
@@ -156,7 +156,7 @@ https://huggingface.co/spaces/Tune-A-Video-library/Tune-A-Video-inference:
|
|
156 |
- diffusion model
|
157 |
- text2video
|
158 |
- video2video
|
159 |
-
|
160 |
arxiv:
|
161 |
- https://arxiv.org/abs/2302.01721
|
162 |
created: "2023-02-06T09:16:46.000Z"
|
@@ -167,7 +167,7 @@ https://huggingface.co/spaces/TEXTurePaper/TEXTure:
|
|
167 |
- text-guided
|
168 |
- texture
|
169 |
- 3D
|
170 |
-
|
171 |
arxiv:
|
172 |
- https://arxiv.org/abs/2301.13826
|
173 |
created: "2023-02-01T06:18:56.000Z"
|
@@ -176,7 +176,7 @@ https://huggingface.co/spaces/AttendAndExcite/Attend-and-Excite:
|
|
176 |
tags:
|
177 |
- diffusion model
|
178 |
- text2image
|
179 |
-
|
180 |
arxiv:
|
181 |
- https://arxiv.org/abs/2212.11565
|
182 |
created: "2023-01-30T08:10:09.000Z"
|
@@ -186,7 +186,7 @@ https://huggingface.co/spaces/Tune-A-Video-library/Tune-A-Video-Training-UI:
|
|
186 |
- diffusion model
|
187 |
- text2video
|
188 |
- video2video
|
189 |
-
|
190 |
arxiv:
|
191 |
- https://arxiv.org/abs/2301.11320
|
192 |
created: "2023-01-28T11:11:22.000Z"
|
@@ -196,7 +196,7 @@ https://huggingface.co/spaces/facebook/CutLER:
|
|
196 |
- object detection
|
197 |
- segmentation
|
198 |
- unsupervised
|
199 |
-
|
200 |
arxiv:
|
201 |
- https://arxiv.org/abs/2301.11320
|
202 |
created: "2023-01-28T04:37:53.000Z"
|
@@ -206,7 +206,7 @@ https://huggingface.co/spaces/facebook/MaskCut:
|
|
206 |
- object detection
|
207 |
- segmentation
|
208 |
- unsupervised
|
209 |
-
|
210 |
created: "2023-01-20T05:11:08.000Z"
|
211 |
github:
|
212 |
- https://github.com/huggingface/diffusers/tree/main/examples/dreambooth
|
@@ -215,7 +215,7 @@ https://huggingface.co/spaces/lora-library/LoRA-DreamBooth-Training-UI:
|
|
215 |
- DreamBooth
|
216 |
- diffusion model
|
217 |
- text2image
|
218 |
-
|
219 |
created: "2022-12-11T07:01:37.000Z"
|
220 |
github:
|
221 |
- https://github.com/cloneofsimo/lora
|
@@ -224,7 +224,7 @@ https://huggingface.co/spaces/hysts/LoRA-SD-training:
|
|
224 |
- DreamBooth
|
225 |
- diffusion model
|
226 |
- text2image
|
227 |
-
|
228 |
arxiv:
|
229 |
- https://arxiv.org/abs/2211.17115
|
230 |
created: "2022-12-03T09:13:05.000Z"
|
@@ -233,14 +233,14 @@ https://huggingface.co/spaces/hysts/multiresolution-textual-inversion:
|
|
233 |
tags:
|
234 |
- diffusion model
|
235 |
- text2image
|
236 |
-
|
237 |
created: "2022-08-12T10:03:37.000Z"
|
238 |
tags:
|
239 |
- diffusion model
|
240 |
- anime
|
241 |
- illustration
|
242 |
- personal project
|
243 |
-
|
244 |
arxiv:
|
245 |
- https://arxiv.org/abs/2205.15868
|
246 |
created: "2022-07-18T09:06:22.000Z"
|
@@ -248,25 +248,25 @@ https://huggingface.co/spaces/THUDM/CogVideo:
|
|
248 |
- https://github.com/THUDM/CogVideo
|
249 |
tags:
|
250 |
- text2video
|
251 |
-
|
252 |
created: "2022-07-13T17:51:20.000Z"
|
253 |
github:
|
254 |
- https://github.com/TachibanaYoshino/AnimeGANv3
|
255 |
tags:
|
256 |
- sketch
|
257 |
-
|
258 |
created: "2022-07-13T11:47:41.000Z"
|
259 |
tags:
|
260 |
- list of papers
|
261 |
-
|
262 |
created: "2022-07-11T12:16:32.000Z"
|
263 |
tags:
|
264 |
- list of papers
|
265 |
-
|
266 |
created: "2022-07-10T21:40:05.000Z"
|
267 |
tags:
|
268 |
- list of papers
|
269 |
-
|
270 |
arxiv:
|
271 |
- https://arxiv.org/abs/2204.14217
|
272 |
created: "2022-06-20T16:06:52.000Z"
|
@@ -274,7 +274,7 @@ https://huggingface.co/spaces/THUDM/CogView2:
|
|
274 |
- https://github.com/THUDM/CogView2
|
275 |
tags:
|
276 |
- text2image
|
277 |
-
|
278 |
created: "2022-06-10T23:12:02.000Z"
|
279 |
github:
|
280 |
- https://github.com/ljsabc/MangaLineExtraction_PyTorch
|
@@ -282,7 +282,7 @@ https://huggingface.co/spaces/hysts/MangaLineExtraction_PyTorch:
|
|
282 |
- anime
|
283 |
- illustration
|
284 |
- sketch
|
285 |
-
|
286 |
arxiv:
|
287 |
- https://arxiv.org/abs/2204.12484
|
288 |
created: "2022-06-08T10:38:50.000Z"
|
@@ -290,7 +290,7 @@ https://huggingface.co/spaces/hysts/ViTPose_video:
|
|
290 |
- https://github.com/ViTAE-Transformer/ViTPose
|
291 |
tags:
|
292 |
- human pose estimation
|
293 |
-
|
294 |
arxiv:
|
295 |
- https://arxiv.org/abs/2205.15996
|
296 |
created: "2022-06-04T08:12:18.000Z"
|
@@ -298,7 +298,7 @@ https://huggingface.co/spaces/hysts/Text2Human:
|
|
298 |
- https://github.com/yumingj/Text2Human
|
299 |
tags:
|
300 |
- text2image
|
301 |
-
|
302 |
arxiv:
|
303 |
- https://arxiv.org/abs/2204.12484
|
304 |
created: "2022-05-29T06:25:28.000Z"
|
@@ -306,7 +306,7 @@ https://huggingface.co/spaces/Gradio-Blocks/ViTPose:
|
|
306 |
- https://github.com/ViTAE-Transformer/ViTPose
|
307 |
tags:
|
308 |
- human pose estimation
|
309 |
-
|
310 |
arxiv:
|
311 |
- http://arxiv.org/abs/2107.00420
|
312 |
created: "2022-05-29T03:51:21.000Z"
|
@@ -314,7 +314,7 @@ https://huggingface.co/spaces/Gradio-Blocks/CBNetV2:
|
|
314 |
- https://github.com/VDIGPKU/CBNetV2
|
315 |
tags:
|
316 |
- image classification
|
317 |
-
|
318 |
created: "2022-05-29T02:11:52.000Z"
|
319 |
github:
|
320 |
- https://github.com/open-mmlab/mmdetection
|
@@ -323,7 +323,7 @@ https://huggingface.co/spaces/hysts/mmdetection:
|
|
323 |
- instance segmentation
|
324 |
- panoptic segmentation
|
325 |
- segmentation
|
326 |
-
|
327 |
arxiv:
|
328 |
- https://arxiv.org/abs/2112.05142
|
329 |
created: "2022-05-26T21:38:44.000Z"
|
@@ -333,7 +333,7 @@ https://huggingface.co/spaces/Gradio-Blocks/HairCLIP:
|
|
333 |
- GAN
|
334 |
- image editing
|
335 |
- face
|
336 |
-
|
337 |
arxiv:
|
338 |
- https://arxiv.org/abs/2203.13248
|
339 |
created: "2022-05-20T22:17:37.000Z"
|
@@ -343,7 +343,7 @@ https://huggingface.co/spaces/Gradio-Blocks/DualStyleGAN:
|
|
343 |
- GAN
|
344 |
- image editing
|
345 |
- face
|
346 |
-
|
347 |
arxiv:
|
348 |
- https://arxiv.org/abs/2204.11823
|
349 |
created: "2022-05-20T13:19:56.000Z"
|
@@ -351,7 +351,7 @@ https://huggingface.co/spaces/Gradio-Blocks/StyleGAN-Human:
|
|
351 |
- https://github.com/stylegan-human/StyleGAN-Human
|
352 |
tags:
|
353 |
- GAN
|
354 |
-
|
355 |
arxiv:
|
356 |
- https://arxiv.org/abs/2204.11823
|
357 |
created: "2022-04-22T09:48:26.000Z"
|
@@ -359,7 +359,7 @@ https://huggingface.co/spaces/hysts/StyleGAN-Human-Interpolation:
|
|
359 |
- https://github.com/stylegan-human/StyleGAN-Human
|
360 |
tags:
|
361 |
- GAN
|
362 |
-
|
363 |
arxiv:
|
364 |
- https://arxiv.org/abs/2204.11823
|
365 |
created: "2022-04-22T08:28:33.000Z"
|
@@ -367,7 +367,7 @@ https://huggingface.co/spaces/hysts/StyleGAN-Human:
|
|
367 |
- https://github.com/stylegan-human/StyleGAN-Human
|
368 |
tags:
|
369 |
- GAN
|
370 |
-
|
371 |
arxiv:
|
372 |
- https://arxiv.org/abs/2101.02477
|
373 |
created: "2022-04-18T18:53:22.000Z"
|
@@ -375,7 +375,7 @@ https://huggingface.co/spaces/hysts/gan-control:
|
|
375 |
- https://github.com/amazon-research/gan-control
|
376 |
tags:
|
377 |
- GAN
|
378 |
-
|
379 |
created: "2022-04-18T17:27:04.000Z"
|
380 |
github:
|
381 |
- https://github.com/kha-white/manga-ocr
|
@@ -383,7 +383,7 @@ https://huggingface.co/spaces/hysts/Manga-OCR:
|
|
383 |
- anime
|
384 |
- illustration
|
385 |
- OCR
|
386 |
-
|
387 |
created: "2022-04-18T15:59:06.000Z"
|
388 |
github:
|
389 |
- https://github.com/atksh/onnx-facial-lmk-detector
|
@@ -391,7 +391,7 @@ https://huggingface.co/spaces/hysts/atksh-onnx-facial-lmk-detector:
|
|
391 |
- face
|
392 |
- face detection
|
393 |
- face landmark detection
|
394 |
-
|
395 |
arxiv:
|
396 |
- https://arxiv.org/abs/1710.00925
|
397 |
created: "2022-04-15T20:03:00.000Z"
|
@@ -400,7 +400,7 @@ https://huggingface.co/spaces/hysts/Hopenet:
|
|
400 |
tags:
|
401 |
- face
|
402 |
- head pose estimation
|
403 |
-
|
404 |
arxiv:
|
405 |
- https://arxiv.org/abs/2006.10204
|
406 |
created: "2022-04-15T19:57:59.000Z"
|
@@ -409,7 +409,7 @@ https://huggingface.co/spaces/hysts/mediapipe-pose-estimation:
|
|
409 |
tags:
|
410 |
- mediapipe
|
411 |
- human pose estimation
|
412 |
-
|
413 |
arxiv:
|
414 |
- https://arxiv.org/abs/1907.05047
|
415 |
created: "2022-04-15T19:56:19.000Z"
|
@@ -419,7 +419,7 @@ https://huggingface.co/spaces/hysts/mediapipe-face-detection:
|
|
419 |
- mediapipe
|
420 |
- face
|
421 |
- face detection
|
422 |
-
|
423 |
arxiv:
|
424 |
- https://arxiv.org/abs/1907.06724
|
425 |
created: "2022-04-15T19:53:57.000Z"
|
@@ -429,7 +429,7 @@ https://huggingface.co/spaces/hysts/mediapipe-face-mesh:
|
|
429 |
- mediapipe
|
430 |
- face
|
431 |
- face landmark detection
|
432 |
-
|
433 |
arxiv:
|
434 |
- https://arxiv.org/abs/2104.05703
|
435 |
created: "2022-04-15T16:10:50.000Z"
|
@@ -439,51 +439,51 @@ https://huggingface.co/spaces/hysts/Anime2Sketch:
|
|
439 |
- anime
|
440 |
- illustration
|
441 |
- sketch
|
442 |
-
|
443 |
created: "2022-04-14T22:00:57.000Z"
|
444 |
tags:
|
445 |
- anime
|
446 |
- illustration
|
447 |
- TADNE
|
448 |
-
|
449 |
created: "2022-04-14T18:09:44.000Z"
|
450 |
tags:
|
451 |
- anime
|
452 |
- illustration
|
453 |
- TADNE
|
454 |
-
|
455 |
created: "2022-04-14T15:06:35.000Z"
|
456 |
tags:
|
457 |
- anime
|
458 |
- illustration
|
459 |
- TADNE
|
460 |
-
|
461 |
created: "2022-04-10T12:05:21.000Z"
|
462 |
tags:
|
463 |
- anime
|
464 |
- illustration
|
465 |
- TADNE
|
466 |
-
|
467 |
created: "2022-04-10T04:54:39.000Z"
|
468 |
tags:
|
469 |
- anime
|
470 |
- illustration
|
471 |
- TADNE
|
472 |
-
|
473 |
created: "2022-04-09T17:20:11.000Z"
|
474 |
github:
|
475 |
- https://github.com/ibug-group/emotion_recognition
|
476 |
tags:
|
477 |
- face
|
478 |
- emotion recognition
|
479 |
-
|
480 |
created: "2022-04-09T17:18:40.000Z"
|
481 |
github:
|
482 |
- https://github.com/ibug-group/face_alignment
|
483 |
tags:
|
484 |
- face
|
485 |
- face landmark detection
|
486 |
-
|
487 |
arxiv:
|
488 |
- https://arxiv.org/abs/2106.11145
|
489 |
created: "2022-04-09T16:04:16.000Z"
|
@@ -492,7 +492,7 @@ https://huggingface.co/spaces/hysts/ibug-fpage:
|
|
492 |
tags:
|
493 |
- face
|
494 |
- age estimation
|
495 |
-
|
496 |
arxiv:
|
497 |
- https://arxiv.org/abs/2102.02717
|
498 |
created: "2022-04-09T13:09:31.000Z"
|
@@ -502,14 +502,14 @@ https://huggingface.co/spaces/hysts/ibug-face_parsing:
|
|
502 |
- face
|
503 |
- face parsing
|
504 |
- segmentation
|
505 |
-
|
506 |
created: "2022-04-09T12:26:23.000Z"
|
507 |
github:
|
508 |
- https://github.com/ibug-group/face_detection
|
509 |
tags:
|
510 |
- face
|
511 |
- face detection
|
512 |
-
|
513 |
arxiv:
|
514 |
- https://arxiv.org/abs/2105.04714
|
515 |
created: "2022-04-06T12:07:02.000Z"
|
@@ -518,7 +518,7 @@ https://huggingface.co/spaces/hysts/insightface-SCRFD:
|
|
518 |
tags:
|
519 |
- face
|
520 |
- face detection
|
521 |
-
|
522 |
arxiv:
|
523 |
- https://arxiv.org/abs/2105.04714
|
524 |
created: "2022-04-06T08:39:44.000Z"
|
@@ -526,7 +526,7 @@ https://huggingface.co/spaces/hysts/insightface-person-detection:
|
|
526 |
- https://github.com/deepinsight/insightface/tree/master/examples/person_detection
|
527 |
tags:
|
528 |
- person detection
|
529 |
-
|
530 |
arxiv:
|
531 |
- https://arxiv.org/abs/1907.11922
|
532 |
created: "2022-04-06T02:56:32.000Z"
|
@@ -536,7 +536,7 @@ https://huggingface.co/spaces/hysts/CelebAMask-HQ-Face-Parsing:
|
|
536 |
- face
|
537 |
- face parsing
|
538 |
- segmentation
|
539 |
-
|
540 |
arxiv:
|
541 |
- https://arxiv.org/abs/2104.04767
|
542 |
created: "2022-04-05T09:23:01.000Z"
|
@@ -544,7 +544,7 @@ https://huggingface.co/spaces/hysts/MobileStyleGAN:
|
|
544 |
- https://github.com/bes-dev/MobileStyleGAN.pytorch
|
545 |
tags:
|
546 |
- GAN
|
547 |
-
|
548 |
arxiv:
|
549 |
- https://arxiv.org/abs/2202.12211
|
550 |
created: "2022-04-05T02:40:27.000Z"
|
@@ -552,7 +552,7 @@ https://huggingface.co/spaces/hysts/Self-Distilled-StyleGAN:
|
|
552 |
- https://github.com/self-distilled-stylegan/self-distilled-internet-photos
|
553 |
tags:
|
554 |
- GAN
|
555 |
-
|
556 |
arxiv:
|
557 |
- https://arxiv.org/abs/1912.04958
|
558 |
created: "2022-04-03T19:53:52.000Z"
|
@@ -560,7 +560,7 @@ https://huggingface.co/spaces/hysts/StyleGAN2:
|
|
560 |
- https://github.com/NVlabs/stylegan3
|
561 |
tags:
|
562 |
- GAN
|
563 |
-
|
564 |
arxiv:
|
565 |
- https://arxiv.org/abs/2111.01007
|
566 |
created: "2022-04-03T17:28:35.000Z"
|
@@ -568,7 +568,7 @@ https://huggingface.co/spaces/hysts/projected_gan:
|
|
568 |
- https://github.com/autonomousvision/projected_gan
|
569 |
tags:
|
570 |
- GAN
|
571 |
-
|
572 |
arxiv:
|
573 |
- https://arxiv.org/abs/2106.12423
|
574 |
created: "2022-04-03T15:56:35.000Z"
|
@@ -576,7 +576,7 @@ https://huggingface.co/spaces/hysts/StyleGAN3:
|
|
576 |
- https://github.com/NVlabs/stylegan3
|
577 |
tags:
|
578 |
- GAN
|
579 |
-
|
580 |
arxiv:
|
581 |
- https://arxiv.org/abs/2202.00273
|
582 |
created: "2022-03-31T19:20:05.000Z"
|
@@ -584,7 +584,7 @@ https://huggingface.co/spaces/hysts/StyleGAN-XL:
|
|
584 |
- https://github.com/autonomousvision/stylegan_xl
|
585 |
tags:
|
586 |
- GAN
|
587 |
-
|
588 |
arxiv:
|
589 |
- https://arxiv.org/abs/1703.07332
|
590 |
created: "2022-03-28T11:33:31.000Z"
|
@@ -594,7 +594,7 @@ https://huggingface.co/spaces/hysts/1adrianb-face-alignment:
|
|
594 |
- face
|
595 |
- face detection
|
596 |
- face landmark detection
|
597 |
-
|
598 |
arxiv:
|
599 |
- https://arxiv.org/abs/2203.13248
|
600 |
created: "2022-03-23T20:42:59.000Z"
|
@@ -602,7 +602,7 @@ https://huggingface.co/spaces/hysts/DualStyleGAN:
|
|
602 |
- https://github.com/williamyang1991/DualStyleGAN
|
603 |
tags:
|
604 |
- GAN
|
605 |
-
|
606 |
arxiv:
|
607 |
- https://arxiv.org/abs/2112.10762
|
608 |
created: "2022-03-15T14:27:48.000Z"
|
@@ -610,14 +610,14 @@ https://huggingface.co/spaces/hysts/StyleSwin:
|
|
610 |
- https://github.com/microsoft/StyleSwin
|
611 |
tags:
|
612 |
- GAN
|
613 |
-
|
614 |
created: "2022-02-28T00:45:11.000Z"
|
615 |
github:
|
616 |
- https://github.com/yu4u/age-estimation-pytorch
|
617 |
tags:
|
618 |
- face
|
619 |
- age estimation
|
620 |
-
|
621 |
created: "2022-01-24T09:09:55.000Z"
|
622 |
github:
|
623 |
- https://github.com/kanosawa/anime_face_landmark_detection
|
@@ -625,7 +625,7 @@ https://huggingface.co/spaces/hysts/anime_face_landmark_detection:
|
|
625 |
- anime
|
626 |
- illustration
|
627 |
- anime face landmark detection
|
628 |
-
|
629 |
created: "2022-01-24T08:24:09.000Z"
|
630 |
github:
|
631 |
- https://github.com/nagadomi/lbpcascade_animeface
|
@@ -633,7 +633,7 @@ https://huggingface.co/spaces/hysts/lbpcascade_animeface:
|
|
633 |
- anime
|
634 |
- illustration
|
635 |
- anime face detection
|
636 |
-
|
637 |
created: "2022-01-24T05:55:41.000Z"
|
638 |
github:
|
639 |
- https://github.com/zymk9/yolov5_anime
|
@@ -641,7 +641,7 @@ https://huggingface.co/spaces/hysts/yolov5_anime:
|
|
641 |
- anime
|
642 |
- illustration
|
643 |
- anime face detection
|
644 |
-
|
645 |
arxiv:
|
646 |
- https://arxiv.org/abs/2108.01819
|
647 |
created: "2022-01-24T03:14:52.000Z"
|
@@ -651,7 +651,7 @@ https://huggingface.co/spaces/hysts/bizarre-pose-estimator-segmenter:
|
|
651 |
- anime
|
652 |
- illustration
|
653 |
- anime character segmentation
|
654 |
-
|
655 |
created: "2022-01-24T00:01:55.000Z"
|
656 |
github:
|
657 |
- https://github.com/zymk9/Yet-Another-Anime-Segmenter
|
@@ -659,7 +659,7 @@ https://huggingface.co/spaces/hysts/Yet-Another-Anime-Segmenter:
|
|
659 |
- anime
|
660 |
- illustration
|
661 |
- anime character segmentation
|
662 |
-
|
663 |
arxiv:
|
664 |
- https://arxiv.org/abs/2108.01819
|
665 |
created: "2022-01-23T13:26:52.000Z"
|
@@ -669,7 +669,7 @@ https://huggingface.co/spaces/hysts/bizarre-pose-estimator-tagger:
|
|
669 |
- anime
|
670 |
- illustration
|
671 |
- anime label prediction
|
672 |
-
|
673 |
created: "2022-01-23T12:04:20.000Z"
|
674 |
github:
|
675 |
- https://github.com/KichangKim/DeepDanbooru
|
@@ -678,7 +678,7 @@ https://huggingface.co/spaces/hysts/DeepDanbooru:
|
|
678 |
- illustration
|
679 |
- anime label prediction
|
680 |
- Danbooru
|
681 |
-
|
682 |
created: "2022-01-23T10:46:30.000Z"
|
683 |
github:
|
684 |
- https://github.com/RF5/danbooru-pretrained
|
@@ -687,22 +687,22 @@ https://huggingface.co/spaces/hysts/danbooru-pretrained:
|
|
687 |
- illustration
|
688 |
- anime label prediction
|
689 |
- Danbooru
|
690 |
-
|
691 |
created: "2021-12-02T19:33:20.000Z"
|
692 |
tags:
|
693 |
- GAN
|
694 |
- personal project
|
695 |
-
|
696 |
created: "2021-11-22T04:05:46.000Z"
|
697 |
tags:
|
698 |
- GAN
|
699 |
- personal project
|
700 |
-
|
701 |
created: "2021-11-22T02:10:56.000Z"
|
702 |
tags:
|
703 |
- GAN
|
704 |
- personal project
|
705 |
-
|
706 |
created: "2021-11-15T02:15:45.000Z"
|
707 |
github:
|
708 |
- https://github.com/hysts/anime-face-detector
|
|
|
1 |
+
hysts/SD-XL:
|
2 |
arxiv:
|
3 |
- https://arxiv.org/abs/2307.01952
|
4 |
created: "2023-07-27T05:10:31.000Z"
|
|
|
7 |
tags:
|
8 |
- diffusion model
|
9 |
- text to image
|
10 |
+
huggingface-projects/llama-2-7b-chat:
|
11 |
arxiv:
|
12 |
- https://arxiv.org/abs/2307.09288
|
13 |
created: "2023-07-19T03:33:27.000Z"
|
|
|
17 |
- chatbot
|
18 |
- LLM
|
19 |
- Llama 2
|
20 |
+
hysts/Kandinsky-2-1:
|
21 |
created: "2023-07-14T12:34:56.000Z"
|
22 |
tags:
|
23 |
- diffusion model
|
24 |
- text to image
|
25 |
+
hysts/Kandinsky-2-2:
|
26 |
created: "2023-07-14T09:48:26.000Z"
|
27 |
tags:
|
28 |
- diffusion model
|
29 |
- text to image
|
30 |
+
hysts/zeroscope-v2:
|
31 |
created: "2023-06-27T02:57:53.000Z"
|
32 |
tags:
|
33 |
- diffusion model
|
34 |
- text to video
|
35 |
+
ICML2023/ICML2023_papers:
|
36 |
created: "2023-06-15T17:16:16.000Z"
|
37 |
tags:
|
38 |
- list of papers
|
39 |
+
hysts/Shap-E:
|
40 |
arxiv:
|
41 |
- https://arxiv.org/abs/2305.02463
|
42 |
created: "2023-05-05T07:40:18.000Z"
|
|
|
46 |
- diffusion model
|
47 |
- text to 3D
|
48 |
- image to 3D
|
49 |
+
hysts/ControlNet-v1-1:
|
50 |
arxiv:
|
51 |
- https://arxiv.org/abs/2302.05543
|
52 |
created: "2023-04-15T01:26:54.000Z"
|
|
|
57 |
- diffusion model
|
58 |
- text2image
|
59 |
- image2image
|
60 |
+
DeepFloyd/IF:
|
61 |
created: "2023-03-30T01:36:56.000Z"
|
62 |
github:
|
63 |
- https://github.com/deep-floyd/IF
|
64 |
tags:
|
65 |
- diffusion model
|
66 |
- text2image
|
67 |
+
damo-vilab/modelscope-text-to-video-synthesis:
|
68 |
arxiv:
|
69 |
- https://arxiv.org/abs/2303.08320
|
70 |
created: "2023-03-19T04:30:43.000Z"
|
|
|
73 |
tags:
|
74 |
- diffusion model
|
75 |
- text2video
|
76 |
+
thu-ml/unidiffuser:
|
77 |
arxiv:
|
78 |
- https://arxiv.org/abs/2303.06555
|
79 |
created: "2023-03-14T05:06:57.000Z"
|
|
|
84 |
- text2image
|
85 |
- image2image
|
86 |
- image captioning
|
87 |
+
ELITE-library/ELITE:
|
88 |
arxiv:
|
89 |
- https://arxiv.org/abs/2302.13848
|
90 |
created: "2023-03-09T03:28:49.000Z"
|
|
|
93 |
tags:
|
94 |
- diffusion model
|
95 |
- text2image
|
96 |
+
hysts/cv_diffusion_text-to-image-synthesis_tiny:
|
97 |
created: "2023-03-08T05:44:53.000Z"
|
98 |
tags:
|
99 |
- diffusion model
|
100 |
- text2image
|
101 |
+
hysts/DDNM-HQ:
|
102 |
arxiv:
|
103 |
- https://arxiv.org/abs/2303.00354
|
104 |
- https://arxiv.org/abs/2212.00490
|
|
|
108 |
tags:
|
109 |
- diffusion model
|
110 |
- super resolution
|
111 |
+
hysts/ControlNet-with-Anything-v4:
|
112 |
arxiv:
|
113 |
- https://arxiv.org/abs/2302.05543
|
114 |
created: "2023-02-16T05:05:27.000Z"
|
|
|
118 |
- diffusion model
|
119 |
- text2image
|
120 |
- image2image
|
121 |
+
hysts/ControlNet:
|
122 |
arxiv:
|
123 |
- https://arxiv.org/abs/2302.05543
|
124 |
created: "2023-02-14T01:03:00.000Z"
|
|
|
128 |
- diffusion model
|
129 |
- text2image
|
130 |
- image2image
|
131 |
+
hysts/PnP-diffusion-features:
|
132 |
arxiv:
|
133 |
- https://arxiv.org/abs/2211.12572
|
134 |
created: "2023-02-13T06:20:32.000Z"
|
|
|
137 |
tags:
|
138 |
- diffusion model
|
139 |
- image2image
|
140 |
+
hysts/BLIP2-with-transformers:
|
141 |
arxiv:
|
142 |
- https://arxiv.org/abs/2301.12597
|
143 |
created: "2023-02-10T09:45:07.000Z"
|
|
|
146 |
tags:
|
147 |
- image captioning
|
148 |
- VQA
|
149 |
+
Tune-A-Video-library/Tune-A-Video-inference:
|
150 |
arxiv:
|
151 |
- https://arxiv.org/abs/2212.11565
|
152 |
created: "2023-02-09T02:39:44.000Z"
|
|
|
156 |
- diffusion model
|
157 |
- text2video
|
158 |
- video2video
|
159 |
+
TEXTurePaper/TEXTure:
|
160 |
arxiv:
|
161 |
- https://arxiv.org/abs/2302.01721
|
162 |
created: "2023-02-06T09:16:46.000Z"
|
|
|
167 |
- text-guided
|
168 |
- texture
|
169 |
- 3D
|
170 |
+
AttendAndExcite/Attend-and-Excite:
|
171 |
arxiv:
|
172 |
- https://arxiv.org/abs/2301.13826
|
173 |
created: "2023-02-01T06:18:56.000Z"
|
|
|
176 |
tags:
|
177 |
- diffusion model
|
178 |
- text2image
|
179 |
+
Tune-A-Video-library/Tune-A-Video-Training-UI:
|
180 |
arxiv:
|
181 |
- https://arxiv.org/abs/2212.11565
|
182 |
created: "2023-01-30T08:10:09.000Z"
|
|
|
186 |
- diffusion model
|
187 |
- text2video
|
188 |
- video2video
|
189 |
+
facebook/CutLER:
|
190 |
arxiv:
|
191 |
- https://arxiv.org/abs/2301.11320
|
192 |
created: "2023-01-28T11:11:22.000Z"
|
|
|
196 |
- object detection
|
197 |
- segmentation
|
198 |
- unsupervised
|
199 |
+
facebook/MaskCut:
|
200 |
arxiv:
|
201 |
- https://arxiv.org/abs/2301.11320
|
202 |
created: "2023-01-28T04:37:53.000Z"
|
|
|
206 |
- object detection
|
207 |
- segmentation
|
208 |
- unsupervised
|
209 |
+
lora-library/LoRA-DreamBooth-Training-UI:
|
210 |
created: "2023-01-20T05:11:08.000Z"
|
211 |
github:
|
212 |
- https://github.com/huggingface/diffusers/tree/main/examples/dreambooth
|
|
|
215 |
- DreamBooth
|
216 |
- diffusion model
|
217 |
- text2image
|
218 |
+
hysts/LoRA-SD-training:
|
219 |
created: "2022-12-11T07:01:37.000Z"
|
220 |
github:
|
221 |
- https://github.com/cloneofsimo/lora
|
|
|
224 |
- DreamBooth
|
225 |
- diffusion model
|
226 |
- text2image
|
227 |
+
hysts/multiresolution-textual-inversion:
|
228 |
arxiv:
|
229 |
- https://arxiv.org/abs/2211.17115
|
230 |
created: "2022-12-03T09:13:05.000Z"
|
|
|
233 |
tags:
|
234 |
- diffusion model
|
235 |
- text2image
|
236 |
+
hysts/diffusers-anime-faces:
|
237 |
created: "2022-08-12T10:03:37.000Z"
|
238 |
tags:
|
239 |
- diffusion model
|
240 |
- anime
|
241 |
- illustration
|
242 |
- personal project
|
243 |
+
THUDM/CogVideo:
|
244 |
arxiv:
|
245 |
- https://arxiv.org/abs/2205.15868
|
246 |
created: "2022-07-18T09:06:22.000Z"
|
|
|
248 |
- https://github.com/THUDM/CogVideo
|
249 |
tags:
|
250 |
- text2video
|
251 |
+
hysts/AnimeGANv3_PortraitSketch:
|
252 |
created: "2022-07-13T17:51:20.000Z"
|
253 |
github:
|
254 |
- https://github.com/TachibanaYoshino/AnimeGANv3
|
255 |
tags:
|
256 |
- sketch
|
257 |
+
hysts/ICML2022_papers:
|
258 |
created: "2022-07-13T11:47:41.000Z"
|
259 |
tags:
|
260 |
- list of papers
|
261 |
+
hysts/CVPR2022_papers:
|
262 |
created: "2022-07-11T12:16:32.000Z"
|
263 |
tags:
|
264 |
- list of papers
|
265 |
+
NAACL2022/papers:
|
266 |
created: "2022-07-10T21:40:05.000Z"
|
267 |
tags:
|
268 |
- list of papers
|
269 |
+
THUDM/CogView2:
|
270 |
arxiv:
|
271 |
- https://arxiv.org/abs/2204.14217
|
272 |
created: "2022-06-20T16:06:52.000Z"
|
|
|
274 |
- https://github.com/THUDM/CogView2
|
275 |
tags:
|
276 |
- text2image
|
277 |
+
hysts/MangaLineExtraction_PyTorch:
|
278 |
created: "2022-06-10T23:12:02.000Z"
|
279 |
github:
|
280 |
- https://github.com/ljsabc/MangaLineExtraction_PyTorch
|
|
|
282 |
- anime
|
283 |
- illustration
|
284 |
- sketch
|
285 |
+
hysts/ViTPose_video:
|
286 |
arxiv:
|
287 |
- https://arxiv.org/abs/2204.12484
|
288 |
created: "2022-06-08T10:38:50.000Z"
|
|
|
290 |
- https://github.com/ViTAE-Transformer/ViTPose
|
291 |
tags:
|
292 |
- human pose estimation
|
293 |
+
hysts/Text2Human:
|
294 |
arxiv:
|
295 |
- https://arxiv.org/abs/2205.15996
|
296 |
created: "2022-06-04T08:12:18.000Z"
|
|
|
298 |
- https://github.com/yumingj/Text2Human
|
299 |
tags:
|
300 |
- text2image
|
301 |
+
Gradio-Blocks/ViTPose:
|
302 |
arxiv:
|
303 |
- https://arxiv.org/abs/2204.12484
|
304 |
created: "2022-05-29T06:25:28.000Z"
|
|
|
306 |
- https://github.com/ViTAE-Transformer/ViTPose
|
307 |
tags:
|
308 |
- human pose estimation
|
309 |
+
Gradio-Blocks/CBNetV2:
|
310 |
arxiv:
|
311 |
- http://arxiv.org/abs/2107.00420
|
312 |
created: "2022-05-29T03:51:21.000Z"
|
|
|
314 |
- https://github.com/VDIGPKU/CBNetV2
|
315 |
tags:
|
316 |
- image classification
|
317 |
+
hysts/mmdetection:
|
318 |
created: "2022-05-29T02:11:52.000Z"
|
319 |
github:
|
320 |
- https://github.com/open-mmlab/mmdetection
|
|
|
323 |
- instance segmentation
|
324 |
- panoptic segmentation
|
325 |
- segmentation
|
326 |
+
Gradio-Blocks/HairCLIP:
|
327 |
arxiv:
|
328 |
- https://arxiv.org/abs/2112.05142
|
329 |
created: "2022-05-26T21:38:44.000Z"
|
|
|
333 |
- GAN
|
334 |
- image editing
|
335 |
- face
|
336 |
+
Gradio-Blocks/DualStyleGAN:
|
337 |
arxiv:
|
338 |
- https://arxiv.org/abs/2203.13248
|
339 |
created: "2022-05-20T22:17:37.000Z"
|
|
|
343 |
- GAN
|
344 |
- image editing
|
345 |
- face
|
346 |
+
Gradio-Blocks/StyleGAN-Human:
|
347 |
arxiv:
|
348 |
- https://arxiv.org/abs/2204.11823
|
349 |
created: "2022-05-20T13:19:56.000Z"
|
|
|
351 |
- https://github.com/stylegan-human/StyleGAN-Human
|
352 |
tags:
|
353 |
- GAN
|
354 |
+
hysts/StyleGAN-Human-Interpolation:
|
355 |
arxiv:
|
356 |
- https://arxiv.org/abs/2204.11823
|
357 |
created: "2022-04-22T09:48:26.000Z"
|
|
|
359 |
- https://github.com/stylegan-human/StyleGAN-Human
|
360 |
tags:
|
361 |
- GAN
|
362 |
+
hysts/StyleGAN-Human:
|
363 |
arxiv:
|
364 |
- https://arxiv.org/abs/2204.11823
|
365 |
created: "2022-04-22T08:28:33.000Z"
|
|
|
367 |
- https://github.com/stylegan-human/StyleGAN-Human
|
368 |
tags:
|
369 |
- GAN
|
370 |
+
hysts/gan-control:
|
371 |
arxiv:
|
372 |
- https://arxiv.org/abs/2101.02477
|
373 |
created: "2022-04-18T18:53:22.000Z"
|
|
|
375 |
- https://github.com/amazon-research/gan-control
|
376 |
tags:
|
377 |
- GAN
|
378 |
+
hysts/Manga-OCR:
|
379 |
created: "2022-04-18T17:27:04.000Z"
|
380 |
github:
|
381 |
- https://github.com/kha-white/manga-ocr
|
|
|
383 |
- anime
|
384 |
- illustration
|
385 |
- OCR
|
386 |
+
hysts/atksh-onnx-facial-lmk-detector:
|
387 |
created: "2022-04-18T15:59:06.000Z"
|
388 |
github:
|
389 |
- https://github.com/atksh/onnx-facial-lmk-detector
|
|
|
391 |
- face
|
392 |
- face detection
|
393 |
- face landmark detection
|
394 |
+
hysts/Hopenet:
|
395 |
arxiv:
|
396 |
- https://arxiv.org/abs/1710.00925
|
397 |
created: "2022-04-15T20:03:00.000Z"
|
|
|
400 |
tags:
|
401 |
- face
|
402 |
- head pose estimation
|
403 |
+
hysts/mediapipe-pose-estimation:
|
404 |
arxiv:
|
405 |
- https://arxiv.org/abs/2006.10204
|
406 |
created: "2022-04-15T19:57:59.000Z"
|
|
|
409 |
tags:
|
410 |
- mediapipe
|
411 |
- human pose estimation
|
412 |
+
hysts/mediapipe-face-detection:
|
413 |
arxiv:
|
414 |
- https://arxiv.org/abs/1907.05047
|
415 |
created: "2022-04-15T19:56:19.000Z"
|
|
|
419 |
- mediapipe
|
420 |
- face
|
421 |
- face detection
|
422 |
+
hysts/mediapipe-face-mesh:
|
423 |
arxiv:
|
424 |
- https://arxiv.org/abs/1907.06724
|
425 |
created: "2022-04-15T19:53:57.000Z"
|
|
|
429 |
- mediapipe
|
430 |
- face
|
431 |
- face landmark detection
|
432 |
+
hysts/Anime2Sketch:
|
433 |
arxiv:
|
434 |
- https://arxiv.org/abs/2104.05703
|
435 |
created: "2022-04-15T16:10:50.000Z"
|
|
|
439 |
- anime
|
440 |
- illustration
|
441 |
- sketch
|
442 |
+
hysts/TADNE-image-search-with-DeepDanbooru:
|
443 |
created: "2022-04-14T22:00:57.000Z"
|
444 |
tags:
|
445 |
- anime
|
446 |
- illustration
|
447 |
- TADNE
|
448 |
+
hysts/TADNE-image-selector:
|
449 |
created: "2022-04-14T18:09:44.000Z"
|
450 |
tags:
|
451 |
- anime
|
452 |
- illustration
|
453 |
- TADNE
|
454 |
+
hysts/TADNE-image-viewer:
|
455 |
created: "2022-04-14T15:06:35.000Z"
|
456 |
tags:
|
457 |
- anime
|
458 |
- illustration
|
459 |
- TADNE
|
460 |
+
hysts/TADNE-interpolation:
|
461 |
created: "2022-04-10T12:05:21.000Z"
|
462 |
tags:
|
463 |
- anime
|
464 |
- illustration
|
465 |
- TADNE
|
466 |
+
hysts/TADNE:
|
467 |
created: "2022-04-10T04:54:39.000Z"
|
468 |
tags:
|
469 |
- anime
|
470 |
- illustration
|
471 |
- TADNE
|
472 |
+
hysts/ibug-emotion_recognition:
|
473 |
created: "2022-04-09T17:20:11.000Z"
|
474 |
github:
|
475 |
- https://github.com/ibug-group/emotion_recognition
|
476 |
tags:
|
477 |
- face
|
478 |
- emotion recognition
|
479 |
+
hysts/ibug-face_alignment:
|
480 |
created: "2022-04-09T17:18:40.000Z"
|
481 |
github:
|
482 |
- https://github.com/ibug-group/face_alignment
|
483 |
tags:
|
484 |
- face
|
485 |
- face landmark detection
|
486 |
+
hysts/ibug-fpage:
|
487 |
arxiv:
|
488 |
- https://arxiv.org/abs/2106.11145
|
489 |
created: "2022-04-09T16:04:16.000Z"
|
|
|
492 |
tags:
|
493 |
- face
|
494 |
- age estimation
|
495 |
+
hysts/ibug-face_parsing:
|
496 |
arxiv:
|
497 |
- https://arxiv.org/abs/2102.02717
|
498 |
created: "2022-04-09T13:09:31.000Z"
|
|
|
502 |
- face
|
503 |
- face parsing
|
504 |
- segmentation
|
505 |
+
hysts/ibug-face_detection:
|
506 |
created: "2022-04-09T12:26:23.000Z"
|
507 |
github:
|
508 |
- https://github.com/ibug-group/face_detection
|
509 |
tags:
|
510 |
- face
|
511 |
- face detection
|
512 |
+
hysts/insightface-SCRFD:
|
513 |
arxiv:
|
514 |
- https://arxiv.org/abs/2105.04714
|
515 |
created: "2022-04-06T12:07:02.000Z"
|
|
|
518 |
tags:
|
519 |
- face
|
520 |
- face detection
|
521 |
+
hysts/insightface-person-detection:
|
522 |
arxiv:
|
523 |
- https://arxiv.org/abs/2105.04714
|
524 |
created: "2022-04-06T08:39:44.000Z"
|
|
|
526 |
- https://github.com/deepinsight/insightface/tree/master/examples/person_detection
|
527 |
tags:
|
528 |
- person detection
|
529 |
+
hysts/CelebAMask-HQ-Face-Parsing:
|
530 |
arxiv:
|
531 |
- https://arxiv.org/abs/1907.11922
|
532 |
created: "2022-04-06T02:56:32.000Z"
|
|
|
536 |
- face
|
537 |
- face parsing
|
538 |
- segmentation
|
539 |
+
hysts/MobileStyleGAN:
|
540 |
arxiv:
|
541 |
- https://arxiv.org/abs/2104.04767
|
542 |
created: "2022-04-05T09:23:01.000Z"
|
|
|
544 |
- https://github.com/bes-dev/MobileStyleGAN.pytorch
|
545 |
tags:
|
546 |
- GAN
|
547 |
+
hysts/Self-Distilled-StyleGAN:
|
548 |
arxiv:
|
549 |
- https://arxiv.org/abs/2202.12211
|
550 |
created: "2022-04-05T02:40:27.000Z"
|
|
|
552 |
- https://github.com/self-distilled-stylegan/self-distilled-internet-photos
|
553 |
tags:
|
554 |
- GAN
|
555 |
+
hysts/StyleGAN2:
|
556 |
arxiv:
|
557 |
- https://arxiv.org/abs/1912.04958
|
558 |
created: "2022-04-03T19:53:52.000Z"
|
|
|
560 |
- https://github.com/NVlabs/stylegan3
|
561 |
tags:
|
562 |
- GAN
|
563 |
+
hysts/projected_gan:
|
564 |
arxiv:
|
565 |
- https://arxiv.org/abs/2111.01007
|
566 |
created: "2022-04-03T17:28:35.000Z"
|
|
|
568 |
- https://github.com/autonomousvision/projected_gan
|
569 |
tags:
|
570 |
- GAN
|
571 |
+
hysts/StyleGAN3:
|
572 |
arxiv:
|
573 |
- https://arxiv.org/abs/2106.12423
|
574 |
created: "2022-04-03T15:56:35.000Z"
|
|
|
576 |
- https://github.com/NVlabs/stylegan3
|
577 |
tags:
|
578 |
- GAN
|
579 |
+
hysts/StyleGAN-XL:
|
580 |
arxiv:
|
581 |
- https://arxiv.org/abs/2202.00273
|
582 |
created: "2022-03-31T19:20:05.000Z"
|
|
|
584 |
- https://github.com/autonomousvision/stylegan_xl
|
585 |
tags:
|
586 |
- GAN
|
587 |
+
hysts/1adrianb-face-alignment:
|
588 |
arxiv:
|
589 |
- https://arxiv.org/abs/1703.07332
|
590 |
created: "2022-03-28T11:33:31.000Z"
|
|
|
594 |
- face
|
595 |
- face detection
|
596 |
- face landmark detection
|
597 |
+
hysts/DualStyleGAN:
|
598 |
arxiv:
|
599 |
- https://arxiv.org/abs/2203.13248
|
600 |
created: "2022-03-23T20:42:59.000Z"
|
|
|
602 |
- https://github.com/williamyang1991/DualStyleGAN
|
603 |
tags:
|
604 |
- GAN
|
605 |
+
hysts/StyleSwin:
|
606 |
arxiv:
|
607 |
- https://arxiv.org/abs/2112.10762
|
608 |
created: "2022-03-15T14:27:48.000Z"
|
|
|
610 |
- https://github.com/microsoft/StyleSwin
|
611 |
tags:
|
612 |
- GAN
|
613 |
+
hysts/age-estimation-APPA-REAL:
|
614 |
created: "2022-02-28T00:45:11.000Z"
|
615 |
github:
|
616 |
- https://github.com/yu4u/age-estimation-pytorch
|
617 |
tags:
|
618 |
- face
|
619 |
- age estimation
|
620 |
+
hysts/anime_face_landmark_detection:
|
621 |
created: "2022-01-24T09:09:55.000Z"
|
622 |
github:
|
623 |
- https://github.com/kanosawa/anime_face_landmark_detection
|
|
|
625 |
- anime
|
626 |
- illustration
|
627 |
- anime face landmark detection
|
628 |
+
hysts/lbpcascade_animeface:
|
629 |
created: "2022-01-24T08:24:09.000Z"
|
630 |
github:
|
631 |
- https://github.com/nagadomi/lbpcascade_animeface
|
|
|
633 |
- anime
|
634 |
- illustration
|
635 |
- anime face detection
|
636 |
+
hysts/yolov5_anime:
|
637 |
created: "2022-01-24T05:55:41.000Z"
|
638 |
github:
|
639 |
- https://github.com/zymk9/yolov5_anime
|
|
|
641 |
- anime
|
642 |
- illustration
|
643 |
- anime face detection
|
644 |
+
hysts/bizarre-pose-estimator-segmenter:
|
645 |
arxiv:
|
646 |
- https://arxiv.org/abs/2108.01819
|
647 |
created: "2022-01-24T03:14:52.000Z"
|
|
|
651 |
- anime
|
652 |
- illustration
|
653 |
- anime character segmentation
|
654 |
+
hysts/Yet-Another-Anime-Segmenter:
|
655 |
created: "2022-01-24T00:01:55.000Z"
|
656 |
github:
|
657 |
- https://github.com/zymk9/Yet-Another-Anime-Segmenter
|
|
|
659 |
- anime
|
660 |
- illustration
|
661 |
- anime character segmentation
|
662 |
+
hysts/bizarre-pose-estimator-tagger:
|
663 |
arxiv:
|
664 |
- https://arxiv.org/abs/2108.01819
|
665 |
created: "2022-01-23T13:26:52.000Z"
|
|
|
669 |
- anime
|
670 |
- illustration
|
671 |
- anime label prediction
|
672 |
+
hysts/DeepDanbooru:
|
673 |
created: "2022-01-23T12:04:20.000Z"
|
674 |
github:
|
675 |
- https://github.com/KichangKim/DeepDanbooru
|
|
|
678 |
- illustration
|
679 |
- anime label prediction
|
680 |
- Danbooru
|
681 |
+
hysts/danbooru-pretrained:
|
682 |
created: "2022-01-23T10:46:30.000Z"
|
683 |
github:
|
684 |
- https://github.com/RF5/danbooru-pretrained
|
|
|
687 |
- illustration
|
688 |
- anime label prediction
|
689 |
- Danbooru
|
690 |
+
hysts/stylegan3-anime-face-exp002:
|
691 |
created: "2021-12-02T19:33:20.000Z"
|
692 |
tags:
|
693 |
- GAN
|
694 |
- personal project
|
695 |
+
hysts/stylegan3-food101:
|
696 |
created: "2021-11-22T04:05:46.000Z"
|
697 |
tags:
|
698 |
- GAN
|
699 |
- personal project
|
700 |
+
hysts/stylegan3-anime-face-exp001:
|
701 |
created: "2021-11-22T02:10:56.000Z"
|
702 |
tags:
|
703 |
- GAN
|
704 |
- personal project
|
705 |
+
hysts/anime-face-detector:
|
706 |
created: "2021-11-15T02:15:45.000Z"
|
707 |
github:
|
708 |
- https://github.com/hysts/anime-face-detector
|
tools/add_creation_timestamps.py
CHANGED
@@ -17,12 +17,13 @@ def add_creation_timestamps() -> None:
|
|
17 |
with open(yaml_path) as f:
|
18 |
data = yaml.safe_load(f)
|
19 |
|
20 |
-
|
21 |
-
for
|
22 |
-
info = data[
|
23 |
if 'created' in info:
|
24 |
continue
|
25 |
with tempfile.TemporaryDirectory() as temp_dir:
|
|
|
26 |
repo = git.Repo.clone_from(url, temp_dir)
|
27 |
commits = list(repo.iter_commits())
|
28 |
initial_commit = commits[-1]
|
|
|
17 |
with open(yaml_path) as f:
|
18 |
data = yaml.safe_load(f)
|
19 |
|
20 |
+
space_ids = list(data.keys())
|
21 |
+
for space_id in tqdm.auto.tqdm(space_ids):
|
22 |
+
info = data[space_id]
|
23 |
if 'created' in info:
|
24 |
continue
|
25 |
with tempfile.TemporaryDirectory() as temp_dir:
|
26 |
+
url = f'https://huggingface.co/spaces/{space_id}'
|
27 |
repo = git.Repo.clone_from(url, temp_dir)
|
28 |
commits = list(repo.iter_commits())
|
29 |
initial_commit = commits[-1]
|