mattdeitke
commited on
Commit
•
5c02574
1
Parent(s):
96eeaba
refactor so source files are separate from api files
Browse files- .github/workflows/ci.yaml +0 -51
- .isort.cfg +0 -3
- .pylintrc +0 -331
- Makefile +0 -28
- github/{github-urls.parquet → github.parquet} +2 -2
- mypy.ini +0 -5
- objaverse_xl/__init__.py +0 -1
- objaverse_xl/abstract.py +0 -97
- objaverse_xl/github.py +0 -597
- objaverse_xl/objaverse_v1.py +0 -537
- objaverse_xl/objaverse_xl_downloader.py +0 -136
- objaverse_xl/smithsonian.py +0 -316
- objaverse_xl/thingiverse.py +0 -380
- objaverse_xl/utils.py +0 -45
- requirements-test.txt +0 -8
- requirements.txt +0 -7
- scripts/rendering/README.md +0 -149
- scripts/rendering/blender_script.py +0 -899
- scripts/rendering/example-objects.json +0 -52
- scripts/rendering/main.py +0 -414
- scripts/rendering/start_x_server.py +0 -272
- setup.py +0 -27
- objaverse_v1/object-metadata.parquet → sketchfab/sketchfab.parquet +2 -2
- scripts/rendering/io_scene_usdz.zip → smithsonian/smithsonian.parquet +2 -2
- tests/test_api.py +0 -553
- thingiverse/thingiverse-objects.parquet +0 -3
- smithsonian/object-metadata.parquet → thingiverse/thingiverse.parquet +2 -2
.github/workflows/ci.yaml
DELETED
@@ -1,51 +0,0 @@
|
|
1 |
-
name: Continuous integration
|
2 |
-
|
3 |
-
on:
|
4 |
-
push:
|
5 |
-
branches:
|
6 |
-
- main
|
7 |
-
pull_request:
|
8 |
-
branches:
|
9 |
-
- main
|
10 |
-
|
11 |
-
jobs:
|
12 |
-
lint:
|
13 |
-
runs-on: ubuntu-latest
|
14 |
-
steps:
|
15 |
-
- uses: actions/checkout@v2
|
16 |
-
- name: Set up Python 3.8
|
17 |
-
uses: actions/setup-python@v2
|
18 |
-
with:
|
19 |
-
python-version: 3.8
|
20 |
-
- name: Install
|
21 |
-
run: |
|
22 |
-
python3 -m venv .env
|
23 |
-
source .env/bin/activate
|
24 |
-
python -m pip install -U pip
|
25 |
-
make install-dev
|
26 |
-
- name: Lint
|
27 |
-
run: |
|
28 |
-
source .env/bin/activate
|
29 |
-
make lint
|
30 |
-
tests:
|
31 |
-
runs-on: ubuntu-latest
|
32 |
-
strategy:
|
33 |
-
matrix:
|
34 |
-
python-version: [3.8]
|
35 |
-
|
36 |
-
steps:
|
37 |
-
- uses: actions/checkout@v2
|
38 |
-
- name: Set up Python ${{ matrix.python-version }}
|
39 |
-
uses: actions/setup-python@v2
|
40 |
-
with:
|
41 |
-
python-version: ${{ matrix.python-version }}
|
42 |
-
- name: Install
|
43 |
-
run: |
|
44 |
-
python3 -m venv .env
|
45 |
-
source .env/bin/activate
|
46 |
-
make install
|
47 |
-
make install-dev
|
48 |
-
- name: Unit tests
|
49 |
-
run: |
|
50 |
-
source .env/bin/activate
|
51 |
-
make test
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
.isort.cfg
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
[settings]
|
2 |
-
profile = black
|
3 |
-
known_third_party = pandas,numpy # Modify with your actual third party libraries
|
|
|
|
|
|
|
|
.pylintrc
DELETED
@@ -1,331 +0,0 @@
|
|
1 |
-
[MASTER]
|
2 |
-
|
3 |
-
# Specify a configuration file.
|
4 |
-
#rcfile=
|
5 |
-
|
6 |
-
# Python code to execute, usually for sys.path manipulation such as
|
7 |
-
# pygtk.require().
|
8 |
-
#init-hook=
|
9 |
-
|
10 |
-
# Profiled execution.
|
11 |
-
profile=no
|
12 |
-
|
13 |
-
# Add files or directories to the blacklist. They should be base names, not
|
14 |
-
# paths.
|
15 |
-
ignore=CVS
|
16 |
-
|
17 |
-
# Pickle collected data for later comparisons.
|
18 |
-
persistent=yes
|
19 |
-
|
20 |
-
# List of plugins (as comma separated values of python modules names) to load,
|
21 |
-
# usually to register additional checkers.
|
22 |
-
load-plugins=
|
23 |
-
|
24 |
-
|
25 |
-
[MESSAGES CONTROL]
|
26 |
-
|
27 |
-
# Enable the message, report, category or checker with the given id(s). You can
|
28 |
-
# either give multiple identifier separated by comma (,) or put this option
|
29 |
-
# multiple time. See also the "--disable" option for examples.
|
30 |
-
enable=indexing-exception,old-raise-syntax
|
31 |
-
|
32 |
-
# Disable the message, report, category or checker with the given id(s). You
|
33 |
-
# can either give multiple identifiers separated by comma (,) or put this
|
34 |
-
# option multiple times (only on the command line, not in the configuration
|
35 |
-
# file where it should appear only once).You can also use "--disable=all" to
|
36 |
-
# disable everything first and then reenable specific checks. For example, if
|
37 |
-
# you want to run only the similarities checker, you can use "--disable=all
|
38 |
-
# --enable=similarities". If you want to run only the classes checker, but have
|
39 |
-
# no Warning level messages displayed, use"--disable=all --enable=classes
|
40 |
-
# --disable=W"
|
41 |
-
disable=design,similarities,no-self-use,attribute-defined-outside-init,locally-disabled,star-args,pointless-except,bad-option-value,global-statement,fixme,suppressed-message,useless-suppression,locally-enabled,no-member,no-name-in-module,import-error,unsubscriptable-object,unbalanced-tuple-unpacking,undefined-variable,not-context-manager,no-else-return,wrong-import-order,unnecessary-pass,logging-fstring-interpolation,logging-format-interpolation,C0330
|
42 |
-
|
43 |
-
|
44 |
-
# Set the cache size for astng objects.
|
45 |
-
cache-size=500
|
46 |
-
|
47 |
-
|
48 |
-
[REPORTS]
|
49 |
-
|
50 |
-
# Set the output format. Available formats are text, parseable, colorized, msvs
|
51 |
-
# (visual studio) and html. You can also give a reporter class, eg
|
52 |
-
# mypackage.mymodule.MyReporterClass.
|
53 |
-
output-format=text
|
54 |
-
|
55 |
-
# Put messages in a separate file for each module / package specified on the
|
56 |
-
# command line instead of printing them on stdout. Reports (if any) will be
|
57 |
-
# written in a file name "pylint_global.[txt|html]".
|
58 |
-
files-output=no
|
59 |
-
|
60 |
-
# Tells whether to display a full report or only the messages
|
61 |
-
reports=no
|
62 |
-
|
63 |
-
# Python expression which should return a note less than 10 (10 is the highest
|
64 |
-
# note). You have access to the variables errors warning, statement which
|
65 |
-
# respectively contain the number of errors / warnings messages and the total
|
66 |
-
# number of statements analyzed. This is used by the global evaluation report
|
67 |
-
# (RP0004).
|
68 |
-
evaluation=10.0 - ((float(5 * error + warning + refactor + convention) / statement) * 10)
|
69 |
-
|
70 |
-
# Add a comment according to your evaluation note. This is used by the global
|
71 |
-
# evaluation report (RP0004).
|
72 |
-
comment=no
|
73 |
-
|
74 |
-
# Template used to display messages. This is a python new-style format string
|
75 |
-
# used to format the message information. See doc for all details
|
76 |
-
#msg-template=
|
77 |
-
|
78 |
-
|
79 |
-
[TYPECHECK]
|
80 |
-
|
81 |
-
# Tells whether missing members accessed in mixin class should be ignored. A
|
82 |
-
# mixin class is detected if its name ends with "mixin" (case insensitive).
|
83 |
-
ignore-mixin-members=yes
|
84 |
-
|
85 |
-
# List of classes names for which member attributes should not be checked
|
86 |
-
# (useful for classes with attributes dynamically set).
|
87 |
-
ignored-classes=SQLObject
|
88 |
-
|
89 |
-
# When zope mode is activated, add a predefined set of Zope acquired attributes
|
90 |
-
# to generated-members.
|
91 |
-
zope=no
|
92 |
-
|
93 |
-
# List of members which are set dynamically and missed by pylint inference
|
94 |
-
# system, and so shouldn't trigger E0201 when accessed. Python regular
|
95 |
-
# expressions are accepted.
|
96 |
-
generated-members=REQUEST,acl_users,aq_parent
|
97 |
-
|
98 |
-
# List of decorators that create context managers from functions, such as
|
99 |
-
# contextlib.contextmanager.
|
100 |
-
contextmanager-decorators=contextlib.contextmanager,contextlib2.contextmanager
|
101 |
-
|
102 |
-
|
103 |
-
[VARIABLES]
|
104 |
-
|
105 |
-
# Tells whether we should check for unused import in __init__ files.
|
106 |
-
init-import=no
|
107 |
-
|
108 |
-
# A regular expression matching the beginning of the name of dummy variables
|
109 |
-
# (i.e. not used).
|
110 |
-
dummy-variables-rgx=^\*{0,2}(_$|unused_|dummy_)
|
111 |
-
|
112 |
-
# List of additional names supposed to be defined in builtins. Remember that
|
113 |
-
# you should avoid to define new builtins when possible.
|
114 |
-
additional-builtins=
|
115 |
-
|
116 |
-
|
117 |
-
[BASIC]
|
118 |
-
|
119 |
-
# Required attributes for module, separated by a comma
|
120 |
-
required-attributes=
|
121 |
-
|
122 |
-
# List of builtins function names that should not be used, separated by a comma
|
123 |
-
bad-functions=apply,input,reduce
|
124 |
-
|
125 |
-
|
126 |
-
# Disable the report(s) with the given id(s).
|
127 |
-
# All non-Google reports are disabled by default.
|
128 |
-
disable-report=R0001,R0002,R0003,R0004,R0101,R0102,R0201,R0202,R0220,R0401,R0402,R0701,R0801,R0901,R0902,R0903,R0904,R0911,R0912,R0913,R0914,R0915,R0921,R0922,R0923
|
129 |
-
|
130 |
-
# Regular expression which should only match correct module names
|
131 |
-
module-rgx=(([a-z_][a-z0-9_]*)|([A-Z][a-zA-Z0-9]+))$
|
132 |
-
|
133 |
-
# Regular expression which should only match correct module level names
|
134 |
-
const-rgx=^(_?[A-Z][A-Z0-9_]*|__[a-z0-9_]+__|_?[a-z][a-z0-9_]*)$
|
135 |
-
|
136 |
-
# Regular expression which should only match correct class names
|
137 |
-
class-rgx=^_?[A-Z][a-zA-Z0-9]*$
|
138 |
-
|
139 |
-
# Regular expression which should only match correct function names
|
140 |
-
function-rgx=^(?:(?P<camel_case>_?[A-Z][a-zA-Z0-9]*)|(?P<snake_case>_?[a-z][a-z0-9_]*))$
|
141 |
-
|
142 |
-
# Regular expression which should only match correct method names
|
143 |
-
method-rgx=^(?:(?P<exempt>__[a-z0-9_]+__|next)|(?P<camel_case>_{0,2}[A-Z][a-zA-Z0-9]*)|(?P<snake_case>_{0,2}[a-z][a-z0-9_]*))$
|
144 |
-
|
145 |
-
# Regular expression which should only match correct instance attribute names
|
146 |
-
attr-rgx=^_{0,2}[a-z][a-z0-9_]*$
|
147 |
-
|
148 |
-
# Regular expression which should only match correct argument names
|
149 |
-
argument-rgx=^[a-z][a-z0-9_]*$
|
150 |
-
|
151 |
-
# Regular expression which should only match correct variable names
|
152 |
-
variable-rgx=^[a-z][a-z0-9_]*$
|
153 |
-
|
154 |
-
# Regular expression which should only match correct attribute names in class
|
155 |
-
# bodies
|
156 |
-
class-attribute-rgx=^(_?[A-Z][A-Z0-9_]*|__[a-z0-9_]+__|_?[a-z][a-z0-9_]*)$
|
157 |
-
|
158 |
-
# Regular expression which should only match correct list comprehension /
|
159 |
-
# generator expression variable names
|
160 |
-
inlinevar-rgx=^[a-z][a-z0-9_]*$
|
161 |
-
|
162 |
-
# Good variable names which should always be accepted, separated by a comma
|
163 |
-
good-names=main,_
|
164 |
-
|
165 |
-
# Bad variable names which should always be refused, separated by a comma
|
166 |
-
bad-names=
|
167 |
-
|
168 |
-
# Regular expression which should only match function or class names that do
|
169 |
-
# not require a docstring.
|
170 |
-
no-docstring-rgx=(__.*__|main)
|
171 |
-
|
172 |
-
# Minimum line length for functions/classes that require docstrings, shorter
|
173 |
-
# ones are exempt.
|
174 |
-
docstring-min-length=10
|
175 |
-
|
176 |
-
|
177 |
-
[FORMAT]
|
178 |
-
|
179 |
-
# Maximum number of characters on a single line.
|
180 |
-
max-line-length=120
|
181 |
-
|
182 |
-
# Regexp for a line that is allowed to be longer than the limit.
|
183 |
-
ignore-long-lines=(?x)
|
184 |
-
(^\s*(import|from)\s
|
185 |
-
|\$Id:\s\/\/depot\/.+#\d+\s\$
|
186 |
-
|^[a-zA-Z_][a-zA-Z0-9_]*\s*=\s*("[^"]\S+"|'[^']\S+')
|
187 |
-
|^\s*\#\ LINT\.ThenChange
|
188 |
-
|^[^#]*\#\ type:\ [a-zA-Z_][a-zA-Z0-9_.,[\] ]*$
|
189 |
-
|pylint
|
190 |
-
|"""
|
191 |
-
|\#
|
192 |
-
|lambda
|
193 |
-
|(https?|ftp):)
|
194 |
-
|
195 |
-
# Allow the body of an if to be on the same line as the test if there is no
|
196 |
-
# else.
|
197 |
-
single-line-if-stmt=y
|
198 |
-
|
199 |
-
# List of optional constructs for which whitespace checking is disabled
|
200 |
-
no-space-check=
|
201 |
-
|
202 |
-
# Maximum number of lines in a module
|
203 |
-
max-module-lines=99999
|
204 |
-
|
205 |
-
# String used as indentation unit. This is usually " " (4 spaces) or "\t" (1
|
206 |
-
# tab).
|
207 |
-
indent-string=' '
|
208 |
-
|
209 |
-
|
210 |
-
[SIMILARITIES]
|
211 |
-
|
212 |
-
# Minimum lines number of a similarity.
|
213 |
-
min-similarity-lines=4
|
214 |
-
|
215 |
-
# Ignore comments when computing similarities.
|
216 |
-
ignore-comments=yes
|
217 |
-
|
218 |
-
# Ignore docstrings when computing similarities.
|
219 |
-
ignore-docstrings=yes
|
220 |
-
|
221 |
-
# Ignore imports when computing similarities.
|
222 |
-
ignore-imports=no
|
223 |
-
|
224 |
-
|
225 |
-
[MISCELLANEOUS]
|
226 |
-
|
227 |
-
# List of note tags to take in consideration, separated by a comma.
|
228 |
-
notes=
|
229 |
-
|
230 |
-
|
231 |
-
[IMPORTS]
|
232 |
-
|
233 |
-
# Deprecated modules which should not be used, separated by a comma
|
234 |
-
deprecated-modules=regsub,TERMIOS,Bastion,rexec,sets
|
235 |
-
|
236 |
-
# Create a graph of every (i.e. internal and external) dependencies in the
|
237 |
-
# given file (report RP0402 must not be disabled)
|
238 |
-
import-graph=
|
239 |
-
|
240 |
-
# Create a graph of external dependencies in the given file (report RP0402 must
|
241 |
-
# not be disabled)
|
242 |
-
ext-import-graph=
|
243 |
-
|
244 |
-
# Create a graph of internal dependencies in the given file (report RP0402 must
|
245 |
-
# not be disabled)
|
246 |
-
int-import-graph=
|
247 |
-
|
248 |
-
extension-pkg-whitelist=_jsonnet
|
249 |
-
|
250 |
-
|
251 |
-
[CLASSES]
|
252 |
-
|
253 |
-
# List of interface methods to ignore, separated by a comma. This is used for
|
254 |
-
# instance to not check methods defines in Zope's Interface base class.
|
255 |
-
ignore-iface-methods=isImplementedBy,deferred,extends,names,namesAndDescriptions,queryDescriptionFor,getBases,getDescriptionFor,getDoc,getName,getTaggedValue,getTaggedValueTags,isEqualOrExtendedBy,setTaggedValue,isImplementedByInstancesOf,adaptWith,is_implemented_by
|
256 |
-
|
257 |
-
# List of method names used to declare (i.e. assign) instance attributes.
|
258 |
-
defining-attr-methods=__init__,__new__,setUp
|
259 |
-
|
260 |
-
# List of valid names for the first argument in a class method.
|
261 |
-
valid-classmethod-first-arg=cls,class_
|
262 |
-
|
263 |
-
# List of valid names for the first argument in a metaclass class method.
|
264 |
-
valid-metaclass-classmethod-first-arg=mcs
|
265 |
-
|
266 |
-
|
267 |
-
[DESIGN]
|
268 |
-
|
269 |
-
# Maximum number of arguments for function / method
|
270 |
-
max-args=5
|
271 |
-
|
272 |
-
# Argument names that match this expression will be ignored. Default to name
|
273 |
-
# with leading underscore
|
274 |
-
ignored-argument-names=_.*
|
275 |
-
|
276 |
-
# Maximum number of locals for function / method body
|
277 |
-
max-locals=15
|
278 |
-
|
279 |
-
# Maximum number of return / yield for function / method body
|
280 |
-
max-returns=6
|
281 |
-
|
282 |
-
# Maximum number of branch for function / method body
|
283 |
-
max-branches=12
|
284 |
-
|
285 |
-
# Maximum number of statements in function / method body
|
286 |
-
max-statements=50
|
287 |
-
|
288 |
-
# Maximum number of parents for a class (see R0901).
|
289 |
-
max-parents=7
|
290 |
-
|
291 |
-
# Maximum number of attributes for a class (see R0902).
|
292 |
-
max-attributes=7
|
293 |
-
|
294 |
-
# Minimum number of public methods for a class (see R0903).
|
295 |
-
min-public-methods=2
|
296 |
-
|
297 |
-
# Maximum number of public methods for a class (see R0904).
|
298 |
-
max-public-methods=20
|
299 |
-
|
300 |
-
|
301 |
-
[EXCEPTIONS]
|
302 |
-
|
303 |
-
# Exceptions that will emit a warning when being caught. Defaults to
|
304 |
-
# "Exception"
|
305 |
-
overgeneral-exceptions=Exception,StandardError,BaseException
|
306 |
-
|
307 |
-
|
308 |
-
[AST]
|
309 |
-
|
310 |
-
# Maximum line length for lambdas
|
311 |
-
short-func-length=1
|
312 |
-
|
313 |
-
# List of module members that should be marked as deprecated.
|
314 |
-
# All of the string functions are listed in 4.1.4 Deprecated string functions
|
315 |
-
# in the Python 2.4 docs.
|
316 |
-
deprecated-members=string.atof,string.atoi,string.atol,string.capitalize,string.expandtabs,string.find,string.rfind,string.index,string.rindex,string.count,string.lower,string.split,string.rsplit,string.splitfields,string.join,string.joinfields,string.lstrip,string.rstrip,string.strip,string.swapcase,string.translate,string.upper,string.ljust,string.rjust,string.center,string.zfill,string.replace,sys.exitfunc
|
317 |
-
|
318 |
-
|
319 |
-
[DOCSTRING]
|
320 |
-
|
321 |
-
# List of exceptions that do not need to be mentioned in the Raises section of
|
322 |
-
# a docstring.
|
323 |
-
ignore-exceptions=AssertionError,NotImplementedError,StopIteration,TypeError
|
324 |
-
|
325 |
-
|
326 |
-
|
327 |
-
[TOKENS]
|
328 |
-
|
329 |
-
# Number of spaces of indent required when the last token on the preceding line
|
330 |
-
# is an open (, [, or {.
|
331 |
-
indent-after-paren=4
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Makefile
DELETED
@@ -1,28 +0,0 @@
|
|
1 |
-
install: ## [Local development] Upgrade pip, install requirements, install package.
|
2 |
-
python -m pip install -U pip
|
3 |
-
python -m pip install -e .
|
4 |
-
|
5 |
-
install-dev: ## [Local development] Install test requirements
|
6 |
-
python -m pip install -r requirements-test.txt
|
7 |
-
|
8 |
-
lint: ## [Local development] Run mypy, pylint and black
|
9 |
-
python -m mypy objaverse_xl
|
10 |
-
python -m pylint objaverse_xl
|
11 |
-
python -m black --check objaverse_xl
|
12 |
-
python -m isort --check-only objaverse_xl
|
13 |
-
python -m black --check scripts --exclude scripts/rendering/blender-3.2.2-linux-x64/
|
14 |
-
python -m isort --check-only scripts/**/*.py --skip scripts/rendering/blender-3.2.2-linux-x64/
|
15 |
-
|
16 |
-
format: ## [Local development] Auto-format python code using black, don't include blender
|
17 |
-
python -m isort objaverse_xl
|
18 |
-
python -m black objaverse_xl
|
19 |
-
python -m isort scripts/**/*.py --skip scripts/rendering/blender-3.2.2-linux-x64/
|
20 |
-
python -m black scripts --exclude scripts/rendering/blender-3.2.2-linux-x64/
|
21 |
-
|
22 |
-
test: ## [Local development] Run unit tests
|
23 |
-
JUPYTER_PLATFORM_DIRS=1 python -m pytest -x -s -v tests
|
24 |
-
|
25 |
-
.PHONY: help
|
26 |
-
|
27 |
-
help: # Run `make help` to get help on the make commands
|
28 |
-
@grep -E '^[0-9a-zA-Z_-]+:.*?## .*$$' $(MAKEFILE_LIST) | sort | awk 'BEGIN {FS = ":.*?## "}; {printf "\033[36m%-30s\033[0m %s\n", $$1, $$2}'
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
github/{github-urls.parquet → github.parquet}
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f14bf4daceac97072eee5721489ad160089b9fc1514e6a9e096574c3e4276da6
|
3 |
+
size 797466051
|
mypy.ini
DELETED
@@ -1,5 +0,0 @@
|
|
1 |
-
# Global options:
|
2 |
-
|
3 |
-
[mypy]
|
4 |
-
python_version = 3.8
|
5 |
-
ignore_missing_imports = True
|
|
|
|
|
|
|
|
|
|
|
|
objaverse_xl/__init__.py
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
"""A package for downloading and processing Objaverse-XL."""
|
|
|
|
objaverse_xl/abstract.py
DELETED
@@ -1,97 +0,0 @@
|
|
1 |
-
"""Abstract class for Objaverse-XL sources."""
|
2 |
-
|
3 |
-
from abc import ABC, abstractmethod
|
4 |
-
from typing import Callable, Dict, Optional
|
5 |
-
|
6 |
-
import pandas as pd
|
7 |
-
|
8 |
-
|
9 |
-
class ObjaverseSource(ABC):
|
10 |
-
"""Abstract class for Objaverse-XL sources."""
|
11 |
-
|
12 |
-
@abstractmethod
|
13 |
-
def get_annotations(self, download_dir: str = "~/.objaverse") -> pd.DataFrame:
|
14 |
-
"""Loads the 3D object metadata as a Pandas DataFrame.
|
15 |
-
|
16 |
-
Args:
|
17 |
-
download_dir (str, optional): Directory to download the parquet metadata
|
18 |
-
file. Supports all file systems supported by fsspec. Defaults to
|
19 |
-
"~/.objaverse".
|
20 |
-
|
21 |
-
Returns:
|
22 |
-
pd.DataFrame: Metadata of the 3D objects as a Pandas DataFrame with columns
|
23 |
-
for the object "fileIdentifier", "license", "source", "fileType",
|
24 |
-
"sha256", and "metadata".
|
25 |
-
"""
|
26 |
-
|
27 |
-
@abstractmethod
|
28 |
-
def download_objects(
|
29 |
-
self,
|
30 |
-
objects: pd.DataFrame,
|
31 |
-
download_dir: str = "~/.objaverse",
|
32 |
-
processes: Optional[int] = None,
|
33 |
-
handle_found_object: Optional[Callable] = None,
|
34 |
-
handle_modified_object: Optional[Callable] = None,
|
35 |
-
handle_missing_object: Optional[Callable] = None,
|
36 |
-
**kwargs
|
37 |
-
) -> Dict[str, str]:
|
38 |
-
"""Downloads all objects from the source.
|
39 |
-
|
40 |
-
Args:
|
41 |
-
objects (pd.DataFrame): Objects to download. Must have columns for
|
42 |
-
the object "fileIdentifier" and "sha256". Use the `get_annotations`
|
43 |
-
function to get the metadata.
|
44 |
-
processes (Optional[int], optional): Number of processes to use for
|
45 |
-
downloading. If None, will use the number of CPUs on the machine.
|
46 |
-
Defaults to None.
|
47 |
-
download_dir (str, optional): Directory to download the objects to.
|
48 |
-
Supports all file systems supported by fsspec. Defaults to
|
49 |
-
"~/.objaverse".
|
50 |
-
save_repo_format (Optional[Literal["zip", "tar", "tar.gz", "files"]],
|
51 |
-
optional): Format to save the repository. If None, the repository will
|
52 |
-
not be saved. If "files" is specified, each file will be saved
|
53 |
-
individually. Otherwise, the repository can be saved as a "zip", "tar",
|
54 |
-
or "tar.gz" file. Defaults to None.
|
55 |
-
handle_found_object (Optional[Callable], optional): Called when an object is
|
56 |
-
successfully found and downloaded. Here, the object has the same sha256
|
57 |
-
as the one that was downloaded with Objaverse-XL. If None, the object
|
58 |
-
will be downloaded, but nothing will be done with it. Args for the
|
59 |
-
function include:
|
60 |
-
- local_path (str): Local path to the downloaded 3D object.
|
61 |
-
- file_identifier (str): File identifier of the 3D object.
|
62 |
-
- sha256 (str): SHA256 of the contents of the 3D object.
|
63 |
-
- metadata (Dict[Hashable, Any]): Metadata about the 3D object,
|
64 |
-
including the GitHub organization and repo names.
|
65 |
-
Return is not used. Defaults to None.
|
66 |
-
handle_modified_object (Optional[Callable], optional): Called when a
|
67 |
-
modified object is found and downloaded. Here, the object is
|
68 |
-
successfully downloaded, but it has a different sha256 than the one that
|
69 |
-
was downloaded with Objaverse-XL. This is not expected to happen very
|
70 |
-
often, because the same commit hash is used for each repo. If None, the
|
71 |
-
object will be downloaded, but nothing will be done with it. Args for
|
72 |
-
the function include:
|
73 |
-
- local_path (str): Local path to the downloaded 3D object.
|
74 |
-
- file_identifier (str): File identifier of the 3D object.
|
75 |
-
- new_sha256 (str): SHA256 of the contents of the newly downloaded 3D
|
76 |
-
object.
|
77 |
-
- old_sha256 (str): Expected SHA256 of the contents of the 3D object as
|
78 |
-
it was when it was downloaded with Objaverse-XL.
|
79 |
-
- metadata (Dict[Hashable, Any]): Metadata about the 3D object, which is
|
80 |
-
particular to the souce.
|
81 |
-
Return is not used. Defaults to None.
|
82 |
-
handle_missing_object (Optional[Callable], optional): Called when an object
|
83 |
-
that is in Objaverse-XL is not found. Here, it is likely that the
|
84 |
-
repository was deleted or renamed. If None, nothing will be done with
|
85 |
-
the missing object.
|
86 |
-
Args for the function include:
|
87 |
-
- file_identifier (str): File identifier of the 3D object.
|
88 |
-
- sha256 (str): SHA256 of the contents of the original 3D object.
|
89 |
-
- metadata (Dict[Hashable, Any]): Metadata about the 3D object, which is
|
90 |
-
particular to the source.
|
91 |
-
Return is not used. Defaults to None.
|
92 |
-
|
93 |
-
Returns:
|
94 |
-
Dict[str, str]: Mapping of file identifiers to local paths of the downloaded
|
95 |
-
3D objects.
|
96 |
-
"""
|
97 |
-
pass
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
objaverse_xl/github.py
DELETED
@@ -1,597 +0,0 @@
|
|
1 |
-
"""Script to download objects from GitHub."""
|
2 |
-
|
3 |
-
import json
|
4 |
-
import multiprocessing
|
5 |
-
import os
|
6 |
-
import shutil
|
7 |
-
import subprocess
|
8 |
-
import tarfile
|
9 |
-
import tempfile
|
10 |
-
from multiprocessing import Pool
|
11 |
-
from typing import Callable, Dict, List, Literal, Optional
|
12 |
-
|
13 |
-
import fsspec
|
14 |
-
import pandas as pd
|
15 |
-
import requests
|
16 |
-
from loguru import logger
|
17 |
-
from tqdm import tqdm
|
18 |
-
|
19 |
-
from objaverse_xl.abstract import ObjaverseSource
|
20 |
-
from objaverse_xl.utils import get_file_hash
|
21 |
-
|
22 |
-
FILE_EXTENSIONS = [
|
23 |
-
".obj",
|
24 |
-
".glb",
|
25 |
-
".gltf",
|
26 |
-
".usdz",
|
27 |
-
".usd",
|
28 |
-
".fbx",
|
29 |
-
".stl",
|
30 |
-
".usda",
|
31 |
-
".dae",
|
32 |
-
".ply",
|
33 |
-
".abc",
|
34 |
-
".blend",
|
35 |
-
]
|
36 |
-
|
37 |
-
|
38 |
-
class GitHubDownloader(ObjaverseSource):
|
39 |
-
"""Script to download objects from GitHub."""
|
40 |
-
|
41 |
-
def get_annotations(self, download_dir: str = "~/.objaverse") -> pd.DataFrame:
|
42 |
-
"""Loads the GitHub 3D object metadata as a Pandas DataFrame.
|
43 |
-
|
44 |
-
Args:
|
45 |
-
download_dir (str, optional): Directory to download the parquet metadata
|
46 |
-
file. Supports all file systems supported by fsspec. Defaults to
|
47 |
-
"~/.objaverse".
|
48 |
-
|
49 |
-
Returns:
|
50 |
-
pd.DataFrame: GitHub 3D object metadata as a Pandas DataFrame with columns
|
51 |
-
for the object "fileIdentifier", "license", "source", "fileType",
|
52 |
-
"sha256", and "metadata".
|
53 |
-
"""
|
54 |
-
filename = os.path.join(download_dir, "github", "github-urls.parquet")
|
55 |
-
fs, path = fsspec.core.url_to_fs(filename)
|
56 |
-
fs.makedirs(os.path.dirname(path), exist_ok=True)
|
57 |
-
|
58 |
-
# download the parquet file if it doesn't exist
|
59 |
-
if not fs.exists(path):
|
60 |
-
url = "https://huggingface.co/datasets/allenai/objaverse-xl/resolve/main/github/github-urls.parquet"
|
61 |
-
|
62 |
-
response = requests.get(url)
|
63 |
-
response.raise_for_status()
|
64 |
-
with fs.open(path, "wb") as file:
|
65 |
-
file.write(response.content)
|
66 |
-
|
67 |
-
# load the parquet file with fsspec
|
68 |
-
with fs.open(path) as f:
|
69 |
-
df = pd.read_parquet(f)
|
70 |
-
|
71 |
-
df["metadata"] = "{}"
|
72 |
-
|
73 |
-
return df
|
74 |
-
|
75 |
-
def _get_repo_id_with_hash(self, item: pd.Series) -> str:
|
76 |
-
org, repo = item["fileIdentifier"].split("/")[3:5]
|
77 |
-
commit_hash = item["fileIdentifier"].split("/")[6]
|
78 |
-
return f"{org}/{repo}/{commit_hash}"
|
79 |
-
|
80 |
-
def _git_shallow_clone(self, repo_url: str, target_directory: str) -> bool:
|
81 |
-
"""Helper function to shallow clone a repo with git.
|
82 |
-
|
83 |
-
Args:
|
84 |
-
repo_url (str): URL of the repo to clone.
|
85 |
-
target_directory (str): Directory to clone the repo to.
|
86 |
-
|
87 |
-
Returns:
|
88 |
-
bool: True if the clone was successful, False otherwise.
|
89 |
-
"""
|
90 |
-
return self._run_command_with_check(
|
91 |
-
["git", "clone", "--depth", "1", repo_url, target_directory],
|
92 |
-
)
|
93 |
-
|
94 |
-
def _run_command_with_check(
|
95 |
-
self, command: List[str], cwd: Optional[str] = None
|
96 |
-
) -> bool:
|
97 |
-
"""Helper function to run a command and check if it was successful.
|
98 |
-
|
99 |
-
Args:
|
100 |
-
command (List[str]): Command to run.
|
101 |
-
cwd (Optional[str], optional): Current working directory to run the command
|
102 |
-
in. Defaults to None.
|
103 |
-
|
104 |
-
Returns:
|
105 |
-
bool: True if the command was successful, False otherwise.
|
106 |
-
"""
|
107 |
-
try:
|
108 |
-
subprocess.run(
|
109 |
-
command,
|
110 |
-
cwd=cwd,
|
111 |
-
check=True,
|
112 |
-
stdout=subprocess.DEVNULL,
|
113 |
-
stderr=subprocess.DEVNULL,
|
114 |
-
)
|
115 |
-
return True
|
116 |
-
except subprocess.CalledProcessError as e:
|
117 |
-
logger.error("Error:", e)
|
118 |
-
logger.error(e.stdout)
|
119 |
-
logger.error(e.stderr)
|
120 |
-
return False
|
121 |
-
|
122 |
-
def _process_repo(
|
123 |
-
self,
|
124 |
-
repo_id: str,
|
125 |
-
fs: fsspec.AbstractFileSystem,
|
126 |
-
base_dir: str,
|
127 |
-
save_repo_format: Optional[Literal["zip", "tar", "tar.gz", "files"]],
|
128 |
-
expected_objects: Dict[str, str],
|
129 |
-
handle_found_object: Optional[Callable],
|
130 |
-
handle_modified_object: Optional[Callable],
|
131 |
-
handle_missing_object: Optional[Callable],
|
132 |
-
handle_new_object: Optional[Callable],
|
133 |
-
commit_hash: Optional[str],
|
134 |
-
) -> Dict[str, str]:
|
135 |
-
"""Process a single repo.
|
136 |
-
|
137 |
-
Args:
|
138 |
-
repo_id (str): GitHub repo ID in the format of organization/repo.
|
139 |
-
fs (fsspec.AbstractFileSystem): File system to use for saving the repo.
|
140 |
-
base_dir (str): Base directory to save the repo to.
|
141 |
-
expected_objects (Dict[str, str]): Dictionary of objects that one expects to
|
142 |
-
find in the repo. Keys are the "fileIdentifier" (i.e., the GitHub URL in
|
143 |
-
this case) and values are the "sha256" of the objects.
|
144 |
-
{and the rest of the args are the same as download_objects}
|
145 |
-
|
146 |
-
Returns:
|
147 |
-
Dict[str, str]: A dictionary that maps from the "fileIdentifier" to the path
|
148 |
-
of the downloaded object.
|
149 |
-
"""
|
150 |
-
# NOTE: assuming that the user has already checked that the repo doesn't exist,
|
151 |
-
org, repo = repo_id.split("/")
|
152 |
-
|
153 |
-
out = {}
|
154 |
-
with tempfile.TemporaryDirectory() as temp_dir:
|
155 |
-
# clone the repo to a temp directory
|
156 |
-
target_directory = os.path.join(temp_dir, repo)
|
157 |
-
successful_clone = self._git_shallow_clone(
|
158 |
-
f"https://github.com/{org}/{repo}.git", target_directory
|
159 |
-
)
|
160 |
-
if not successful_clone:
|
161 |
-
logger.error(f"Could not clone {repo_id}")
|
162 |
-
if handle_missing_object is not None:
|
163 |
-
for github_url, sha256 in expected_objects.items():
|
164 |
-
handle_missing_object(
|
165 |
-
file_identifier=github_url,
|
166 |
-
sha256=sha256,
|
167 |
-
metadata=dict(github_organization=org, github_repo=repo),
|
168 |
-
)
|
169 |
-
return {}
|
170 |
-
|
171 |
-
# use the commit hash if specified
|
172 |
-
repo_commit_hash = self._get_commit_hash_from_local_git_dir(
|
173 |
-
target_directory
|
174 |
-
)
|
175 |
-
if commit_hash is not None:
|
176 |
-
keep_going = True
|
177 |
-
if repo_commit_hash != commit_hash:
|
178 |
-
# run git reset --hard && git checkout 37f4d8d287e201ce52c048bf74d46d6a09d26b2c
|
179 |
-
if not self._run_command_with_check(
|
180 |
-
["git", "fetch", "origin", commit_hash], target_directory
|
181 |
-
):
|
182 |
-
logger.error(
|
183 |
-
f"Error in git fetch! Sticking with {repo_commit_hash=} instead of {commit_hash=}"
|
184 |
-
)
|
185 |
-
keep_going = False
|
186 |
-
|
187 |
-
if keep_going and not self._run_command_with_check(
|
188 |
-
["git", "reset", "--hard"], target_directory
|
189 |
-
):
|
190 |
-
logger.error(
|
191 |
-
f"Error in git reset! Sticking with {repo_commit_hash=} instead of {commit_hash=}"
|
192 |
-
)
|
193 |
-
keep_going = False
|
194 |
-
|
195 |
-
if keep_going:
|
196 |
-
if self._run_command_with_check(
|
197 |
-
["git", "checkout", commit_hash], target_directory
|
198 |
-
):
|
199 |
-
repo_commit_hash = commit_hash
|
200 |
-
else:
|
201 |
-
logger.error(
|
202 |
-
f"Error in git checkout! Sticking with {repo_commit_hash=} instead of {commit_hash=}"
|
203 |
-
)
|
204 |
-
|
205 |
-
# pull the lfs files
|
206 |
-
self._pull_lfs_files(target_directory)
|
207 |
-
|
208 |
-
# get all the files in the repo
|
209 |
-
files = self._list_files(target_directory)
|
210 |
-
files_with_3d_extension = [
|
211 |
-
file
|
212 |
-
for file in files
|
213 |
-
if any(file.lower().endswith(ext) for ext in FILE_EXTENSIONS)
|
214 |
-
]
|
215 |
-
|
216 |
-
# get the sha256 for each file
|
217 |
-
file_hashes = []
|
218 |
-
for file in tqdm(files_with_3d_extension, desc="Handling 3D object files"):
|
219 |
-
file_hash = get_file_hash(file)
|
220 |
-
# remove the temp_dir from the file path
|
221 |
-
github_url = file.replace(
|
222 |
-
target_directory,
|
223 |
-
f"https://github.com/{org}/{repo}/blob/{repo_commit_hash}",
|
224 |
-
)
|
225 |
-
file_hashes.append(dict(sha256=file_hash, fileIdentifier=github_url))
|
226 |
-
|
227 |
-
# handle the object under different conditions
|
228 |
-
if github_url in expected_objects:
|
229 |
-
out[github_url] = file[len(target_directory) + 1 :]
|
230 |
-
if expected_objects[github_url] == file_hash:
|
231 |
-
if handle_found_object is not None:
|
232 |
-
handle_found_object(
|
233 |
-
local_path=file,
|
234 |
-
file_identifier=github_url,
|
235 |
-
sha256=file_hash,
|
236 |
-
metadata=dict(
|
237 |
-
github_organization=org, github_repo=repo
|
238 |
-
),
|
239 |
-
)
|
240 |
-
else:
|
241 |
-
if handle_modified_object is not None:
|
242 |
-
handle_modified_object(
|
243 |
-
local_path=file,
|
244 |
-
file_identifier=github_url,
|
245 |
-
new_sha256=file_hash,
|
246 |
-
old_sha256=expected_objects[github_url],
|
247 |
-
metadata=dict(
|
248 |
-
github_organization=org, github_repo=repo
|
249 |
-
),
|
250 |
-
)
|
251 |
-
elif handle_new_object is not None:
|
252 |
-
handle_new_object(
|
253 |
-
local_path=file,
|
254 |
-
file_identifier=github_url,
|
255 |
-
sha256=file_hash,
|
256 |
-
metadata=dict(github_organization=org, github_repo=repo),
|
257 |
-
)
|
258 |
-
|
259 |
-
# save the file hashes to a json file
|
260 |
-
with open(
|
261 |
-
os.path.join(target_directory, ".objaverse-file-hashes.json"),
|
262 |
-
"w",
|
263 |
-
encoding="utf-8",
|
264 |
-
) as f:
|
265 |
-
json.dump(file_hashes, f, indent=2)
|
266 |
-
|
267 |
-
# remove the .git directory
|
268 |
-
shutil.rmtree(os.path.join(target_directory, ".git"))
|
269 |
-
|
270 |
-
if save_repo_format is None:
|
271 |
-
# remove the paths, since it's not downloaded
|
272 |
-
out = {}
|
273 |
-
else:
|
274 |
-
logger.debug(f"Saving {org}/{repo} as {save_repo_format}")
|
275 |
-
# save the repo to a zip file
|
276 |
-
if save_repo_format == "zip":
|
277 |
-
shutil.make_archive(target_directory, "zip", target_directory)
|
278 |
-
elif save_repo_format == "tar":
|
279 |
-
with tarfile.open(
|
280 |
-
os.path.join(temp_dir, f"{repo}.tar"), "w"
|
281 |
-
) as tar:
|
282 |
-
tar.add(target_directory, arcname=repo)
|
283 |
-
elif save_repo_format == "tar.gz":
|
284 |
-
with tarfile.open(
|
285 |
-
os.path.join(temp_dir, f"{repo}.tar.gz"), "w:gz"
|
286 |
-
) as tar:
|
287 |
-
tar.add(target_directory, arcname=repo)
|
288 |
-
elif save_repo_format == "files":
|
289 |
-
pass
|
290 |
-
else:
|
291 |
-
raise ValueError(
|
292 |
-
f"save_repo_format must be one of zip, tar, tar.gz, files. Got {save_repo_format}"
|
293 |
-
)
|
294 |
-
|
295 |
-
dirname = os.path.join(base_dir, "repos", org)
|
296 |
-
fs.makedirs(dirname, exist_ok=True)
|
297 |
-
if save_repo_format != "files":
|
298 |
-
# move the repo to the correct location (with put)
|
299 |
-
fs.put(
|
300 |
-
os.path.join(temp_dir, f"{repo}.{save_repo_format}"),
|
301 |
-
os.path.join(dirname, f"{repo}.{save_repo_format}"),
|
302 |
-
)
|
303 |
-
|
304 |
-
for file_identifier in out.copy():
|
305 |
-
out[file_identifier] = os.path.join(
|
306 |
-
dirname, f"{repo}.{save_repo_format}", out[file_identifier]
|
307 |
-
)
|
308 |
-
else:
|
309 |
-
# move the repo to the correct location (with put)
|
310 |
-
fs.put(target_directory, dirname, recursive=True)
|
311 |
-
|
312 |
-
for file_identifier in out.copy():
|
313 |
-
out[file_identifier] = os.path.join(
|
314 |
-
dirname, repo, out[file_identifier]
|
315 |
-
)
|
316 |
-
|
317 |
-
# get each object that was missing from the expected objects
|
318 |
-
if handle_missing_object is not None:
|
319 |
-
obtained_urls = {x["fileIdentifier"] for x in file_hashes}
|
320 |
-
for github_url, sha256 in expected_objects.items():
|
321 |
-
if github_url not in obtained_urls:
|
322 |
-
handle_missing_object(
|
323 |
-
file_identifier=github_url,
|
324 |
-
sha256=sha256,
|
325 |
-
metadata=dict(github_organization=org, github_repo=repo),
|
326 |
-
)
|
327 |
-
|
328 |
-
return out
|
329 |
-
|
330 |
-
def _list_files(self, root_dir: str) -> List[str]:
|
331 |
-
return [
|
332 |
-
os.path.join(root, f)
|
333 |
-
for root, dirs, files in os.walk(root_dir)
|
334 |
-
for f in files
|
335 |
-
]
|
336 |
-
|
337 |
-
def _pull_lfs_files(self, repo_dir: str) -> None:
|
338 |
-
if self._has_lfs_files(repo_dir):
|
339 |
-
subprocess.run(["git", "lfs", "pull"], cwd=repo_dir, check=True)
|
340 |
-
|
341 |
-
def _has_lfs_files(self, repo_dir: str) -> bool:
|
342 |
-
gitattributes_path = os.path.join(repo_dir, ".gitattributes")
|
343 |
-
if not os.path.exists(gitattributes_path):
|
344 |
-
return False
|
345 |
-
with open(gitattributes_path, "r", encoding="utf-8") as f:
|
346 |
-
for line in f:
|
347 |
-
if "filter=lfs" in line:
|
348 |
-
return True
|
349 |
-
return False
|
350 |
-
|
351 |
-
def _get_commit_hash_from_local_git_dir(self, local_git_dir: str) -> str:
|
352 |
-
# get the git hash of the repo
|
353 |
-
result = subprocess.run(
|
354 |
-
["git", "rev-parse", "HEAD"],
|
355 |
-
cwd=local_git_dir,
|
356 |
-
capture_output=True,
|
357 |
-
check=True,
|
358 |
-
)
|
359 |
-
commit_hash = result.stdout.strip().decode("utf-8")
|
360 |
-
return commit_hash
|
361 |
-
|
362 |
-
def _parallel_process_repo(self, args) -> Dict[str, str]:
|
363 |
-
"""Helper function to process a repo in parallel.
|
364 |
-
|
365 |
-
Note: This function is used to parallelize the processing of repos. It is not
|
366 |
-
intended to be called directly.
|
367 |
-
|
368 |
-
Args:
|
369 |
-
args (Tuple): Tuple of arguments to pass to _process_repo.
|
370 |
-
|
371 |
-
Returns:
|
372 |
-
Dict[str, str]: A dictionary that maps from the "fileIdentifier" to the path
|
373 |
-
of the downloaded object.
|
374 |
-
"""
|
375 |
-
|
376 |
-
(
|
377 |
-
repo_id_hash,
|
378 |
-
fs,
|
379 |
-
base_dir,
|
380 |
-
save_repo_format,
|
381 |
-
expected_objects,
|
382 |
-
handle_found_object,
|
383 |
-
handle_modified_object,
|
384 |
-
handle_missing_object,
|
385 |
-
handle_new_object,
|
386 |
-
) = args
|
387 |
-
repo_id = "/".join(repo_id_hash.split("/")[:2])
|
388 |
-
commit_hash = repo_id_hash.split("/")[2]
|
389 |
-
return self._process_repo(
|
390 |
-
repo_id=repo_id,
|
391 |
-
fs=fs,
|
392 |
-
base_dir=base_dir,
|
393 |
-
save_repo_format=save_repo_format,
|
394 |
-
expected_objects=expected_objects,
|
395 |
-
handle_found_object=handle_found_object,
|
396 |
-
handle_modified_object=handle_modified_object,
|
397 |
-
handle_missing_object=handle_missing_object,
|
398 |
-
handle_new_object=handle_new_object,
|
399 |
-
commit_hash=commit_hash,
|
400 |
-
)
|
401 |
-
|
402 |
-
def _process_group(self, group):
|
403 |
-
key, group_df = group
|
404 |
-
return key, group_df.set_index("fileIdentifier")["sha256"].to_dict()
|
405 |
-
|
406 |
-
def download_objects(
|
407 |
-
self,
|
408 |
-
objects: pd.DataFrame,
|
409 |
-
download_dir: Optional[str] = "~/.objaverse",
|
410 |
-
processes: Optional[int] = None,
|
411 |
-
handle_found_object: Optional[Callable] = None,
|
412 |
-
handle_modified_object: Optional[Callable] = None,
|
413 |
-
handle_missing_object: Optional[Callable] = None,
|
414 |
-
**kwargs,
|
415 |
-
) -> Dict[str, str]:
|
416 |
-
"""Download the specified GitHub objects.
|
417 |
-
|
418 |
-
Args:
|
419 |
-
objects (pd.DataFrame): GitHub objects to download. Must have columns for
|
420 |
-
the object "fileIdentifier" and "sha256". Use the `get_annotations`
|
421 |
-
function to get the metadata.
|
422 |
-
download_dir (Optional[str], optional): Directory to download the GitHub
|
423 |
-
objects to. Supports all file systems supported by fsspec. If None, the
|
424 |
-
repository will not be saved (note that save_repo_format must also be
|
425 |
-
None in this case, otherwise a ValueError is raised). Defaults to
|
426 |
-
"~/.objaverse".
|
427 |
-
processes (Optional[int], optional): Number of processes to use for
|
428 |
-
downloading. If None, will use the number of CPUs on the machine.
|
429 |
-
Defaults to None.
|
430 |
-
handle_found_object (Optional[Callable], optional): Called when an object is
|
431 |
-
successfully found and downloaded. Here, the object has the same sha256
|
432 |
-
as the one that was downloaded with Objaverse-XL. If None, the object
|
433 |
-
will be downloaded, but nothing will be done with it. Args for the
|
434 |
-
function include:
|
435 |
-
- local_path (str): Local path to the downloaded 3D object.
|
436 |
-
- file_identifier (str): GitHub URL of the 3D object.
|
437 |
-
- sha256 (str): SHA256 of the contents of the 3D object.
|
438 |
-
- metadata (Dict[str, Any]): Metadata about the 3D object, including the
|
439 |
-
GitHub organization and repo names.
|
440 |
-
Return is not used. Defaults to None.
|
441 |
-
handle_modified_object (Optional[Callable], optional): Called when a
|
442 |
-
modified object is found and downloaded. Here, the object is
|
443 |
-
successfully downloaded, but it has a different sha256 than the one that
|
444 |
-
was downloaded with Objaverse-XL. This is not expected to happen very
|
445 |
-
often, because the same commit hash is used for each repo. If None, the
|
446 |
-
object will be downloaded, but nothing will be done with it. Args for
|
447 |
-
the function include:
|
448 |
-
- local_path (str): Local path to the downloaded 3D object.
|
449 |
-
- file_identifier (str): GitHub URL of the 3D object.
|
450 |
-
- new_sha256 (str): SHA256 of the contents of the newly downloaded 3D
|
451 |
-
object.
|
452 |
-
- old_sha256 (str): Expected SHA256 of the contents of the 3D object as
|
453 |
-
it was when it was downloaded with Objaverse-XL.
|
454 |
-
- metadata (Dict[str, Any]): Metadata about the 3D object, including the
|
455 |
-
GitHub organization and repo names.
|
456 |
-
Return is not used. Defaults to None.
|
457 |
-
handle_missing_object (Optional[Callable], optional): Called when an object
|
458 |
-
that is in Objaverse-XL is not found. Here, it is likely that the
|
459 |
-
repository was deleted or renamed. If None, nothing will be done with
|
460 |
-
the missing object. Args for the function include:
|
461 |
-
- file_identifier (str): GitHub URL of the 3D object.
|
462 |
-
- sha256 (str): SHA256 of the contents of the original 3D object.
|
463 |
-
- metadata (Dict[str, Any]): Metadata about the 3D object, including the
|
464 |
-
GitHub organization and repo names.
|
465 |
-
Return is not used. Defaults to None.
|
466 |
-
save_repo_format (Optional[Literal["zip", "tar", "tar.gz", "files"]],
|
467 |
-
optional): Format to save the repository. If None, the repository will
|
468 |
-
not be saved. If "files" is specified, each file will be saved
|
469 |
-
individually. Otherwise, the repository can be saved as a "zip", "tar",
|
470 |
-
or "tar.gz" file. Defaults to None.
|
471 |
-
handle_new_object (Optional[Callable], optional): Called when a new object
|
472 |
-
is found. Here, the object is not used in Objaverse-XL, but is still
|
473 |
-
downloaded with the repository. The object may have not been used
|
474 |
-
because it does not successfully import into Blender. If None, the
|
475 |
-
object will be downloaded, but nothing will be done with it. Args for
|
476 |
-
the function include:
|
477 |
-
- local_path (str): Local path to the downloaded 3D object.
|
478 |
-
- file_identifier (str): GitHub URL of the 3D object.
|
479 |
-
- sha256 (str): SHA256 of the contents of the 3D object.
|
480 |
-
- metadata (Dict[str, Any]): Metadata about the 3D object, including the
|
481 |
-
GitHub organization and repo names.
|
482 |
-
Return is not used. Defaults to None.
|
483 |
-
|
484 |
-
Raises:
|
485 |
-
ValueError: If download_dir is None and save_repo_format is not None.
|
486 |
-
Otherwise, we don't know where to save the repo!
|
487 |
-
|
488 |
-
Returns:
|
489 |
-
Dict[str, str]: A dictionary that maps from the "fileIdentifier" to the path
|
490 |
-
of the downloaded object.
|
491 |
-
"""
|
492 |
-
save_repo_format = kwargs.get("save_repo_format", None)
|
493 |
-
handle_new_object = kwargs.get("handle_new_object", None)
|
494 |
-
|
495 |
-
if processes is None:
|
496 |
-
processes = multiprocessing.cpu_count()
|
497 |
-
if download_dir is None:
|
498 |
-
if save_repo_format is not None:
|
499 |
-
raise ValueError(
|
500 |
-
f"If {save_repo_format=} is not None, {download_dir=} must be specified."
|
501 |
-
)
|
502 |
-
# path doesn't matter if we're not saving the repo
|
503 |
-
download_dir = "~/.objaverse"
|
504 |
-
|
505 |
-
base_download_dir = os.path.join(download_dir, "github")
|
506 |
-
fs, path = fsspec.core.url_to_fs(base_download_dir)
|
507 |
-
fs.makedirs(path, exist_ok=True)
|
508 |
-
|
509 |
-
# Getting immediate subdirectories of root_path
|
510 |
-
if save_repo_format == "files":
|
511 |
-
downloaded_repo_dirs = fs.glob(base_download_dir + "/repos/*/*/")
|
512 |
-
downloaded_repo_ids = {
|
513 |
-
"/".join(x.split("/")[-2:]) for x in downloaded_repo_dirs
|
514 |
-
}
|
515 |
-
else:
|
516 |
-
downloaded_repo_dirs = fs.glob(
|
517 |
-
base_download_dir + f"/repos/*/*.{save_repo_format}"
|
518 |
-
)
|
519 |
-
downloaded_repo_ids = set()
|
520 |
-
for x in downloaded_repo_dirs:
|
521 |
-
org, repo = x.split("/")[-2:]
|
522 |
-
repo = repo[: -len(f".{save_repo_format}")]
|
523 |
-
repo_id = f"{org}/{repo}"
|
524 |
-
downloaded_repo_ids.add(repo_id)
|
525 |
-
|
526 |
-
# make copy of objects
|
527 |
-
objects = objects.copy()
|
528 |
-
|
529 |
-
# get the unique repoIds
|
530 |
-
objects["repoIdHash"] = objects.apply(self._get_repo_id_with_hash, axis=1)
|
531 |
-
repo_id_hashes = set(objects["repoIdHash"].unique().tolist())
|
532 |
-
repo_ids = {
|
533 |
-
"/".join(repo_id_hash.split("/")[:2]) for repo_id_hash in repo_id_hashes
|
534 |
-
}
|
535 |
-
assert len(repo_id_hashes) == len(repo_ids), (
|
536 |
-
f"More than 1 commit hash per repoId!"
|
537 |
-
f" {len(repo_id_hashes)=}, {len(repo_ids)=}"
|
538 |
-
)
|
539 |
-
|
540 |
-
logger.info(
|
541 |
-
f"Provided {len(repo_ids)} repoIds with {len(objects)} objects to process."
|
542 |
-
)
|
543 |
-
|
544 |
-
# remove repoIds that have already been downloaded
|
545 |
-
repo_ids_to_download = repo_ids - downloaded_repo_ids
|
546 |
-
repo_id_hashes_to_download = [
|
547 |
-
repo_id_hash
|
548 |
-
for repo_id_hash in repo_id_hashes
|
549 |
-
if "/".join(repo_id_hash.split("/")[:2]) in repo_ids_to_download
|
550 |
-
]
|
551 |
-
|
552 |
-
logger.info(
|
553 |
-
f"Found {len(repo_ids_to_download)} repoIds not yet downloaded. Downloading now..."
|
554 |
-
)
|
555 |
-
|
556 |
-
# get the objects to download
|
557 |
-
groups = list(objects.groupby("repoIdHash"))
|
558 |
-
with Pool(processes=processes) as pool:
|
559 |
-
out_list = list(
|
560 |
-
tqdm(
|
561 |
-
pool.imap_unordered(self._process_group, groups),
|
562 |
-
total=len(groups),
|
563 |
-
desc="Grouping objects by repository",
|
564 |
-
)
|
565 |
-
)
|
566 |
-
objects_per_repo_id_hash = dict(out_list)
|
567 |
-
|
568 |
-
all_args = [
|
569 |
-
(
|
570 |
-
repo_id_hash,
|
571 |
-
fs,
|
572 |
-
path,
|
573 |
-
save_repo_format,
|
574 |
-
objects_per_repo_id_hash[repo_id_hash],
|
575 |
-
handle_found_object,
|
576 |
-
handle_missing_object,
|
577 |
-
handle_modified_object,
|
578 |
-
handle_new_object,
|
579 |
-
)
|
580 |
-
for repo_id_hash in repo_id_hashes_to_download
|
581 |
-
]
|
582 |
-
|
583 |
-
with Pool(processes=processes) as pool:
|
584 |
-
# use tqdm to show progress
|
585 |
-
out = list(
|
586 |
-
tqdm(
|
587 |
-
pool.imap_unordered(self._parallel_process_repo, all_args),
|
588 |
-
total=len(all_args),
|
589 |
-
desc="Downloading repositories",
|
590 |
-
)
|
591 |
-
)
|
592 |
-
|
593 |
-
out_dict = {}
|
594 |
-
for x in out:
|
595 |
-
out_dict.update(x)
|
596 |
-
|
597 |
-
return out_dict
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
objaverse_xl/objaverse_v1.py
DELETED
@@ -1,537 +0,0 @@
|
|
1 |
-
"""Script to download objects from Objaverse 1.0."""
|
2 |
-
|
3 |
-
import gzip
|
4 |
-
import json
|
5 |
-
import multiprocessing
|
6 |
-
import os
|
7 |
-
import tempfile
|
8 |
-
import urllib.request
|
9 |
-
from multiprocessing import Pool
|
10 |
-
from typing import Any, Callable, Dict, List, Optional, Tuple
|
11 |
-
|
12 |
-
import fsspec
|
13 |
-
import pandas as pd
|
14 |
-
import requests
|
15 |
-
from loguru import logger
|
16 |
-
from tqdm import tqdm
|
17 |
-
|
18 |
-
from objaverse_xl.abstract import ObjaverseSource
|
19 |
-
from objaverse_xl.utils import get_file_hash
|
20 |
-
|
21 |
-
|
22 |
-
class SketchfabDownloader(ObjaverseSource):
|
23 |
-
"""A class for downloading and processing Objaverse 1.0."""
|
24 |
-
|
25 |
-
def get_annotations(self, download_dir: str = "~/.objaverse") -> pd.DataFrame:
|
26 |
-
"""Load the annotations from the given directory.
|
27 |
-
|
28 |
-
Args:
|
29 |
-
download_dir (str, optional): The directory to load the annotations from.
|
30 |
-
Supports all file systems supported by fsspec. Defaults to
|
31 |
-
"~/.objaverse".
|
32 |
-
|
33 |
-
Returns:
|
34 |
-
pd.DataFrame: The annotations, which includes the columns "thingId", "fileId",
|
35 |
-
"filename", and "license".
|
36 |
-
"""
|
37 |
-
remote_url = "https://huggingface.co/datasets/allenai/objaverse-xl/resolve/main/objaverse_v1/object-metadata.parquet"
|
38 |
-
download_path = os.path.join(
|
39 |
-
download_dir, "hf-objaverse-v1", "thingiverse-objects.parquet"
|
40 |
-
)
|
41 |
-
fs, path = fsspec.core.url_to_fs(download_path)
|
42 |
-
|
43 |
-
if not fs.exists(path):
|
44 |
-
fs.makedirs(os.path.dirname(path), exist_ok=True)
|
45 |
-
logger.info(f"Downloading {remote_url} to {download_path}")
|
46 |
-
response = requests.get(remote_url)
|
47 |
-
response.raise_for_status()
|
48 |
-
with fs.open(path, "wb") as file:
|
49 |
-
file.write(response.content)
|
50 |
-
|
51 |
-
# read the file with pandas and fsspec
|
52 |
-
with fs.open(download_path, "rb") as f:
|
53 |
-
annotations_df = pd.read_parquet(f)
|
54 |
-
|
55 |
-
annotations_df["metadata"] = "{}"
|
56 |
-
|
57 |
-
return annotations_df
|
58 |
-
|
59 |
-
def load_full_annotations(
|
60 |
-
self,
|
61 |
-
uids: Optional[List[str]] = None,
|
62 |
-
download_dir: str = "~/.objaverse",
|
63 |
-
) -> Dict[str, Any]:
|
64 |
-
"""Load the full metadata of all objects in the dataset.
|
65 |
-
|
66 |
-
Args:
|
67 |
-
uids: A list of uids with which to load metadata. If None, it loads
|
68 |
-
the metadata for all uids.
|
69 |
-
download_dir: The base directory to download the annotations to. Supports all
|
70 |
-
file systems supported by fsspec. Defaults to "~/.objaverse".
|
71 |
-
|
72 |
-
Returns:
|
73 |
-
A dictionary of the metadata for each object. The keys are the uids and the
|
74 |
-
values are the metadata for that object.
|
75 |
-
"""
|
76 |
-
# make the metadata dir if it doesn't exist
|
77 |
-
metadata_path = os.path.join(download_dir, "hf-objaverse-v1", "metadata")
|
78 |
-
fs, _ = fsspec.core.url_to_fs(metadata_path)
|
79 |
-
fs.makedirs(metadata_path, exist_ok=True)
|
80 |
-
|
81 |
-
# get the dir ids that need to be loaded if only downloading a subset of uids
|
82 |
-
object_paths = self._load_object_paths(download_dir=download_dir)
|
83 |
-
dir_ids = (
|
84 |
-
{object_paths[uid].split("/")[1] for uid in uids}
|
85 |
-
if uids is not None
|
86 |
-
else {f"{i // 1000:03d}-{i % 1000:03d}" for i in range(160)}
|
87 |
-
)
|
88 |
-
|
89 |
-
# get the existing metadata files
|
90 |
-
existing_metadata_files = fs.glob(
|
91 |
-
os.path.join(metadata_path, "*.json.gz"), refresh=True
|
92 |
-
)
|
93 |
-
existing_dir_ids = {
|
94 |
-
file.split("/")[-1].split(".")[0]
|
95 |
-
for file in existing_metadata_files
|
96 |
-
if file.endswith(".json.gz") # note partial files end with .json.gz.tmp
|
97 |
-
}
|
98 |
-
downloaded_dir_ids = existing_dir_ids.intersection(dir_ids)
|
99 |
-
logger.info(
|
100 |
-
f"Found {len(downloaded_dir_ids)} metadata files already downloaded"
|
101 |
-
)
|
102 |
-
|
103 |
-
# download the metadata from the missing dir_ids
|
104 |
-
dir_ids_to_download = dir_ids - existing_dir_ids
|
105 |
-
logger.info(f"Downloading {len(dir_ids_to_download)} metadata files")
|
106 |
-
|
107 |
-
# download the metadata file if it doesn't exist
|
108 |
-
if len(dir_ids_to_download) > 0:
|
109 |
-
for i_id in tqdm(dir_ids_to_download, desc="Downloading metadata files"):
|
110 |
-
# get the path to the json file
|
111 |
-
path = os.path.join(metadata_path, f"{i_id}.json.gz")
|
112 |
-
|
113 |
-
# get the url to the remote json file
|
114 |
-
hf_url = f"https://huggingface.co/datasets/allenai/objaverse/resolve/main/metadata/{i_id}.json.gz"
|
115 |
-
|
116 |
-
# download the file to a tmp path to avoid partial downloads on interruption
|
117 |
-
tmp_path = f"{path}.tmp"
|
118 |
-
with fs.open(tmp_path, "wb") as f:
|
119 |
-
with urllib.request.urlopen(hf_url) as response:
|
120 |
-
f.write(response.read())
|
121 |
-
fs.rename(tmp_path, path)
|
122 |
-
|
123 |
-
out = {}
|
124 |
-
for i_id in tqdm(dir_ids, desc="Reading metadata files"):
|
125 |
-
# get the path to the json file
|
126 |
-
path = os.path.join(metadata_path, f"{i_id}.json.gz")
|
127 |
-
|
128 |
-
# read the json file of the metadata chunk
|
129 |
-
with fs.open(path, "rb") as f:
|
130 |
-
with gzip.GzipFile(fileobj=f) as gfile:
|
131 |
-
content = gfile.read()
|
132 |
-
data = json.loads(content)
|
133 |
-
|
134 |
-
# filter the data to only include the uids we want
|
135 |
-
if uids is not None:
|
136 |
-
data = {uid: data[uid] for uid in uids if uid in data}
|
137 |
-
|
138 |
-
# add the data to the out dict
|
139 |
-
out.update(data)
|
140 |
-
|
141 |
-
return out
|
142 |
-
|
143 |
-
def _load_object_paths(self, download_dir: str) -> Dict[str, str]:
|
144 |
-
"""Load the object paths from the dataset.
|
145 |
-
|
146 |
-
The object paths specify the location of where the object is located in the
|
147 |
-
Hugging Face repo.
|
148 |
-
|
149 |
-
Returns:
|
150 |
-
A dictionary mapping the uid to the object path.
|
151 |
-
"""
|
152 |
-
object_paths_file = "object-paths.json.gz"
|
153 |
-
local_path = os.path.join(download_dir, "hf-objaverse-v1", object_paths_file)
|
154 |
-
|
155 |
-
# download the object_paths file if it doesn't exist
|
156 |
-
fs, path = fsspec.core.url_to_fs(local_path)
|
157 |
-
if not fs.exists(path):
|
158 |
-
hf_url = f"https://huggingface.co/datasets/allenai/objaverse/resolve/main/{object_paths_file}"
|
159 |
-
fs.makedirs(os.path.dirname(path), exist_ok=True)
|
160 |
-
|
161 |
-
# download the file to a tmp path to avoid partial downloads on interruption
|
162 |
-
tmp_path = f"{path}.tmp"
|
163 |
-
with fs.open(tmp_path, "wb") as f:
|
164 |
-
with urllib.request.urlopen(hf_url) as response:
|
165 |
-
f.write(response.read())
|
166 |
-
fs.rename(tmp_path, path)
|
167 |
-
|
168 |
-
# read the object_paths
|
169 |
-
with fs.open(path, "rb") as f:
|
170 |
-
with gzip.GzipFile(fileobj=f) as gfile:
|
171 |
-
content = gfile.read()
|
172 |
-
object_paths = json.loads(content)
|
173 |
-
|
174 |
-
return object_paths
|
175 |
-
|
176 |
-
def load_uids(self, download_dir: str = "~/.objaverse") -> List[str]:
|
177 |
-
"""Load the uids from the dataset.
|
178 |
-
|
179 |
-
Returns:
|
180 |
-
A list of all the UIDs from the dataset.
|
181 |
-
"""
|
182 |
-
return list(self._load_object_paths(download_dir=download_dir).keys())
|
183 |
-
|
184 |
-
def _download_object(
|
185 |
-
self,
|
186 |
-
file_identifier: str,
|
187 |
-
hf_object_path: str,
|
188 |
-
download_dir: Optional[str],
|
189 |
-
expected_sha256: str,
|
190 |
-
handle_found_object: Optional[Callable] = None,
|
191 |
-
handle_modified_object: Optional[Callable] = None,
|
192 |
-
) -> Tuple[str, Optional[str]]:
|
193 |
-
"""Download the object for the given uid.
|
194 |
-
|
195 |
-
Args:
|
196 |
-
file_identifier: The file identifier of the object.
|
197 |
-
hf_object_path: The path to the object in the Hugging Face repo. Here,
|
198 |
-
hf_object_path is the part that comes after "main" in the Hugging Face
|
199 |
-
repo url:
|
200 |
-
https://huggingface.co/datasets/allenai/objaverse/resolve/main/{hf_object_path}
|
201 |
-
download_dir: The base directory to download the object to. Supports all
|
202 |
-
file systems supported by fsspec. Defaults to "~/.objaverse".
|
203 |
-
expected_sha256 (str): The expected SHA256 of the contents of the downloade
|
204 |
-
object.
|
205 |
-
handle_found_object (Optional[Callable]): Called when an object is
|
206 |
-
successfully found and downloaded. Here, the object has the same sha256
|
207 |
-
as the one that was downloaded with Objaverse-XL. If None, the object
|
208 |
-
will be downloaded, but nothing will be done with it. Args for the
|
209 |
-
function include:
|
210 |
-
- local_path (str): Local path to the downloaded 3D object.
|
211 |
-
- file_identifier (str): GitHub URL of the 3D object.
|
212 |
-
- sha256 (str): SHA256 of the contents of the 3D object.
|
213 |
-
- metadata (Dict[str, Any]): Metadata about the 3D object, including the
|
214 |
-
GitHub organization and repo names.
|
215 |
-
Return is not used.
|
216 |
-
handle_modified_object (Optional[Callable]): Called when a modified object
|
217 |
-
is found and downloaded. Here, the object is successfully downloaded,
|
218 |
-
but it has a different sha256 than the one that was downloaded with
|
219 |
-
Objaverse-XL. This is not expected to happen very often, because the
|
220 |
-
same commit hash is used for each repo. If None, the object will be
|
221 |
-
downloaded, but nothing will be done with it. Args for the function
|
222 |
-
include:
|
223 |
-
- local_path (str): Local path to the downloaded 3D object.
|
224 |
-
- file_identifier (str): GitHub URL of the 3D object.
|
225 |
-
- new_sha256 (str): SHA256 of the contents of the newly downloaded 3D
|
226 |
-
object.
|
227 |
-
- old_sha256 (str): Expected SHA256 of the contents of the 3D object as
|
228 |
-
it was when it was downloaded with Objaverse-XL.
|
229 |
-
- metadata (Dict[str, Any]): Metadata about the 3D object, including the
|
230 |
-
GitHub organization and repo names.
|
231 |
-
Return is not used.
|
232 |
-
handle_missing_object (Optional[Callable]): Called when an object that is in
|
233 |
-
Objaverse-XL is not found. Here, it is likely that the repository was
|
234 |
-
deleted or renamed. If None, nothing will be done with the missing
|
235 |
-
object. Args for the function include:
|
236 |
-
- file_identifier (str): GitHub URL of the 3D object.
|
237 |
-
- sha256 (str): SHA256 of the contents of the original 3D object.
|
238 |
-
- metadata (Dict[str, Any]): Metadata about the 3D object, including the
|
239 |
-
GitHub organization and repo names.
|
240 |
-
Return is not used.
|
241 |
-
|
242 |
-
|
243 |
-
Returns:
|
244 |
-
A tuple of the uid and the path to where the downloaded object. If
|
245 |
-
download_dir is None, the path will be None.
|
246 |
-
"""
|
247 |
-
hf_url = f"https://huggingface.co/datasets/allenai/objaverse/resolve/main/{hf_object_path}"
|
248 |
-
|
249 |
-
with tempfile.TemporaryDirectory() as temp_dir:
|
250 |
-
# download the file locally
|
251 |
-
temp_path = os.path.join(temp_dir, hf_object_path)
|
252 |
-
os.makedirs(os.path.dirname(temp_path), exist_ok=True)
|
253 |
-
temp_path_tmp = f"{temp_path}.tmp"
|
254 |
-
with open(temp_path_tmp, "wb") as file:
|
255 |
-
with urllib.request.urlopen(hf_url) as response:
|
256 |
-
file.write(response.read())
|
257 |
-
os.rename(temp_path_tmp, temp_path)
|
258 |
-
|
259 |
-
# get the sha256 of the downloaded file
|
260 |
-
sha256 = get_file_hash(temp_path)
|
261 |
-
|
262 |
-
if sha256 == expected_sha256:
|
263 |
-
if handle_found_object is not None:
|
264 |
-
handle_found_object(
|
265 |
-
local_path=temp_path,
|
266 |
-
file_identifier=file_identifier,
|
267 |
-
sha256=sha256,
|
268 |
-
metadata={},
|
269 |
-
)
|
270 |
-
else:
|
271 |
-
if handle_modified_object is not None:
|
272 |
-
handle_modified_object(
|
273 |
-
local_path=temp_path,
|
274 |
-
file_identifier=file_identifier,
|
275 |
-
new_sha256=sha256,
|
276 |
-
old_sha256=expected_sha256,
|
277 |
-
metadata={},
|
278 |
-
)
|
279 |
-
|
280 |
-
if download_dir is not None:
|
281 |
-
filename = os.path.join(download_dir, "hf-objaverse-v1", hf_object_path)
|
282 |
-
fs, path = fsspec.core.url_to_fs(filename)
|
283 |
-
fs.makedirs(os.path.dirname(path), exist_ok=True)
|
284 |
-
fs.put(temp_path, path)
|
285 |
-
else:
|
286 |
-
path = None
|
287 |
-
|
288 |
-
return file_identifier, path
|
289 |
-
|
290 |
-
def _parallel_download_object(self, args):
|
291 |
-
# workaround since starmap doesn't work well with tqdm
|
292 |
-
return self._download_object(*args)
|
293 |
-
|
294 |
-
def _get_uid(self, item: pd.Series) -> str:
|
295 |
-
file_identifier = item["fileIdentifier"]
|
296 |
-
return file_identifier.split("/")[-1]
|
297 |
-
|
298 |
-
def uid_to_file_identifier(self, uid: str) -> str:
|
299 |
-
"""Convert the uid to the file identifier.
|
300 |
-
|
301 |
-
Args:
|
302 |
-
uid (str): The uid of the object.
|
303 |
-
|
304 |
-
Returns:
|
305 |
-
The file identifier of the object.
|
306 |
-
"""
|
307 |
-
return f"https://sketchfab.com/3d-models/{uid}"
|
308 |
-
|
309 |
-
def file_identifier_to_uid(self, file_identifier: str) -> str:
|
310 |
-
"""Convert the file identifier to the uid.
|
311 |
-
|
312 |
-
Args:
|
313 |
-
file_identifier (str): The file identifier of the object.
|
314 |
-
|
315 |
-
Returns:
|
316 |
-
The uid of the object.
|
317 |
-
"""
|
318 |
-
return file_identifier.split("/")[-1]
|
319 |
-
|
320 |
-
def download_objects(
|
321 |
-
self,
|
322 |
-
objects: pd.DataFrame,
|
323 |
-
download_dir: Optional[str] = "~/.objaverse",
|
324 |
-
processes: Optional[int] = None,
|
325 |
-
handle_found_object: Optional[Callable] = None,
|
326 |
-
handle_modified_object: Optional[Callable] = None,
|
327 |
-
handle_missing_object: Optional[Callable] = None,
|
328 |
-
**kwargs,
|
329 |
-
) -> Dict[str, str]:
|
330 |
-
"""Return the path to the object files for the given uids.
|
331 |
-
|
332 |
-
If the object is not already downloaded, it will be downloaded.
|
333 |
-
|
334 |
-
Args:
|
335 |
-
objects (pd.DataFrame): Objects to download. Must have columns for
|
336 |
-
the object "fileIdentifier" and "sha256". Use the `get_annotations`
|
337 |
-
function to get the metadata.
|
338 |
-
download_dir (Optional[str], optional): The base directory to download the
|
339 |
-
object to. Supports all file systems supported by fsspec. If None, the
|
340 |
-
objects will be removed after downloading. Defaults to "~/.objaverse".
|
341 |
-
processes (Optional[int], optional): The number of processes to use to
|
342 |
-
download the objects. If None, the number of processes will be set to
|
343 |
-
the number of CPUs on the machine (multiprocessing.cpu_count()).
|
344 |
-
Defaults to None.
|
345 |
-
handle_found_object (Optional[Callable], optional): Called when an object is
|
346 |
-
successfully found and downloaded. Here, the object has the same sha256
|
347 |
-
as the one that was downloaded with Objaverse-XL. If None, the object
|
348 |
-
will be downloaded, but nothing will be done with it. Args for the
|
349 |
-
function include:
|
350 |
-
- local_path (str): Local path to the downloaded 3D object.
|
351 |
-
- file_identifier (str): File identifier of the 3D object.
|
352 |
-
- sha256 (str): SHA256 of the contents of the 3D object.
|
353 |
-
- metadata (Dict[Hashable, Any]): Metadata about the 3D object,
|
354 |
-
including the GitHub organization and repo names.
|
355 |
-
Return is not used. Defaults to None.
|
356 |
-
handle_modified_object (Optional[Callable], optional): Called when a
|
357 |
-
modified object is found and downloaded. Here, the object is
|
358 |
-
successfully downloaded, but it has a different sha256 than the one that
|
359 |
-
was downloaded with Objaverse-XL. This is not expected to happen very
|
360 |
-
often, because the same commit hash is used for each repo. If None, the
|
361 |
-
object will be downloaded, but nothing will be done with it. Args for
|
362 |
-
the function include:
|
363 |
-
- local_path (str): Local path to the downloaded 3D object.
|
364 |
-
- file_identifier (str): File identifier of the 3D object.
|
365 |
-
- new_sha256 (str): SHA256 of the contents of the newly downloaded 3D
|
366 |
-
object.
|
367 |
-
- old_sha256 (str): Expected SHA256 of the contents of the 3D object as
|
368 |
-
it was when it was downloaded with Objaverse-XL.
|
369 |
-
- metadata (Dict[Hashable, Any]): Metadata about the 3D object, which is
|
370 |
-
particular to the souce.
|
371 |
-
Return is not used. Defaults to None.
|
372 |
-
handle_missing_object (Optional[Callable], optional): Called when an object
|
373 |
-
that is in Objaverse-XL is not found. Here, it is likely that the
|
374 |
-
repository was deleted or renamed. If None, nothing will be done with
|
375 |
-
the missing object.
|
376 |
-
Args for the function include:
|
377 |
-
- file_identifier (str): File identifier of the 3D object.
|
378 |
-
- sha256 (str): SHA256 of the contents of the original 3D object.
|
379 |
-
- metadata (Dict[Hashable, Any]): Metadata about the 3D object, which is
|
380 |
-
particular to the source.
|
381 |
-
Return is not used. Defaults to None.
|
382 |
-
|
383 |
-
|
384 |
-
Returns:
|
385 |
-
A dictionary mapping the object fileIdentifier to the local path of where
|
386 |
-
the object downloaded.
|
387 |
-
"""
|
388 |
-
hf_object_paths = self._load_object_paths(
|
389 |
-
download_dir=download_dir if download_dir is not None else "~/.objaverse"
|
390 |
-
)
|
391 |
-
if processes is None:
|
392 |
-
processes = multiprocessing.cpu_count()
|
393 |
-
|
394 |
-
# make a copy of the objects so we don't modify the original
|
395 |
-
objects = objects.copy()
|
396 |
-
objects["uid"] = objects.apply(self._get_uid, axis=1)
|
397 |
-
uids_to_sha256 = dict(zip(objects["uid"], objects["sha256"]))
|
398 |
-
uids_set = set(uids_to_sha256.keys())
|
399 |
-
|
400 |
-
# create a new df where the uids are the index
|
401 |
-
objects_uid_index = objects.set_index("uid")
|
402 |
-
|
403 |
-
out = {}
|
404 |
-
objects_to_download = []
|
405 |
-
if download_dir is None:
|
406 |
-
for _, item in objects.iterrows():
|
407 |
-
uid = item["uid"]
|
408 |
-
if uid not in hf_object_paths:
|
409 |
-
logger.error(f"Could not find object with uid {uid}!")
|
410 |
-
if handle_missing_object is not None:
|
411 |
-
handle_missing_object(
|
412 |
-
file_identifier=item["fileIdentifier"],
|
413 |
-
sha256=item["sha256"],
|
414 |
-
metadata={},
|
415 |
-
)
|
416 |
-
continue
|
417 |
-
objects_to_download.append(
|
418 |
-
(item["fileIdentifier"], hf_object_paths[uid], item["sha256"])
|
419 |
-
)
|
420 |
-
else:
|
421 |
-
versioned_dirname = os.path.join(download_dir, "hf-objaverse-v1")
|
422 |
-
fs, path = fsspec.core.url_to_fs(versioned_dirname)
|
423 |
-
|
424 |
-
# Get the existing file paths. This is much faster than calling fs.exists() for each
|
425 |
-
# file. `glob()` is like walk, but returns a list of files instead of the nested
|
426 |
-
# directory structure. glob() is also faster than find() / walk() since it doesn't
|
427 |
-
# need to traverse the entire directory structure.
|
428 |
-
existing_file_paths = fs.glob(
|
429 |
-
os.path.join(path, "glbs", "*", "*.glb"), refresh=True
|
430 |
-
)
|
431 |
-
existing_uids = {
|
432 |
-
file.split("/")[-1].split(".")[0]
|
433 |
-
for file in existing_file_paths
|
434 |
-
if file.endswith(".glb") # note partial files end with .glb.tmp
|
435 |
-
}
|
436 |
-
|
437 |
-
# add the existing downloaded uids to the return dict
|
438 |
-
already_downloaded_uids = uids_set.intersection(existing_uids)
|
439 |
-
for uid in already_downloaded_uids:
|
440 |
-
hf_object_path = hf_object_paths[uid]
|
441 |
-
fs_abs_object_path = os.path.join(versioned_dirname, hf_object_path)
|
442 |
-
out[self.uid_to_file_identifier(uid)] = fs_abs_object_path
|
443 |
-
|
444 |
-
logger.info(
|
445 |
-
f"Found {len(already_downloaded_uids)} objects already downloaded"
|
446 |
-
)
|
447 |
-
|
448 |
-
# get the uids that need to be downloaded
|
449 |
-
remaining_uids = uids_set - existing_uids
|
450 |
-
for uid in remaining_uids:
|
451 |
-
item = objects_uid_index.loc[uid]
|
452 |
-
if uid not in hf_object_paths:
|
453 |
-
logger.error(f"Could not find object with uid {uid}. Skipping it.")
|
454 |
-
if handle_missing_object is not None:
|
455 |
-
handle_missing_object(
|
456 |
-
file_identifier=item["fileIdentifier"],
|
457 |
-
sha256=item["sha256"],
|
458 |
-
metadata={},
|
459 |
-
)
|
460 |
-
continue
|
461 |
-
objects_to_download.append(
|
462 |
-
(item["fileIdentifier"], hf_object_paths[uid], item["sha256"])
|
463 |
-
)
|
464 |
-
|
465 |
-
logger.info(
|
466 |
-
f"Downloading {len(objects_to_download)} new objects across {processes} processes"
|
467 |
-
)
|
468 |
-
|
469 |
-
# check if all objects are already downloaded
|
470 |
-
if len(objects_to_download) == 0:
|
471 |
-
return out
|
472 |
-
|
473 |
-
args = [
|
474 |
-
(
|
475 |
-
file_identifier,
|
476 |
-
hf_object_path,
|
477 |
-
download_dir,
|
478 |
-
sha256,
|
479 |
-
handle_found_object,
|
480 |
-
handle_modified_object,
|
481 |
-
)
|
482 |
-
for file_identifier, hf_object_path, sha256 in objects_to_download
|
483 |
-
]
|
484 |
-
|
485 |
-
# download the objects in parallel
|
486 |
-
with Pool(processes) as pool:
|
487 |
-
new_object_downloads = list(
|
488 |
-
tqdm(
|
489 |
-
pool.imap_unordered(self._parallel_download_object, args),
|
490 |
-
total=len(args),
|
491 |
-
)
|
492 |
-
)
|
493 |
-
|
494 |
-
for file_identifier, local_path in new_object_downloads:
|
495 |
-
out[file_identifier] = local_path
|
496 |
-
|
497 |
-
return out
|
498 |
-
|
499 |
-
def load_lvis_annotations(
|
500 |
-
self,
|
501 |
-
download_dir: str = "~/.objaverse",
|
502 |
-
) -> Dict[str, List[str]]:
|
503 |
-
"""Load the LVIS annotations.
|
504 |
-
|
505 |
-
If the annotations are not already downloaded, they will be downloaded.
|
506 |
-
|
507 |
-
Args:
|
508 |
-
download_dir: The base directory to download the annotations to. Supports all
|
509 |
-
file systems supported by fsspec. Defaults to "~/.objaverse".
|
510 |
-
|
511 |
-
Returns:
|
512 |
-
A dictionary mapping the LVIS category to the list of uids in that category.
|
513 |
-
"""
|
514 |
-
hf_url = "https://huggingface.co/datasets/allenai/objaverse/resolve/main/lvis-annotations.json.gz"
|
515 |
-
|
516 |
-
download_path = os.path.join(
|
517 |
-
download_dir, "hf-objaverse-v1", "lvis-annotations.json.gz"
|
518 |
-
)
|
519 |
-
|
520 |
-
# use fsspec
|
521 |
-
fs, path = fsspec.core.url_to_fs(download_path)
|
522 |
-
if not fs.exists(path):
|
523 |
-
# make dir if it doesn't exist
|
524 |
-
fs.makedirs(os.path.dirname(path), exist_ok=True)
|
525 |
-
|
526 |
-
# download the file
|
527 |
-
with fs.open(path, "wb") as f:
|
528 |
-
with urllib.request.urlopen(hf_url) as response:
|
529 |
-
f.write(response.read())
|
530 |
-
|
531 |
-
# load the gzip file
|
532 |
-
with fs.open(path, "rb") as f:
|
533 |
-
with gzip.GzipFile(fileobj=f) as gfile:
|
534 |
-
content = gfile.read()
|
535 |
-
data = json.loads(content)
|
536 |
-
|
537 |
-
return data
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
objaverse_xl/objaverse_xl_downloader.py
DELETED
@@ -1,136 +0,0 @@
|
|
1 |
-
"""Downloads 3D objects from all Objaverse-XL sources."""
|
2 |
-
|
3 |
-
from typing import Callable, Dict, Optional
|
4 |
-
|
5 |
-
import pandas as pd
|
6 |
-
|
7 |
-
from objaverse_xl.abstract import ObjaverseSource
|
8 |
-
from objaverse_xl.github import GitHubDownloader
|
9 |
-
from objaverse_xl.objaverse_v1 import SketchfabDownloader
|
10 |
-
from objaverse_xl.smithsonian import SmithsonianDownloader
|
11 |
-
from objaverse_xl.thingiverse import ThingiverseDownloader
|
12 |
-
|
13 |
-
|
14 |
-
class ObjaverseXLDownloader(ObjaverseSource):
|
15 |
-
"""Downloads 3D objects from all Objaverse-XL sources."""
|
16 |
-
|
17 |
-
def __init__(self):
|
18 |
-
super().__init__()
|
19 |
-
|
20 |
-
self.downloaders = {
|
21 |
-
"github": GitHubDownloader(),
|
22 |
-
"thingiverse": ThingiverseDownloader(),
|
23 |
-
"smithsonian": SmithsonianDownloader(),
|
24 |
-
"sketchfab": SketchfabDownloader(),
|
25 |
-
}
|
26 |
-
|
27 |
-
def get_annotations(self, download_dir: str = "~/.objaverse") -> pd.DataFrame:
|
28 |
-
"""Loads the 3D object metadata as a Pandas DataFrame.
|
29 |
-
|
30 |
-
Args:
|
31 |
-
download_dir (str, optional): Directory to download the parquet metadata
|
32 |
-
file. Supports all file systems supported by fsspec. Defaults to
|
33 |
-
"~/.objaverse".
|
34 |
-
|
35 |
-
Returns:
|
36 |
-
pd.DataFrame: Metadata of the 3D objects as a Pandas DataFrame with columns
|
37 |
-
for the object "fileIdentifier", "license", "source", "fileType",
|
38 |
-
"sha256", and "metadata".
|
39 |
-
"""
|
40 |
-
annotations = [
|
41 |
-
downloader.get_annotations(download_dir)
|
42 |
-
for downloader in self.downloaders.values()
|
43 |
-
]
|
44 |
-
return pd.concat(annotations, ignore_index=True)
|
45 |
-
|
46 |
-
def download_objects(
|
47 |
-
self,
|
48 |
-
objects: pd.DataFrame,
|
49 |
-
download_dir: str = "~/.objaverse",
|
50 |
-
processes: Optional[int] = None,
|
51 |
-
handle_found_object: Optional[Callable] = None,
|
52 |
-
handle_modified_object: Optional[Callable] = None,
|
53 |
-
handle_missing_object: Optional[Callable] = None,
|
54 |
-
**kwargs,
|
55 |
-
) -> Dict[str, str]:
|
56 |
-
"""Downloads all objects from the source.
|
57 |
-
|
58 |
-
Args:
|
59 |
-
objects (pd.DataFrame): Objects to download. Must have columns for
|
60 |
-
the object "fileIdentifier" and "sha256". Use the `get_annotations`
|
61 |
-
function to get the metadata.
|
62 |
-
download_dir (str, optional): Directory to download the objects to.
|
63 |
-
Supports all file systems supported by fsspec. Defaults to
|
64 |
-
"~/.objaverse".
|
65 |
-
processes (Optional[int], optional): Number of processes to use for
|
66 |
-
downloading. If None, will use the number of CPUs on the machine.
|
67 |
-
Defaults to None.
|
68 |
-
save_repo_format (Optional[Literal["zip", "tar", "tar.gz", "files"]],
|
69 |
-
optional): Format to save the repository. If None, the repository will
|
70 |
-
not be saved. If "files" is specified, each file will be saved
|
71 |
-
individually. Otherwise, the repository can be saved as a "zip", "tar",
|
72 |
-
or "tar.gz" file. Defaults to None.
|
73 |
-
handle_found_object (Optional[Callable], optional): Called when an object is
|
74 |
-
successfully found and downloaded. Here, the object has the same sha256
|
75 |
-
as the one that was downloaded with Objaverse-XL. If None, the object
|
76 |
-
will be downloaded, but nothing will be done with it. Args for the
|
77 |
-
function include:
|
78 |
-
- local_path (str): Local path to the downloaded 3D object.
|
79 |
-
- file_identifier (str): File identifier of the 3D object.
|
80 |
-
- sha256 (str): SHA256 of the contents of the 3D object.
|
81 |
-
- metadata (Dict[Hashable, Any]): Metadata about the 3D object,
|
82 |
-
including the GitHub organization and repo names.
|
83 |
-
Return is not used. Defaults to None.
|
84 |
-
handle_modified_object (Optional[Callable], optional): Called when a
|
85 |
-
modified object is found and downloaded. Here, the object is
|
86 |
-
successfully downloaded, but it has a different sha256 than the one that
|
87 |
-
was downloaded with Objaverse-XL. This is not expected to happen very
|
88 |
-
often, because the same commit hash is used for each repo. If None, the
|
89 |
-
object will be downloaded, but nothing will be done with it. Args for
|
90 |
-
the function include:
|
91 |
-
- local_path (str): Local path to the downloaded 3D object.
|
92 |
-
- file_identifier (str): File identifier of the 3D object.
|
93 |
-
- new_sha256 (str): SHA256 of the contents of the newly downloaded 3D
|
94 |
-
object.
|
95 |
-
- old_sha256 (str): Expected SHA256 of the contents of the 3D object as
|
96 |
-
it was when it was downloaded with Objaverse-XL.
|
97 |
-
- metadata (Dict[Hashable, Any]): Metadata about the 3D object, which is
|
98 |
-
particular to the souce.
|
99 |
-
Return is not used. Defaults to None.
|
100 |
-
handle_missing_object (Optional[Callable], optional): Called when an object
|
101 |
-
that is in Objaverse-XL is not found. Here, it is likely that the
|
102 |
-
repository was deleted or renamed. If None, nothing will be done with
|
103 |
-
the missing object.
|
104 |
-
Args for the function include:
|
105 |
-
- file_identifier (str): File identifier of the 3D object.
|
106 |
-
- sha256 (str): SHA256 of the contents of the original 3D object.
|
107 |
-
- metadata (Dict[Hashable, Any]): Metadata about the 3D object, which is
|
108 |
-
particular to the source.
|
109 |
-
Return is not used. Defaults to None.
|
110 |
-
|
111 |
-
Returns:
|
112 |
-
Dict[str, str]: Mapping of file identifiers to local paths of the downloaded
|
113 |
-
3D objects.
|
114 |
-
"""
|
115 |
-
sources = set(objects["source"].unique().tolist())
|
116 |
-
all_sources = {"github", "thingiverse", "smithsonian", "sketchfab"}
|
117 |
-
|
118 |
-
if not sources.issubset(all_sources):
|
119 |
-
raise ValueError(
|
120 |
-
f"Invalid sources: {sources}. Must be a subset of {all_sources}."
|
121 |
-
)
|
122 |
-
|
123 |
-
downloaded_objects = {}
|
124 |
-
for source in sources:
|
125 |
-
source_downloads = self.downloaders[source].download_objects(
|
126 |
-
objects[objects["source"] == source],
|
127 |
-
download_dir,
|
128 |
-
processes,
|
129 |
-
handle_found_object,
|
130 |
-
handle_modified_object,
|
131 |
-
handle_missing_object,
|
132 |
-
**kwargs,
|
133 |
-
)
|
134 |
-
downloaded_objects.update(source_downloads)
|
135 |
-
|
136 |
-
return downloaded_objects
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
objaverse_xl/smithsonian.py
DELETED
@@ -1,316 +0,0 @@
|
|
1 |
-
"""Script to download 3D objects from the Smithsonian Institution."""
|
2 |
-
|
3 |
-
import multiprocessing
|
4 |
-
import os
|
5 |
-
import tempfile
|
6 |
-
from multiprocessing import Pool
|
7 |
-
from typing import Callable, Dict, Optional, Tuple
|
8 |
-
|
9 |
-
import fsspec
|
10 |
-
import pandas as pd
|
11 |
-
import requests
|
12 |
-
from loguru import logger
|
13 |
-
from tqdm import tqdm
|
14 |
-
|
15 |
-
from objaverse_xl.abstract import ObjaverseSource
|
16 |
-
from objaverse_xl.utils import get_file_hash, get_uid_from_str
|
17 |
-
|
18 |
-
|
19 |
-
class SmithsonianDownloader(ObjaverseSource):
|
20 |
-
"""Script to download objects from the Smithsonian Institute."""
|
21 |
-
|
22 |
-
def get_annotations(self, download_dir: str = "~/.objaverse") -> pd.DataFrame:
|
23 |
-
"""Loads the Smithsonian Object Metadata dataset as a Pandas DataFrame.
|
24 |
-
|
25 |
-
Args:
|
26 |
-
download_dir (str, optional): Directory to download the parquet metadata file.
|
27 |
-
Supports all file systems supported by fsspec. Defaults to "~/.objaverse".
|
28 |
-
|
29 |
-
Returns:
|
30 |
-
pd.DataFrame: Smithsonian Object Metadata dataset as a Pandas DataFrame with
|
31 |
-
columns for the object "title", "url", "quality", "file_type", "uid", and
|
32 |
-
"license". The quality is always Medium and the file_type is always glb.
|
33 |
-
"""
|
34 |
-
filename = os.path.join(download_dir, "smithsonian", "object-metadata.parquet")
|
35 |
-
fs, path = fsspec.core.url_to_fs(filename)
|
36 |
-
fs.makedirs(os.path.dirname(path), exist_ok=True)
|
37 |
-
|
38 |
-
# download the parquet file if it doesn't exist
|
39 |
-
if not fs.exists(path):
|
40 |
-
url = "https://huggingface.co/datasets/allenai/objaverse-xl/resolve/main/smithsonian/object-metadata.parquet"
|
41 |
-
response = requests.get(url)
|
42 |
-
response.raise_for_status()
|
43 |
-
with fs.open(path, "wb") as file:
|
44 |
-
file.write(response.content)
|
45 |
-
|
46 |
-
# load the parquet file with fsspec
|
47 |
-
with fs.open(path) as f:
|
48 |
-
df = pd.read_parquet(f)
|
49 |
-
|
50 |
-
return df
|
51 |
-
|
52 |
-
def _download_smithsonian_object(
|
53 |
-
self,
|
54 |
-
file_identifier: str,
|
55 |
-
download_dir: Optional[str],
|
56 |
-
expected_sha256: str,
|
57 |
-
handle_found_object: Optional[Callable],
|
58 |
-
handle_modified_object: Optional[Callable],
|
59 |
-
handle_missing_object: Optional[Callable],
|
60 |
-
) -> Tuple[str, Optional[str]]:
|
61 |
-
"""Downloads a Smithsonian Object from a URL.
|
62 |
-
|
63 |
-
Overwrites the file if it already exists and assumes this was previous checked.
|
64 |
-
|
65 |
-
Args:
|
66 |
-
file_identifier (str): URL to download the Smithsonian Object from.
|
67 |
-
download_dir (Optional[str]): Directory to download the Smithsonian Object
|
68 |
-
to. Supports all file systems supported by fsspec. If None, the
|
69 |
-
Smithsonian Object will be deleted after it is downloaded and processed
|
70 |
-
with the handler functions.
|
71 |
-
expected_sha256 (str): The expected SHA256 of the contents of the downloaded
|
72 |
-
object.
|
73 |
-
handle_found_object (Optional[Callable]): Called when an object is
|
74 |
-
successfully found and downloaded. Here, the object has the same sha256
|
75 |
-
as the one that was downloaded with Objaverse-XL. If None, the object
|
76 |
-
will be downloaded, but nothing will be done with it. Args for the
|
77 |
-
function include:
|
78 |
-
- local_path (str): Local path to the downloaded 3D object.
|
79 |
-
- file_identifier (str): File identifier of the 3D object.
|
80 |
-
- sha256 (str): SHA256 of the contents of the 3D object.
|
81 |
-
- metadata (Dict[str, Any]): Metadata about the 3D object, including the
|
82 |
-
GitHub organization and repo names.
|
83 |
-
Return is not used.
|
84 |
-
handle_modified_object (Optional[Callable]): Called when a modified object
|
85 |
-
is found and downloaded. Here, the object is successfully downloaded,
|
86 |
-
but it has a different sha256 than the one that was downloaded with
|
87 |
-
Objaverse-XL. This is not expected to happen very often, because the
|
88 |
-
same commit hash is used for each repo. If None, the object will be
|
89 |
-
downloaded, but nothing will be done with it. Args for the function
|
90 |
-
include:
|
91 |
-
- local_path (str): Local path to the downloaded 3D object.
|
92 |
-
- file_identifier (str): File identifier of the 3D object.
|
93 |
-
- new_sha256 (str): SHA256 of the contents of the newly downloaded 3D
|
94 |
-
object.
|
95 |
-
- old_sha256 (str): Expected SHA256 of the contents of the 3D object as
|
96 |
-
it was when it was downloaded with Objaverse-XL.
|
97 |
-
- metadata (Dict[str, Any]): Metadata about the 3D object, including the
|
98 |
-
GitHub organization and repo names.
|
99 |
-
Return is not used.
|
100 |
-
handle_missing_object (Optional[Callable]): Called when an object that is in
|
101 |
-
Objaverse-XL is not found. Here, it is likely that the repository was
|
102 |
-
deleted or renamed. If None, nothing will be done with the missing
|
103 |
-
object. Args for the function include:
|
104 |
-
- file_identifier (str): File identifier of the 3D object.
|
105 |
-
- sha256 (str): SHA256 of the contents of the original 3D object.
|
106 |
-
- metadata (Dict[str, Any]): Metadata about the 3D object, including the
|
107 |
-
GitHub organization and repo names.
|
108 |
-
Return is not used.
|
109 |
-
|
110 |
-
|
111 |
-
Returns:
|
112 |
-
Tuple[str, Optional[str]]: Tuple of the URL and the path to the downloaded
|
113 |
-
Smithsonian Object. If the Smithsonian Object was not downloaded, the path
|
114 |
-
will be None.
|
115 |
-
"""
|
116 |
-
uid = get_uid_from_str(file_identifier)
|
117 |
-
|
118 |
-
with tempfile.TemporaryDirectory() as temp_dir:
|
119 |
-
temp_path = os.path.join(temp_dir, f"{uid}.glb")
|
120 |
-
temp_path_tmp = f"{temp_path}.tmp"
|
121 |
-
|
122 |
-
response = requests.get(file_identifier)
|
123 |
-
|
124 |
-
# check if the path is valid
|
125 |
-
if response.status_code == 404:
|
126 |
-
logger.warning(f"404 for {file_identifier}")
|
127 |
-
if handle_missing_object is not None:
|
128 |
-
handle_missing_object(
|
129 |
-
file_identifier=file_identifier,
|
130 |
-
sha256=expected_sha256,
|
131 |
-
metadata={},
|
132 |
-
)
|
133 |
-
return file_identifier, None
|
134 |
-
|
135 |
-
with open(temp_path_tmp, "wb") as file:
|
136 |
-
for chunk in response.iter_content(chunk_size=8192):
|
137 |
-
file.write(chunk)
|
138 |
-
|
139 |
-
# rename to temp_path
|
140 |
-
os.rename(temp_path_tmp, temp_path)
|
141 |
-
|
142 |
-
# check the sha256
|
143 |
-
sha256 = get_file_hash(temp_path)
|
144 |
-
|
145 |
-
if sha256 == expected_sha256:
|
146 |
-
if handle_found_object is not None:
|
147 |
-
handle_found_object(
|
148 |
-
local_path=temp_path,
|
149 |
-
file_identifier=file_identifier,
|
150 |
-
sha256=sha256,
|
151 |
-
metadata={},
|
152 |
-
)
|
153 |
-
else:
|
154 |
-
if handle_modified_object is not None:
|
155 |
-
handle_modified_object(
|
156 |
-
local_path=temp_path,
|
157 |
-
file_identifier=file_identifier,
|
158 |
-
new_sha256=sha256,
|
159 |
-
old_sha256=expected_sha256,
|
160 |
-
metadata={},
|
161 |
-
)
|
162 |
-
|
163 |
-
if download_dir is not None:
|
164 |
-
filename = os.path.join(
|
165 |
-
download_dir, "smithsonian", "objects", f"{uid}.glb"
|
166 |
-
)
|
167 |
-
fs, path = fsspec.core.url_to_fs(filename)
|
168 |
-
fs.makedirs(os.path.dirname(path), exist_ok=True)
|
169 |
-
fs.put(temp_path, path)
|
170 |
-
else:
|
171 |
-
path = None
|
172 |
-
|
173 |
-
return file_identifier, path
|
174 |
-
|
175 |
-
def _parallel_download_object(self, args):
|
176 |
-
# workaround since starmap doesn't work well with tqdm
|
177 |
-
return self._download_smithsonian_object(*args)
|
178 |
-
|
179 |
-
def download_objects(
|
180 |
-
self,
|
181 |
-
objects: pd.DataFrame,
|
182 |
-
download_dir: Optional[str] = "~/.objaverse",
|
183 |
-
processes: Optional[int] = None,
|
184 |
-
handle_found_object: Optional[Callable] = None,
|
185 |
-
handle_modified_object: Optional[Callable] = None,
|
186 |
-
handle_missing_object: Optional[Callable] = None,
|
187 |
-
**kwargs,
|
188 |
-
) -> Dict[str, str]:
|
189 |
-
"""Downloads all Smithsonian Objects.
|
190 |
-
|
191 |
-
Args:
|
192 |
-
objects (pd.DataFrmae): Objects to download. Must have columns for
|
193 |
-
the object "fileIdentifier" and "sha256". Use the `get_annotations`
|
194 |
-
function to get the metadata.
|
195 |
-
download_dir (Optional[str], optional): Directory to download the
|
196 |
-
Smithsonian Objects to. Supports all file systems supported by fsspec.
|
197 |
-
If None, the Smithsonian Objects will be deleted after they are
|
198 |
-
downloaded and processed with the handler functions. Defaults to
|
199 |
-
"~/.objaverse".
|
200 |
-
processes (Optional[int], optional): Number of processes to use for
|
201 |
-
downloading the Smithsonian Objects. If None, the number of processes
|
202 |
-
will be set to the number of CPUs on the machine
|
203 |
-
(multiprocessing.cpu_count()). Defaults to None.
|
204 |
-
handle_found_object (Optional[Callable], optional): Called when an object is
|
205 |
-
successfully found and downloaded. Here, the object has the same sha256
|
206 |
-
as the one that was downloaded with Objaverse-XL. If None, the object
|
207 |
-
will be downloaded, but nothing will be done with it. Args for the
|
208 |
-
function include:
|
209 |
-
- local_path (str): Local path to the downloaded 3D object.
|
210 |
-
- file_identifier (str): File identifier of the 3D object.
|
211 |
-
- sha256 (str): SHA256 of the contents of the 3D object.
|
212 |
-
- metadata (Dict[Hashable, Any]): Metadata about the 3D object,
|
213 |
-
including the GitHub organization and repo names.
|
214 |
-
Return is not used. Defaults to None.
|
215 |
-
handle_modified_object (Optional[Callable], optional): Called when a
|
216 |
-
modified object is found and downloaded. Here, the object is
|
217 |
-
successfully downloaded, but it has a different sha256 than the one that
|
218 |
-
was downloaded with Objaverse-XL. This is not expected to happen very
|
219 |
-
often, because the same commit hash is used for each repo. If None, the
|
220 |
-
object will be downloaded, but nothing will be done with it. Args for
|
221 |
-
the function include:
|
222 |
-
- local_path (str): Local path to the downloaded 3D object.
|
223 |
-
- file_identifier (str): File identifier of the 3D object.
|
224 |
-
- new_sha256 (str): SHA256 of the contents of the newly downloaded 3D
|
225 |
-
object.
|
226 |
-
- old_sha256 (str): Expected SHA256 of the contents of the 3D object as
|
227 |
-
it was when it was downloaded with Objaverse-XL.
|
228 |
-
- metadata (Dict[Hashable, Any]): Metadata about the 3D object, which is
|
229 |
-
particular to the souce.
|
230 |
-
Return is not used. Defaults to None.
|
231 |
-
handle_missing_object (Optional[Callable], optional): Called when an object
|
232 |
-
that is in Objaverse-XL is not found. Here, it is likely that the
|
233 |
-
repository was deleted or renamed. If None, nothing will be done with
|
234 |
-
the missing object.
|
235 |
-
Args for the function include:
|
236 |
-
- file_identifier (str): File identifier of the 3D object.
|
237 |
-
- sha256 (str): SHA256 of the contents of the original 3D object.
|
238 |
-
- metadata (Dict[Hashable, Any]): Metadata about the 3D object, which is
|
239 |
-
particular to the source.
|
240 |
-
Return is not used. Defaults to None.
|
241 |
-
|
242 |
-
Returns:
|
243 |
-
Dict[str, str]: A dictionary mapping from the fileIdentifier to the
|
244 |
-
download_path.
|
245 |
-
"""
|
246 |
-
if processes is None:
|
247 |
-
processes = multiprocessing.cpu_count()
|
248 |
-
|
249 |
-
out = {}
|
250 |
-
objects_to_download = []
|
251 |
-
if download_dir is not None:
|
252 |
-
objects_dir = os.path.join(download_dir, "smithsonian", "objects")
|
253 |
-
fs, path = fsspec.core.url_to_fs(objects_dir)
|
254 |
-
fs.makedirs(path, exist_ok=True)
|
255 |
-
|
256 |
-
# get the existing glb files
|
257 |
-
existing_glb_files = fs.glob(
|
258 |
-
os.path.join(objects_dir, "*.glb"), refresh=True
|
259 |
-
)
|
260 |
-
existing_uids = set(
|
261 |
-
os.path.basename(file).split(".")[0] for file in existing_glb_files
|
262 |
-
)
|
263 |
-
|
264 |
-
# find the urls that need to be downloaded
|
265 |
-
already_downloaded_objects = set()
|
266 |
-
for _, item in objects.iterrows():
|
267 |
-
file_identifier = item["fileIdentifier"]
|
268 |
-
uid = get_uid_from_str(file_identifier)
|
269 |
-
if uid not in existing_uids:
|
270 |
-
objects_to_download.append(item)
|
271 |
-
else:
|
272 |
-
already_downloaded_objects.add(file_identifier)
|
273 |
-
out[file_identifier] = os.path.join(
|
274 |
-
os.path.expanduser(objects_dir), f"{uid}.glb"
|
275 |
-
)
|
276 |
-
else:
|
277 |
-
existing_uids = set()
|
278 |
-
objects_to_download = [item for _, item in objects.iterrows()]
|
279 |
-
already_downloaded_objects = set()
|
280 |
-
out = {}
|
281 |
-
|
282 |
-
logger.info(
|
283 |
-
f"Found {len(already_downloaded_objects)} Smithsonian Objects already downloaded"
|
284 |
-
)
|
285 |
-
logger.info(
|
286 |
-
f"Downloading {len(objects_to_download)} Smithsonian Objects with {processes} processes"
|
287 |
-
)
|
288 |
-
|
289 |
-
if len(objects_to_download) == 0:
|
290 |
-
return out
|
291 |
-
|
292 |
-
args = [
|
293 |
-
[
|
294 |
-
item["fileIdentifier"],
|
295 |
-
download_dir,
|
296 |
-
item["sha256"],
|
297 |
-
handle_found_object,
|
298 |
-
handle_modified_object,
|
299 |
-
handle_missing_object,
|
300 |
-
]
|
301 |
-
for item in objects_to_download
|
302 |
-
]
|
303 |
-
with Pool(processes=processes) as pool:
|
304 |
-
results = list(
|
305 |
-
tqdm(
|
306 |
-
pool.imap_unordered(self._parallel_download_object, args),
|
307 |
-
total=len(objects_to_download),
|
308 |
-
desc="Downloading Smithsonian Objects",
|
309 |
-
)
|
310 |
-
)
|
311 |
-
|
312 |
-
for file_identifier, download_path in results:
|
313 |
-
if download_path is not None:
|
314 |
-
out[file_identifier] = download_path
|
315 |
-
|
316 |
-
return out
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
objaverse_xl/thingiverse.py
DELETED
@@ -1,380 +0,0 @@
|
|
1 |
-
"""Script to download objects from Thingiverse."""
|
2 |
-
|
3 |
-
import multiprocessing
|
4 |
-
import os
|
5 |
-
import tempfile
|
6 |
-
import time
|
7 |
-
from multiprocessing import Pool
|
8 |
-
from typing import Callable, Dict, Optional, Tuple
|
9 |
-
|
10 |
-
import fsspec
|
11 |
-
import pandas as pd
|
12 |
-
import requests
|
13 |
-
from loguru import logger
|
14 |
-
from tqdm import tqdm
|
15 |
-
|
16 |
-
from objaverse_xl.abstract import ObjaverseSource
|
17 |
-
from objaverse_xl.utils import get_file_hash
|
18 |
-
|
19 |
-
|
20 |
-
class ThingiverseDownloader(ObjaverseSource):
|
21 |
-
"""Script to download objects from Thingiverse."""
|
22 |
-
|
23 |
-
def get_annotations(self, download_dir: str = "~/.objaverse") -> pd.DataFrame:
|
24 |
-
"""Load the annotations from the given directory.
|
25 |
-
|
26 |
-
Args:
|
27 |
-
download_dir (str, optional): The directory to load the annotations from.
|
28 |
-
Supports all file systems supported by fsspec. Defaults to
|
29 |
-
"~/.objaverse".
|
30 |
-
|
31 |
-
Returns:
|
32 |
-
pd.DataFrame: The annotations, which includes the columns "thingId", "fileId",
|
33 |
-
"filename", and "license".
|
34 |
-
"""
|
35 |
-
remote_url = "https://huggingface.co/datasets/allenai/objaverse-xl/resolve/main/thingiverse/thingiverse-objects.parquet"
|
36 |
-
download_path = os.path.join(
|
37 |
-
download_dir, "thingiverse", "thingiverse-objects.parquet"
|
38 |
-
)
|
39 |
-
fs, path = fsspec.core.url_to_fs(download_path)
|
40 |
-
|
41 |
-
if not fs.exists(path):
|
42 |
-
fs.makedirs(os.path.dirname(path), exist_ok=True)
|
43 |
-
logger.info(f"Downloading {remote_url} to {download_path}")
|
44 |
-
response = requests.get(remote_url)
|
45 |
-
response.raise_for_status()
|
46 |
-
with fs.open(path, "wb") as file:
|
47 |
-
file.write(response.content)
|
48 |
-
|
49 |
-
# read the file with pandas and fsspec
|
50 |
-
with fs.open(download_path, "rb") as f:
|
51 |
-
annotations_df = pd.read_parquet(f)
|
52 |
-
|
53 |
-
return annotations_df
|
54 |
-
|
55 |
-
def _get_response_with_retries(
|
56 |
-
self, url: str, max_retries: int = 3, retry_delay: int = 5
|
57 |
-
) -> Optional[requests.models.Response]:
|
58 |
-
"""Get a response from a URL with retries.
|
59 |
-
|
60 |
-
Args:
|
61 |
-
url (str): The URL to get a response from.
|
62 |
-
max_retries (int, optional): The maximum number of retries. Defaults to 3.
|
63 |
-
retry_delay (int, optional): The delay between retries in seconds. Defaults to 5.
|
64 |
-
|
65 |
-
Returns:
|
66 |
-
Optional[requests.models.Response]: The response from the URL. If there was an error, returns None.
|
67 |
-
"""
|
68 |
-
|
69 |
-
for i in range(max_retries):
|
70 |
-
try:
|
71 |
-
response = requests.get(url, stream=True)
|
72 |
-
# if successful, break out of loop
|
73 |
-
if response.status_code not in {200, 404}:
|
74 |
-
time.sleep(retry_delay)
|
75 |
-
continue
|
76 |
-
break
|
77 |
-
except ConnectionError:
|
78 |
-
if i < max_retries - 1: # i.e. not on the last try
|
79 |
-
time.sleep(retry_delay)
|
80 |
-
else:
|
81 |
-
return None
|
82 |
-
|
83 |
-
return response
|
84 |
-
|
85 |
-
def _download_item(
|
86 |
-
self,
|
87 |
-
thingi_file_id: str,
|
88 |
-
thingi_thing_id: str,
|
89 |
-
file_identifier: str,
|
90 |
-
download_dir: Optional[str],
|
91 |
-
expected_sha256: str,
|
92 |
-
handle_found_object: Optional[Callable],
|
93 |
-
handle_modified_object: Optional[Callable],
|
94 |
-
handle_missing_object: Optional[Callable],
|
95 |
-
) -> Tuple[str, Optional[str]]:
|
96 |
-
"""Download the given item.
|
97 |
-
|
98 |
-
Args:
|
99 |
-
thingi_file_id (str): The Thingiverse file ID of the object.
|
100 |
-
thingi_thing_id (str): The Thingiverse thing ID of the object.
|
101 |
-
file_identifier (str): File identifier of the Thingiverse object.
|
102 |
-
download_dir (Optional[str]): Directory to download the Smithsonian Object
|
103 |
-
to. Supports all file systems supported by fsspec. If None, the
|
104 |
-
Smithsonian Object will be deleted after it is downloaded and processed
|
105 |
-
with the handler functions.
|
106 |
-
expected_sha256 (str): The expected SHA256 of the contents of the downloaded
|
107 |
-
object.
|
108 |
-
handle_found_object (Optional[Callable]): Called when an object is
|
109 |
-
successfully found and downloaded. Here, the object has the same sha256
|
110 |
-
as the one that was downloaded with Objaverse-XL. If None, the object
|
111 |
-
will be downloaded, but nothing will be done with it. Args for the
|
112 |
-
function include:
|
113 |
-
- local_path (str): Local path to the downloaded 3D object.
|
114 |
-
- file_identifier (str): File identifier of the 3D object.
|
115 |
-
- sha256 (str): SHA256 of the contents of the 3D object.
|
116 |
-
- metadata (Dict[str, Any]): Metadata about the 3D object, including the
|
117 |
-
GitHub organization and repo names.
|
118 |
-
Return is not used.
|
119 |
-
handle_modified_object (Optional[Callable]): Called when a modified object
|
120 |
-
is found and downloaded. Here, the object is successfully downloaded,
|
121 |
-
but it has a different sha256 than the one that was downloaded with
|
122 |
-
Objaverse-XL. This is not expected to happen very often, because the
|
123 |
-
same commit hash is used for each repo. If None, the object will be
|
124 |
-
downloaded, but nothing will be done with it. Args for the function
|
125 |
-
include:
|
126 |
-
- local_path (str): Local path to the downloaded 3D object.
|
127 |
-
- file_identifier (str): File identifier of the 3D object.
|
128 |
-
- new_sha256 (str): SHA256 of the contents of the newly downloaded 3D
|
129 |
-
object.
|
130 |
-
- old_sha256 (str): Expected SHA256 of the contents of the 3D object as
|
131 |
-
it was when it was downloaded with Objaverse-XL.
|
132 |
-
- metadata (Dict[str, Any]): Metadata about the 3D object, including the
|
133 |
-
GitHub organization and repo names.
|
134 |
-
Return is not used.
|
135 |
-
handle_missing_object (Optional[Callable]): Called when an object that is in
|
136 |
-
Objaverse-XL is not found. Here, it is likely that the repository was
|
137 |
-
deleted or renamed. If None, nothing will be done with the missing
|
138 |
-
object. Args for the function include:
|
139 |
-
- file_identifier (str): File identifier of the 3D object.
|
140 |
-
- sha256 (str): SHA256 of the contents of the original 3D object.
|
141 |
-
- metadata (Dict[str, Any]): Metadata about the 3D object, including the
|
142 |
-
GitHub organization and repo names.
|
143 |
-
Return is not used.
|
144 |
-
|
145 |
-
|
146 |
-
Returns:
|
147 |
-
Optional[str]: The path to the downloaded file. If there was an error or 404,
|
148 |
-
returns None.
|
149 |
-
"""
|
150 |
-
url = f"https://www.thingiverse.com/download:{thingi_file_id}"
|
151 |
-
response = self._get_response_with_retries(url)
|
152 |
-
filename = f"thing-{thingi_thing_id}-file-{thingi_file_id}.stl"
|
153 |
-
|
154 |
-
if response is None:
|
155 |
-
logger.warning(
|
156 |
-
f"Thingiverse file ID {thingi_file_id} could not get response from {url}"
|
157 |
-
)
|
158 |
-
# NOTE: the object is probably not missing, but the request failed
|
159 |
-
return file_identifier, None
|
160 |
-
|
161 |
-
# Check if the request was successful
|
162 |
-
if response.status_code == 404:
|
163 |
-
logger.warning(
|
164 |
-
f"Thingiverse file ID {thingi_file_id} (404) could not find file"
|
165 |
-
)
|
166 |
-
if handle_missing_object is not None:
|
167 |
-
handle_missing_object(
|
168 |
-
file_identifier=file_identifier, sha256=expected_sha256, metadata={}
|
169 |
-
)
|
170 |
-
return file_identifier, None
|
171 |
-
|
172 |
-
with tempfile.TemporaryDirectory() as temp_dir:
|
173 |
-
temp_path = os.path.join(temp_dir, filename)
|
174 |
-
temp_path_tmp = temp_path + ".tmp"
|
175 |
-
|
176 |
-
with open(temp_path_tmp, "wb") as file:
|
177 |
-
for chunk in response.iter_content(chunk_size=8192):
|
178 |
-
file.write(chunk)
|
179 |
-
|
180 |
-
# rename to temp_path
|
181 |
-
os.rename(temp_path_tmp, temp_path)
|
182 |
-
|
183 |
-
# check the sha256
|
184 |
-
sha256 = get_file_hash(temp_path)
|
185 |
-
|
186 |
-
if sha256 == expected_sha256:
|
187 |
-
if handle_found_object is not None:
|
188 |
-
handle_found_object(
|
189 |
-
local_path=temp_path,
|
190 |
-
file_identifier=file_identifier,
|
191 |
-
sha256=sha256,
|
192 |
-
metadata={},
|
193 |
-
)
|
194 |
-
else:
|
195 |
-
if handle_modified_object is not None:
|
196 |
-
handle_modified_object(
|
197 |
-
local_path=temp_path,
|
198 |
-
file_identifier=file_identifier,
|
199 |
-
new_sha256=sha256,
|
200 |
-
old_sha256=expected_sha256,
|
201 |
-
metadata={},
|
202 |
-
)
|
203 |
-
|
204 |
-
if download_dir is not None:
|
205 |
-
filename = os.path.join(download_dir, filename)
|
206 |
-
fs, path = fsspec.core.url_to_fs(filename)
|
207 |
-
fs.makedirs(os.path.dirname(path), exist_ok=True)
|
208 |
-
fs.put(temp_path, path)
|
209 |
-
else:
|
210 |
-
path = None
|
211 |
-
|
212 |
-
return file_identifier, path
|
213 |
-
|
214 |
-
def _parallel_download_item(self, args):
|
215 |
-
return self._download_item(*args)
|
216 |
-
|
217 |
-
def get_file_id_from_file_identifier(self, file_identifier: str) -> str:
|
218 |
-
"""Get the thingiverse file ID from the Objaverse-XL file identifier.
|
219 |
-
|
220 |
-
Args:
|
221 |
-
file_identifier (str): The Objaverse-XL file identifier.
|
222 |
-
|
223 |
-
Returns:
|
224 |
-
str: The Thingiverse file ID.
|
225 |
-
"""
|
226 |
-
return file_identifier.split("fileId=")[-1]
|
227 |
-
|
228 |
-
def get_thing_id_from_file_identifier(self, file_identifier: str) -> str:
|
229 |
-
"""Get the thingiverse thing ID from the Objaverse-XL file identifier.
|
230 |
-
|
231 |
-
Args:
|
232 |
-
file_identifier (str): The Objaverse-XL file identifier.
|
233 |
-
|
234 |
-
Returns:
|
235 |
-
str: The Thingiverse thing ID.
|
236 |
-
"""
|
237 |
-
return file_identifier.split("/")[-2].split(":")[1]
|
238 |
-
|
239 |
-
def download_objects(
|
240 |
-
self,
|
241 |
-
objects: pd.DataFrame,
|
242 |
-
download_dir: Optional[str] = "~/.objaverse",
|
243 |
-
processes: Optional[int] = None,
|
244 |
-
handle_found_object: Optional[Callable] = None,
|
245 |
-
handle_modified_object: Optional[Callable] = None,
|
246 |
-
handle_missing_object: Optional[Callable] = None,
|
247 |
-
**kwargs,
|
248 |
-
) -> Dict[str, str]:
|
249 |
-
"""Download the objects from the given list of things and files.
|
250 |
-
|
251 |
-
Args:
|
252 |
-
objects (pd.DataFrame): Thingiverse objects to download. Must have columns
|
253 |
-
for the object "fileIdentifier" and "sha256". Use the `get_annotations`
|
254 |
-
function to get the metadata.
|
255 |
-
download_dir (str, optional): The directory to save the files to. Supports
|
256 |
-
all file systems supported by fsspec. Defaults to "~/.objaverse-xl".
|
257 |
-
processes (int, optional): The number of processes to use. If None, maps to
|
258 |
-
use all available CPUs using multiprocessing.cpu_count(). Defaults to
|
259 |
-
None.
|
260 |
-
handle_found_object (Optional[Callable], optional): Called when an object is
|
261 |
-
successfully found and downloaded. Here, the object has the same sha256
|
262 |
-
as the one that was downloaded with Objaverse-XL. If None, the object
|
263 |
-
will be downloaded, but nothing will be done with it. Args for the
|
264 |
-
function include:
|
265 |
-
- local_path (str): Local path to the downloaded 3D object.
|
266 |
-
- file_identifier (str): File identifier of the 3D object.
|
267 |
-
- sha256 (str): SHA256 of the contents of the 3D object.
|
268 |
-
- metadata (Dict[Hashable, Any]): Metadata about the 3D object,
|
269 |
-
including the GitHub organization and repo names.
|
270 |
-
Return is not used. Defaults to None.
|
271 |
-
handle_modified_object (Optional[Callable], optional): Called when a
|
272 |
-
modified object is found and downloaded. Here, the object is
|
273 |
-
successfully downloaded, but it has a different sha256 than the one that
|
274 |
-
was downloaded with Objaverse-XL. This is not expected to happen very
|
275 |
-
often, because the same commit hash is used for each repo. If None, the
|
276 |
-
object will be downloaded, but nothing will be done with it. Args for
|
277 |
-
the function include:
|
278 |
-
- local_path (str): Local path to the downloaded 3D object.
|
279 |
-
- file_identifier (str): File identifier of the 3D object.
|
280 |
-
- new_sha256 (str): SHA256 of the contents of the newly downloaded 3D
|
281 |
-
object.
|
282 |
-
- old_sha256 (str): Expected SHA256 of the contents of the 3D object as
|
283 |
-
it was when it was downloaded with Objaverse-XL.
|
284 |
-
- metadata (Dict[Hashable, Any]): Metadata about the 3D object, which is
|
285 |
-
particular to the souce.
|
286 |
-
Return is not used. Defaults to None.
|
287 |
-
handle_missing_object (Optional[Callable], optional): Called when an object
|
288 |
-
that is in Objaverse-XL is not found. Here, it is likely that the
|
289 |
-
repository was deleted or renamed. If None, nothing will be done with
|
290 |
-
the missing object.
|
291 |
-
Args for the function include:
|
292 |
-
- file_identifier (str): File identifier of the 3D object.
|
293 |
-
- sha256 (str): SHA256 of the contents of the original 3D object.
|
294 |
-
- metadata (Dict[Hashable, Any]): Metadata about the 3D object, which is
|
295 |
-
particular to the source.
|
296 |
-
Return is not used. Defaults to None.
|
297 |
-
|
298 |
-
Returns:
|
299 |
-
Dict[str, str]: A dictionary mapping from the fileIdentifier to the path of
|
300 |
-
the downloaded file.
|
301 |
-
"""
|
302 |
-
if processes is None:
|
303 |
-
processes = multiprocessing.cpu_count()
|
304 |
-
|
305 |
-
objects = objects.copy()
|
306 |
-
objects["thingiFileId"] = objects["fileIdentifier"].apply(
|
307 |
-
self.get_file_id_from_file_identifier
|
308 |
-
)
|
309 |
-
objects["thingiThingId"] = objects["fileIdentifier"].apply(
|
310 |
-
self.get_thing_id_from_file_identifier
|
311 |
-
)
|
312 |
-
|
313 |
-
# create the download directory
|
314 |
-
out = {}
|
315 |
-
if download_dir is not None:
|
316 |
-
download_dir = os.path.join(download_dir, "thingiverse")
|
317 |
-
fs, path = fsspec.core.url_to_fs(download_dir)
|
318 |
-
fs.makedirs(path, exist_ok=True)
|
319 |
-
|
320 |
-
# check to filter out files that already exist
|
321 |
-
existing_files = fs.glob(os.path.join(download_dir, "*.stl"), refresh=True)
|
322 |
-
existing_file_ids = {
|
323 |
-
os.path.basename(file).split(".")[0].split("-")[-1]
|
324 |
-
for file in existing_files
|
325 |
-
}
|
326 |
-
|
327 |
-
# filter out existing files
|
328 |
-
items_to_download = []
|
329 |
-
already_downloaded_count = 0
|
330 |
-
for _, item in objects.iterrows():
|
331 |
-
if item["thingiFileId"] in existing_file_ids:
|
332 |
-
already_downloaded_count += 1
|
333 |
-
out[item["fileIdentifier"]] = os.path.join(
|
334 |
-
os.path.expanduser(download_dir),
|
335 |
-
f"thing-{item['thingiThingId']}-file-{item['thingiFileId']}.stl",
|
336 |
-
)
|
337 |
-
else:
|
338 |
-
items_to_download.append(item)
|
339 |
-
|
340 |
-
logger.info(
|
341 |
-
f"Found {already_downloaded_count} Thingiverse objects downloaded"
|
342 |
-
)
|
343 |
-
else:
|
344 |
-
items_to_download = [item for _, item in objects.iterrows()]
|
345 |
-
|
346 |
-
logger.info(
|
347 |
-
f"Downloading {len(items_to_download)} Thingiverse objects with {processes=}"
|
348 |
-
)
|
349 |
-
if len(items_to_download) == 0:
|
350 |
-
return out
|
351 |
-
|
352 |
-
# download the files
|
353 |
-
args = [
|
354 |
-
(
|
355 |
-
item["thingiFileId"],
|
356 |
-
item["thingiThingId"],
|
357 |
-
item["fileIdentifier"],
|
358 |
-
download_dir,
|
359 |
-
item["sha256"],
|
360 |
-
handle_found_object,
|
361 |
-
handle_modified_object,
|
362 |
-
handle_missing_object,
|
363 |
-
)
|
364 |
-
for item in items_to_download
|
365 |
-
]
|
366 |
-
|
367 |
-
with Pool(processes=processes) as pool:
|
368 |
-
results = list(
|
369 |
-
tqdm(
|
370 |
-
pool.imap_unordered(self._parallel_download_item, args),
|
371 |
-
total=len(args),
|
372 |
-
desc="Downloading Thingiverse Objects",
|
373 |
-
)
|
374 |
-
)
|
375 |
-
|
376 |
-
for file_identifier, download_path in results:
|
377 |
-
if download_path is not None:
|
378 |
-
out[file_identifier] = download_path
|
379 |
-
|
380 |
-
return out
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
objaverse_xl/utils.py
DELETED
@@ -1,45 +0,0 @@
|
|
1 |
-
"""Utility functions for the objaverse_xl package."""
|
2 |
-
|
3 |
-
import hashlib
|
4 |
-
import os
|
5 |
-
import uuid
|
6 |
-
|
7 |
-
|
8 |
-
def get_uid_from_str(string: str) -> str:
|
9 |
-
"""Generates a UUID from a string.
|
10 |
-
|
11 |
-
Args:
|
12 |
-
string (str): String to generate a UUID from.
|
13 |
-
|
14 |
-
Returns:
|
15 |
-
str: UUID generated from the string.
|
16 |
-
"""
|
17 |
-
namespace = uuid.NAMESPACE_DNS
|
18 |
-
return str(uuid.uuid5(namespace, string))
|
19 |
-
|
20 |
-
|
21 |
-
def get_file_hash(file_path: str) -> str:
|
22 |
-
"""Get the sha256 hash of a file.
|
23 |
-
|
24 |
-
Args:
|
25 |
-
file_path (str): Path to the file.
|
26 |
-
|
27 |
-
Returns:
|
28 |
-
str: sha256 hash of the file.
|
29 |
-
"""
|
30 |
-
# Check if the path is a symbolic link
|
31 |
-
if os.path.islink(file_path):
|
32 |
-
# Resolve the symbolic link
|
33 |
-
resolved_path = os.readlink(file_path)
|
34 |
-
# Check if the resolved path exists
|
35 |
-
if not os.path.exists(resolved_path):
|
36 |
-
raise FileNotFoundError(
|
37 |
-
f"The symbolic link points to a file that doesn't exist: {resolved_path}"
|
38 |
-
)
|
39 |
-
sha256 = hashlib.sha256()
|
40 |
-
# Read the file from the path
|
41 |
-
with open(file_path, "rb") as f:
|
42 |
-
# Loop till the end of the file
|
43 |
-
for byte_block in iter(lambda: f.read(4096), b""):
|
44 |
-
sha256.update(byte_block)
|
45 |
-
return sha256.hexdigest()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
requirements-test.txt
DELETED
@@ -1,8 +0,0 @@
|
|
1 |
-
black==22.3.0
|
2 |
-
mypy==0.942
|
3 |
-
pylint==2.13.4
|
4 |
-
pytest-cov==3.0.0
|
5 |
-
pytest-xdist==2.5.0
|
6 |
-
pytest==7.0.1
|
7 |
-
isort==5.12.0
|
8 |
-
types-requests==2.31.0.2
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
requirements.txt
DELETED
@@ -1,7 +0,0 @@
|
|
1 |
-
requests
|
2 |
-
pandas
|
3 |
-
pyarrow
|
4 |
-
tqdm
|
5 |
-
loguru
|
6 |
-
fsspec>=2022.11.0
|
7 |
-
gputil==1.4.0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
scripts/rendering/README.md
DELETED
@@ -1,149 +0,0 @@
|
|
1 |
-
# 🪐 Objaverse-XL Rendering Script
|
2 |
-
|
3 |
-
![Blender generated with MidJourney](https://github.com/allenai/objaverse-xl/assets/28768645/69064f78-a752-40d6-bd36-ea7c15ffa1ec)
|
4 |
-
|
5 |
-
Scripts for rendering Objaverse-XL with [Blender](https://www.blender.org/). Rendering is the process of taking pictures of the 3D objects. These images can then be used for training AI models.
|
6 |
-
|
7 |
-
## 🖥️ Setup
|
8 |
-
|
9 |
-
1. Clone the repository and enter the rendering directory:
|
10 |
-
|
11 |
-
```bash
|
12 |
-
git clone https://github.com/allenai/objaverse-xl.git && \
|
13 |
-
cd objaverse-xl/scripts/rendering
|
14 |
-
```
|
15 |
-
|
16 |
-
2. Download Blender:
|
17 |
-
|
18 |
-
```bash
|
19 |
-
wget https://download.blender.org/release/Blender3.2/blender-3.2.2-linux-x64.tar.xz && \
|
20 |
-
tar -xf blender-3.2.2-linux-x64.tar.xz && \
|
21 |
-
rm blender-3.2.2-linux-x64.tar.xz
|
22 |
-
```
|
23 |
-
|
24 |
-
3. If you're on a headless Linux server, install Xorg and start it:
|
25 |
-
|
26 |
-
```bash
|
27 |
-
sudo apt-get install xserver-xorg -y && \
|
28 |
-
sudo python3 start_x_server.py start
|
29 |
-
```
|
30 |
-
|
31 |
-
4. Install the Python dependencies. Note that Python >3.8 is required:
|
32 |
-
|
33 |
-
```bash
|
34 |
-
cd ../.. && \
|
35 |
-
pip install -r requirements.txt && \
|
36 |
-
pip install -e . && \
|
37 |
-
cd scripts/rendering
|
38 |
-
```
|
39 |
-
|
40 |
-
## 📸 Usage
|
41 |
-
|
42 |
-
### 🐥 Minimal Example
|
43 |
-
|
44 |
-
After setup, we can start to render objects using the `main.py` script:
|
45 |
-
|
46 |
-
```bash
|
47 |
-
python3 main.py
|
48 |
-
```
|
49 |
-
|
50 |
-
After running this, you should see 10 zip files located in `~/.objaverse/github/renders`. Each zip file corresponds to the rendering of a unique object, in this case from [our example 3D objects repo](https://github.com/mattdeitke/objaverse-xl-test-files):
|
51 |
-
|
52 |
-
```bash
|
53 |
-
> ls ~/.objaverse/github/renders
|
54 |
-
0fde27a0-99f0-5029-8e20-be9b8ecabb59.zip 54f7478b-4983-5541-8cf7-1ab2e39a842e.zip 93499b75-3ee0-5069-8f4b-1bab60d2e6d6.zip
|
55 |
-
21dd4d7b-b203-5d00-b325-0c041f43524e.zip 5babbc61-d4e1-5b5c-9b47-44994bbf958e.zip ab30e24f-1046-5257-8806-2e346f4efebe.zip
|
56 |
-
415ca2d5-9d87-568c-a5ff-73048a084229.zip 5f6d2547-3661-54d5-9895-bebc342c753d.zip
|
57 |
-
44414a2a-e8f0-5a5f-bb58-6be50d8fd034.zip 8a170083-0529-547f-90ec-ebc32eafe594.zip
|
58 |
-
```
|
59 |
-
|
60 |
-
If we unzip one of the zip files:
|
61 |
-
|
62 |
-
```bash
|
63 |
-
> cd ~/.objaverse/github/renders
|
64 |
-
> unzip 0fde27a0-99f0-5029-8e20-be9b8ecabb59.zip
|
65 |
-
```
|
66 |
-
|
67 |
-
we will see that there is a new `0fde27a0-99f0-5029-8e20-be9b8ecabb59` directory. If we look in that directory, we'll find the following files:
|
68 |
-
|
69 |
-
```bash
|
70 |
-
> ls 0fde27a0-99f0-5029-8e20-be9b8ecabb59
|
71 |
-
000.npy 001.npy 002.npy 003.npy 004.npy 005.npy 006.npy 007.npy 008.npy 009.npy 010.npy 011.npy metadata.json
|
72 |
-
000.png 001.png 002.png 003.png 004.png 005.png 006.png 007.png 008.png 009.png 010.png 011.png
|
73 |
-
```
|
74 |
-
|
75 |
-
Here, we see that there are 12 renders `[000-011].png`. Each render will look something like one of the 4 images shown below, but likely with the camera at a different location as its location is randomized during rendering:
|
76 |
-
|
77 |
-
![temp](https://github.com/allenai/objaverse-xl/assets/28768645/69d79e26-4df1-4bd2-854c-7d3c888adae7)
|
78 |
-
|
79 |
-
Additionally, there are 12 npy files `[000-011].npy`, which include information about the camera's pose for a given render. We can read the npy files using:
|
80 |
-
|
81 |
-
```python
|
82 |
-
import numpy as np
|
83 |
-
array = np.load("000.npy")
|
84 |
-
```
|
85 |
-
|
86 |
-
where array is now a 3x4 [camera matrix](https://en.wikipedia.org/wiki/Camera_matrix) that looks something like:
|
87 |
-
|
88 |
-
```python
|
89 |
-
array([[6.07966840e-01, 7.93962419e-01, 3.18103019e-08, 2.10451518e-07],
|
90 |
-
[4.75670159e-01, -3.64238620e-01, 8.00667346e-01, -5.96046448e-08],
|
91 |
-
[6.35699809e-01, -4.86779213e-01, -5.99109232e-01, -1.66008198e+00]])
|
92 |
-
```
|
93 |
-
|
94 |
-
Finally, we also have a `metadata.json` file, which contains metadata about the object and scene:
|
95 |
-
|
96 |
-
```json
|
97 |
-
{
|
98 |
-
"animation_count": 0,
|
99 |
-
"armature_count": 0,
|
100 |
-
"edge_count": 2492,
|
101 |
-
"file_identifier": "https://github.com/mattdeitke/objaverse-xl-test-files/blob/ead0bed6a76012452273bbe18d12e4d68a881956/example.abc",
|
102 |
-
"file_size": 108916,
|
103 |
-
"lamp_count": 1,
|
104 |
-
"linked_files": [],
|
105 |
-
"material_count": 0,
|
106 |
-
"mesh_count": 3,
|
107 |
-
"missing_textures": {
|
108 |
-
"count": 0,
|
109 |
-
"file_path_to_color": {},
|
110 |
-
"files": []
|
111 |
-
},
|
112 |
-
"object_count": 8,
|
113 |
-
"poly_count": 984,
|
114 |
-
"random_color": null,
|
115 |
-
"save_uid": "0fde27a0-99f0-5029-8e20-be9b8ecabb59",
|
116 |
-
"scene_size": {
|
117 |
-
"bbox_max": [
|
118 |
-
4.999998569488525,
|
119 |
-
6.0,
|
120 |
-
1.0
|
121 |
-
],
|
122 |
-
"bbox_min": [
|
123 |
-
-4.999995231628418,
|
124 |
-
-6.0,
|
125 |
-
-1.0
|
126 |
-
]
|
127 |
-
},
|
128 |
-
"sha256": "879bc9d2d85e4f3866f0cfef41f5236f9fff5f973380461af9f69cdbed53a0da",
|
129 |
-
"shape_key_count": 0,
|
130 |
-
"vert_count": 2032
|
131 |
-
}
|
132 |
-
```
|
133 |
-
|
134 |
-
### 🎛 Configuration
|
135 |
-
|
136 |
-
### 🧑🔬️ Experimental Features
|
137 |
-
|
138 |
-
USDZ support is experimental. Since Blender does not natively support usdz, we use [this Blender addon](https://github.com/robmcrosby/BlenderUSDZ), but it doesn't work with all types of USDZs. If you have a better solution, PRs are very much welcome 😄!
|
139 |
-
|
140 |
-
## 👋 Our Team
|
141 |
-
|
142 |
-
Objaverse-XL is an open-source project managed by the [PRIOR team](//prior.allenai.org) at the [Allen Institute for AI](//allenai.org) (AI2).
|
143 |
-
AI2 is a non-profit institute with the mission to contribute to humanity through high-impact AI research and engineering.
|
144 |
-
|
145 |
-
<br />
|
146 |
-
|
147 |
-
<a href="//prior.allenai.org">
|
148 |
-
<p align="center"><img width="100%" src="https://raw.githubusercontent.com/allenai/ai2thor/main/doc/static/ai2-prior.svg" /></p>
|
149 |
-
</a>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
scripts/rendering/blender_script.py
DELETED
@@ -1,899 +0,0 @@
|
|
1 |
-
"""Blender script to render images of 3D models."""
|
2 |
-
|
3 |
-
import argparse
|
4 |
-
import json
|
5 |
-
import math
|
6 |
-
import os
|
7 |
-
import random
|
8 |
-
import sys
|
9 |
-
from typing import Any, Callable, Dict, Generator, List, Literal, Optional, Set, Tuple
|
10 |
-
|
11 |
-
import bpy
|
12 |
-
import numpy as np
|
13 |
-
from mathutils import Matrix, Vector
|
14 |
-
|
15 |
-
IMPORT_FUNCTIONS: Dict[str, Callable] = {
|
16 |
-
"obj": bpy.ops.import_scene.obj,
|
17 |
-
"glb": bpy.ops.import_scene.gltf,
|
18 |
-
"gltf": bpy.ops.import_scene.gltf,
|
19 |
-
"usd": bpy.ops.import_scene.usd,
|
20 |
-
"fbx": bpy.ops.import_scene.fbx,
|
21 |
-
"stl": bpy.ops.import_mesh.stl,
|
22 |
-
"usda": bpy.ops.import_scene.usda,
|
23 |
-
"dae": bpy.ops.wm.collada_import,
|
24 |
-
"ply": bpy.ops.import_mesh.ply,
|
25 |
-
"abc": bpy.ops.wm.alembic_import,
|
26 |
-
"blend": bpy.ops.wm.append,
|
27 |
-
}
|
28 |
-
|
29 |
-
|
30 |
-
def reset_cameras() -> None:
|
31 |
-
"""Resets the cameras in the scene to a single default camera."""
|
32 |
-
# Delete all existing cameras
|
33 |
-
bpy.ops.object.select_all(action="DESELECT")
|
34 |
-
bpy.ops.object.select_by_type(type="CAMERA")
|
35 |
-
bpy.ops.object.delete()
|
36 |
-
|
37 |
-
# Create a new camera with default properties
|
38 |
-
bpy.ops.object.camera_add()
|
39 |
-
|
40 |
-
# Rename the new camera to 'NewDefaultCamera'
|
41 |
-
new_camera = bpy.context.active_object
|
42 |
-
new_camera.name = "Camera"
|
43 |
-
|
44 |
-
# Set the new camera as the active camera for the scene
|
45 |
-
scene.camera = new_camera
|
46 |
-
|
47 |
-
|
48 |
-
def sample_point_on_sphere(radius: float) -> Tuple[float, float, float]:
|
49 |
-
"""Samples a point on a sphere with the given radius.
|
50 |
-
|
51 |
-
Args:
|
52 |
-
radius (float): Radius of the sphere.
|
53 |
-
|
54 |
-
Returns:
|
55 |
-
Tuple[float, float, float]: A point on the sphere.
|
56 |
-
"""
|
57 |
-
theta = random.random() * 2 * math.pi
|
58 |
-
phi = math.acos(2 * random.random() - 1)
|
59 |
-
return (
|
60 |
-
radius * math.sin(phi) * math.cos(theta),
|
61 |
-
radius * math.sin(phi) * math.sin(theta),
|
62 |
-
radius * math.cos(phi),
|
63 |
-
)
|
64 |
-
|
65 |
-
|
66 |
-
def _sample_spherical(
|
67 |
-
radius_min: float = 1.5,
|
68 |
-
radius_max: float = 2.0,
|
69 |
-
maxz: float = 1.6,
|
70 |
-
minz: float = -0.75,
|
71 |
-
) -> np.ndarray:
|
72 |
-
"""Sample a random point in a spherical shell.
|
73 |
-
|
74 |
-
Args:
|
75 |
-
radius_min (float): Minimum radius of the spherical shell.
|
76 |
-
radius_max (float): Maximum radius of the spherical shell.
|
77 |
-
maxz (float): Maximum z value of the spherical shell.
|
78 |
-
minz (float): Minimum z value of the spherical shell.
|
79 |
-
|
80 |
-
Returns:
|
81 |
-
np.ndarray: A random (x, y, z) point in the spherical shell.
|
82 |
-
"""
|
83 |
-
correct = False
|
84 |
-
vec = np.array([0, 0, 0])
|
85 |
-
while not correct:
|
86 |
-
vec = np.random.uniform(-1, 1, 3)
|
87 |
-
# vec[2] = np.abs(vec[2])
|
88 |
-
radius = np.random.uniform(radius_min, radius_max, 1)
|
89 |
-
vec = vec / np.linalg.norm(vec, axis=0) * radius[0]
|
90 |
-
if maxz > vec[2] > minz:
|
91 |
-
correct = True
|
92 |
-
return vec
|
93 |
-
|
94 |
-
|
95 |
-
def randomize_camera(
|
96 |
-
radius_min: float = 1.5,
|
97 |
-
radius_max: float = 2.2,
|
98 |
-
maxz: float = 2.2,
|
99 |
-
minz: float = -2.2,
|
100 |
-
only_northern_hemisphere: bool = False,
|
101 |
-
) -> bpy.types.Object:
|
102 |
-
"""Randomizes the camera location and rotation inside of a spherical shell.
|
103 |
-
|
104 |
-
Args:
|
105 |
-
radius_min (float, optional): Minimum radius of the spherical shell. Defaults to
|
106 |
-
1.5.
|
107 |
-
radius_max (float, optional): Maximum radius of the spherical shell. Defaults to
|
108 |
-
2.0.
|
109 |
-
maxz (float, optional): Maximum z value of the spherical shell. Defaults to 1.6.
|
110 |
-
minz (float, optional): Minimum z value of the spherical shell. Defaults to
|
111 |
-
-0.75.
|
112 |
-
only_northern_hemisphere (bool, optional): Whether to only sample points in the
|
113 |
-
northern hemisphere. Defaults to False.
|
114 |
-
|
115 |
-
Returns:
|
116 |
-
bpy.types.Object: The camera object.
|
117 |
-
"""
|
118 |
-
|
119 |
-
x, y, z = _sample_spherical(
|
120 |
-
radius_min=radius_min, radius_max=radius_max, maxz=maxz, minz=minz
|
121 |
-
)
|
122 |
-
camera = bpy.data.objects["Camera"]
|
123 |
-
|
124 |
-
# only positive z
|
125 |
-
if only_northern_hemisphere:
|
126 |
-
z = abs(z)
|
127 |
-
|
128 |
-
camera.location = Vector(np.array([x, y, z]))
|
129 |
-
|
130 |
-
direction = -camera.location
|
131 |
-
rot_quat = direction.to_track_quat("-Z", "Y")
|
132 |
-
camera.rotation_euler = rot_quat.to_euler()
|
133 |
-
|
134 |
-
return camera
|
135 |
-
|
136 |
-
|
137 |
-
def _set_camera_at_size(i: int, scale: float = 1.5) -> bpy.types.Object:
|
138 |
-
"""Debugging function to set the camera on the 6 faces of a cube.
|
139 |
-
|
140 |
-
Args:
|
141 |
-
i (int): Index of the face of the cube.
|
142 |
-
scale (float, optional): Scale of the cube. Defaults to 1.5.
|
143 |
-
|
144 |
-
Returns:
|
145 |
-
bpy.types.Object: The camera object.
|
146 |
-
"""
|
147 |
-
if i == 0:
|
148 |
-
x, y, z = scale, 0, 0
|
149 |
-
elif i == 1:
|
150 |
-
x, y, z = -scale, 0, 0
|
151 |
-
elif i == 2:
|
152 |
-
x, y, z = 0, scale, 0
|
153 |
-
elif i == 3:
|
154 |
-
x, y, z = 0, -scale, 0
|
155 |
-
elif i == 4:
|
156 |
-
x, y, z = 0, 0, scale
|
157 |
-
elif i == 5:
|
158 |
-
x, y, z = 0, 0, -scale
|
159 |
-
else:
|
160 |
-
raise ValueError(f"Invalid index: i={i}, must be int in range [0, 5].")
|
161 |
-
camera = bpy.data.objects["Camera"]
|
162 |
-
camera.location = Vector(np.array([x, y, z]))
|
163 |
-
direction = -camera.location
|
164 |
-
rot_quat = direction.to_track_quat("-Z", "Y")
|
165 |
-
camera.rotation_euler = rot_quat.to_euler()
|
166 |
-
return camera
|
167 |
-
|
168 |
-
|
169 |
-
def _create_light(
|
170 |
-
name: str,
|
171 |
-
light_type: Literal["POINT", "SUN", "SPOT", "AREA"],
|
172 |
-
location: Tuple[float, float, float],
|
173 |
-
rotation: Tuple[float, float, float],
|
174 |
-
energy: float,
|
175 |
-
use_shadow: bool = False,
|
176 |
-
specular_factor: float = 1.0,
|
177 |
-
):
|
178 |
-
"""Creates a light object.
|
179 |
-
|
180 |
-
Args:
|
181 |
-
name (str): Name of the light object.
|
182 |
-
light_type (Literal["POINT", "SUN", "SPOT", "AREA"]): Type of the light.
|
183 |
-
location (Tuple[float, float, float]): Location of the light.
|
184 |
-
rotation (Tuple[float, float, float]): Rotation of the light.
|
185 |
-
energy (float): Energy of the light.
|
186 |
-
use_shadow (bool, optional): Whether to use shadows. Defaults to False.
|
187 |
-
specular_factor (float, optional): Specular factor of the light. Defaults to 1.0.
|
188 |
-
|
189 |
-
Returns:
|
190 |
-
bpy.types.Object: The light object.
|
191 |
-
"""
|
192 |
-
|
193 |
-
light_data = bpy.data.lights.new(name=name, type=light_type)
|
194 |
-
light_object = bpy.data.objects.new(name, light_data)
|
195 |
-
bpy.context.collection.objects.link(light_object)
|
196 |
-
light_object.location = location
|
197 |
-
light_object.rotation_euler = rotation
|
198 |
-
light_data.use_shadow = use_shadow
|
199 |
-
light_data.specular_factor = specular_factor
|
200 |
-
light_data.energy = energy
|
201 |
-
return light_object
|
202 |
-
|
203 |
-
|
204 |
-
def randomize_lighting() -> Dict[str, bpy.types.Object]:
|
205 |
-
"""Randomizes the lighting in the scene.
|
206 |
-
|
207 |
-
Returns:
|
208 |
-
Dict[str, bpy.types.Object]: Dictionary of the lights in the scene. The keys are
|
209 |
-
"key_light", "fill_light", "rim_light", and "bottom_light".
|
210 |
-
"""
|
211 |
-
|
212 |
-
# Clear existing lights
|
213 |
-
bpy.ops.object.select_all(action="DESELECT")
|
214 |
-
bpy.ops.object.select_by_type(type="LIGHT")
|
215 |
-
bpy.ops.object.delete()
|
216 |
-
|
217 |
-
# Create key light
|
218 |
-
key_light = _create_light(
|
219 |
-
name="Key_Light",
|
220 |
-
light_type="SUN",
|
221 |
-
location=(0, 0, 0),
|
222 |
-
rotation=(0.785398, 0, -0.785398),
|
223 |
-
energy=random.choice([3, 4, 5]),
|
224 |
-
)
|
225 |
-
|
226 |
-
# Create fill light
|
227 |
-
fill_light = _create_light(
|
228 |
-
name="Fill_Light",
|
229 |
-
light_type="SUN",
|
230 |
-
location=(0, 0, 0),
|
231 |
-
rotation=(0.785398, 0, 2.35619),
|
232 |
-
energy=random.choice([2, 3, 4]),
|
233 |
-
)
|
234 |
-
|
235 |
-
# Create rim light
|
236 |
-
rim_light = _create_light(
|
237 |
-
name="Rim_Light",
|
238 |
-
light_type="SUN",
|
239 |
-
location=(0, 0, 0),
|
240 |
-
rotation=(-0.785398, 0, -3.92699),
|
241 |
-
energy=random.choice([3, 4, 5]),
|
242 |
-
)
|
243 |
-
|
244 |
-
# Create bottom light
|
245 |
-
bottom_light = _create_light(
|
246 |
-
name="Bottom_Light",
|
247 |
-
light_type="SUN",
|
248 |
-
location=(0, 0, 0),
|
249 |
-
rotation=(3.14159, 0, 0),
|
250 |
-
energy=random.choice([1, 2, 3]),
|
251 |
-
)
|
252 |
-
|
253 |
-
return dict(
|
254 |
-
key_light=key_light,
|
255 |
-
fill_light=fill_light,
|
256 |
-
rim_light=rim_light,
|
257 |
-
bottom_light=bottom_light,
|
258 |
-
)
|
259 |
-
|
260 |
-
|
261 |
-
def reset_scene() -> None:
|
262 |
-
"""Resets the scene to a clean state.
|
263 |
-
|
264 |
-
Returns:
|
265 |
-
None
|
266 |
-
"""
|
267 |
-
# delete everything that isn't part of a camera or a light
|
268 |
-
for obj in bpy.data.objects:
|
269 |
-
if obj.type not in {"CAMERA", "LIGHT"}:
|
270 |
-
bpy.data.objects.remove(obj, do_unlink=True)
|
271 |
-
|
272 |
-
# delete all the materials
|
273 |
-
for material in bpy.data.materials:
|
274 |
-
bpy.data.materials.remove(material, do_unlink=True)
|
275 |
-
|
276 |
-
# delete all the textures
|
277 |
-
for texture in bpy.data.textures:
|
278 |
-
bpy.data.textures.remove(texture, do_unlink=True)
|
279 |
-
|
280 |
-
# delete all the images
|
281 |
-
for image in bpy.data.images:
|
282 |
-
bpy.data.images.remove(image, do_unlink=True)
|
283 |
-
|
284 |
-
|
285 |
-
def load_object(object_path: str) -> None:
|
286 |
-
"""Loads a model with a supported file extension into the scene.
|
287 |
-
|
288 |
-
Args:
|
289 |
-
object_path (str): Path to the model file.
|
290 |
-
|
291 |
-
Raises:
|
292 |
-
ValueError: If the file extension is not supported.
|
293 |
-
|
294 |
-
Returns:
|
295 |
-
None
|
296 |
-
"""
|
297 |
-
file_extension = object_path.split(".")[-1].lower()
|
298 |
-
if file_extension is None:
|
299 |
-
raise ValueError(f"Unsupported file type: {object_path}")
|
300 |
-
|
301 |
-
if file_extension == "usdz":
|
302 |
-
# install usdz io package
|
303 |
-
dirname = os.path.dirname(os.path.realpath(__file__))
|
304 |
-
usdz_package = os.path.join(dirname, "io_scene_usdz.zip")
|
305 |
-
bpy.ops.preferences.addon_install(filepath=usdz_package)
|
306 |
-
# enable it
|
307 |
-
addon_name = "io_scene_usdz"
|
308 |
-
bpy.ops.preferences.addon_enable(module=addon_name)
|
309 |
-
# import the usdz
|
310 |
-
from io_scene_usdz.import_usdz import import_usdz
|
311 |
-
|
312 |
-
import_usdz(context, filepath=object_path, materials=True, animations=True)
|
313 |
-
return None
|
314 |
-
|
315 |
-
# load from existing import functions
|
316 |
-
import_function = IMPORT_FUNCTIONS[file_extension]
|
317 |
-
|
318 |
-
if file_extension == "blend":
|
319 |
-
import_function(directory=object_path, link=False)
|
320 |
-
elif file_extension in {"glb", "gltf"}:
|
321 |
-
import_function(filepath=object_path, merge_vertices=True)
|
322 |
-
else:
|
323 |
-
import_function(filepath=object_path)
|
324 |
-
|
325 |
-
|
326 |
-
def scene_bbox(
|
327 |
-
single_obj: Optional[bpy.types.Object] = None, ignore_matrix: bool = False
|
328 |
-
) -> Tuple[Vector, Vector]:
|
329 |
-
"""Returns the bounding box of the scene.
|
330 |
-
|
331 |
-
Taken from Shap-E rendering script
|
332 |
-
(https://github.com/openai/shap-e/blob/main/shap_e/rendering/blender/blender_script.py#L68-L82)
|
333 |
-
|
334 |
-
Args:
|
335 |
-
single_obj (Optional[bpy.types.Object], optional): If not None, only computes
|
336 |
-
the bounding box for the given object. Defaults to None.
|
337 |
-
ignore_matrix (bool, optional): Whether to ignore the object's matrix. Defaults
|
338 |
-
to False.
|
339 |
-
|
340 |
-
Raises:
|
341 |
-
RuntimeError: If there are no objects in the scene.
|
342 |
-
|
343 |
-
Returns:
|
344 |
-
Tuple[Vector, Vector]: The minimum and maximum coordinates of the bounding box.
|
345 |
-
"""
|
346 |
-
bbox_min = (math.inf,) * 3
|
347 |
-
bbox_max = (-math.inf,) * 3
|
348 |
-
found = False
|
349 |
-
for obj in get_scene_meshes() if single_obj is None else [single_obj]:
|
350 |
-
found = True
|
351 |
-
for coord in obj.bound_box:
|
352 |
-
coord = Vector(coord)
|
353 |
-
if not ignore_matrix:
|
354 |
-
coord = obj.matrix_world @ coord
|
355 |
-
bbox_min = tuple(min(x, y) for x, y in zip(bbox_min, coord))
|
356 |
-
bbox_max = tuple(max(x, y) for x, y in zip(bbox_max, coord))
|
357 |
-
|
358 |
-
if not found:
|
359 |
-
raise RuntimeError("no objects in scene to compute bounding box for")
|
360 |
-
|
361 |
-
return Vector(bbox_min), Vector(bbox_max)
|
362 |
-
|
363 |
-
|
364 |
-
def get_scene_root_objects() -> Generator[bpy.types.Object, None, None]:
|
365 |
-
"""Returns all root objects in the scene.
|
366 |
-
|
367 |
-
Yields:
|
368 |
-
Generator[bpy.types.Object, None, None]: Generator of all root objects in the
|
369 |
-
scene.
|
370 |
-
"""
|
371 |
-
for obj in bpy.context.scene.objects.values():
|
372 |
-
if not obj.parent:
|
373 |
-
yield obj
|
374 |
-
|
375 |
-
|
376 |
-
def get_scene_meshes() -> Generator[bpy.types.Object, None, None]:
|
377 |
-
"""Returns all meshes in the scene.
|
378 |
-
|
379 |
-
Yields:
|
380 |
-
Generator[bpy.types.Object, None, None]: Generator of all meshes in the scene.
|
381 |
-
"""
|
382 |
-
for obj in bpy.context.scene.objects.values():
|
383 |
-
if isinstance(obj.data, (bpy.types.Mesh)):
|
384 |
-
yield obj
|
385 |
-
|
386 |
-
|
387 |
-
def get_3x4_RT_matrix_from_blender(cam: bpy.types.Object) -> Matrix:
|
388 |
-
"""Returns the 3x4 RT matrix from the given camera.
|
389 |
-
|
390 |
-
Taken from Zero123, which in turn was taken from
|
391 |
-
https://github.com/panmari/stanford-shapenet-renderer/blob/master/render_blender.py
|
392 |
-
|
393 |
-
Args:
|
394 |
-
cam (bpy.types.Object): The camera object.
|
395 |
-
|
396 |
-
Returns:
|
397 |
-
Matrix: The 3x4 RT matrix from the given camera.
|
398 |
-
"""
|
399 |
-
# Use matrix_world instead to account for all constraints
|
400 |
-
location, rotation = cam.matrix_world.decompose()[0:2]
|
401 |
-
R_world2bcam = rotation.to_matrix().transposed()
|
402 |
-
|
403 |
-
# Use location from matrix_world to account for constraints:
|
404 |
-
T_world2bcam = -1 * R_world2bcam @ location
|
405 |
-
|
406 |
-
# put into 3x4 matrix
|
407 |
-
RT = Matrix(
|
408 |
-
(
|
409 |
-
R_world2bcam[0][:] + (T_world2bcam[0],),
|
410 |
-
R_world2bcam[1][:] + (T_world2bcam[1],),
|
411 |
-
R_world2bcam[2][:] + (T_world2bcam[2],),
|
412 |
-
)
|
413 |
-
)
|
414 |
-
return RT
|
415 |
-
|
416 |
-
|
417 |
-
def delete_invisible_objects() -> None:
|
418 |
-
"""Deletes all invisible objects in the scene.
|
419 |
-
|
420 |
-
Returns:
|
421 |
-
None
|
422 |
-
"""
|
423 |
-
bpy.ops.object.select_all(action="DESELECT")
|
424 |
-
for obj in scene.objects:
|
425 |
-
if obj.hide_viewport or obj.hide_render:
|
426 |
-
obj.hide_viewport = False
|
427 |
-
obj.hide_render = False
|
428 |
-
obj.hide_select = False
|
429 |
-
obj.select_set(True)
|
430 |
-
bpy.ops.object.delete()
|
431 |
-
|
432 |
-
# Delete invisible collections
|
433 |
-
invisible_collections = [col for col in bpy.data.collections if col.hide_viewport]
|
434 |
-
for col in invisible_collections:
|
435 |
-
bpy.data.collections.remove(col)
|
436 |
-
|
437 |
-
|
438 |
-
def normalize_scene() -> None:
|
439 |
-
"""Normalizes the scene by scaling and translating it to fit in a unit cube centered
|
440 |
-
at the origin.
|
441 |
-
|
442 |
-
Mostly taken from the Point-E / Shap-E rendering script
|
443 |
-
(https://github.com/openai/point-e/blob/main/point_e/evals/scripts/blender_script.py#L97-L112),
|
444 |
-
but fix for multiple root objects: (see bug report here:
|
445 |
-
https://github.com/openai/shap-e/pull/60).
|
446 |
-
|
447 |
-
Returns:
|
448 |
-
None
|
449 |
-
"""
|
450 |
-
if len(list(get_scene_root_objects())) > 1:
|
451 |
-
# create an empty object to be used as a parent for all root objects
|
452 |
-
parent_empty = bpy.data.objects.new("ParentEmpty", None)
|
453 |
-
bpy.context.scene.collection.objects.link(parent_empty)
|
454 |
-
|
455 |
-
# parent all root objects to the empty object
|
456 |
-
for obj in get_scene_root_objects():
|
457 |
-
if obj != parent_empty:
|
458 |
-
obj.parent = parent_empty
|
459 |
-
|
460 |
-
bbox_min, bbox_max = scene_bbox()
|
461 |
-
scale = 1 / max(bbox_max - bbox_min)
|
462 |
-
for obj in get_scene_root_objects():
|
463 |
-
obj.scale = obj.scale * scale
|
464 |
-
|
465 |
-
# Apply scale to matrix_world.
|
466 |
-
bpy.context.view_layer.update()
|
467 |
-
bbox_min, bbox_max = scene_bbox()
|
468 |
-
offset = -(bbox_min + bbox_max) / 2
|
469 |
-
for obj in get_scene_root_objects():
|
470 |
-
obj.matrix_world.translation += offset
|
471 |
-
bpy.ops.object.select_all(action="DESELECT")
|
472 |
-
|
473 |
-
# unparent the camera
|
474 |
-
bpy.data.objects["Camera"].parent = None
|
475 |
-
|
476 |
-
|
477 |
-
def delete_missing_textures() -> Dict[str, Any]:
|
478 |
-
"""Deletes all missing textures in the scene.
|
479 |
-
|
480 |
-
Returns:
|
481 |
-
Dict[str, Any]: Dictionary with keys "count", "files", and "file_path_to_color".
|
482 |
-
"count" is the number of missing textures, "files" is a list of the missing
|
483 |
-
texture file paths, and "file_path_to_color" is a dictionary mapping the
|
484 |
-
missing texture file paths to a random color.
|
485 |
-
"""
|
486 |
-
missing_file_count = 0
|
487 |
-
out_files = []
|
488 |
-
file_path_to_color = {}
|
489 |
-
|
490 |
-
# Check all materials in the scene
|
491 |
-
for material in bpy.data.materials:
|
492 |
-
if material.use_nodes:
|
493 |
-
for node in material.node_tree.nodes:
|
494 |
-
if node.type == "TEX_IMAGE":
|
495 |
-
image = node.image
|
496 |
-
if image is not None:
|
497 |
-
file_path = bpy.path.abspath(image.filepath)
|
498 |
-
if file_path == "":
|
499 |
-
# means it's embedded
|
500 |
-
continue
|
501 |
-
|
502 |
-
if not os.path.exists(file_path):
|
503 |
-
# Find the connected Principled BSDF node
|
504 |
-
connected_node = node.outputs[0].links[0].to_node
|
505 |
-
|
506 |
-
if connected_node.type == "BSDF_PRINCIPLED":
|
507 |
-
if file_path not in file_path_to_color:
|
508 |
-
# Set a random color for the unique missing file path
|
509 |
-
random_color = [random.random() for _ in range(3)]
|
510 |
-
file_path_to_color[file_path] = random_color + [1]
|
511 |
-
|
512 |
-
connected_node.inputs[
|
513 |
-
"Base Color"
|
514 |
-
].default_value = file_path_to_color[file_path]
|
515 |
-
|
516 |
-
# Delete the TEX_IMAGE node
|
517 |
-
material.node_tree.nodes.remove(node)
|
518 |
-
missing_file_count += 1
|
519 |
-
out_files.append(image.filepath)
|
520 |
-
return {
|
521 |
-
"count": missing_file_count,
|
522 |
-
"files": out_files,
|
523 |
-
"file_path_to_color": file_path_to_color,
|
524 |
-
}
|
525 |
-
|
526 |
-
|
527 |
-
def _get_random_color() -> Tuple[float, float, float, float]:
|
528 |
-
"""Generates a random RGB-A color.
|
529 |
-
|
530 |
-
The alpha value is always 1.
|
531 |
-
|
532 |
-
Returns:
|
533 |
-
Tuple[float, float, float, float]: A random RGB-A color. Each value is in the
|
534 |
-
range [0, 1].
|
535 |
-
"""
|
536 |
-
return (random.random(), random.random(), random.random(), 1)
|
537 |
-
|
538 |
-
|
539 |
-
def _apply_color_to_object(
|
540 |
-
obj: bpy.types.Object, color: Tuple[float, float, float, float]
|
541 |
-
) -> None:
|
542 |
-
"""Applies the given color to the object.
|
543 |
-
|
544 |
-
Args:
|
545 |
-
obj (bpy.types.Object): The object to apply the color to.
|
546 |
-
color (Tuple[float, float, float, float]): The color to apply to the object.
|
547 |
-
|
548 |
-
Returns:
|
549 |
-
None
|
550 |
-
"""
|
551 |
-
mat = bpy.data.materials.new(name=f"RandomMaterial_{obj.name}")
|
552 |
-
mat.use_nodes = True
|
553 |
-
nodes = mat.node_tree.nodes
|
554 |
-
principled_bsdf = nodes.get("Principled BSDF")
|
555 |
-
if principled_bsdf:
|
556 |
-
principled_bsdf.inputs["Base Color"].default_value = color
|
557 |
-
obj.data.materials.append(mat)
|
558 |
-
|
559 |
-
|
560 |
-
def apply_single_random_color_to_all_objects() -> Tuple[float, float, float, float]:
|
561 |
-
"""Applies a single random color to all objects in the scene.
|
562 |
-
|
563 |
-
Returns:
|
564 |
-
Tuple[float, float, float, float]: The random color that was applied to all
|
565 |
-
objects.
|
566 |
-
"""
|
567 |
-
rand_color = _get_random_color()
|
568 |
-
for obj in bpy.context.scene.objects:
|
569 |
-
if obj.type == "MESH":
|
570 |
-
_apply_color_to_object(obj, rand_color)
|
571 |
-
return rand_color
|
572 |
-
|
573 |
-
|
574 |
-
class MetadataExtractor:
|
575 |
-
"""Class to extract metadata from a Blender scene."""
|
576 |
-
|
577 |
-
def __init__(
|
578 |
-
self, object_path: str, scene: bpy.types.Scene, bdata: bpy.types.BlendData
|
579 |
-
) -> None:
|
580 |
-
"""Initializes the MetadataExtractor.
|
581 |
-
|
582 |
-
Args:
|
583 |
-
object_path (str): Path to the object file.
|
584 |
-
scene (bpy.types.Scene): The current scene object from `bpy.context.scene`.
|
585 |
-
bdata (bpy.types.BlendData): The current blender data from `bpy.data`.
|
586 |
-
|
587 |
-
Returns:
|
588 |
-
None
|
589 |
-
"""
|
590 |
-
self.object_path = object_path
|
591 |
-
self.scene = scene
|
592 |
-
self.bdata = bdata
|
593 |
-
|
594 |
-
def get_poly_count(self) -> int:
|
595 |
-
"""Returns the total number of polygons in the scene."""
|
596 |
-
total_poly_count = 0
|
597 |
-
for obj in self.scene.objects:
|
598 |
-
if obj.type == "MESH":
|
599 |
-
total_poly_count += len(obj.data.polygons)
|
600 |
-
return total_poly_count
|
601 |
-
|
602 |
-
def get_vertex_count(self) -> int:
|
603 |
-
"""Returns the total number of vertices in the scene."""
|
604 |
-
total_vertex_count = 0
|
605 |
-
for obj in self.scene.objects:
|
606 |
-
if obj.type == "MESH":
|
607 |
-
total_vertex_count += len(obj.data.vertices)
|
608 |
-
return total_vertex_count
|
609 |
-
|
610 |
-
def get_edge_count(self) -> int:
|
611 |
-
"""Returns the total number of edges in the scene."""
|
612 |
-
total_edge_count = 0
|
613 |
-
for obj in self.scene.objects:
|
614 |
-
if obj.type == "MESH":
|
615 |
-
total_edge_count += len(obj.data.edges)
|
616 |
-
return total_edge_count
|
617 |
-
|
618 |
-
def get_lamp_count(self) -> int:
|
619 |
-
"""Returns the number of lamps in the scene."""
|
620 |
-
return sum(1 for obj in self.scene.objects if obj.type == "LIGHT")
|
621 |
-
|
622 |
-
def get_mesh_count(self) -> int:
|
623 |
-
"""Returns the number of meshes in the scene."""
|
624 |
-
return sum(1 for obj in self.scene.objects if obj.type == "MESH")
|
625 |
-
|
626 |
-
def get_material_count(self) -> int:
|
627 |
-
"""Returns the number of materials in the scene."""
|
628 |
-
return len(self.bdata.materials)
|
629 |
-
|
630 |
-
def get_object_count(self) -> int:
|
631 |
-
"""Returns the number of objects in the scene."""
|
632 |
-
return len(self.bdata.objects)
|
633 |
-
|
634 |
-
def get_animation_count(self) -> int:
|
635 |
-
"""Returns the number of animations in the scene."""
|
636 |
-
return len(self.bdata.actions)
|
637 |
-
|
638 |
-
def get_linked_files(self) -> List[str]:
|
639 |
-
"""Returns the filepaths of all linked files."""
|
640 |
-
image_filepaths = self._get_image_filepaths()
|
641 |
-
material_filepaths = self._get_material_filepaths()
|
642 |
-
linked_libraries_filepaths = self._get_linked_libraries_filepaths()
|
643 |
-
|
644 |
-
all_filepaths = (
|
645 |
-
image_filepaths | material_filepaths | linked_libraries_filepaths
|
646 |
-
)
|
647 |
-
if "" in all_filepaths:
|
648 |
-
all_filepaths.remove("")
|
649 |
-
return list(all_filepaths)
|
650 |
-
|
651 |
-
def _get_image_filepaths(self) -> Set[str]:
|
652 |
-
"""Returns the filepaths of all images used in the scene."""
|
653 |
-
filepaths = set()
|
654 |
-
for image in self.bdata.images:
|
655 |
-
if image.source == "FILE":
|
656 |
-
filepaths.add(bpy.path.abspath(image.filepath))
|
657 |
-
return filepaths
|
658 |
-
|
659 |
-
def _get_material_filepaths(self) -> Set[str]:
|
660 |
-
"""Returns the filepaths of all images used in materials."""
|
661 |
-
filepaths = set()
|
662 |
-
for material in self.bdata.materials:
|
663 |
-
if material.use_nodes:
|
664 |
-
for node in material.node_tree.nodes:
|
665 |
-
if node.type == "TEX_IMAGE":
|
666 |
-
image = node.image
|
667 |
-
if image is not None:
|
668 |
-
filepaths.add(bpy.path.abspath(image.filepath))
|
669 |
-
return filepaths
|
670 |
-
|
671 |
-
def _get_linked_libraries_filepaths(self) -> Set[str]:
|
672 |
-
"""Returns the filepaths of all linked libraries."""
|
673 |
-
filepaths = set()
|
674 |
-
for library in self.bdata.libraries:
|
675 |
-
filepaths.add(bpy.path.abspath(library.filepath))
|
676 |
-
return filepaths
|
677 |
-
|
678 |
-
def get_scene_size(self) -> Dict[str, list]:
|
679 |
-
"""Returns the size of the scene bounds in meters."""
|
680 |
-
bbox_min, bbox_max = scene_bbox()
|
681 |
-
return {"bbox_max": list(bbox_max), "bbox_min": list(bbox_min)}
|
682 |
-
|
683 |
-
def get_shape_key_count(self) -> int:
|
684 |
-
"""Returns the number of shape keys in the scene."""
|
685 |
-
total_shape_key_count = 0
|
686 |
-
for obj in self.scene.objects:
|
687 |
-
if obj.type == "MESH":
|
688 |
-
shape_keys = obj.data.shape_keys
|
689 |
-
if shape_keys is not None:
|
690 |
-
total_shape_key_count += (
|
691 |
-
len(shape_keys.key_blocks) - 1
|
692 |
-
) # Subtract 1 to exclude the Basis shape key
|
693 |
-
return total_shape_key_count
|
694 |
-
|
695 |
-
def get_armature_count(self) -> int:
|
696 |
-
"""Returns the number of armatures in the scene."""
|
697 |
-
total_armature_count = 0
|
698 |
-
for obj in self.scene.objects:
|
699 |
-
if obj.type == "ARMATURE":
|
700 |
-
total_armature_count += 1
|
701 |
-
return total_armature_count
|
702 |
-
|
703 |
-
def read_file_size(self) -> int:
|
704 |
-
"""Returns the size of the file in bytes."""
|
705 |
-
return os.path.getsize(self.object_path)
|
706 |
-
|
707 |
-
def get_metadata(self) -> Dict[str, Any]:
|
708 |
-
"""Returns the metadata of the scene.
|
709 |
-
|
710 |
-
Returns:
|
711 |
-
Dict[str, Any]: Dictionary of the metadata with keys for "file_size",
|
712 |
-
"poly_count", "vert_count", "edge_count", "material_count", "object_count",
|
713 |
-
"lamp_count", "mesh_count", "animation_count", "linked_files", "scene_size",
|
714 |
-
"shape_key_count", and "armature_count".
|
715 |
-
"""
|
716 |
-
return {
|
717 |
-
"file_size": self.read_file_size(),
|
718 |
-
"poly_count": self.get_poly_count(),
|
719 |
-
"vert_count": self.get_vertex_count(),
|
720 |
-
"edge_count": self.get_edge_count(),
|
721 |
-
"material_count": self.get_material_count(),
|
722 |
-
"object_count": self.get_object_count(),
|
723 |
-
"lamp_count": self.get_lamp_count(),
|
724 |
-
"mesh_count": self.get_mesh_count(),
|
725 |
-
"animation_count": self.get_animation_count(),
|
726 |
-
"linked_files": self.get_linked_files(),
|
727 |
-
"scene_size": self.get_scene_size(),
|
728 |
-
"shape_key_count": self.get_shape_key_count(),
|
729 |
-
"armature_count": self.get_armature_count(),
|
730 |
-
}
|
731 |
-
|
732 |
-
|
733 |
-
def render_object(
|
734 |
-
object_file: str,
|
735 |
-
num_renders: int,
|
736 |
-
only_northern_hemisphere: bool,
|
737 |
-
output_dir: str,
|
738 |
-
) -> None:
|
739 |
-
"""Saves rendered images with its camera matrix and metadata of the object.
|
740 |
-
|
741 |
-
Args:
|
742 |
-
object_file (str): Path to the object file.
|
743 |
-
num_renders (int): Number of renders to save of the object.
|
744 |
-
only_northern_hemisphere (bool): Whether to only render sides of the object that
|
745 |
-
are in the northern hemisphere. This is useful for rendering objects that
|
746 |
-
are photogrammetrically scanned, as the bottom of the object often has
|
747 |
-
holes.
|
748 |
-
output_dir (str): Path to the directory where the rendered images and metadata
|
749 |
-
will be saved.
|
750 |
-
|
751 |
-
Returns:
|
752 |
-
None
|
753 |
-
"""
|
754 |
-
os.makedirs(output_dir, exist_ok=True)
|
755 |
-
|
756 |
-
# load the object
|
757 |
-
if object_file.endswith(".blend"):
|
758 |
-
bpy.ops.object.mode_set(mode="OBJECT")
|
759 |
-
reset_cameras()
|
760 |
-
delete_invisible_objects()
|
761 |
-
else:
|
762 |
-
reset_scene()
|
763 |
-
load_object(object_file)
|
764 |
-
|
765 |
-
# Set up cameras
|
766 |
-
cam = scene.objects["Camera"]
|
767 |
-
cam.data.lens = 35
|
768 |
-
cam.data.sensor_width = 32
|
769 |
-
|
770 |
-
# Set up camera constraints
|
771 |
-
cam_constraint = cam.constraints.new(type="TRACK_TO")
|
772 |
-
cam_constraint.track_axis = "TRACK_NEGATIVE_Z"
|
773 |
-
cam_constraint.up_axis = "UP_Y"
|
774 |
-
empty = bpy.data.objects.new("Empty", None)
|
775 |
-
scene.collection.objects.link(empty)
|
776 |
-
cam_constraint.target = empty
|
777 |
-
|
778 |
-
# Extract the metadata. This must be done before normalizing the scene to get
|
779 |
-
# accurate bounding box information.
|
780 |
-
metadata_extractor = MetadataExtractor(
|
781 |
-
object_path=object_file, scene=scene, bdata=bpy.data
|
782 |
-
)
|
783 |
-
metadata = metadata_extractor.get_metadata()
|
784 |
-
|
785 |
-
# delete all objects that are not meshes
|
786 |
-
if object_file.lower().endswith(".usdz"):
|
787 |
-
# don't delete missing textures on usdz files, lots of them are embedded
|
788 |
-
missing_textures = None
|
789 |
-
else:
|
790 |
-
missing_textures = delete_missing_textures()
|
791 |
-
metadata["missing_textures"] = missing_textures
|
792 |
-
|
793 |
-
# possibly apply a random color to all objects
|
794 |
-
if object_file.endswith(".stl") or object_file.endswith(".ply"):
|
795 |
-
assert len(bpy.context.selected_objects) == 1
|
796 |
-
rand_color = apply_single_random_color_to_all_objects()
|
797 |
-
metadata["random_color"] = rand_color
|
798 |
-
else:
|
799 |
-
metadata["random_color"] = None
|
800 |
-
|
801 |
-
# save metadata
|
802 |
-
metadata_path = os.path.join(output_dir, "metadata.json")
|
803 |
-
os.makedirs(os.path.dirname(metadata_path), exist_ok=True)
|
804 |
-
with open(metadata_path, "w", encoding="utf-8") as f:
|
805 |
-
json.dump(metadata, f, sort_keys=True, indent=2)
|
806 |
-
|
807 |
-
# normalize the scene
|
808 |
-
normalize_scene()
|
809 |
-
|
810 |
-
# randomize the lighting
|
811 |
-
randomize_lighting()
|
812 |
-
|
813 |
-
# render the images
|
814 |
-
for i in range(num_renders):
|
815 |
-
# set camera
|
816 |
-
camera = randomize_camera(
|
817 |
-
only_northern_hemisphere=only_northern_hemisphere,
|
818 |
-
)
|
819 |
-
|
820 |
-
# render the image
|
821 |
-
render_path = os.path.join(output_dir, f"{i:03d}.png")
|
822 |
-
scene.render.filepath = render_path
|
823 |
-
bpy.ops.render.render(write_still=True)
|
824 |
-
|
825 |
-
# save camera RT matrix
|
826 |
-
rt_matrix = get_3x4_RT_matrix_from_blender(camera)
|
827 |
-
rt_matrix_path = os.path.join(output_dir, f"{i:03d}.npy")
|
828 |
-
np.save(rt_matrix_path, rt_matrix)
|
829 |
-
|
830 |
-
|
831 |
-
if __name__ == "__main__":
|
832 |
-
parser = argparse.ArgumentParser()
|
833 |
-
parser.add_argument(
|
834 |
-
"--object_path",
|
835 |
-
type=str,
|
836 |
-
required=True,
|
837 |
-
help="Path to the object file",
|
838 |
-
)
|
839 |
-
parser.add_argument(
|
840 |
-
"--output_dir",
|
841 |
-
type=str,
|
842 |
-
required=True,
|
843 |
-
help="Path to the directory where the rendered images and metadata will be saved.",
|
844 |
-
)
|
845 |
-
parser.add_argument(
|
846 |
-
"--engine",
|
847 |
-
type=str,
|
848 |
-
default="BLENDER_EEVEE",
|
849 |
-
choices=["CYCLES", "BLENDER_EEVEE"],
|
850 |
-
)
|
851 |
-
parser.add_argument(
|
852 |
-
"--only_northern_hemisphere",
|
853 |
-
action="store_true",
|
854 |
-
help="Only render the northern hemisphere of the object.",
|
855 |
-
default=False,
|
856 |
-
)
|
857 |
-
parser.add_argument(
|
858 |
-
"--num_renders",
|
859 |
-
type=int,
|
860 |
-
default=12,
|
861 |
-
help="Number of renders to save of the object.",
|
862 |
-
)
|
863 |
-
argv = sys.argv[sys.argv.index("--") + 1 :]
|
864 |
-
args = parser.parse_args(argv)
|
865 |
-
|
866 |
-
context = bpy.context
|
867 |
-
scene = context.scene
|
868 |
-
render = scene.render
|
869 |
-
|
870 |
-
# Set render settings
|
871 |
-
render.engine = args.engine
|
872 |
-
render.image_settings.file_format = "PNG"
|
873 |
-
render.image_settings.color_mode = "RGBA"
|
874 |
-
render.resolution_x = 512
|
875 |
-
render.resolution_y = 512
|
876 |
-
render.resolution_percentage = 100
|
877 |
-
|
878 |
-
# Set cycles settings
|
879 |
-
scene.cycles.device = "GPU"
|
880 |
-
scene.cycles.samples = 128
|
881 |
-
scene.cycles.diffuse_bounces = 1
|
882 |
-
scene.cycles.glossy_bounces = 1
|
883 |
-
scene.cycles.transparent_max_bounces = 3
|
884 |
-
scene.cycles.transmission_bounces = 3
|
885 |
-
scene.cycles.filter_width = 0.01
|
886 |
-
scene.cycles.use_denoising = True
|
887 |
-
scene.render.film_transparent = True
|
888 |
-
bpy.context.preferences.addons["cycles"].preferences.get_devices()
|
889 |
-
bpy.context.preferences.addons[
|
890 |
-
"cycles"
|
891 |
-
].preferences.compute_device_type = "CUDA" # or "OPENCL"
|
892 |
-
|
893 |
-
# Render the images
|
894 |
-
render_object(
|
895 |
-
object_file=args.object_path,
|
896 |
-
num_renders=args.num_renders,
|
897 |
-
only_northern_hemisphere=args.only_northern_hemisphere,
|
898 |
-
output_dir=args.output_dir,
|
899 |
-
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
scripts/rendering/example-objects.json
DELETED
@@ -1,52 +0,0 @@
|
|
1 |
-
[
|
2 |
-
{
|
3 |
-
"sha256": "d2b9a5d7c47dc93526082c9b630157ab6bce4fd8669610d942176f4a36444e71",
|
4 |
-
"fileIdentifier": "https://github.com/mattdeitke/objaverse-xl-test-files/blob/ead0bed6a76012452273bbe18d12e4d68a881956/example.obj",
|
5 |
-
"source": "github"
|
6 |
-
},
|
7 |
-
{
|
8 |
-
"sha256": "f15f7541614940dedd1ce373dee9cde1cc63471db84081259a44488aca267408",
|
9 |
-
"fileIdentifier": "https://github.com/mattdeitke/objaverse-xl-test-files/blob/ead0bed6a76012452273bbe18d12e4d68a881956/example-linked.gltf",
|
10 |
-
"source": "github"
|
11 |
-
},
|
12 |
-
{
|
13 |
-
"sha256": "cd567dfd1605a5fd60af1b8ac67e44b6af25b9e3f160da66047bb1187d02a071",
|
14 |
-
"fileIdentifier": "https://github.com/mattdeitke/objaverse-xl-test-files/blob/ead0bed6a76012452273bbe18d12e4d68a881956/example.dae",
|
15 |
-
"source": "github"
|
16 |
-
},
|
17 |
-
{
|
18 |
-
"sha256": "879bc9d2d85e4f3866f0cfef41f5236f9fff5f973380461af9f69cdbed53a0da",
|
19 |
-
"fileIdentifier": "https://github.com/mattdeitke/objaverse-xl-test-files/blob/ead0bed6a76012452273bbe18d12e4d68a881956/example.abc",
|
20 |
-
"source": "github"
|
21 |
-
},
|
22 |
-
{
|
23 |
-
"sha256": "ac69da6df0c83b593902c71238a7721f575233723730150230da95d583466562",
|
24 |
-
"fileIdentifier": "https://github.com/mattdeitke/objaverse-xl-test-files/blob/ead0bed6a76012452273bbe18d12e4d68a881956/example.blend",
|
25 |
-
"source": "github"
|
26 |
-
},
|
27 |
-
{
|
28 |
-
"sha256": "bc4cc8a78bc57d3c41ddaca9135d2b4e66b59e97846e2b796e5cc6b09c62b273",
|
29 |
-
"fileIdentifier": "https://github.com/mattdeitke/objaverse-xl-test-files/blob/ead0bed6a76012452273bbe18d12e4d68a881956/example-embedded.gltf",
|
30 |
-
"source": "github"
|
31 |
-
},
|
32 |
-
{
|
33 |
-
"sha256": "04e6377317d6818e32c5cbd1951e76deb3641bbf4f6db6933046221d5fbf1c5c",
|
34 |
-
"fileIdentifier": "https://github.com/mattdeitke/objaverse-xl-test-files/blob/ead0bed6a76012452273bbe18d12e4d68a881956/example.glb",
|
35 |
-
"source": "github"
|
36 |
-
},
|
37 |
-
{
|
38 |
-
"sha256": "7037575f47816118e5a34e7c0da9927e1be7be3f5b4adfac337710822eb50fa9",
|
39 |
-
"fileIdentifier": "https://github.com/mattdeitke/objaverse-xl-test-files/blob/ead0bed6a76012452273bbe18d12e4d68a881956/example.fbx",
|
40 |
-
"source": "github"
|
41 |
-
},
|
42 |
-
{
|
43 |
-
"sha256": "10fd5944c3fe32cb25fdcf46319cbda250b98bd22fd87fc41de5769b118d5262",
|
44 |
-
"fileIdentifier": "https://github.com/mattdeitke/objaverse-xl-test-files/blob/ead0bed6a76012452273bbe18d12e4d68a881956/example.stl",
|
45 |
-
"source": "github"
|
46 |
-
},
|
47 |
-
{
|
48 |
-
"sha256": "aa5c41753ad8a91562122d8cafbe1a940146fc3d227b7086eb8c830da36cb42b",
|
49 |
-
"fileIdentifier": "https://github.com/mattdeitke/objaverse-xl-test-files/blob/ead0bed6a76012452273bbe18d12e4d68a881956/example.ply",
|
50 |
-
"source": "github"
|
51 |
-
}
|
52 |
-
]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
scripts/rendering/main.py
DELETED
@@ -1,414 +0,0 @@
|
|
1 |
-
import glob
|
2 |
-
import json
|
3 |
-
import multiprocessing
|
4 |
-
import os
|
5 |
-
import platform
|
6 |
-
import random
|
7 |
-
import subprocess
|
8 |
-
import tempfile
|
9 |
-
import time
|
10 |
-
import zipfile
|
11 |
-
from functools import partial
|
12 |
-
from typing import Any, Dict, List, Literal, Optional, Union
|
13 |
-
|
14 |
-
import fire
|
15 |
-
import fsspec
|
16 |
-
import GPUtil
|
17 |
-
import pandas as pd
|
18 |
-
from loguru import logger
|
19 |
-
|
20 |
-
from objaverse_xl.objaverse_xl_downloader import ObjaverseXLDownloader
|
21 |
-
from objaverse_xl.utils import get_uid_from_str
|
22 |
-
|
23 |
-
|
24 |
-
def log_processed_object(csv_filename: str, *args) -> None:
|
25 |
-
"""Log when an object is done being used.
|
26 |
-
|
27 |
-
Args:
|
28 |
-
csv_filename (str): Name of the CSV file to save the logs to.
|
29 |
-
*args: Arguments to save to the CSV file.
|
30 |
-
|
31 |
-
Returns:
|
32 |
-
None
|
33 |
-
"""
|
34 |
-
args = ",".join([str(arg) for arg in args])
|
35 |
-
# log that this object was rendered successfully
|
36 |
-
# saving locally to avoid excessive writes to the cloud
|
37 |
-
dirname = os.path.expanduser(f"~/.objaverse/github/logs/")
|
38 |
-
os.makedirs(dirname, exist_ok=True)
|
39 |
-
with open(os.path.join(dirname, csv_filename), "a", encoding="utf-8") as f:
|
40 |
-
f.write(f"{time.time()},{args}\n")
|
41 |
-
|
42 |
-
|
43 |
-
def zipdir(path: str, ziph: zipfile.ZipFile) -> None:
|
44 |
-
"""Zip up a directory with an arcname structure.
|
45 |
-
|
46 |
-
Args:
|
47 |
-
path (str): Path to the directory to zip.
|
48 |
-
ziph (zipfile.ZipFile): ZipFile handler object to write to.
|
49 |
-
|
50 |
-
Returns:
|
51 |
-
None
|
52 |
-
"""
|
53 |
-
# ziph is zipfile handle
|
54 |
-
for root, dirs, files in os.walk(path):
|
55 |
-
for file in files:
|
56 |
-
# this ensures the structure inside the zip starts at folder/
|
57 |
-
arcname = os.path.join(os.path.basename(root), file)
|
58 |
-
ziph.write(os.path.join(root, file), arcname=arcname)
|
59 |
-
|
60 |
-
|
61 |
-
def handle_found_object(
|
62 |
-
local_path: str,
|
63 |
-
file_identifier: str,
|
64 |
-
sha256: str,
|
65 |
-
metadata: Dict[str, Any],
|
66 |
-
num_renders: int,
|
67 |
-
render_dir: str,
|
68 |
-
only_northern_hemisphere: bool,
|
69 |
-
gpu_devices: Union[int, List[int]],
|
70 |
-
render_timeout: int,
|
71 |
-
successful_log_file: Optional[str] = "handle-found-object-successful.csv",
|
72 |
-
failed_log_file: Optional[str] = "handle-found-object-failed.csv",
|
73 |
-
) -> bool:
|
74 |
-
"""Called when an object is successfully found and downloaded.
|
75 |
-
|
76 |
-
Here, the object has the same sha256 as the one that was downloaded with
|
77 |
-
Objaverse-XL. If None, the object will be downloaded, but nothing will be done with
|
78 |
-
it.
|
79 |
-
|
80 |
-
Args:
|
81 |
-
local_path (str): Local path to the downloaded 3D object.
|
82 |
-
file_identifier (str): GitHub URL of the 3D object.
|
83 |
-
sha256 (str): SHA256 of the contents of the 3D object.
|
84 |
-
metadata (Dict[str, Any]): Metadata about the 3D object, including keys for
|
85 |
-
`github_organization` and `github_repo`.
|
86 |
-
num_renders (int): Number of renders to save of the object.
|
87 |
-
render_dir (str): Directory where the objects will be rendered.
|
88 |
-
only_northern_hemisphere (bool): Only render the northern hemisphere of the
|
89 |
-
object.
|
90 |
-
gpu_devices (Union[int, List[int]]): GPU device(s) to use for rendering. If
|
91 |
-
an int, the GPU device will be randomly selected from 0 to gpu_devices - 1.
|
92 |
-
If a list, the GPU device will be randomly selected from the list.
|
93 |
-
If 0, the CPU will be used for rendering.
|
94 |
-
render_timeout (int): Number of seconds to wait for the rendering job to
|
95 |
-
complete.
|
96 |
-
successful_log_file (str): Name of the log file to save successful renders to.
|
97 |
-
failed_log_file (str): Name of the log file to save failed renders to.
|
98 |
-
|
99 |
-
Returns: True if the object was rendered successfully, False otherwise.
|
100 |
-
"""
|
101 |
-
save_uid = get_uid_from_str(file_identifier)
|
102 |
-
args = f"--object_path '{local_path}' --num_renders {num_renders}"
|
103 |
-
|
104 |
-
# get the GPU to use for rendering
|
105 |
-
using_gpu: bool = True
|
106 |
-
gpu_i = 0
|
107 |
-
if isinstance(gpu_devices, int) and gpu_devices > 0:
|
108 |
-
num_gpus = gpu_devices
|
109 |
-
gpu_i = random.randint(0, num_gpus - 1)
|
110 |
-
elif isinstance(gpu_devices, list):
|
111 |
-
gpu_i = random.choice(gpu_devices)
|
112 |
-
elif isinstance(gpu_devices, int) and gpu_devices == 0:
|
113 |
-
using_gpu = False
|
114 |
-
else:
|
115 |
-
raise ValueError(
|
116 |
-
f"gpu_devices must be an int > 0, 0, or a list of ints. Got {gpu_devices}."
|
117 |
-
)
|
118 |
-
|
119 |
-
with tempfile.TemporaryDirectory() as temp_dir:
|
120 |
-
# get the target directory for the rendering job
|
121 |
-
target_directory = os.path.join(temp_dir, save_uid)
|
122 |
-
os.makedirs(target_directory, exist_ok=True)
|
123 |
-
args += f" --output_dir {target_directory}"
|
124 |
-
|
125 |
-
# check for Linux / Ubuntu or MacOS
|
126 |
-
if platform.system() == "Linux" and using_gpu:
|
127 |
-
args += " --engine BLENDER_EEVEE"
|
128 |
-
elif platform.system() == "Darwin" or (
|
129 |
-
platform.system() == "Linux" and not using_gpu
|
130 |
-
):
|
131 |
-
# As far as I know, MacOS does not support BLENER_EEVEE, which uses GPU
|
132 |
-
# rendering. Generally, I'd only recommend using MacOS for debugging and
|
133 |
-
# small rendering jobs, since CYCLES is much slower than BLENDER_EEVEE.
|
134 |
-
args += " --engine CYCLES"
|
135 |
-
else:
|
136 |
-
raise NotImplementedError(f"Platform {platform.system()} is not supported.")
|
137 |
-
|
138 |
-
# check if we should only render the northern hemisphere
|
139 |
-
if only_northern_hemisphere:
|
140 |
-
args += " --only_northern_hemisphere"
|
141 |
-
|
142 |
-
# get the command to run
|
143 |
-
command = f"blender-3.2.2-linux-x64/blender --background --python blender_script.py -- {args}"
|
144 |
-
if using_gpu:
|
145 |
-
command = f"export DISPLAY=:0.{gpu_i} && {command}"
|
146 |
-
|
147 |
-
# render the object (put in dev null)
|
148 |
-
subprocess.run(
|
149 |
-
["bash", "-c", command],
|
150 |
-
timeout=render_timeout,
|
151 |
-
check=False,
|
152 |
-
stdout=subprocess.DEVNULL,
|
153 |
-
stderr=subprocess.DEVNULL,
|
154 |
-
)
|
155 |
-
|
156 |
-
# check that the renders were saved successfully
|
157 |
-
png_files = glob.glob(os.path.join(target_directory, "*.png"))
|
158 |
-
metadata_files = glob.glob(os.path.join(target_directory, "*.json"))
|
159 |
-
npy_files = glob.glob(os.path.join(target_directory, "*.npy"))
|
160 |
-
if (
|
161 |
-
(len(png_files) != num_renders)
|
162 |
-
or (len(npy_files) != num_renders)
|
163 |
-
or (len(metadata_files) != 1)
|
164 |
-
):
|
165 |
-
logger.error(
|
166 |
-
f"Found object {file_identifier} was not rendered successfully!"
|
167 |
-
)
|
168 |
-
if failed_log_file is not None:
|
169 |
-
log_processed_object(
|
170 |
-
failed_log_file,
|
171 |
-
file_identifier,
|
172 |
-
sha256,
|
173 |
-
)
|
174 |
-
return False
|
175 |
-
|
176 |
-
# update the metadata
|
177 |
-
metadata_path = os.path.join(target_directory, "metadata.json")
|
178 |
-
with open(metadata_path, "r", encoding="utf-8") as f:
|
179 |
-
metadata_file = json.load(f)
|
180 |
-
metadata_file["sha256"] = sha256
|
181 |
-
metadata_file["file_identifier"] = file_identifier
|
182 |
-
metadata_file["save_uid"] = save_uid
|
183 |
-
metadata_file["metadata"] = metadata
|
184 |
-
with open(metadata_path, "w", encoding="utf-8") as f:
|
185 |
-
json.dump(metadata, f, indent=2, sort_keys=True)
|
186 |
-
|
187 |
-
# Make a zip of the target_directory.
|
188 |
-
# Keeps the {save_uid} directory structure when unzipped
|
189 |
-
with zipfile.ZipFile(
|
190 |
-
f"{target_directory}.zip", "w", zipfile.ZIP_DEFLATED
|
191 |
-
) as ziph:
|
192 |
-
zipdir(target_directory, ziph)
|
193 |
-
|
194 |
-
# move the zip to the render_dir
|
195 |
-
fs, path = fsspec.core.url_to_fs(render_dir)
|
196 |
-
|
197 |
-
# move the zip to the render_dir
|
198 |
-
fs.makedirs(os.path.join(path, "github", "renders"), exist_ok=True)
|
199 |
-
fs.put(
|
200 |
-
os.path.join(f"{target_directory}.zip"),
|
201 |
-
os.path.join(path, "github", "renders", f"{save_uid}.zip"),
|
202 |
-
)
|
203 |
-
|
204 |
-
# log that this object was rendered successfully
|
205 |
-
if successful_log_file is not None:
|
206 |
-
log_processed_object(successful_log_file, file_identifier, sha256)
|
207 |
-
|
208 |
-
return True
|
209 |
-
|
210 |
-
|
211 |
-
def handle_new_object(
|
212 |
-
local_path: str,
|
213 |
-
file_identifier: str,
|
214 |
-
sha256: str,
|
215 |
-
metadata: Dict[str, Any],
|
216 |
-
log_file: str = "handle-new-object.csv",
|
217 |
-
) -> None:
|
218 |
-
"""Called when a new object is found.
|
219 |
-
|
220 |
-
Here, the object is not used in Objaverse-XL, but is still downloaded with the
|
221 |
-
repository. The object may have not been used because it does not successfully
|
222 |
-
import into Blender. If None, the object will be downloaded, but nothing will be
|
223 |
-
done with it.
|
224 |
-
|
225 |
-
Args:
|
226 |
-
local_path (str): Local path to the downloaded 3D object.
|
227 |
-
file_identifier (str): GitHub URL of the 3D object.
|
228 |
-
sha256 (str): SHA256 of the contents of the 3D object.
|
229 |
-
metadata (Dict[str, Any]): Metadata about the 3D object, including the GitHub
|
230 |
-
organization and repo names.
|
231 |
-
log_file (str): Name of the log file to save the handle_new_object logs to.
|
232 |
-
|
233 |
-
Returns:
|
234 |
-
None
|
235 |
-
"""
|
236 |
-
# log the new object
|
237 |
-
log_processed_object(log_file, file_identifier, sha256)
|
238 |
-
|
239 |
-
|
240 |
-
def handle_modified_object(
|
241 |
-
local_path: str,
|
242 |
-
file_identifier: str,
|
243 |
-
new_sha256: str,
|
244 |
-
old_sha256: str,
|
245 |
-
metadata: Dict[str, Any],
|
246 |
-
num_renders: int,
|
247 |
-
render_dir: str,
|
248 |
-
only_northern_hemisphere: bool,
|
249 |
-
gpu_devices: Union[int, List[int]],
|
250 |
-
render_timeout: int,
|
251 |
-
) -> None:
|
252 |
-
"""Called when a modified object is found and downloaded.
|
253 |
-
|
254 |
-
Here, the object is successfully downloaded, but it has a different sha256 than the
|
255 |
-
one that was downloaded with Objaverse-XL. This is not expected to happen very
|
256 |
-
often, because the same commit hash is used for each repo. If None, the object will
|
257 |
-
be downloaded, but nothing will be done with it.
|
258 |
-
|
259 |
-
Args:
|
260 |
-
local_path (str): Local path to the downloaded 3D object.
|
261 |
-
file_identifier (str): GitHub URL of the 3D object.
|
262 |
-
new_sha256 (str): SHA256 of the contents of the newly downloaded 3D object.
|
263 |
-
old_sha256 (str): Expected SHA256 of the contents of the 3D object as it was
|
264 |
-
when it was downloaded with Objaverse-XL.
|
265 |
-
metadata (Dict[str, Any]): Metadata about the 3D object, including the GitHub
|
266 |
-
organization and repo names.
|
267 |
-
num_renders (int): Number of renders to save of the object.
|
268 |
-
render_dir (str): Directory where the objects will be rendered.
|
269 |
-
only_northern_hemisphere (bool): Only render the northern hemisphere of the
|
270 |
-
object.
|
271 |
-
gpu_devices (Union[int, List[int]]): GPU device(s) to use for rendering. If
|
272 |
-
an int, the GPU device will be randomly selected from 0 to gpu_devices - 1.
|
273 |
-
If a list, the GPU device will be randomly selected from the list.
|
274 |
-
If 0, the CPU will be used for rendering.
|
275 |
-
render_timeout (int): Number of seconds to wait for the rendering job to
|
276 |
-
complete.
|
277 |
-
|
278 |
-
Returns:
|
279 |
-
None
|
280 |
-
"""
|
281 |
-
success = handle_found_object(
|
282 |
-
local_path=local_path,
|
283 |
-
file_identifier=file_identifier,
|
284 |
-
sha256=new_sha256,
|
285 |
-
metadata=metadata,
|
286 |
-
num_renders=num_renders,
|
287 |
-
render_dir=render_dir,
|
288 |
-
only_northern_hemisphere=only_northern_hemisphere,
|
289 |
-
gpu_devices=gpu_devices,
|
290 |
-
render_timeout=render_timeout,
|
291 |
-
successful_log_file=None,
|
292 |
-
failed_log_file=None,
|
293 |
-
)
|
294 |
-
|
295 |
-
if success:
|
296 |
-
log_processed_object(
|
297 |
-
"handle-modified-object-successful.csv",
|
298 |
-
file_identifier,
|
299 |
-
old_sha256,
|
300 |
-
new_sha256,
|
301 |
-
)
|
302 |
-
else:
|
303 |
-
log_processed_object(
|
304 |
-
"handle-modified-object-failed.csv",
|
305 |
-
file_identifier,
|
306 |
-
old_sha256,
|
307 |
-
new_sha256,
|
308 |
-
)
|
309 |
-
|
310 |
-
|
311 |
-
def handle_missing_object(
|
312 |
-
github_url: str,
|
313 |
-
sha256: str,
|
314 |
-
metadata: Dict[str, Any],
|
315 |
-
log_file: str = "handle-missing-object.csv",
|
316 |
-
) -> None:
|
317 |
-
"""Called when an object that is in Objaverse-XL is not found.
|
318 |
-
|
319 |
-
Here, it is likely that the repository was deleted or renamed. If None, nothing
|
320 |
-
will be done with the missing object.
|
321 |
-
|
322 |
-
Args:
|
323 |
-
github_url (str): GitHub URL of the 3D object.
|
324 |
-
sha256 (str): SHA256 of the contents of the original 3D object.
|
325 |
-
metadata (Dict[str, Any]): Metadata about the 3D object, including the GitHub
|
326 |
-
organization and repo names.
|
327 |
-
log_file (str): Name of the log file to save missing renders to.
|
328 |
-
|
329 |
-
Returns:
|
330 |
-
None
|
331 |
-
"""
|
332 |
-
# log the missing object
|
333 |
-
log_processed_object(log_file, github_url, sha256)
|
334 |
-
|
335 |
-
|
336 |
-
def get_example_objects() -> pd.DataFrame:
|
337 |
-
"""Returns a DataFrame of example objects to use for debugging."""
|
338 |
-
return pd.read_json("example-objects.json", orient="records")
|
339 |
-
|
340 |
-
|
341 |
-
def render_objects(
|
342 |
-
render_dir: str = "~/.objaverse",
|
343 |
-
num_renders: int = 12,
|
344 |
-
processes: Optional[int] = None,
|
345 |
-
save_repo_format: Optional[Literal["zip", "tar", "tar.gz"]] = None,
|
346 |
-
only_northern_hemisphere: bool = False,
|
347 |
-
render_timeout: int = 300,
|
348 |
-
gpu_devices: Optional[Union[int, List[int]]] = None,
|
349 |
-
) -> None:
|
350 |
-
"""Renders all GitHub objects in the Objaverse-XL dataset.
|
351 |
-
|
352 |
-
Args:
|
353 |
-
render_dir (str): Directory where the objects will be rendered.
|
354 |
-
num_renders (int): Number of renders to save of the object.
|
355 |
-
processes (Optional[int]): Number of processes to use for downloading the
|
356 |
-
objects. If None, defaults to multiprocessing.cpu_count() * 3.
|
357 |
-
save_repo_format (Optional[Literal["zip", "tar", "tar.gz"]]): If not None,
|
358 |
-
the GitHub repo will be deleted after rendering each object from it.
|
359 |
-
only_northern_hemisphere (bool): Only render the northern hemisphere of the
|
360 |
-
object. Useful for rendering objects that are obtained from photogrammetry,
|
361 |
-
since the southern hemisphere is often has holes.
|
362 |
-
render_timeout (int): Number of seconds to wait for the rendering job to
|
363 |
-
complete.
|
364 |
-
gpu_devices (Optional[Union[int, List[int]]]): GPU device(s) to use for
|
365 |
-
rendering. If an int, the GPU device will be randomly selected from 0 to
|
366 |
-
gpu_devices - 1. If a list, the GPU device will be randomly selected from
|
367 |
-
the list. If 0, the CPU will be used for rendering. If None, defaults to
|
368 |
-
use all available GPUs.
|
369 |
-
|
370 |
-
Returns:
|
371 |
-
None
|
372 |
-
"""
|
373 |
-
if platform.system() not in ["Linux", "Darwin"]:
|
374 |
-
raise NotImplementedError(
|
375 |
-
f"Platform {platform.system()} is not supported. Use Linux or MacOS."
|
376 |
-
)
|
377 |
-
|
378 |
-
# get the gpu devices to use
|
379 |
-
parsed_gpu_devices: Union[int, List[int]] = 0
|
380 |
-
if gpu_devices is None:
|
381 |
-
parsed_gpu_devices = len(GPUtil.getGPUs())
|
382 |
-
|
383 |
-
if processes is None:
|
384 |
-
processes = multiprocessing.cpu_count() * 3
|
385 |
-
|
386 |
-
objects = get_example_objects()
|
387 |
-
objaverse_xl_downloader = ObjaverseXLDownloader()
|
388 |
-
objaverse_xl_downloader.download_objects(
|
389 |
-
objects=objects,
|
390 |
-
processes=processes,
|
391 |
-
save_repo_format=save_repo_format,
|
392 |
-
download_dir=render_dir, # only used when save_repo_format is not None
|
393 |
-
handle_found_object=partial(
|
394 |
-
handle_found_object,
|
395 |
-
render_dir=render_dir,
|
396 |
-
num_renders=num_renders,
|
397 |
-
only_northern_hemisphere=only_northern_hemisphere,
|
398 |
-
gpu_devices=parsed_gpu_devices,
|
399 |
-
render_timeout=render_timeout,
|
400 |
-
),
|
401 |
-
handle_new_object=handle_new_object,
|
402 |
-
handle_modified_object=partial(
|
403 |
-
handle_modified_object,
|
404 |
-
render_dir=render_dir,
|
405 |
-
num_renders=num_renders,
|
406 |
-
gpu_devices=parsed_gpu_devices,
|
407 |
-
only_northern_hemisphere=only_northern_hemisphere,
|
408 |
-
),
|
409 |
-
handle_missing_object=handle_missing_object,
|
410 |
-
)
|
411 |
-
|
412 |
-
|
413 |
-
if __name__ == "__main__":
|
414 |
-
fire.Fire(render_objects)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
scripts/rendering/start_x_server.py
DELETED
@@ -1,272 +0,0 @@
|
|
1 |
-
# Taken from https://github.com/allenai/ai2thor/blob/main/scripts/ai2thor-xorg
|
2 |
-
# Starts an x-server to support running Blender on a headless machine with
|
3 |
-
# dedicated NVIDIA GPUs
|
4 |
-
|
5 |
-
import argparse
|
6 |
-
|
7 |
-
#!/usr/bin/env python3
|
8 |
-
import os
|
9 |
-
import platform
|
10 |
-
import re
|
11 |
-
import shlex
|
12 |
-
import signal
|
13 |
-
import subprocess
|
14 |
-
import sys
|
15 |
-
import time
|
16 |
-
|
17 |
-
# Turning off automatic black formatting for this script as it breaks quotes.
|
18 |
-
# fmt: off
|
19 |
-
from typing import List
|
20 |
-
|
21 |
-
PID_FILE = "/var/run/ai2thor-xorg.pid"
|
22 |
-
CONFIG_FILE = "/tmp/ai2thor-xorg.conf"
|
23 |
-
|
24 |
-
DEFAULT_HEIGHT = 768
|
25 |
-
DEFAULT_WIDTH = 1024
|
26 |
-
|
27 |
-
|
28 |
-
def process_alive(pid):
|
29 |
-
"""
|
30 |
-
Use kill(0) to determine if pid is alive
|
31 |
-
:param pid: process id
|
32 |
-
:rtype: bool
|
33 |
-
"""
|
34 |
-
try:
|
35 |
-
os.kill(pid, 0)
|
36 |
-
except OSError:
|
37 |
-
return False
|
38 |
-
|
39 |
-
return True
|
40 |
-
|
41 |
-
|
42 |
-
def find_devices(excluded_device_ids):
|
43 |
-
devices = []
|
44 |
-
id_counter = 0
|
45 |
-
for r in pci_records():
|
46 |
-
if r.get("Vendor", "") == "NVIDIA Corporation" and r["Class"] in [
|
47 |
-
"VGA compatible controller",
|
48 |
-
"3D controller",
|
49 |
-
]:
|
50 |
-
bus_id = "PCI:" + ":".join(
|
51 |
-
map(lambda x: str(int(x, 16)), re.split(r"[:\.]", r["Slot"]))
|
52 |
-
)
|
53 |
-
|
54 |
-
if id_counter not in excluded_device_ids:
|
55 |
-
devices.append(bus_id)
|
56 |
-
|
57 |
-
id_counter += 1
|
58 |
-
|
59 |
-
if not devices:
|
60 |
-
print("Error: ai2thor-xorg requires at least one NVIDIA device")
|
61 |
-
sys.exit(1)
|
62 |
-
|
63 |
-
return devices
|
64 |
-
|
65 |
-
def active_display_bus_ids():
|
66 |
-
# this determines whether a monitor is connected to the GPU
|
67 |
-
# if one is, the following Option is added for the Screen "UseDisplayDevice" "None"
|
68 |
-
command = "nvidia-smi --query-gpu=pci.bus_id,display_active --format=csv,noheader"
|
69 |
-
active_bus_ids = set()
|
70 |
-
result = subprocess.run(command, shell=True, stdout=subprocess.PIPE)
|
71 |
-
if result.returncode == 0:
|
72 |
-
for line in result.stdout.decode().strip().split("\n"):
|
73 |
-
nvidia_bus_id, display_status = re.split(r",\s?", line.strip())
|
74 |
-
bus_id = "PCI:" + ":".join(
|
75 |
-
map(lambda x: str(int(x, 16)), re.split(r"[:\.]", nvidia_bus_id)[1:])
|
76 |
-
)
|
77 |
-
if display_status.lower() == "enabled":
|
78 |
-
active_bus_ids.add(bus_id)
|
79 |
-
|
80 |
-
return active_bus_ids
|
81 |
-
|
82 |
-
def pci_records():
|
83 |
-
records = []
|
84 |
-
command = shlex.split("lspci -vmm")
|
85 |
-
output = subprocess.check_output(command).decode()
|
86 |
-
|
87 |
-
for devices in output.strip().split("\n\n"):
|
88 |
-
record = {}
|
89 |
-
records.append(record)
|
90 |
-
for row in devices.split("\n"):
|
91 |
-
key, value = row.split("\t")
|
92 |
-
record[key.split(":")[0]] = value
|
93 |
-
|
94 |
-
return records
|
95 |
-
|
96 |
-
|
97 |
-
def read_pid():
|
98 |
-
if os.path.isfile(PID_FILE):
|
99 |
-
with open(PID_FILE) as f:
|
100 |
-
return int(f.read())
|
101 |
-
else:
|
102 |
-
return None
|
103 |
-
|
104 |
-
|
105 |
-
def start(display: str, excluded_device_ids: List[int], width: int, height: int):
|
106 |
-
pid = read_pid()
|
107 |
-
|
108 |
-
if pid and process_alive(pid):
|
109 |
-
print("Error: ai2thor-xorg is already running with pid: %s" % pid)
|
110 |
-
sys.exit(1)
|
111 |
-
|
112 |
-
with open(CONFIG_FILE, "w") as f:
|
113 |
-
f.write(generate_xorg_conf(excluded_device_ids, width=width, height=height))
|
114 |
-
|
115 |
-
log_file = "/var/log/ai2thor-xorg.%s.log" % display
|
116 |
-
error_log_file = "/var/log/ai2thor-xorg-error.%s.log" % display
|
117 |
-
command = shlex.split(
|
118 |
-
"Xorg -quiet -maxclients 1024 -noreset +extension GLX +extension RANDR +extension RENDER -logfile %s -config %s :%s"
|
119 |
-
% (log_file, CONFIG_FILE, display)
|
120 |
-
)
|
121 |
-
|
122 |
-
pid = None
|
123 |
-
with open(error_log_file, "w") as error_log_f:
|
124 |
-
proc = subprocess.Popen(command, stdin=subprocess.DEVNULL, stdout=subprocess.DEVNULL, stderr=error_log_f)
|
125 |
-
pid = proc.pid
|
126 |
-
try:
|
127 |
-
proc.wait(timeout=0.25)
|
128 |
-
except subprocess.TimeoutExpired:
|
129 |
-
pass
|
130 |
-
|
131 |
-
if pid and process_alive(pid):
|
132 |
-
with open(PID_FILE, "w") as f:
|
133 |
-
f.write(str(proc.pid))
|
134 |
-
else:
|
135 |
-
print("Error: error with command '%s'" % " ".join(command))
|
136 |
-
with open(error_log_file, "r") as f:
|
137 |
-
print(f.read())
|
138 |
-
|
139 |
-
|
140 |
-
def print_config(excluded_device_ids: List[int], width: int, height: int):
|
141 |
-
print(generate_xorg_conf(excluded_device_ids, width=width, height=height))
|
142 |
-
|
143 |
-
|
144 |
-
def stop():
|
145 |
-
pid = read_pid()
|
146 |
-
if pid and process_alive(pid):
|
147 |
-
os.kill(pid, signal.SIGTERM)
|
148 |
-
|
149 |
-
for i in range(10):
|
150 |
-
time.sleep(0.2)
|
151 |
-
if not process_alive(pid):
|
152 |
-
os.unlink(PID_FILE)
|
153 |
-
break
|
154 |
-
|
155 |
-
|
156 |
-
def generate_xorg_conf(
|
157 |
-
excluded_device_ids: List[int], width: int, height: int
|
158 |
-
):
|
159 |
-
devices = find_devices(excluded_device_ids)
|
160 |
-
active_display_devices = active_display_bus_ids()
|
161 |
-
|
162 |
-
xorg_conf = []
|
163 |
-
|
164 |
-
device_section = """
|
165 |
-
Section "Device"
|
166 |
-
Identifier "Device{device_id}"
|
167 |
-
Driver "nvidia"
|
168 |
-
VendorName "NVIDIA Corporation"
|
169 |
-
BusID "{bus_id}"
|
170 |
-
EndSection
|
171 |
-
"""
|
172 |
-
server_layout_section = """
|
173 |
-
Section "ServerLayout"
|
174 |
-
Identifier "Layout0"
|
175 |
-
{screen_records}
|
176 |
-
EndSection
|
177 |
-
"""
|
178 |
-
screen_section = """
|
179 |
-
Section "Screen"
|
180 |
-
Identifier "Screen{screen_id}"
|
181 |
-
Device "Device{device_id}"
|
182 |
-
DefaultDepth 24
|
183 |
-
Option "AllowEmptyInitialConfiguration" "True"
|
184 |
-
Option "Interactive" "False"
|
185 |
-
{extra_options}
|
186 |
-
SubSection "Display"
|
187 |
-
Depth 24
|
188 |
-
Virtual {width} {height}
|
189 |
-
EndSubSection
|
190 |
-
EndSection
|
191 |
-
"""
|
192 |
-
screen_records = []
|
193 |
-
for i, bus_id in enumerate(devices):
|
194 |
-
extra_options = ""
|
195 |
-
if bus_id in active_display_devices:
|
196 |
-
# See https://github.com/allenai/ai2thor/pull/990
|
197 |
-
# when a monitor is connected, this option must be used otherwise
|
198 |
-
# Xorg will fail to start
|
199 |
-
extra_options = 'Option "UseDisplayDevice" "None"'
|
200 |
-
xorg_conf.append(device_section.format(device_id=i, bus_id=bus_id))
|
201 |
-
xorg_conf.append(screen_section.format(device_id=i, screen_id=i, width=width, height=height, extra_options=extra_options))
|
202 |
-
screen_records.append(
|
203 |
-
'Screen {screen_id} "Screen{screen_id}" 0 0'.format(screen_id=i)
|
204 |
-
)
|
205 |
-
|
206 |
-
xorg_conf.append(
|
207 |
-
server_layout_section.format(screen_records="\n ".join(screen_records))
|
208 |
-
)
|
209 |
-
|
210 |
-
output = "\n".join(xorg_conf)
|
211 |
-
return output
|
212 |
-
|
213 |
-
|
214 |
-
# fmt: on
|
215 |
-
|
216 |
-
if __name__ == "__main__":
|
217 |
-
if os.geteuid() != 0:
|
218 |
-
path = os.path.abspath(__file__)
|
219 |
-
print("Executing ai2thor-xorg with sudo")
|
220 |
-
args = ["--", path] + sys.argv[1:]
|
221 |
-
os.execvp("sudo", args)
|
222 |
-
|
223 |
-
if platform.system() != "Linux":
|
224 |
-
print("Error: Can only run ai2thor-xorg on linux")
|
225 |
-
sys.exit(1)
|
226 |
-
|
227 |
-
parser = argparse.ArgumentParser()
|
228 |
-
parser.add_argument(
|
229 |
-
"--exclude-device",
|
230 |
-
help="exclude a specific GPU device",
|
231 |
-
action="append",
|
232 |
-
type=int,
|
233 |
-
default=[],
|
234 |
-
)
|
235 |
-
parser.add_argument(
|
236 |
-
"--width",
|
237 |
-
help="width of the screen to start (should be greater than the maximum"
|
238 |
-
f" width of any ai2thor instance you will start) [default: {DEFAULT_WIDTH}]",
|
239 |
-
type=int,
|
240 |
-
default=DEFAULT_WIDTH,
|
241 |
-
)
|
242 |
-
parser.add_argument(
|
243 |
-
"--height",
|
244 |
-
help="height of the screen to start (should be greater than the maximum"
|
245 |
-
f" height of any ai2thor instance you will start) [default: {DEFAULT_HEIGHT}]",
|
246 |
-
type=int,
|
247 |
-
default=DEFAULT_HEIGHT,
|
248 |
-
)
|
249 |
-
parser.add_argument(
|
250 |
-
"command",
|
251 |
-
help="command to be executed",
|
252 |
-
choices=["start", "stop", "print-config"],
|
253 |
-
)
|
254 |
-
parser.add_argument(
|
255 |
-
"display", help="display to be used", nargs="?", type=int, default=0
|
256 |
-
)
|
257 |
-
args = parser.parse_args()
|
258 |
-
if args.command == "start":
|
259 |
-
start(
|
260 |
-
display=args.display,
|
261 |
-
excluded_device_ids=args.exclude_device,
|
262 |
-
height=args.height,
|
263 |
-
width=args.width,
|
264 |
-
)
|
265 |
-
elif args.command == "stop":
|
266 |
-
stop()
|
267 |
-
elif args.command == "print-config":
|
268 |
-
print_config(
|
269 |
-
excluded_device_ids=args.exclude_device,
|
270 |
-
width=args.width,
|
271 |
-
height=args.height,
|
272 |
-
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
setup.py
DELETED
@@ -1,27 +0,0 @@
|
|
1 |
-
from setuptools import find_packages, setup
|
2 |
-
|
3 |
-
# Read the README content
|
4 |
-
with open("README.md", "r", encoding="utf-8") as fh:
|
5 |
-
long_description = fh.read()
|
6 |
-
|
7 |
-
# Read requirements from requirements.txt
|
8 |
-
with open("requirements.txt", "r", encoding="utf-8") as f:
|
9 |
-
requirements = [line.strip() for line in f.readlines()]
|
10 |
-
|
11 |
-
setup(
|
12 |
-
name="objaverse_xl",
|
13 |
-
version="0.1.1",
|
14 |
-
author="Allen Institute for AI",
|
15 |
-
author_email="mattd@allenai.org",
|
16 |
-
description="Objaverse-XL is an open dataset of over 10 million 3D objects",
|
17 |
-
long_description=long_description,
|
18 |
-
long_description_content_type="text/markdown",
|
19 |
-
url="https://huggingface.co/datasets/allenai/objaverse-xl",
|
20 |
-
packages=find_packages(),
|
21 |
-
classifiers=[
|
22 |
-
"Programming Language :: Python :: 3",
|
23 |
-
"License :: OSI Approved :: Apache Software License",
|
24 |
-
],
|
25 |
-
install_requires=requirements,
|
26 |
-
extras_require={"s3": ["s3fs==0.4.2"]},
|
27 |
-
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
objaverse_v1/object-metadata.parquet → sketchfab/sketchfab.parquet
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6462ef2ecf39069e210f6df1416e8a28ca931039e3c7fe6aa3dca593115c64e3
|
3 |
+
size 79625309
|
scripts/rendering/io_scene_usdz.zip → smithsonian/smithsonian.parquet
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e8dc6d4eb8035f174dc518449f6eea3e5d5d55084b9891371db8ae0141259ae5
|
3 |
+
size 313114
|
tests/test_api.py
DELETED
@@ -1,553 +0,0 @@
|
|
1 |
-
import os
|
2 |
-
import shutil
|
3 |
-
|
4 |
-
import fsspec
|
5 |
-
import pandas as pd
|
6 |
-
|
7 |
-
from objaverse_xl.github import GitHubDownloader
|
8 |
-
from objaverse_xl.objaverse_v1 import SketchfabDownloader
|
9 |
-
from objaverse_xl.thingiverse import ThingiverseDownloader
|
10 |
-
from objaverse_xl.smithsonian import SmithsonianDownloader
|
11 |
-
|
12 |
-
|
13 |
-
def test_github_process_repo():
|
14 |
-
github_downloader = GitHubDownloader()
|
15 |
-
download_dir = "~/.objaverse-tests"
|
16 |
-
base_download_dir = os.path.join(download_dir, "github")
|
17 |
-
fs, path = fsspec.core.url_to_fs(base_download_dir)
|
18 |
-
fs.makedirs(path, exist_ok=True)
|
19 |
-
|
20 |
-
new_objects = []
|
21 |
-
handle_new_object = (
|
22 |
-
lambda local_path, file_identifier, sha256, metadata: new_objects.append(
|
23 |
-
dict(
|
24 |
-
local_path=local_path,
|
25 |
-
file_identifier=file_identifier,
|
26 |
-
sha256=sha256,
|
27 |
-
metadata=metadata,
|
28 |
-
)
|
29 |
-
)
|
30 |
-
)
|
31 |
-
|
32 |
-
for save_repo_format in ["tar", "tar.gz", "zip", "files"]:
|
33 |
-
shutil.rmtree(os.path.join(path, "repos"), ignore_errors=True)
|
34 |
-
out = github_downloader._process_repo(
|
35 |
-
repo_id="mattdeitke/objaverse-xl-test-files",
|
36 |
-
fs=fs,
|
37 |
-
base_dir=path,
|
38 |
-
save_repo_format=save_repo_format, # type: ignore
|
39 |
-
expected_objects=dict(),
|
40 |
-
handle_found_object=None,
|
41 |
-
handle_modified_object=None,
|
42 |
-
handle_missing_object=None,
|
43 |
-
handle_new_object=handle_new_object,
|
44 |
-
commit_hash="6928b08a2501aa7a4a4aabac1f888b66e7782056",
|
45 |
-
)
|
46 |
-
|
47 |
-
# test that the sha256's are correct
|
48 |
-
assert len(out) == 0
|
49 |
-
sha256s = [x["sha256"] for x in new_objects]
|
50 |
-
for sha256 in [
|
51 |
-
"d2b9a5d7c47dc93526082c9b630157ab6bce4fd8669610d942176f4a36444e71",
|
52 |
-
"04e6377317d6818e32c5cbd1951e76deb3641bbf4f6db6933046221d5fbf1c5c",
|
53 |
-
"7037575f47816118e5a34e7c0da9927e1be7be3f5b4adfac337710822eb50fa9",
|
54 |
-
]:
|
55 |
-
assert sha256 in sha256s, f"{sha256=} not in {sha256s=}"
|
56 |
-
github_urls = [x["file_identifier"] for x in new_objects]
|
57 |
-
for github_url in [
|
58 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.fbx",
|
59 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.glb",
|
60 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.obj",
|
61 |
-
]:
|
62 |
-
assert github_url in github_urls, f"{github_url=} not in {github_urls=}"
|
63 |
-
|
64 |
-
# test that the files are correct
|
65 |
-
if save_repo_format != "files":
|
66 |
-
assert fs.exists(
|
67 |
-
os.path.join(
|
68 |
-
path,
|
69 |
-
"repos",
|
70 |
-
"mattdeitke",
|
71 |
-
f"objaverse-xl-test-files.{save_repo_format}",
|
72 |
-
)
|
73 |
-
)
|
74 |
-
else:
|
75 |
-
assert fs.exists(
|
76 |
-
os.path.join(
|
77 |
-
base_download_dir, "repos", "mattdeitke", "objaverse-xl-test-files"
|
78 |
-
)
|
79 |
-
)
|
80 |
-
|
81 |
-
|
82 |
-
def test_github_handle_new_object():
|
83 |
-
github_downloader = GitHubDownloader()
|
84 |
-
found_objects = []
|
85 |
-
handle_found_object = (
|
86 |
-
lambda local_path, file_identifier, sha256, metadata: found_objects.append(
|
87 |
-
dict(
|
88 |
-
local_path=local_path,
|
89 |
-
file_identifier=file_identifier,
|
90 |
-
sha256=sha256,
|
91 |
-
metadata=metadata,
|
92 |
-
)
|
93 |
-
)
|
94 |
-
)
|
95 |
-
|
96 |
-
missing_objects = []
|
97 |
-
handle_missing_object = (
|
98 |
-
lambda file_identifier, sha256, metadata: missing_objects.append(
|
99 |
-
dict(
|
100 |
-
file_identifier=file_identifier,
|
101 |
-
sha256=sha256,
|
102 |
-
metadata=metadata,
|
103 |
-
)
|
104 |
-
)
|
105 |
-
)
|
106 |
-
|
107 |
-
new_objects = []
|
108 |
-
handle_new_object = (
|
109 |
-
lambda local_path, file_identifier, sha256, metadata: new_objects.append(
|
110 |
-
dict(
|
111 |
-
local_path=local_path,
|
112 |
-
file_identifier=file_identifier,
|
113 |
-
sha256=sha256,
|
114 |
-
metadata=metadata,
|
115 |
-
)
|
116 |
-
)
|
117 |
-
)
|
118 |
-
|
119 |
-
modified_objects = []
|
120 |
-
handle_modified_object = lambda local_path, file_identifier, new_sha256, old_sha256, metadata: modified_objects.append(
|
121 |
-
dict(
|
122 |
-
local_path=local_path,
|
123 |
-
file_identifier=file_identifier,
|
124 |
-
new_sha256=new_sha256,
|
125 |
-
old_sha256=old_sha256,
|
126 |
-
metadata=metadata,
|
127 |
-
)
|
128 |
-
)
|
129 |
-
|
130 |
-
download_dir = "~/.objaverse-tests"
|
131 |
-
base_download_dir = os.path.join(download_dir, "github")
|
132 |
-
fs, path = fsspec.core.url_to_fs(base_download_dir)
|
133 |
-
fs.makedirs(path, exist_ok=True)
|
134 |
-
|
135 |
-
shutil.rmtree(os.path.join(path, "repos"), ignore_errors=True)
|
136 |
-
out = github_downloader._process_repo(
|
137 |
-
repo_id="mattdeitke/objaverse-xl-test-files",
|
138 |
-
fs=fs,
|
139 |
-
base_dir=path,
|
140 |
-
save_repo_format=None,
|
141 |
-
expected_objects=dict(),
|
142 |
-
handle_found_object=handle_found_object,
|
143 |
-
handle_modified_object=handle_modified_object,
|
144 |
-
handle_missing_object=handle_missing_object,
|
145 |
-
handle_new_object=handle_new_object,
|
146 |
-
commit_hash="6928b08a2501aa7a4a4aabac1f888b66e7782056",
|
147 |
-
)
|
148 |
-
|
149 |
-
assert len(out) == 0
|
150 |
-
assert len(new_objects) == 3
|
151 |
-
assert len(found_objects) == 0
|
152 |
-
assert len(modified_objects) == 0
|
153 |
-
assert len(missing_objects) == 0
|
154 |
-
|
155 |
-
|
156 |
-
def test_github_handle_found_object():
|
157 |
-
github_downloader = GitHubDownloader()
|
158 |
-
found_objects = []
|
159 |
-
handle_found_object = (
|
160 |
-
lambda local_path, file_identifier, sha256, metadata: found_objects.append(
|
161 |
-
dict(
|
162 |
-
local_path=local_path,
|
163 |
-
file_identifier=file_identifier,
|
164 |
-
sha256=sha256,
|
165 |
-
metadata=metadata,
|
166 |
-
)
|
167 |
-
)
|
168 |
-
)
|
169 |
-
|
170 |
-
missing_objects = []
|
171 |
-
handle_missing_object = (
|
172 |
-
lambda file_identifier, sha256, metadata: missing_objects.append(
|
173 |
-
dict(
|
174 |
-
file_identifier=file_identifier,
|
175 |
-
sha256=sha256,
|
176 |
-
metadata=metadata,
|
177 |
-
)
|
178 |
-
)
|
179 |
-
)
|
180 |
-
|
181 |
-
new_objects = []
|
182 |
-
handle_new_object = (
|
183 |
-
lambda local_path, file_identifier, sha256, metadata: new_objects.append(
|
184 |
-
dict(
|
185 |
-
local_path=local_path,
|
186 |
-
file_identifier=file_identifier,
|
187 |
-
sha256=sha256,
|
188 |
-
metadata=metadata,
|
189 |
-
)
|
190 |
-
)
|
191 |
-
)
|
192 |
-
|
193 |
-
modified_objects = []
|
194 |
-
handle_modified_object = lambda local_path, file_identifier, new_sha256, old_sha256, metadata: modified_objects.append(
|
195 |
-
dict(
|
196 |
-
local_path=local_path,
|
197 |
-
file_identifier=file_identifier,
|
198 |
-
new_sha256=new_sha256,
|
199 |
-
old_sha256=old_sha256,
|
200 |
-
metadata=metadata,
|
201 |
-
)
|
202 |
-
)
|
203 |
-
|
204 |
-
download_dir = "~/.objaverse-tests"
|
205 |
-
base_download_dir = os.path.join(download_dir, "github")
|
206 |
-
fs, path = fsspec.core.url_to_fs(base_download_dir)
|
207 |
-
fs.makedirs(path, exist_ok=True)
|
208 |
-
|
209 |
-
shutil.rmtree(os.path.join(path, "repos"), ignore_errors=True)
|
210 |
-
out = github_downloader._process_repo(
|
211 |
-
repo_id="mattdeitke/objaverse-xl-test-files",
|
212 |
-
fs=fs,
|
213 |
-
base_dir=path,
|
214 |
-
save_repo_format=None,
|
215 |
-
expected_objects={
|
216 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.fbx": "7037575f47816118e5a34e7c0da9927e1be7be3f5b4adfac337710822eb50fa9",
|
217 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.glb": "04e6377317d6818e32c5cbd1951e76deb3641bbf4f6db6933046221d5fbf1c5c",
|
218 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.obj": "d2b9a5d7c47dc93526082c9b630157ab6bce4fd8669610d942176f4a36444e71",
|
219 |
-
},
|
220 |
-
handle_found_object=handle_found_object,
|
221 |
-
handle_modified_object=handle_modified_object,
|
222 |
-
handle_missing_object=handle_missing_object,
|
223 |
-
handle_new_object=handle_new_object,
|
224 |
-
commit_hash="6928b08a2501aa7a4a4aabac1f888b66e7782056",
|
225 |
-
)
|
226 |
-
|
227 |
-
assert len(out) == 0
|
228 |
-
assert len(found_objects) == 3
|
229 |
-
assert len(missing_objects) == 0
|
230 |
-
assert len(new_objects) == 0
|
231 |
-
assert len(modified_objects) == 0
|
232 |
-
|
233 |
-
|
234 |
-
def test_github_handle_modified_object():
|
235 |
-
github_downloader = GitHubDownloader()
|
236 |
-
found_objects = []
|
237 |
-
handle_found_object = (
|
238 |
-
lambda local_path, file_identifier, sha256, metadata: found_objects.append(
|
239 |
-
dict(
|
240 |
-
local_path=local_path,
|
241 |
-
file_identifier=file_identifier,
|
242 |
-
sha256=sha256,
|
243 |
-
metadata=metadata,
|
244 |
-
)
|
245 |
-
)
|
246 |
-
)
|
247 |
-
|
248 |
-
missing_objects = []
|
249 |
-
handle_missing_object = (
|
250 |
-
lambda file_identifier, sha256, metadata: missing_objects.append(
|
251 |
-
dict(
|
252 |
-
file_identifier=file_identifier,
|
253 |
-
sha256=sha256,
|
254 |
-
metadata=metadata,
|
255 |
-
)
|
256 |
-
)
|
257 |
-
)
|
258 |
-
|
259 |
-
new_objects = []
|
260 |
-
handle_new_object = (
|
261 |
-
lambda local_path, file_identifier, sha256, metadata: new_objects.append(
|
262 |
-
dict(
|
263 |
-
local_path=local_path,
|
264 |
-
file_identifier=file_identifier,
|
265 |
-
sha256=sha256,
|
266 |
-
metadata=metadata,
|
267 |
-
)
|
268 |
-
)
|
269 |
-
)
|
270 |
-
|
271 |
-
modified_objects = []
|
272 |
-
handle_modified_object = lambda local_path, file_identifier, new_sha256, old_sha256, metadata: modified_objects.append(
|
273 |
-
dict(
|
274 |
-
local_path=local_path,
|
275 |
-
file_identifier=file_identifier,
|
276 |
-
new_sha256=new_sha256,
|
277 |
-
old_sha256=old_sha256,
|
278 |
-
metadata=metadata,
|
279 |
-
)
|
280 |
-
)
|
281 |
-
|
282 |
-
download_dir = "~/.objaverse-tests"
|
283 |
-
base_download_dir = os.path.join(download_dir, "github")
|
284 |
-
fs, path = fsspec.core.url_to_fs(base_download_dir)
|
285 |
-
fs.makedirs(path, exist_ok=True)
|
286 |
-
|
287 |
-
shutil.rmtree(os.path.join(path, "repos"), ignore_errors=True)
|
288 |
-
out = github_downloader._process_repo(
|
289 |
-
repo_id="mattdeitke/objaverse-xl-test-files",
|
290 |
-
fs=fs,
|
291 |
-
base_dir=path,
|
292 |
-
save_repo_format=None,
|
293 |
-
expected_objects={
|
294 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.fbx": "7037575f47816118e5a34e7c0da9927e1be7be3f5b4adfac337710822eb50fa9<modified>",
|
295 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.glb": "04e6377317d6818e32c5cbd1951e76deb3641bbf4f6db6933046221d5fbf1c5c",
|
296 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.obj": "d2b9a5d7c47dc93526082c9b630157ab6bce4fd8669610d942176f4a36444e71",
|
297 |
-
},
|
298 |
-
handle_found_object=handle_found_object,
|
299 |
-
handle_modified_object=handle_modified_object,
|
300 |
-
handle_missing_object=handle_missing_object,
|
301 |
-
handle_new_object=handle_new_object,
|
302 |
-
commit_hash="6928b08a2501aa7a4a4aabac1f888b66e7782056",
|
303 |
-
)
|
304 |
-
|
305 |
-
assert len(out) == 0
|
306 |
-
assert len(found_objects) == 2
|
307 |
-
assert len(missing_objects) == 0
|
308 |
-
assert len(new_objects) == 0
|
309 |
-
assert len(modified_objects) == 1
|
310 |
-
|
311 |
-
|
312 |
-
def test_github_handle_missing_object():
|
313 |
-
github_downloader = GitHubDownloader()
|
314 |
-
found_objects = []
|
315 |
-
handle_found_object = (
|
316 |
-
lambda local_path, file_identifier, sha256, metadata: found_objects.append(
|
317 |
-
dict(
|
318 |
-
local_path=local_path,
|
319 |
-
file_identifier=file_identifier,
|
320 |
-
sha256=sha256,
|
321 |
-
metadata=metadata,
|
322 |
-
)
|
323 |
-
)
|
324 |
-
)
|
325 |
-
|
326 |
-
missing_objects = []
|
327 |
-
handle_missing_object = (
|
328 |
-
lambda file_identifier, sha256, metadata: missing_objects.append(
|
329 |
-
dict(
|
330 |
-
file_identifier=file_identifier,
|
331 |
-
sha256=sha256,
|
332 |
-
metadata=metadata,
|
333 |
-
)
|
334 |
-
)
|
335 |
-
)
|
336 |
-
|
337 |
-
new_objects = []
|
338 |
-
handle_new_object = (
|
339 |
-
lambda local_path, file_identifier, sha256, metadata: new_objects.append(
|
340 |
-
dict(
|
341 |
-
local_path=local_path,
|
342 |
-
file_identifier=file_identifier,
|
343 |
-
sha256=sha256,
|
344 |
-
metadata=metadata,
|
345 |
-
)
|
346 |
-
)
|
347 |
-
)
|
348 |
-
|
349 |
-
modified_objects = []
|
350 |
-
handle_modified_object = lambda local_path, file_identifier, new_sha256, old_sha256, metadata: modified_objects.append(
|
351 |
-
dict(
|
352 |
-
local_path=local_path,
|
353 |
-
file_identifier=file_identifier,
|
354 |
-
new_sha256=new_sha256,
|
355 |
-
old_sha256=old_sha256,
|
356 |
-
metadata=metadata,
|
357 |
-
)
|
358 |
-
)
|
359 |
-
|
360 |
-
download_dir = "~/.objaverse-tests"
|
361 |
-
base_download_dir = os.path.join(download_dir, "github")
|
362 |
-
fs, path = fsspec.core.url_to_fs(base_download_dir)
|
363 |
-
fs.makedirs(path, exist_ok=True)
|
364 |
-
|
365 |
-
shutil.rmtree(os.path.join(path, "repos"), ignore_errors=True)
|
366 |
-
out = github_downloader._process_repo(
|
367 |
-
repo_id="mattdeitke/objaverse-xl-test-files",
|
368 |
-
fs=fs,
|
369 |
-
base_dir=path,
|
370 |
-
save_repo_format=None,
|
371 |
-
expected_objects={
|
372 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.fbx": "7037575f47816118e5a34e7c0da9927e1be7be3f5b4adfac337710822eb50fa9",
|
373 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example-2.fbx": "<fake-missing-object>",
|
374 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.glb": "04e6377317d6818e32c5cbd1951e76deb3641bbf4f6db6933046221d5fbf1c5c",
|
375 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.obj": "d2b9a5d7c47dc93526082c9b630157ab6bce4fd8669610d942176f4a36444e71",
|
376 |
-
},
|
377 |
-
handle_found_object=handle_found_object,
|
378 |
-
handle_modified_object=handle_modified_object,
|
379 |
-
handle_missing_object=handle_missing_object,
|
380 |
-
handle_new_object=handle_new_object,
|
381 |
-
commit_hash="6928b08a2501aa7a4a4aabac1f888b66e7782056",
|
382 |
-
)
|
383 |
-
|
384 |
-
assert len(out) == 0
|
385 |
-
assert len(found_objects) == 3
|
386 |
-
assert len(missing_objects) == 1
|
387 |
-
assert len(new_objects) == 0
|
388 |
-
assert len(modified_objects) == 0
|
389 |
-
|
390 |
-
|
391 |
-
def test_github_handle_missing_object_2():
|
392 |
-
github_downloader = GitHubDownloader()
|
393 |
-
found_objects = []
|
394 |
-
handle_found_object = (
|
395 |
-
lambda local_path, file_identifier, sha256, metadata: found_objects.append(
|
396 |
-
dict(
|
397 |
-
local_path=local_path,
|
398 |
-
file_identifier=file_identifier,
|
399 |
-
sha256=sha256,
|
400 |
-
metadata=metadata,
|
401 |
-
)
|
402 |
-
)
|
403 |
-
)
|
404 |
-
|
405 |
-
missing_objects = []
|
406 |
-
handle_missing_object = (
|
407 |
-
lambda file_identifier, sha256, metadata: missing_objects.append(
|
408 |
-
dict(
|
409 |
-
file_identifier=file_identifier,
|
410 |
-
sha256=sha256,
|
411 |
-
metadata=metadata,
|
412 |
-
)
|
413 |
-
)
|
414 |
-
)
|
415 |
-
|
416 |
-
new_objects = []
|
417 |
-
handle_new_object = (
|
418 |
-
lambda local_path, file_identifier, sha256, metadata: new_objects.append(
|
419 |
-
dict(
|
420 |
-
local_path=local_path,
|
421 |
-
file_identifier=file_identifier,
|
422 |
-
sha256=sha256,
|
423 |
-
metadata=metadata,
|
424 |
-
)
|
425 |
-
)
|
426 |
-
)
|
427 |
-
|
428 |
-
modified_objects = []
|
429 |
-
handle_modified_object = lambda local_path, file_identifier, new_sha256, old_sha256, metadata: modified_objects.append(
|
430 |
-
dict(
|
431 |
-
local_path=local_path,
|
432 |
-
file_identifier=file_identifier,
|
433 |
-
new_sha256=new_sha256,
|
434 |
-
old_sha256=old_sha256,
|
435 |
-
metadata=metadata,
|
436 |
-
)
|
437 |
-
)
|
438 |
-
|
439 |
-
download_dir = "~/.objaverse-tests"
|
440 |
-
base_download_dir = os.path.join(download_dir, "github")
|
441 |
-
fs, path = fsspec.core.url_to_fs(base_download_dir)
|
442 |
-
fs.makedirs(path, exist_ok=True)
|
443 |
-
|
444 |
-
shutil.rmtree(os.path.join(path, "repos"), ignore_errors=True)
|
445 |
-
out = github_downloader._process_repo(
|
446 |
-
repo_id="mattdeitke/objaverse-xl-test-files-does-not-exist",
|
447 |
-
fs=fs,
|
448 |
-
base_dir=path,
|
449 |
-
save_repo_format=None,
|
450 |
-
expected_objects={
|
451 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.fbx": "7037575f47816118e5a34e7c0da9927e1be7be3f5b4adfac337710822eb50fa9",
|
452 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example-2.fbx": "<fake-missing-object>",
|
453 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.glb": "04e6377317d6818e32c5cbd1951e76deb3641bbf4f6db6933046221d5fbf1c5c",
|
454 |
-
"https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.obj": "d2b9a5d7c47dc93526082c9b630157ab6bce4fd8669610d942176f4a36444e71",
|
455 |
-
},
|
456 |
-
handle_found_object=handle_found_object,
|
457 |
-
handle_modified_object=handle_modified_object,
|
458 |
-
handle_missing_object=handle_missing_object,
|
459 |
-
handle_new_object=handle_new_object,
|
460 |
-
commit_hash="6928b08a2501aa7a4a4aabac1f888b66e7782056",
|
461 |
-
)
|
462 |
-
|
463 |
-
assert len(out) == 0
|
464 |
-
assert len(found_objects) == 0
|
465 |
-
assert len(missing_objects) == 4
|
466 |
-
assert len(new_objects) == 0
|
467 |
-
assert len(modified_objects) == 0
|
468 |
-
|
469 |
-
|
470 |
-
def test_github_download_cache():
|
471 |
-
github_downloader = GitHubDownloader()
|
472 |
-
objects = pd.DataFrame(
|
473 |
-
[
|
474 |
-
{
|
475 |
-
"fileIdentifier": "https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.fbx",
|
476 |
-
"license": None,
|
477 |
-
"sha256": "7037575f47816118e5a34e7c0da9927e1be7be3f5b4adfac337710822eb50fa9",
|
478 |
-
},
|
479 |
-
{
|
480 |
-
"fileIdentifier": "https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.glb",
|
481 |
-
"license": None,
|
482 |
-
"sha256": "04e6377317d6818e32c5cbd1951e76deb3641bbf4f6db6933046221d5fbf1c5c",
|
483 |
-
},
|
484 |
-
{
|
485 |
-
"fileIdentifier": "https://github.com/mattdeitke/objaverse-xl-test-files/blob/6928b08a2501aa7a4a4aabac1f888b66e7782056/example.obj",
|
486 |
-
"license": None,
|
487 |
-
"sha256": "d2b9a5d7c47dc93526082c9b630157ab6bce4fd8669610d942176f4a36444e71",
|
488 |
-
},
|
489 |
-
]
|
490 |
-
)
|
491 |
-
|
492 |
-
# remove the repos directory
|
493 |
-
for save_repo_format in ["tar", "tar.gz", "zip", "files"]:
|
494 |
-
repos_dir = "~/.objaverse-tests/github/repos"
|
495 |
-
shutil.rmtree(os.path.expanduser(repos_dir), ignore_errors=True)
|
496 |
-
|
497 |
-
out = github_downloader.download_objects(
|
498 |
-
objects=objects,
|
499 |
-
processes=1,
|
500 |
-
download_dir="~/.objaverse-tests",
|
501 |
-
save_repo_format=save_repo_format, # type: ignore
|
502 |
-
)
|
503 |
-
assert len(out) == 3
|
504 |
-
|
505 |
-
out = github_downloader.download_objects(
|
506 |
-
objects=objects,
|
507 |
-
processes=1,
|
508 |
-
download_dir="~/.objaverse-tests",
|
509 |
-
save_repo_format=save_repo_format, # type: ignore
|
510 |
-
)
|
511 |
-
assert len(out) == 0
|
512 |
-
|
513 |
-
|
514 |
-
def test_annotations():
|
515 |
-
downloaders = [
|
516 |
-
GitHubDownloader(),
|
517 |
-
SketchfabDownloader(),
|
518 |
-
SmithsonianDownloader(),
|
519 |
-
ThingiverseDownloader(),
|
520 |
-
]
|
521 |
-
|
522 |
-
for downloader in downloaders:
|
523 |
-
annotations_df = downloader.get_annotations()
|
524 |
-
|
525 |
-
# make sure the columns are
|
526 |
-
assert set(annotations_df.columns) == set(
|
527 |
-
["fileIdentifier", "source", "license", "fileType", "sha256", "metadata"]
|
528 |
-
)
|
529 |
-
|
530 |
-
|
531 |
-
def test_download_objects():
|
532 |
-
downloaders = [
|
533 |
-
GitHubDownloader(),
|
534 |
-
SketchfabDownloader(),
|
535 |
-
SmithsonianDownloader(),
|
536 |
-
ThingiverseDownloader(),
|
537 |
-
]
|
538 |
-
|
539 |
-
download_dir = "~/.objaverse-tests"
|
540 |
-
|
541 |
-
for downloader in downloaders:
|
542 |
-
shutil.rmtree(os.path.expanduser(download_dir), ignore_errors=True)
|
543 |
-
|
544 |
-
annotations_df = downloader.get_annotations()
|
545 |
-
|
546 |
-
test_objects = annotations_df.head(n=2)
|
547 |
-
|
548 |
-
out = downloader.download_objects(
|
549 |
-
objects=test_objects,
|
550 |
-
download_dir=download_dir,
|
551 |
-
processes=2,
|
552 |
-
)
|
553 |
-
assert isinstance(out, dict), f"{out=}"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
thingiverse/thingiverse-objects.parquet
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:57f0a79e23af18fed59c08efeabbaa8886bfc5240e94119687451190c09d76ab
|
3 |
-
size 292395520
|
|
|
|
|
|
|
|
smithsonian/object-metadata.parquet → thingiverse/thingiverse.parquet
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b4603c1e039b1e963f4848af145abae6e15d6f201c8d075b0c29d6939cc3eeee
|
3 |
+
size 340534075
|