Datasets:

Languages:
English
ArXiv:
License:
mattdeitke's picture
Update README.md
f92d05d unverified
|
raw
history blame
No virus
5.14 kB
# πŸͺ Objaverse-XL Rendering Script
![Blender generated with MidJourney](https://github.com/allenai/objaverse-xl/assets/28768645/69064f78-a752-40d6-bd36-ea7c15ffa1ec)
Scripts for rendering Objaverse-XL with [Blender](https://www.blender.org/). Rendering is the process of taking pictures of the 3D objects. These images can then be used for training AI models.
## πŸ–₯️ Setup
1. Clone the repository and enter the rendering directory:
```bash
git clone https://github.com/allenai/objaverse-xl.git && \
cd objaverse-xl/scripts/rendering
```
2. Download Blender:
```bash
wget https://download.blender.org/release/Blender3.2/blender-3.2.2-linux-x64.tar.xz && \
tar -xf blender-3.2.2-linux-x64.tar.xz && \
rm blender-3.2.2-linux-x64.tar.xz
```
3. If you're on a headless Linux server, install Xorg and start it:
```bash
sudo apt-get install xserver-xorg -y && \
sudo python3 start_x_server.py start
```
4. Install the Python dependencies. Note that Python >3.8 is required:
```bash
cd ../.. && \
pip install -r requirements.txt && \
pip install -e . && \
cd scripts/rendering
```
## πŸ“Έ Usage
### πŸ₯ Minimal Example
After setup, we can start to render objects using the `main.py` script:
```bash
python3 main.py
```
After running this, you should see 10 zip files located in `~/.objaverse/github/renders`. Each zip file corresponds to the rendering of a unique object, in this case from [our example 3D objects repo](https://github.com/mattdeitke/objaverse-xl-test-files):
```bash
> ls ~/.objaverse/github/renders
0fde27a0-99f0-5029-8e20-be9b8ecabb59.zip 54f7478b-4983-5541-8cf7-1ab2e39a842e.zip 93499b75-3ee0-5069-8f4b-1bab60d2e6d6.zip
21dd4d7b-b203-5d00-b325-0c041f43524e.zip 5babbc61-d4e1-5b5c-9b47-44994bbf958e.zip ab30e24f-1046-5257-8806-2e346f4efebe.zip
415ca2d5-9d87-568c-a5ff-73048a084229.zip 5f6d2547-3661-54d5-9895-bebc342c753d.zip
44414a2a-e8f0-5a5f-bb58-6be50d8fd034.zip 8a170083-0529-547f-90ec-ebc32eafe594.zip
```
If we unzip one of the zip files:
```bash
> cd ~/.objaverse/github/renders
> unzip 0fde27a0-99f0-5029-8e20-be9b8ecabb59.zip
```
we will see that there is a new `0fde27a0-99f0-5029-8e20-be9b8ecabb59` directory. If we look in that directory, we'll find the following files:
```bash
> ls 0fde27a0-99f0-5029-8e20-be9b8ecabb59
000.npy 001.npy 002.npy 003.npy 004.npy 005.npy 006.npy 007.npy 008.npy 009.npy 010.npy 011.npy metadata.json
000.png 001.png 002.png 003.png 004.png 005.png 006.png 007.png 008.png 009.png 010.png 011.png
```
Here, we see that there are 12 renders `[000-011].png`. Each render will look something like one of the 4 images shown below, but likely with the camera at a different location as its location is randomized during rendering:
![temp](https://github.com/allenai/objaverse-xl/assets/28768645/69d79e26-4df1-4bd2-854c-7d3c888adae7)
Additionally, there are 12 npy files `[000-011].npy`, which include information about the camera's pose for a given render. We can read the npy files using:
```python
import numpy as np
array = np.load("000.npy")
```
where array is now a 3x4 [camera matrix](https://en.wikipedia.org/wiki/Camera_matrix) that looks something like:
```python
array([[6.07966840e-01, 7.93962419e-01, 3.18103019e-08, 2.10451518e-07],
[4.75670159e-01, -3.64238620e-01, 8.00667346e-01, -5.96046448e-08],
[6.35699809e-01, -4.86779213e-01, -5.99109232e-01, -1.66008198e+00]])
```
Finally, we also have a `metadata.json` file, which contains metadata about the object and scene:
```json
{
"animation_count": 0,
"armature_count": 0,
"edge_count": 2492,
"file_identifier": "https://github.com/mattdeitke/objaverse-xl-test-files/blob/ead0bed6a76012452273bbe18d12e4d68a881956/example.abc",
"file_size": 108916,
"lamp_count": 1,
"linked_files": [],
"material_count": 0,
"mesh_count": 3,
"missing_textures": {
"count": 0,
"file_path_to_color": {},
"files": []
},
"object_count": 8,
"poly_count": 984,
"random_color": null,
"save_uid": "0fde27a0-99f0-5029-8e20-be9b8ecabb59",
"scene_size": {
"bbox_max": [
4.999998569488525,
6.0,
1.0
],
"bbox_min": [
-4.999995231628418,
-6.0,
-1.0
]
},
"sha256": "879bc9d2d85e4f3866f0cfef41f5236f9fff5f973380461af9f69cdbed53a0da",
"shape_key_count": 0,
"vert_count": 2032
}
```
### πŸŽ› Configuration
### πŸ§‘β€πŸ”¬οΈ Experimental Features
USDZ support is experimental. Since Blender does not natively support usdz, we use [this Blender addon](https://github.com/robmcrosby/BlenderUSDZ), but it doesn't work with all types of USDZs. If you have a better solution, PRs are very much welcome πŸ˜„!
## πŸ‘‹ Our Team
Objaverse-XL is an open-source project managed by the [PRIOR team](//prior.allenai.org) at the [Allen Institute for AI](//allenai.org) (AI2).
AI2 is a non-profit institute with the mission to contribute to humanity through high-impact AI research and engineering.
<br />
<a href="//prior.allenai.org">
<p align="center"><img width="100%" src="https://raw.githubusercontent.com/allenai/ai2thor/main/doc/static/ai2-prior.svg" /></p>
</a>