
- Posts: 414
- Joined: Fri May 15, 2020 1:12 pm
- Real Name: Wayne Folta
I just got an iPhone 12 Pro and did a quick 3D scan and wanted to pass on some results for anyone else who is interested.
The iPhone 12 Pro (and iPhone 12 Pro Max) both have a lidar scanner in them, so can more accurately gauge depth in an image. The free app 3D Scanner App is (https://www.3dscannerapp.com) is available on the iPhone App Store. With it, you can scan a scene/object.
After scanning, you can select the scan and pan around in it, then you have a More menu that has some critical options. I'm not sure what Refine Scan does yet. But I was able to use Simplify Mesh followed by Smooth Mesh to come up with a fairly small and useful result. It warned me that these operations would remove the texture, but after doing them I was able to use the Color option to reintroduce the texture back onto the mesh.
You can then Share an OBJ file that will have an accompanying JPEG texture image and a .mtl file to tie the two together. I chose to share as a single Zip and used Air Drop to send it to my laptop. There are other options.
You can import the OBJ file into Blender (I'm using the latest, 2.91) and turn on Viewport Shading and you should see your scene, with texture. (All three files should be in the same directory so Blender sees the two texture-associated files when you load the OBJ.)
Then you can export an Alembic File (.abc) which can be imported into Fusion. I exported the whole "animation" -- nothing was animated even though the default Blender timeline is 250 frames or something like that.
The texture by itself looks weird, but the final product looks the same as in 3D Scanner App and Blender. Scroll down in the attached image to see the Fusion Nodes I used. There may be a better way to do this, but it's what worked for me. I loaded the Alembic file with the 3D node for Alembic files, and then loaded the texture JPEG and applied the texture.
Just a quick scan, not rigorous. You'll notice a strange bump on the chair's arm, and lots of missing background -- the background was mainly included by chance as I went around the chair. You could really polish things, but I figured this might be of general interest to folks. With a phone, you can go all Hollywood and make a reasonable, textured 3D scene and bring it into Fusion.
I didn't get down on the floor to allow it to scan under the chair, so the feet and underlying carpet are part of the texture on the boxy "chair" itself. Further refinement in 3D Scanner App would help a lot -- I can't find any documentation, so it's a bit of a poke-around-n-experiment process, but the tools I've figured out are very nice. I used Low Resolution because of dire warnings that High Resolution should only be used on small objects. Further cleanup could be done in Blender, then bring it into Fusion and work away.
Note that the texture is just an image of the actual object, so has the lighting of the actual scene baked into it. So if you really wanted to relight it, you'd probably want to light the object/scene as diffusely as possible. Maybe even a ring light around the phone's camera lens -- not blocking the lidar, of course -- would be helpful. Or just light the scene in 3D as it's lit when you scan it.
The iPhone 12 Pro (and iPhone 12 Pro Max) both have a lidar scanner in them, so can more accurately gauge depth in an image. The free app 3D Scanner App is (https://www.3dscannerapp.com) is available on the iPhone App Store. With it, you can scan a scene/object.
After scanning, you can select the scan and pan around in it, then you have a More menu that has some critical options. I'm not sure what Refine Scan does yet. But I was able to use Simplify Mesh followed by Smooth Mesh to come up with a fairly small and useful result. It warned me that these operations would remove the texture, but after doing them I was able to use the Color option to reintroduce the texture back onto the mesh.
You can then Share an OBJ file that will have an accompanying JPEG texture image and a .mtl file to tie the two together. I chose to share as a single Zip and used Air Drop to send it to my laptop. There are other options.
You can import the OBJ file into Blender (I'm using the latest, 2.91) and turn on Viewport Shading and you should see your scene, with texture. (All three files should be in the same directory so Blender sees the two texture-associated files when you load the OBJ.)
- Blender View
- Blender View.jpg (750.36 KiB) Viewed 217 times
Then you can export an Alembic File (.abc) which can be imported into Fusion. I exported the whole "animation" -- nothing was animated even though the default Blender timeline is 250 frames or something like that.
- Fusion View
- Fusion View.jpg (716.94 KiB) Viewed 217 times
The texture by itself looks weird, but the final product looks the same as in 3D Scanner App and Blender. Scroll down in the attached image to see the Fusion Nodes I used. There may be a better way to do this, but it's what worked for me. I loaded the Alembic file with the 3D node for Alembic files, and then loaded the texture JPEG and applied the texture.
Just a quick scan, not rigorous. You'll notice a strange bump on the chair's arm, and lots of missing background -- the background was mainly included by chance as I went around the chair. You could really polish things, but I figured this might be of general interest to folks. With a phone, you can go all Hollywood and make a reasonable, textured 3D scene and bring it into Fusion.
I didn't get down on the floor to allow it to scan under the chair, so the feet and underlying carpet are part of the texture on the boxy "chair" itself. Further refinement in 3D Scanner App would help a lot -- I can't find any documentation, so it's a bit of a poke-around-n-experiment process, but the tools I've figured out are very nice. I used Low Resolution because of dire warnings that High Resolution should only be used on small objects. Further cleanup could be done in Blender, then bring it into Fusion and work away.
Note that the texture is just an image of the actual object, so has the lighting of the actual scene baked into it. So if you really wanted to relight it, you'd probably want to light the object/scene as diffusely as possible. Maybe even a ring light around the phone's camera lens -- not blocking the lidar, of course -- would be helpful. Or just light the scene in 3D as it's lit when you scan it.
Resolve Studio 17 latest, Fusion Studio 17 latest, macOS Big Sur latest, MacBook Pro 2020 64GB RAM and Radeon Pro 5600M 8GB VRAM