Replies: 4 comments 8 replies
-
| 
         There are two different types of stereo view synthesis: 
 iw3 is 2. 3d movie(post production) is 1. That is why the right eye frame is much different. Also, there is an option to output various parameters for comparison.  --find-param Line 1742 in e830a9a Adding the option to output in method 1 to this function should achieve it.  | 
  
Beta Was this translation helpful? Give feedback.
-
| 
         I added   | 
  
Beta Was this translation helpful? Give feedback.
-
| 
         I finally got around to testing some of this and uploaded scripts to https://github.com/alexmi256/2D3DTestingScript Here is what I found: Scripts for testing 2D to 3D image generationData used was from live indoor 3D video keyframes and cannot be shared TL;DRFor right image only synthetic view: 
 Summary
 ResultsA lower score is better because it means there is less % of a difference between ground truth and generated image BaselineWith 1093 images from video keyframes converted and compared with imgcompare.image_diff_percent Grayscale
 RGBSame but images were kept as RGB instead of grayscale conversion 
 Tuning OptionsAll of these were done with RGB mode --edge-dilationEdge dilation only applicable for models with  
 --depth-aaEdge dilation only applicable for models with  
 --ttaThis dropped most processing by 2-3/s 
 --methodI used  
 --find-paramsI used  
 
 
 
 --edge-dilation --depth-aa --convergence 0.5 --divergence 4.0
 --edge-dilation --depth-aa --convergence 0.5 --divergence 4.0 --method mlbw_l4s --tta
  | 
  
Beta Was this translation helpful? Give feedback.
-
| 
         All those numbers and results... the listing seems very confusing... lower score is better? While going back and deleting old outdated AI and 3D renders, I came across some Depth Anything (I don't know which model was used) vs Apple's higher resolution Depth Pro comparisons, and with fresh eyes months later, Depth Pro was looking good. Not absolute perfection and the best result with every detail, but overall looked better. I'd be interested in seeing Depth Pro get added to iw3 again... obviously unless there's been some newer model that's surpassed it. If I remember correctly, Depth Pro didn't get fully implemented because of flickering in video? So we don't use it for video then.. but it could be an improvement for photos.. or if there's some new implementation method to fix the flickering it may have had before. Back to the rankings/numbers shown.. I don't see how any Zoe based model rated any good. Zoe causes mid range bulging, foreground crush, etc. So I don't understand how Zoe models are mixed in randomly in those results... And if lower numbers are better, then it shows Depth Anything v2 is worse than v1? Idk about that... And it shows Large is worse than Base is worse than Small?? That's definitely not the case. The results are highly questionable and confusing.  | 
  
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I've extracted the left and right frames from a true stereoscopic movie and I'd like to compare the different models and settings against these.
How is this best achieved?
Are there built in utils/functions to do this? Some of the files in training seem promising and I see some eval functions.
I though I could give iw3 the left image and then just compare the original vs output right image but it doesn't seem to be as straighforward as that.
I saw that there is a
--exportoption which outputs the depth map for the given image at the default or specified resolution but if I do this I would need to compute a depth map for the original left & right images which seems doesn't seem straightforward (most solutions rely on OpenCV and there seem to be variables for this).I also tried outputing a Full SBS image and then overlaying it against the original and noticed differences in both left and right frames with the right frame containing a much larger differece. I would have throught the left frame would be untouched. This doesn't seem to be the case however and I can't tell if it's because of encoding/decoding or if the way the depth map is applied actually generates two new images (this would actually make a lot of sense to me).
Beta Was this translation helpful? Give feedback.
All reactions