What’s the largest dataset you’ve stitched with ClusterODM? What instance sizes worked best for you?
My team has successfully stitched over 7500 images (nearly 200 acres) in size. The set was very difficult to stitch as it was very homogenous in nature. To that, we used 'split': 250 to give the best chance for success.
I’ve heard from @pierotofy, that he’s stitched sets of 13,000+ images.
I’m pretty sure it was split into ~800 image submodels with 120 m of overlap from primary node s-16vcpu-64gb image to ensure enough storage, memory, and so forth for being the primary node. It spun up 16 additional instances of s-8vcpu-32gb.
Hi - Sorry for going back in time so far…but looking for a little guidance on the rationale for split size and overlap? I have been able to successfully split merge very small data sets, but am working my way to a very large data set…right now I am attempting a 721 image set, split into 103 image sets at 25m overlap. It’s been running for 10+hours (92gb allocated…). Any thoughts, recommendations would be greatly appreciated?