What’s the largest dataset you’ve stitched with ClusterODM? What instance sizes worked best for you?
My team has successfully stitched over 7500 images (nearly 200 acres) in size. The set was very difficult to stitch as it was very homogenous in nature. To that, we used 'split': 250 to give the best chance for success.
I’ve heard from @pierotofy, that he’s stitched sets of 13,000+ images.
I’m pretty sure it was split into ~800 image submodels with 120 m of overlap from primary node s-16vcpu-64gb image to ensure enough storage, memory, and so forth for being the primary node. It spun up 16 additional instances of s-8vcpu-32gb.
It turns out the reason is as simple as one might imagine: his bucket was in nyc3, but he tends to have more success spinning up images in sfo2, and he just hasn’t moved his bucket.
Hi - Sorry for going back in time so far…but looking for a little guidance on the rationale for split size and overlap? I have been able to successfully split merge very small data sets, but am working my way to a very large data set…right now I am attempting a 721 image set, split into 103 image sets at 25m overlap. It’s been running for 10+hours (92gb allocated…). Any thoughts, recommendations would be greatly appreciated?