I think the disconnect is that you don't understand the tech enough. When I talk to people about this, over and over they send me one of about 20 videos of an anime girl dancing in place that someone made. They don't understand that an anime girl is literally the only subject that the youtuber or whatever can render at that quality, they don't get that if the camera moved around too much the whole thing would melt down in noise, it's just someone showing off a pretty cheap trick. Haven't you ever wondered why there's 100 of these dancing girl videos and not even one actual film yet?
The flickering was never a problem. At least the luma flickering, that's what's gone here in your clip. It's just a single drag and drop effect locked behind the davinchi resolve paywall, which I paid last month after I paid off the PSVR2. So that particular issue isn't what I was talking about, but I've also solved the other one that I was talking about in the intervening time, which requires another program. Right now my pipeline runs across 6 programs using python scripts and some sections have many stages. The last couple videos you sent are just SD outputs with some compositing and a single deflicker layer. Right now the other flickering, the line pixel displacement over time flickering, I'm solving with an ebsynth midstage, which was the plan from way back. I actually have EBsynth tests from like 2019. Long story short, take these videos with a grain of salt, because last month about 200 DK people announced that they had solved this, and literally not one of them is even close. There are people that are getting close though, at huge functioning companies like microsoft and google, and those are the people I actually do have to worry about. Also some of the legit research scientists, but they have been screwing around with this since 2014 with remarkably little progress.
Here's what you might think is a finished solution at first glance, available free 3 years ago. This is the problem with trying to infer the state of complex technologies from a cursory examination of a short video. From appearances, and even the dialogue of the video, you'd be forgiven for thinking we had this whole thing solved in 2019 right? So why did I spend over 100 hours running tests on this specific solution over the last month? Because it works, but is so work intensive and unstable that I could never complete a pro work with this.
Here is the only video from last month that's even close, and it's not even close to a full solution. They had to do a lot of cheating to make it work, in terms of just stealing frames from anime shows, doing composites and just saying it was AI because it was 15% AI, that sort of thing. But it is closer than anything else out there to the method I've been building. They spent a week or two with a paid crew making a short, dumb joke film, and moved on. That's the top youtube competitor in the world right now in this area, and they aren't even half serious about it.
Try to see it from my perspective -
"Hey I'm Sean, and I just spent 2 years writing a novel about a wizard school"
"Duh, too late, Jake Paul just uploaded a 5 second gif of a wizard school that he worked on for an hour yesterday, geez Sean, I thought you were smarter than this, get your own idea."
"It's not really the same thing, you see I'm working on a serious book where I really have to think about the structure of the plot to make sure it's viable beyond some 5 second flash in the pan image, you know, a novel, not a gif, it's different in so many ways and requires so much more to actually work"
"I don't know what all that means but I'm just saying Jake came up with that wizard school thing and his is already done and published, so it's like Jake Paul 1 Sean 0 right?"
"I'm not writing a 300 page GIF though, so you can understand that it's not really a valid comparison right? Even if I was, you get that a GIF that lasted 20 hours would require a lot more infrastructure than one that only needed to work for a few minutes?"
------
Basically I was past this level seen in the video a long while back, essentially as soon as control net was published anyone could do this part by training a custom model on their footage, and then assigning it to a narrow LORA or textural inversion to get a fidelity output. That's great when you have a free model library that by the way is 99.8 percent waifu drawing styles, and literally almost nothing else. Should we start making movies now that this guy has solved it ahead of me? The movies can't have cars in them, because there are no models out for moving cars, so they can't be retargeted like this. How about having a door open in your movie? This person's method doesn't do doors. It's a one off quick hack that makes this one scene work one time.
What I need for my work is a system that doesn't break down under any circumstance, and can be automated to a waaaaaay larger degree than what you're seeing here, which requires individual work and attention for every single element to function.
Their idea of building a leverage device for scooping
My idea for building a leverage device for scooping.
Their idea of a 4k camera
My actual 4k camera, gimbal, and crane
Same words, very different reality, if you take my point. I think you'll find a lot of 2 minute Harry Potter spoofs on youtube, and it should be easy for you to understand the difference between what you are doing, and what they are doing.