U.S. Patent No. 10,855,990
This is old technology, but was used extensively in JPEG and JPEG2000. All these patents are and have been long expired. There is no novel approach in U.S. Patent No. 10,855,990. More specifically, all the claims they're making in terms of the specific violations of this patent were covered in ITU-13818-2. Though ITU-https://www.itu.int/rec/T-REC-H.263-200501-I/en hammers the last nails in the coffin. I have read and reviewed the claim and the patent and the technologies presented in 10,855,990 are just reiterations of earlier work with scrambled wording to try and give a new name for variable sized macroblocks. They novel approach implemented in H.264, H.265 and H.266 was the method of selecting which specific pattern of "coding units" to apply. I have not checked for reuse of this, but this is neither in 10,855,990 or the claim. So, I believe they checked and found out that there was no violation. Oh and to be clear, they're completely fixated on the sharing coding parameters between blocks. Their approach is almost, barely, kinda novel, but the fact is, I'd make a strong argument that this is obvious, it's basically just macroblock grouping which has been part of standard video coding as far back as MPEG-1 and ASF. And the method applied could easily be argued to be an almost direct copy of LZW compression.
U.S. Patent No. 9,924,193
I couldn't find a copy of the original text (not wearing my glasses) and frankly their description was so TL;DR that they just started making things up. Ok... here's the argument against this. This has been a core features of all DWT based compression methods since the start. It was even the reason we used DWT. JPEG2000 is almost entirely based on what they're claiming here. If I spent an hour on this one, I could tear it to pieces without even trying. And skip mode... what in the world do you think something like Google Earth is?
U.S. Patent No. 9,596,469
Encoding data in a way that would allow independent parallel decoding of different portions, bands, blocks whatever of the image ... blah blah. Back to JPEG-2000 and Google Earth and stuff like that. The first time I saw this personally was at Disney Epcot Center when there was a Cray Supercomputer on display showing off a google earth like experience. The computer was streaming data at different spatial representations in parallel to hundreds of CPU cores who were all decoding and texturizing. The number of patents filed and expired on this one tech is immense. I haven't dug up specifics, but I can guarantee that the JPEG2000 patent pool clearly invalidates this
I just doom scrolled through the rest of this. I highly doubt I'm the only signal processing and video/image compression historian out there. I'm guessing that the LLMs could easily tear this crap apart too. But I'd be willing to make a few bucks as an advisor on this. I've either worked with, against, for, on, etc... on every technology being claimed here and I did this 15-20 years ago... and the tech was already old.