They support two prior versions of OS-X and that's it. So OS-X 10.7, released 3 years ago, is unsupported as of October 2014. I guess that works if you have the attitude of just always updating to the latest OS, but it can be an issue for various enterprise setups that prefer to version freeze for longer times, or for 3rd party software/hardware that doesn't get updated. Also can screw you over if Apple decides to change hardware like with the PPC to Intel change.
Friends and family are surely tired of my tinfoil hat, they just do not seem to care about their privacy. Many say the "I have nothing to hide" line.
Between a bit better language design and superior support and tools, CUDA is way easier to do your work in. We've 4 labs that use CUDA in one fashion or another, none that use OpenCL. A number have tried it (also tried lines like the Cell cards that IBM sold for awhile) but settled on CUDA as being the easiest in terms of development. Open standards are nice and all but they've got shit to do and never enough time to do it, so whatever works the easiest is a win for them.
On a different side of things, I've seen less issues out of nVidia on CUDA than AMD on OpenCL for video editing. Sony Vegas supports both for accelerating video effects and encoding. When I had an AMD card, it was crashes all the time with acceleration on. Sony had to disable acceleration on a number of effects with it. I had to turn it off to have a usable setup. With nVidia, I find problems are very infrequent.
Obviously this is one one data point and I don't know the details of development. However it is one of the few examples I know of a product that supports both APIs.
It's funny because I learned to program in Pascal, that's what high school AP CS was taught in at the time. Then I learned C 6 months later and wondered why anyone would ever use Pascal. Then I saw this thread, and I wonder
The issue is that most of the time people doing this work are on the hardware teams, rather than the software teams. It's a hard world to live in, to be sure, the silicon & pcb guys think you're software (i.e. you write code that executes ON a processor, not code that creates the processor). The software guys don't speak the same language: they're hung up on methodologies, APIs, code style & the business of software, or else are more heavy in to the software research side of pure algoritms. It can be tough to fit in, but the job has to be done, it isn't any less essential, just a bit more niche.
NTSC stuff is so bad when viewed on a large TV. It is amazing how blurry things look when you flip back and forth between the HD and SD channels. That is part of what lead to the rise of big screen TVs was actually having content for them. With NTSC, a large TV just meant a big blurry image. With ATSC it can mean a nice large image.
Why shouldn't they continually improve their products? Even with NTSC sets this was done. New ones would be larger, have better focus, more clearly resolve the signal, have better phosphors, and so on. Why shouldn't this continue? They should keep trying to improve their products as technology allows.
None of that means you need to buy a new toy all the time though. You can stick with what you have until it breaks, or until the new stuff is a big enough leap that you wish to own it.
I think a lot of the whining from people comes down to simple jealousy. They'd like to own the new stuff, but cannot afford it, or do not wish to. So they try and hate on it and act like a luddite. You see it practically any time Slashdot has a story on new technology. People complain about it like it is somehow a bad thing that there might be something new.
Most panels in higher end screens are actually real 120fps panels. However that is just used for 3D and for reduced motion blur. The only set I know that advertises support for 120fps input is Vizio. Others could do it, if they wanted to, however.
As you say, the issue with higher refresh rates isn't in the display technology.
Part of it is just getting people used to the idea I think. We've seen shitty, jerky, frame rates in moves for so long people start to associate that with being "cinematic". People need to get used to the idea that's bullshit and maybe they'll start to like it more.
Hopefully sports and such will get shot at 60fps some day and that may help.
You know, when something says that we are so close to destruction for over half a century... well you have to wonder why anyone would put any stock in it. It is a bit hard to reconcile with being on the edge of destruction, and yet everything continuing to not be destroyed.
Thing thing is if you go and look at benchmarks of the cards in actual games, you find out the 970 wrecks shit, particularly given its price point. The 980 is an overpriced luxury (I say this as a 980 owner) because the 970 gets nearly the same performance for like half the price. The difference with its memory controller just doesn't seem to matter in actual games out there on the market.
And that's the real thing here the the spec head forget: You buy these to run actual software. If it does well on all actual software, then who gives a shit about the details?
So did STTNG.
No you misread. FiOS isn't AS profitable, it's profitable. Someone without a conflict of interest, willing to compete with wireless, could set up a business and make money, give good service, employ people and return value to an investor. Verizon won't, they see it as a cannibalizing their wireless market.
This is an example of all that is wrong with telecom.
Statements like that one unfortunately lives to regret. Lucas dropped the ball on the prequels, but I'm pretty sure JJ got hit in the head once with the ball and suffered long term damage.
He may well have been as smart as he thought (I'm not saying that is the case for sure, mind) but turns out others were smart enough, and more knowledgeable in the ways that mattered.
Hans Reiser is a good example. Man is unquestionably very smart. However, he had the geek hubris that I call SMFU, Smartest Motherfucker in the Universe syndrome. He figured he was so much smarter than everyone else, he could easily get away with his crime. Turns out that the police have some smart people too, and those people know a lot more about criminal investigation than he did.