That said, the front page also has a 'Video Bytes' line half way down full of crap, so I guess someone is really keen on killing the site. Thank $DEITY for Soylent News...
Nokia has also been in the market of selling the infrastructure for mobile networks for a long time. And, unlike the handsets, this is a very profitable place to be. Both Nokia and Ericsson saw the commoditisation of the handset market and Nokia in particular watched their margins evaporate and decided it was time to get out. But because they're now no longer in the public eye, they're perceived as losing. Now their customers are people who make money from the products that they sell, so are willing to pay a reasonable premium because a few minutes of downtime costs far more.
Of course, when Apple decides to concentrate on the high-margin part of a business, no one claims that they're dying, because they concentrate on a consumer-visible part of the market.
There's some overlap. Altera FPGAs have lots of fixed-function blocks on them, ranging from simple block RAMs to fast floating point units. There's a good chance that Intel could reuse some of their existing designs (which, after all, are already optimised for their manufacturing process) from things like AVX units and caches on x86 chips. A lot of the FPGAs also include things like PCIe, USB, Ethernet and so on controllers. Again, Intel makes these in their chipset division and, again, they're optimised for Intel's process so being able to stick them on FPGAs instead of the Altera ones would make sense.
The main reason that you're probably right is that Intel is generally pretty bad at getting their own internal divisions to play nicely together, let alone ones that are used to being in a completely separate company.
My guess would be coarse-grained reconfigurable architectures. Altera FPGAs aren't just FPGAs, they also have a load of fixed-function blocks. The kinds of signal processing that the other poster talks about work because there are various floating point blocks on the FPGA and so you're using the programmable part to connect a sequence of these operations together without any instruction fetch/decode or register renaming overhead (you'd be surprised how much of the die area of a modern CPU is register renaming and how little is ALUs).
FPGAs are great for prototyping (we've built an experimental CPU as a softcore that runs on an Altera FPGA at 100MHz), but there are a lot of applications that could be made faster by being able to wire a set of SSE / AVX execution units together into a fixed chain and just fire data at them.
A Kickstarter-like model would work. Release a single for free, designate an amount that you think the full album is worth. If enough people are willing to pay, then you release the album for free. For the second album, hopefully enough people have copied the first that you don't need to do much to encourage them to pay for the second. As an added bonus, you can reduce your up-front costs by only renting the studio time to record the first track and only record the rest once people have paid for it.
Recording a song (at least, a song that people want to buy) requires talent, creativity, and often expensive instruments and studio time. Copying a song once it's recorded is basically free. Any business model that relies on doing the difficult thing for free and then trying to persuade people to pay for you to do the easy thing is doomed to failure. Imagine if Ford had noticed that people wanted coloured cars and decided to give away unpainted cars and charge for painting them, then bribed politicians to pass laws so that only Ford was allowed to paint cars Ford sold and driving an unpainted car on the road was illegal. It wouldn't take people long to realise that this was a stupid business model and that you could get rid of the laws and charge for the cars, but in the case of copyright people are still trying very hard to make the 'free car, expensive and exclusive paint' model work with different variations.
It's an ABI mismatch, and the summary is nonsense, saying almost the exact opposite of TFA (which I actually read, because the summary is obvious nonsense). The issue is that the Windows ABI defines long double as being a 64-bit floating point value (which is fine, because the only requirement for long double is that it have no less precision than double. If you're using it and expecting some guaranteed precision for vaguely portable code then you're an idiot). For some reason, MinGW (which aims to be ABI-compatible with MS, at least for C libraries) uses 80-bit x87 values for long double, so you get truncation. I forget the exact calling conventions for Windows i386, but I believe that in some cases this will be silently hidden, as the value will be passed in x87 register and so be transparently extended to 80 bits in the caller and truncated in the callee anyway. It's only if it's passed on the stack (or indirectly via a pointer) that it's a problem.
It's not obvious which definition of long double is better. On modern x86, you'll use SSE for 32- and 64-bit values, and may lose precision moving between x87 and SSE registers. You also get worse IEEE compliance out of the x87 unit, which may matter more than the extra 16 bits of precision. 80-bit floats are not available on any platform other than x86 (128-bit is more common, though PowerPC has its own special non-IEEE version of these and on some other platforms they're entirely done in software), so they're a bad choice if you want portable code that generates the same output on different platforms.
The same is true of university exams. My undergraduate exams, for example, mostly required that you answer two of three questions per exam. To get a first (for people outside the UK: the highest classification), you needed to get 70%. Most questions were around 40% knowledge and 60% application of the knowledge. If you could predict the topics that the examiner would pick, then that meant that you could immediately discard a third of the material. To get the top grade, you needed to get 100% in one question and 40% in another. This meant that you could understand a third of the material really well and understand another third well enough to get the repetition marks, but not the understanding ones and still get the top grade. This meant that you could study 50% of the material and still do very well in the exams, as long as you picked the correct 50%. And some of the lecturers were very predictable when setting exams...
What jobs do you imagine existing in 10-20 years that don't require some understanding of programming? I thought my stepfather, as head greenskeeper at a golf course might have had one before he retired, but it turns out that the irrigation system that he had to use came with a domain-specific programming language for controlling it. A lot of farm equipment is moving in the same direction. Office jobs generally require either wasting a lot of time, or learning a bit of scripting (hint: the employees who opt for the first choice are not going to be the ones that keep their jobs for long). Jobs that don't require any programming are the ones that are easy to automate.
But, of course, we don't need to teach our children to write. After all, they can always hire a scribe if they need to and there really aren't enough jobs for scribes to justify teaching it to everyone.
But if I'd rolled out USB A sockets in 1995, I don't think I'd object strongly to replacing the faceplates on the sockets with USB C ones in the next five years, if the wires in the wall could supply the required power.
I have yet to see a USB-C connector yet, and I am usually a first adopter.
No one you know has a MacBook Air? Most of the next generation of mobiles are going to have USB C (Apple and Google are among the bigger backers), so expect to see a lot of them appearing.