Comment Re:When will slashdot follow? (Score 2) 187
but when I became a man, I put away childish things.
Including the fear of being perceived as being childish?
but when I became a man, I put away childish things.
Including the fear of being perceived as being childish?
Probably not the right thread for it, but you're answering questions, so what the hell!
Will the non-javascript classic (the REAL classic version) be staying? I prefer that version, but I sometimes have to browse with scripts disabled in order to get it to work (actually, I usually read from Dillo or Links2 neither of which do JS, because I like the responsiveness). Unfortunately, though this means that I can't metamoderate any more since that requires JavaScript.
What if you don't like your head? Ugly people could have their brains transplanted into pretty people. Think of the potential for sex changes too! Then just as there's a market for stealing kidneys, there will arise a market for stealing whole bodies. You don't have to steal a rich person's digital identity. You can steal their WHOLE LIFE. Mind you, you'd have to study up really hard on every detail of their life and even learn to imitate their accent and speach patterns exactly, so their friends don't catch on. And then, with everyone knowing about the potential for this kind of identity theft, imagine someone undergoes a personality change due to illness -- everyone will assume their brain was replaced by an imposter, because most people are fucking clueless about mental illness.
Your comment is probably the most insightful here.
Even with extreme optimism about neurplasticity and nerve cells sliced in the middle (not separated at the synapses but chopped like a stalk of celery) deciding to attach to other sliced neurons, the idea of taking one spinal cord and gluing it to another spinal cord and having ANYTHING line up right seems absurd to me. Talk about a registration problem! I suspect the only way to do it would be to be to perform microscopic surgery where tiny machines connect neurons one-by-one. Of course, even if we had that technology, there's a whole other problem of knowing which ones to connect to which, which probably doesn't have a real solution as you pointed out.
Yes because waiting so long that vendors develop incompatible libraries and then shipping incredibly buggy system libraries then not fixing them for a whole major version has no effect on whether the apps reliably support the feature, because apps are *magic* and don't need system libraries.
BLE on android is a total clusterfuck and as a result app compatibility across versions is bad.
"great games shipping for linux
Well, I've been playing Creeper World III recently. I loved 1 and 2, and 3 is available native on Linux, where as I had to run 2 in a Windows VM. it's an indie game, so not super fancy but it runs on low-end hardware and the game mechanic is interesting. It also has a lot of built in content, a huge amount of very imaginative user generated content and a remarkably good random level generator.
It's not.
The only people who claim so are, frankly, ignorant.
The ignorance comes because Linux is easier to use for development than the alternatives. You just apt-get install the libraries you need and get hacking.
In order to make something portable you need to do what you have to do on Windows anyway: package all the libraries with your program.
It's just that by default Linux is much easier in that regard.
Except Bluetooth Low energy.
For some reason despite having support in the Kernel, it took them ages to support it on Android. 4.3 or newer only unless you want to also use some phone endor's own libraries for pre 4.3. That's a pretty bad indictment when several vendors had to make their own, incompatible libraries for hardware because Android dropped the ball.
Oh and the 4.x series has some really nasty bugs too and they're only fixed in 5.x. the recommended solution is to turn bluetooth off and on again and if that doesn't work, reboot the phone.
Not an iOS fanboi (I hae no iOS devices), but Android is awful and really driving me up the wall right now.
Do you have a link for that? It's not that I disbelieve you, I strongly suspect that Intel would do that crap. I'd like to read more about it however if you hae a link handy, then stash the link for the next time this benchmark comes up.
Personally, I like the Phoronix Linux benchmarks. They're more meaningful for me since I use Linux and they're all based on GCC which is trustworthy.
http://www.phoronix.com/scan.p...
The i7 4770 ocasionally blows away the FX8350 by a factor of 2, but in many benchmarks they're close, and Intel loses a fair fraction. The 4770 is the best overall performer, but not by all that much. It seems that the choice of CPU is fairly workload dependent.
For servers, I still prefer the supermicro 4s opteron boxes. 64 cores, 512G RAM, 1U. Nice.
Yeah it sounded like that right up until I said other architectures might be competitive. I made a typo. You're the one who decided to be a condescending neckbeard about it.
Dig around. Someone's ported noscript to Android: I'm happily running it.
There is something peculiar about peanuts in the US that seems to make them more allergenic. Peanut allergies occur less in other parts of the world. (Or so some people tell us anyhow.)
One hypothesis is that we just don't have enough parasites here. IgE immune response primarily targets parasites. Without them, it has nothing to do and starts attacking harmless things like food proteins and environmental allergens like pollen. Elsewhere in the world, those who would otherwise be allergic to peanuts are instead having their immune systems busy with the parasites.
Another hypothesis is that there's something growing on the peanuts here that people are actually reacting to, like a mold or fungus. In fact, that is a commonly offered explanation for corn allergies as well.
I don't know the status on GMO and peanuts, but some people seem to think that genetic modifications are correlated with higher rates of allergies. That's probably just noise in the data, assuming there IS any data.
No it's not
Indeed there was a word missing. If you read the post, it'd be clear from the content that I am aware of other architectures since I clearly referred to other architectures.
Nonetheless your condescending attitude neatly matches your nick.
The vleartext is actually:
Tb shpx lbhefrys.
Gung vf nyy.
I think this busts the physics, unless I misunderstand completely. Paging Dr. Shannon...
Nope.
Think about baseband for a moment.
Let's say you hae a bandwidth of 100MHz.
You can basically change from 0v to 1v 100e6 times per second, giving 100Mbit/s.
But you can also introduce more symbols. If you have 10 voltage levels between 0 and 1 V you get 1Gbit
What limits the number of symbols is noise. The datarate is symbol rate * bits per symbol. In the absence of noise, you can transmit an infinite amount of data in a 1Hz channel.
For non baseband signals, they generally use QAM to get symbols spanning the whole phase space around the centre frequency.
"Experience has proved that some people indeed know everything." -- Russell Baker