Want to read Slashdot from your mobile device? Point it at m.slashdot.org and keep reading!


Forgot your password?
Check out the new SourceForge HTML5 internet speed test! No Flash necessary and runs on all devices. ×

Comment Re:Who called it AI? (Score 1) 30

Well, I guess the difference is that while it would be a stretch to call this AI, it is very reasonable to call it machine learning - and that's what they called it. And yeah, I get it, you have some pointless axe to grind about people generally overusing the term AI. Nobody cares.

But I will say that this exact comment, which is a waste of space every time it crops up on this kind of story, at least provided me a laugh this time. Your shitty, pouty reply is just icing on the entertainment cake. I would have felt maybe a bit mean if you'd actually copped to shitposting without reading the summary. But I don't imagine you're that kind of person (ie. someone with a shred of dignity or self awareness).

Comment Re:Very true (Score 1) 2837

I'm not saying I necessarily agree with dude, but the situations are different.

When picking between Hillary and Bernie, voters aren't just deciding which of those two they want to win, they're also trying to help "a Democrat" winning the general election. The idea, commonly presented at the time, was that a protracted race would hurt Hillary as the candidates hit each other, and that in a long fight with Bernie Hillary would be forced further to the left. Thus if you, as a Democratic primary voter, believe Hillary is going to win regardless, you might as well avoid the long fight and just vote for her to help win the general.

You could expect sort-of-related polling effects in a general if there was a third party (ie. I want Hillary to win over Trump, but she's so far ahead I feel safe to use my vote to support the Green Party or whatever) - but I don't think that was much of a factor this time. You may have seen some depressed turnout for Hillary, though, as some may have felt it was a done deal.

Overall, whatever polling may have done, I think the much more explanatory factors are "Many people don't like Hillary" and "People are tired of doing what the media says" (which we saw previously with Brexit).

Comment Couldn't they have addressed the privacy concern? (Score 2) 104

Make the site request permission, like it would for the camera or GPS location.

I don't see this is as some big thing, but just because I can't think of an important use case doesn't mean there isn't a good one somewhere. Surely somebody wanted this for some reason? Anyway, it seems weird to introduce it then take it back over concerns that seem pretty mild, and also pretty easy to address the same way other concerns have been addressed in the past.

Comment Re:I'm speaking as someone who manages IT... (Score 1) 524

I don't think you're being fair here. As before, I believe (based on experience) that Macs overall will require less support given time, and I'm aware of lots of problems with Windows PCs. I just don't believe it's going to stay at an 8 times ratio. Like you say, there's going to be lots of users that will require support no matter what they're given (ie. the kind of people who can't find the start menu). If nothing else, they'll be the leveler here. There's no way you'll keep the "Mac users who need support" at 1/8th as you bring more of "them" into the mix.

I think you're imagining that I'm fighting for some Windows side or something - I'm not, I'm just suggesting that we temper unrealistic expectations.

Moving on, it's "hand waving", not "hand waiving", and I was right to ignore his sad little anecdote. Lots of people who are into computers (of all types) manage to keep their computers going without support; the fact that he was able to keep his home Mac going brings nothing to the conversation. Would you find it interesting to know that my non-techie dad manages to keep a Windows PC going, year-after-year without help? Does that change your view of the whole thing? Of course not; that's the kind of dumb crap you bring up when you have no facts or knowledge, but you feel like you need to write -something- in your condescending garbage post.

Anyway, yeah, my post was a little sarcastic in that previous post, but when buddy is like "facepalm, don't you know Macs don't need any support?", my inclination to be polite kind of evaporates.

Comment I'm speaking as someone who manages IT... (Score 2) 524

...at a medium sized company that supports Windows, Mac, and Linux desktops. I'm more on the programming side, but I stay on top of the support issues for various departments. Macs need tech support largely for the same reason Windows users do: because most users aren't terribly computer savvy, aren't confident enough to just try plugging things in, make dumb mistakes, and generally don't know where to find easy answers.

From my experience, Macs need very little tech support when we give them to, say, the publications department - but become much more problematic for field staff and managers (especially to start) because things aren't where they've grown to expect them to be, because of limited software availability, and because of more limited "local guru helpers" (ie. that guy in cubicle 4 who's into computers).

So when I say that I wouldn't think IBM will see this sort of support benefit ratio as they move to wider roll out, I'm doing so based on experience, and also on a suspicion that IBM has motivation to present this information in an exaggerated way (a suspicion confirmed by insider perspectives in other comments).

But now that I know that you, personally, haven't had problems with your Macs... well that changes everything. Thanks so much.

Comment Were the users randomized? (Score 5, Insightful) 524

I mean, I'm sure our Linux users overall require the least tech support. But that's a function of who they are more than what they're using.

I don't doubt that Macs require less support, but 40% vs 5% says that something else is going on - and I doubt that sort of ratio will hold once people are converted in bulk.

Comment Re:this can't be (Score 3, Interesting) 43

..and Amazon will likely retain its lead in supporting large and/or public facing websites.

But there's a lot of businesses, usually non-IT focused ones, who will remain Microsoft shops. For them, hosting their internal, B2B, and smaller public applications on Azure makes a lot of sense, and it's going to be a growth area for MS for a while. Yes you can host your .NET applications lots of places, but all the Azure nonsense is baked into Visual Studio and so, to the extent it basically works, it's going to be the path of least resistance for a lot of people.

In a lot of ways, MS is in a different market than other cloud vendors.

Comment Don't worry guys, nothing will ever really change (Score 1) 917

In the past, technology destroyed some jobs but created many more. This has been happening for thousands of years and thus will continue to happen forever. People have worried about this in the past and been wrong, thus they are wrong now.

To summarize: technological progress will continue forever, but we'll never need to adapt economic policy because people will always be able to contribute something that machines can't - and those things will employ enough people for near full employment, in perpetuity, regardless of the capabilities of machines. Like most of the other reflexively doubtful posters in this thread, I can't posit what kind of things those will be (personal touch, maybe?) - but I'm sure it'll be something because:

1. People have been wrong previously about this
2. Cotton gins and farmers
3. Socialism is wrong
4. QED

Comment I don't use Netflix for movies much anymore (Score 5, Interesting) 181

I'd much rather Netflix spends their money on TV shows (especially originals) than chasing expensive, popular movies. If I feel I need to watch The Dark Knight again (and I don't expect to) I'll find a way. No - I stay subscribed to them for TV: Stranger Things and House of Cards and Better Call Saul.

Well, that and my kids have been into Digimon lately.

Comment Re:This is not a good next step (Score 1) 63

I don't have anything to add, and I don't mean to be weird... but, uh, thanks for posting. For the first time in a while, I've come away from a discussion on Slashdot with something new to think about.

Your perspective makes sense... and I am going to be wondering all day why wireless mice still perform so poorly.

Comment Re:This is not a good next step (Score 1) 63

You may well be right on graphics and how this will play out; I don't know what kinds of things people will make and what will catch on. I'm sure a lot of my opinion is just informed by what I want personally: the same setup I have now, but wireless.

And yeah, I'd be surprised if the HDMI->WiFi->HDMI type things end up producing something good (even though these guys in particular seem confident). Like, you say, I expect a proper wireless solution will require a custom protocol.

Comment Re:This is not a good next step (Score 1) 63

I agree that eventually a standalone option will be preferable - I just don't think you can make a good enough one now, and I think attempting it is going to mean either huge costs (since you're not reusing current hardware) or huge compromises (ie. terrible performance).

I'm currently running a Vive with a 1080, and I still often can't keep framerate perfect (which you really want) at good supersampling (which makes a huge difference too). Sure you'll get some advantages with a dedicated device and integration, but it won't nearly make up for what you're losing in raw horsepower. Worse, requirements still have a ways to go up before they settle - a HMD really wants ~4K/90FPS at very high quality (well, and very low latency). Even without the extra wrench of 2 perspectives, most normal PC games still can't hit that, even on very large advanced hardware that draws big watts. VR needs more raw power, not less. If there were easy shortcuts to get this stuff, PCs and consoles would be using them; turns out sometimes you just need a bajillion transistors and a big power supply.

While I don't think positioning is the hardest problem for a standalone solution to solve, it is a real problem and you could improve current GearVR solutions 10 fold and still have something that's garbage. Tracking needs to be pretty much perfect or else VR is a barf-fest. Eventually inside-out camera based positioning might be good enough - but it has a tough hill to climb to match current state of the art. SteamVR's spinning lasers and fancy algorithms are magically good - crazy accurate, fast, and even cheap to build. They completely outclass current camera based solutions (like the Oculus uses) even though those solutions are much simpler than inside-out (because it's easier to track a diode you control than random surroundings you don't). Tracking tech really is the magic of Vive, and even a small downgrade could really break the experience.

And yes, currently there's nobody doing wireless video with low enough latency, but that's not because it's an insurmountably hard task - it just needs dedicated work, and VR is a good reason to get around to that work. The hard part - bandwidth - is already there, you just need to trim out some protocol transitions and latency would be very good, perhaps better than wired HDMI by the end (assuming that we can start the wireless chain directly from the GPU). Even current Rube Goldberg setups - HDMI->WiFi->HDMI, often on general purpose computers - are close to good enough. I think this is solvable, but it does remain to be seen whether someone with enough juice (eg. NVidia) will give it a proper go.

Anyway, I'm getting a lot of fun out of VR already, and I think it's going to progress really fast over the next couple years. I'd be pleased as punch of someone came out with a great standalone solution, I just think that's still a ways off.

Comment This is not a good next step (Score 1) 63

The clear next step, to me, is wireless connection to a computer.

A good VR experience takes way more grunt than you're going to get in a low-power head mounted device anytime soon. Current VR experiences benefit from any extra bit of GPU horsepower, compute speed, or basically any improvement they can get. Using a Rift or Vive, you're working with i7s and GTX 1080s, and still you're not able to run consistently at the supersampling levels you want. Ideally, you'd also have a higher resolution screen with less latency than you're getting in current HMDs. To work well, VR will need even more resources going forward, not way less.

Also, making a standalone device, such that you're not able to reuse your current cell phone or home computer, isn't the path to making your overall device less expensive. For right now, a wireless connection to a computer is the most sensible solution from a technical and economic perspective, assuming they want a reasonable quality experience in the end.

Anyway - yes, wireless would be great, positional tracking is absolutely necessary (and the current "professional" Oculus Rift isn't good enough - the Vive is way better - and yes I've used both of them a lot), and it's all too expensive. Eventually obviously we should expect fully standalone devices, but for now it's just not going to happen, so this is not a reasonable step forwards at this time.

Slashdot Top Deals

Machines have less problems. I'd like to be a machine. -- Andy Warhol