With Newtonian physics, if we could find something that wasn't symmetrical with regards to conservation of motion, we have a perpetual motion system and unlimited energy. So far, every process that converts energy from one form to another always has some loss due to friction or heat.
The system bounces microwaves inside a cone. When a photon hits the base of the cone, that pushes the vehicle forward as it bounces backwards. But the conical shape means that the next "bounce" of the photon will be distributed between a sideways movement and a small backwards movement. But because of the circular shape of the cone, the sideways movements cancel out. The forwards force is greater than the backwards force, so the final velocity is forwards.
The information disclosed by Snowden can be reduced down to "The three letter agencies can convert any electronic device with a microphone into a hidden tape recorder" and "anything sent down The Tubes can also be recorded". So they meet in person and just leave their smartphones in the room outside.
Unfortunately for Airbus, it didn't quite work out when an airshow decided to have an aircraft do a low fly-pass in front of the crowds. The combination of low altitude, low speed with flaps and landing gear lowered made the AI think that the pilots wanted the plane to land. So the flight control system cut the engine power in preparation for landing.
I used acrylic plastic in my dental fillings. Hardened using UV light and still going strong after 20 years.
The problem is that blast furnaces aren't simply switched on and off, but have feedback software systems that adjusts fuel feeds, cooling systems and exhaust extraction to achieve the desired temperature while minimizing fuel consumption, cooling and pollution. Much the same way as electronic car ignition. The operating temperature would have to be ramped up and down slowly to avoid any damage through thermal stress.
It's the hardware overrides that would allow the cooling system to be reduced or switched off while the fuel feeds remain on.
We already have the slashdot story generator:
Back in those days, start of the art technology in CPU's were "restricted exports". The USA wanted to show that Communism didn't lead to as many advancements in technology as Capitalism, so they restricted exports on technology such as chip design software, CPU's and other chip logic (remember the A-team trying to block smugglers exporting flip-flop chips? It was that serious). This led to the Eastern European countries doing various work-arounds. They could get gray imports through third-party countries that weren't part of the Western trade block, and weren't part of the USSR either. Or they could set up fake companies in the host country that would export the technology.
Another strategy was to make their own logic chips. However, yields for complex logic such as CPU's, wasn't that good, so they ended up with CPU's with missing instructions. But that wasn't a problem, mathematician/software engineers figured out ways of emulating broken instructions using other instructions. If JMP was broken, then use CLR; BCC. Arithmetic operations like ADD could be replaced by NEG and SUB, and so on... So they ended up with an abstraction layer using assembler macros that provided a set of functioning instructions.
In Norway, you can just go online to the bank website, use an authentication system based on a username, password and your mobile phone.
Then you just use the IBAN/SWIFT system to transfer the money to the account anywhere else in the world, and you can download your transaction history as a spreadsheet file.
Other banks in the UK require you to go into a branch, and have a clerk use a quill pen to fill out an entry in a giant leatherbound ledger book.
But where do you start encrypting? You have to secure your PC against keyloggers and screen-grabbers even if your disk drive and communications are encrypted. And how would be sure that no man-in-the-middle intercept didn't have the processing power to crack the encryption?
It was the dot-com boom era. Established companies (bricks and mortar) were slow to catch onto the the internet presence, so there were thousands of startups which tried to fill that niche (Peapod who did online groceries, companies who get your bills redirected, so they could be opened and scanned in, so all you had to do was write the cheque to the utility company, and Amazon). The people that jumped into the CompSci courses at this time, were referred to as gold-diggers who were only interested in the money from knowing some HTML and scripting,
From the job descriptions available now, a software developer, software engineer or senior software engineer will consult with clients and other engineers, write design specifications, write, design and implement unit tests using Microsoft or Google Test frameworks, do task breakdowns, provide accurate time estimates, implement software modules, provide code reviews, keep bug lists up to date, mentor software engineers through techniques such as pair or extreme programming, participate in Agile and Scrum meetings. Depending on the field of work, it's either going to be "Big Data" with R, Scala, Hadoop, Java, "Embedded Systems" with C, embedded C, assembler, IOS or Android; or "Desktop Systems" with C#, Java. In every case, they wil five years or more experience.
Then the employer would only consider hiring H1B's from other companies and not do the petitioning themselves.
t I was trying to explain that there was a vast number of applications using classic supercomputer type technology, ranging from academic research down to multiplayer games. A modern game console now uses multiple cores, vector processors, vector chaining, kernels (if you consider vertex, fragment, geometry shaders as kernels), client-server communication to update players moves. Even geometry data is streamed across the network as some game MMORG worlds are so vast, all the data couldn't be stored on one disk drive. Then there are techniques of "sharding" games world so that the same geometry can be used, but with different groups of players.
Those numeric and simulation libraries were optimized in conjunnction with the Fortran compiler to take advantage of the hardware. The most obvious example; having fixed sized multi-dimensional arrays as global variables. For regular grids, the compiler can then decide which way to slice that data block up so that every processing node gets assigned a chunk of data. Since each function is not more than a few loop counters reading previous and current state for each grid cell, those get optimized into parallelised for-each calls. Once the scientists and engineers started writing and sharing numeric and simulation libraries and building on top of them, there isn't any need to change.