Follow Slashdot blog updates by subscribing to our blog RSS feed

 



Forgot your password?
typodupeerror
×

Submission + - Is It Time to Call BS on "The Retention Policy Ate My Communications" Excuse?

theodp writes: The FTC is accusing Amazon execs, including founder Jeff Bezos, of using encrypted messaging apps that automatically delete messages to communicate, even after they were notified they were under investigation. The FTC is asking a judge to force Amazon to produce documents related to the company’s failure to preserve Signal messages, the company’s document preservation notices, and its instructions about using disappearing messaging applications. The FTC alleges Amazon execs did this while discussing "sensitive business matters, including antitrust" (instead of using email) to destroy potential evidence. Google also came under fire this week in its antitrust case over an issue about whether it intentionally deleted or failed to retain documents that might have been used as evidence in the trial. Google had a policy of having 'history off' on its chats by default, leaving it to employees [including CEO Sundar Pichai] to determine when to turn it on for relevant conversations (akin to some police bodycam policies). The Department of Justice (DOJ) called the alleged destruction of documents "unequivocal and honestly breathtaking," adding that "there’s no question" executives "intentionally had conversations with history off." "Google’s retention policy leaves a lot to be desired," said the judge, adding disapprovingly that it was “surprising to me that a company would leave it to their employees to decide when to preserve documents." And back in 2018, Facebook acknowledged that a secret Messenger retention policy feature was the cause of CEO Mark Zuckerberg's mysteriously disappearing messages.

Which begs the question — are Mission Impossible-like self-destructing email, messaging, and document policies beneficial to rank-and-file employees, or is this more about a play to "reduce your risk in the event of litigation [...] by permanently deleting old content that you're no longer required to keep," as Microsoft explains? Microsoft goes on to claim that destroying all of your employees' communications — like the University of Washington's just-implemented Microsoft Teams Chat Message 'Retention' Project that calls for destroying all of the university's messages after 30 days with 'no exceptions' (UW also suggests other FOIA-dodging 'best practices') — will also "help your organization to share knowledge effectively and be more agile by ensuring that your users work only with content that's current and relevant to them." However, former Microsoft Researcher Jonathan Grudin (coincidentally a UW affiliate professor) found plenty of pushback on the idea of improving-knowledge-by-deleting-communications when the company unsuccessfully tried to make Microsoft employees eat their own retention policy dogfood that the company was selling to other organizations. Grudin explained in a 2021 interview:

"Now I'll describe a couple unpublished projects. One was an email system. Someone said, 'We call it email retention but really it's email deletion.' We were told that starting the next April, all email a year old would be automatically deleted. IBM had such a system and some of our customers wanted it. I contacted friends at IBM who described it as a nightmare. [...] Why did we think it would be a good idea to use it internally at Microsoft? Some guessed storage costs, but those were dropping daily. Well, companies might have bodies that they'd like to remain buried, conversations that they would prefer not to surface. But you can't legally destroy inculpatory evidence, and an embarrassing remark that makes headlines generally has little weight in court where they look for patterns of behavior over time. The real reason turned out to be discovery costs. Microsoft and many companies are involved in far more legal proceedings than you read about. They have to pay attorneys to read all subpoenaed emails. It reportedly came to about $30 million a year. A team of about 10 people were managing the email deletion project. Some had given up other jobs to work on it, because they loved this idea. Most had information management backgrounds. They believed that only records with business value should be kept. Seeing big email folders 'makes my skin crawl,' one remarked. This view came from an era of paper documents and Rolodexes when filing and finding documents was manual. It was really difficult. It was expensive. Whereas for me and others, email is a Rolodex as well as a source of a lot of information whose future value we don't know."

"I learned that 1000 Microsoft employees were testing the software, a process referred here to as eating dogfood. I asked how it was going for these folks. An information manager beamed and said, 'It's working!' [...] I asked, 'What do the employees using it, think about it?' This surprised the team. It never occurred to them to ask. They were sure that the employees would see the value of email deletion for the company. They were really curious. They did realize that a survey and interview might uncover gripes, but they wanted to find out. [...] The interviews, which of course did find ingenious and time-consuming ways that people were dodging deletion. [...] So what did we find? Well, the cost to the company, in lost time and effort from email deletion, would easily exceed $30 million annually. [..] The deployment was canceled. [...] A partner in a San Francisco law firm heard about my findings and called up. He said that some companies would use email deletion software, whatever the cost. He explained, 'Phillip Morris is in the business of addicting people to something that will kill them. They'll pay what they need to as long as the business is profitable. Once it stops being profitable, they'll stop.'"

Submission + - Trickle-Down Economics: Giving Kids the Best AI Education $5 Gift Cards Can Buy

theodp writes: The past week saw AI-fueled quarterly earnings announcements from the likes of Amazon ($143.5B revenue, $1.86T market cap), Microsoft ($61.9B revenue, $2.93T market cap), and Google ($80.5B revenue, $1.11T market cap). All three are advisors to Code.org's TeachAI initiative, which aims to help teachers incorporate AI into K-12 CS education. Towards that end, TeachAI and the Computer Science Teachers Association (CSTA) are now looking to give kids the best AI education 100 $5 gift cards can buy.

From the CSTA Ohio website: "Shape the future of CS Education in an age of AI! CSTA and TeachAI invite K-12 CS teachers to share their expertise and insights in a short survey. Your responses will be anonymous and contribute to a greater understanding of what’s happening in CS classrooms today. What works? What do we still need to know? What are the risks and benefits? How can we prepare our students to learn and thrive in an age of AI? We will use this information to help us develop guidance for educators that will be released at the CSTA Conference in July 2024. The first 100 respondents will receive a $5 gift card."

TeachAI also receives support from the Bill ($126.8B net worth) and Melinda ($11B net worth) Gates Foundation ($75B endowment). So, is trickle-down economics an effective K-12 AI education funding strategy?

Submission + - Bill Gates Is Still Pulling the Strings at Microsoft

theodp writes: According to Business Insider, reports of the death of Bill Gates' influence at Microsoft have been greatly exaggerated: "Publicly, [Bill] Gates has been almost entirely out of the picture at Microsoft since 2021, following allegations that he had behaved inappropriately toward female employees. In fact, Business Insider has learned, Gates has been quietly orchestrating much of Microsoft's AI revolution from behind the scenes. Current and former executives say Gates remains intimately involved in the company's operations — advising on strategy, reviewing products, recruiting high-level executives, and nurturing Microsoft's crucial relationship with Sam Altman, the cofounder and CEO of OpenAI. In early 2023, when Microsoft debuted a version of its search engine Bing turbocharged by the same technology as ChatGPT, throwing down the gauntlet against competitors like Google, Gates, executives said, was pivotal in setting the plan in motion. While Nadella might be the public face of the company's AI success — the Oz who built the yellow-brick road to a $3 trillion juggernaut — Gates has been the man behind the curtain."

"Today, Gates remains close with Altman, who visits his home a few times a year, and OpenAI seeks his counsel on developments. There's a 'tight coupling' between Gates and OpenAI, a person familiar with the relationship said. 'Sam and Bill are good friends. OpenAI takes his opinion and consult overall seriously.' OpenAI spokesperson Kayla Wood confirmed OpenAI continues to meet with Gates."

Among the three new members named to OpenAI's Board of Directors in March 2024 following Altman's return to the company was Sue Desmond-Hellmann, former CEO of the Bill and Melinda Gates Foundation. And in January, Microsoft exec Dee Templeton joined OpenAI’s board as Microsoft's non-voting observer as part of the OpenAI boardroom revamp. Templeton is an advisor to Microsoft CTO Kevin Scott, who has been credited with brokering the Microsoft-OpenAI partnership. Talking to Bill Gates in a March 2023 podcast, Scott shared a secret: "By the time this podcast airs, OpenAI will have made their announcement to the world about GPT-4 (prompting Gates to declare The Age of AI Has Begun), but I want to sort of set the stage. The unveiling of the first instance of GPT-4 outside of OpenAI was actually to you last August at a dinner that you hosted with Reid [Hoffman] and Sam Altman and Greg Brockman and Satya [Nadella] and a whole bunch of other folks [including Scott]. Earlier that same month, a Microsoft spokesperson downplayed a report by The Information on Gates' role in the Microsoft-OpenAI pact.

Submission + - Microsoft Says Monolithic Architecture Impeded Its K-12 CS Racial Equity Goals

theodp writes: In June 2020, Microsoft CEO Satya Nadella outlined a series of commitments designed to address the racial injustice and inequity experienced by racial and ethnic minorities in the US, including Black and African American communities. "Over the next five years," Nadella wrote, "we will expand in 13 states and the District of Columbia the Microsoft TEALS industry volunteer program to bring computer science education to an additional 620 high schools primarily serving Black and African American students."

But standing in the way of that expansion commitment, according to a new Microsoft blog post on Turning to Microsoft Azure to Put Software Engineers in High School Classrooms, was the TEALS Operational Platform (TOP). "Legacy TOP couldn’t support the directive," Microsoft explains. "TOP was about to be outdated and out of support; it had more than 500 bugs, and accessibility improvements were needed. Furthermore, it couldn’t perform well with all the data it was handling. 'The TOP legacy system was at risk of breaking down, posing significant risk to the program’s operations along with system security, reliability, and availability,' [Sr. Product Manager Emily] Fishkind says. Importantly, it couldn’t scale. The original TOP that had enabled the program’s early development was now holding it back from further growth and impact."

In response, Microsoft recently revamped the TEALS Operational Platform and says the new built-on-Microsoft-Azure TOP vNext platform allows regional managers of TEALS to incorporate many more schools, which will allow the company to get many more Microsoft and other software engineer volunteers into classrooms. "The adoption of microservices architecture in TOP vNext protects TEALS from the risk of total system breakdown, offering a more resilient and stable system," Microsoft explains, describing legacy TOP as an outdated system posing security risks by comparison. "The traditional monolithic architecture, where a single malfunction could cause a complete system failure, has become obsolete. With microservices, TEALS is now immune to total system breakdowns as each component operates independently and can continue functioning even if one component experiences issues. This independence not only increases system reliability but also facilitates efficient testing and maintenance processes."

Submission + - Amazon Board Member Andrew Ng Pushing States for CS+AI HS Graduation Requirement

theodp writes: "We're pleased to welcome Dr. Andrew Ng to our Board of Directors effective April 9, 2024," Amazon announced last week, noting that AI legend Ng is the Managing General Partner of AI Fund, a venture studio that supports entrepreneurs to build AI companies." Among AI Fund's investments is Kira Learning, a STEM-focused K-12 education publisher that aims to transform computer science and AI education for young learners. Kira, where Ng is Chairman, is partnering with the State of Tennessee to implement the state's new computer science high school graduation requirement. Tennessee lawmakers earned praise in a VentureBeat op ed (Schools Should Teach AI to Every Child) penned by Ng and Kira CEO and co-founder Andrea Pasinetti. Ng and Pasinetti were both signatories to a letter issued by the CEOs for CS, part of a campaign coordinated by tech-backed nonprofit Code.org that was credited with pressuring the nation's Governors into signing a compact agreeing to make their states' children more CS-savvy (signers included TN Gov Bill Lee). Ng's and Kira's efforts to convince state officials of the need to make CS a high school graduation requirement don't stop there.

News sites, state legislature websites, and social media have noted the lobbying efforts of CS Forward, an advocacy group that calls for "Computer Science Education In Every State" and notes it "unites non-profits, educators, companies, and community leaders in advocating for state computer science graduation requirements." On its About page, CS Forward explains, "Our mission is to ensure every K-12 student masters CS and AI coding. We're here to influence state legislation, budget allocations and shape public policy." A graphic on the page hints at the advocacy group's ties to Kira Learning, as does the presence of Kira Chairman Ng and Kira co-founder and Director of AI Applications Jagriti Agrawal on CS Forward's Advisory Board (Kira's Head of Public Affairs Dave Brown also does double-duty as the Director of CS Forward). CS Forward recently celebrated a legislative victory that saw the Louisiana House of Representatives vote 102-1 to pass a CS high school graduation requirement. A graphic circulated on Twitter alerting the LA Senate of the near-unanimous vote sported the logos of CS Forward, Kira, Code.org, and Code.org Platinum Supporters ($3+ million) Amazon & Microsoft. California State legislature records note that the same five groups registered support for requiring CS for high school graduation in CA. CS Forward LinkedIn posts cite additional efforts to pass bills expanding CS education and making it a high school graduation requirement in Indiana (mission accomplished), Washington (defeated, despite help from Code.org), Iowa, and Minnesota.

Ng has argued that "Laws to ensure AI applications are safe, fair, and transparent are needed." Which begs the question: Are laws also needed to ensure K-12 AI+CS education lobbying is transparent?

Submission + - Teen Girls Confront an Epidemic of Deepfake Nudes in Schools

theodp writes: At Macworld Expo 1998, the NY Times reported, Steve Jobs unveiled new software that he explained could be used to give your daughters a Cosmopolitan Virtual Makeover (video). A quarter of a century later, the NY Times reports that middle and high school students are using AI software to fabricate explicit nude images of female classmates and share the doctored pictures.

From the article: "Blindsided last year by the sudden popularity of A.I.-powered chatbots like ChatGPT, schools across the United States scurried to contain the text-generating bots in an effort to forestall student cheating. Now a more alarming A.I. image-generating phenomenon is shaking schools. Boys in several states have used widely available 'nudification' apps to pervert real, identifiable photos of their clothed female classmates, shown attending events like school proms, into graphic, convincing-looking images of the girls with exposed A.I.-generated breasts and genitalia. In some cases, boys shared the faked images in the school lunchroom, on the school bus or through group chats on platforms like Snapchat and Instagram, according to school and police reports."

So, to paraphrase the flying car trope: They promised us AI would help all students achieve their educational goals, instead we got deepfake nudes and plagiarism and cheating.

Submission + - Code.org Launches AI Teaching Assistant for Grades 6-10 in Stanford Partnership 2

theodp writes: From a Wednesday press release: "Code.org, in collaboration with The Piech Lab at Stanford University, launched today its AI Teaching Assistant, ushering in a new era of computer science instruction to support teachers in preparing students with the foundational skills necessary to work, live and thrive in an AI world. [...] Launching as a part of Code.org's leading Computer Science Discoveries (CSD) curriculum [for grades 6-10], the tool is designed to bolster teacher confidence in teaching computer science." EdWeek reports that in a limited pilot project involving twenty teachers nationwide, the AI computer science grading tool cut one middle school teacher's grading time in half. Code.org is now inviting an additional 300 teachers to give the tool a try. "Many teachers who lead computer science courses," EdWeek notes, "don’t have a degree in the subject—or even much training on how to teach it—and might be the only educator in their school leading a computer science course."

Stanford's Piech Lab is headed by assistant professor of CS Chris Piech, who also runs the wildly-successful free Code in Place MOOC (30,000+ learners and counting), which teaches fundamentals from Stanford’s flagship introduction to Python course. Prior to coming up with the new AI teaching assistant, which automatically assesses Code.org students' JavaScript game code, Piech worked on a Stanford Research team that partnered with Code.org nearly a decade ago to create algorithms to generate hints for K-12 students trying to solve Code.org's Hour of Code block-based programming puzzles (2015 paper). And several years ago, Piech's lab again teamed with Code.org on Play-to-Grade, which sought to "provide scalable automated grading on all types of coding assignments" by analyzing the game play of Code.org students' projects. Play-to-Grade, a 2022 paper noted, was "supported in part by a Stanford Hoffman-Yee Human Centered AI grant" for AI Tutors to Help Prepare Students for the 21st Century Workforce. That project also aimed to develop a "Super Teaching Assistant" for Piech's Code in Place MOOC. LinkedIn co-founder Reid Hoffman, who was present for the presentation of the 'AI Tutors' work he and his wife funded, is a Code.org Diamond Supporter ($1+ million).

Submission + - AI's Impact on CS Education Likened to Calculator's Impact on Math Education

theodp writes: In Generative AI and CS Education, the new Global Head and VP of Google.org Maggie Johnson writes: "There is a common analogy between calculators and their impact on mathematics education, and generative AI and its impact on CS education. Teachers had to find the right amount of long-hand arithmetic and mathematical problem solving for students to do, in order for them to have the “number sense” to be successful later in algebra and calculus. Too much focus on calculators diminished number sense. We have a similar situation in determining the 'code sense' required for students to be successful in this new realm of automated software engineering. It will take a few iterations to understand exactly what kind of praxis students need in this new era of LLMs to develop sufficient code sense, but now is the time to experiment."

Johnson's CACM article echoes comments she made in a featured talk called The Future of Computational Thinking at last year's Blockly Summit (Blockly is the Google technology that powers drag-and-drop coding IDE's used for K-12 CS education, including Scratch and Code.org). Envisioning a world where AI generates code and humans proofread it, Johnson explained: "One can imagine a future where these generative coding systems become so reliable, so capable, and so secure that the amount of time doing low-level coding really decreases for both students and for professionals. So, we see a shift with students to focus more on reading and understanding and assessing generated code and less about actually writing it. [...] I don't anticipate that the need for understanding code is going to go away entirely right away [...] I think there will still be at least in the near term a need to understand read and understand code so that you can assess the reliabilities, the correctness of generated code. So, I think in the near term there's still going to be a need for that." In the following Q&A, Johnson is caught by surprise when asked whether there will even be a need for Blockly at all in the AI-driven world she describes, which she concedes there may not be.

Johnson's call to embrace AI to "raise the level of abstraction for software engineers" to boost their productivity comes as she exits the Board of Code.org, the tech-backed K-12 CS education nonprofit that pushed coding — including Java — into K-12 schools, but deviated a bit from their 'rigorous CS' mission last year to launch a new TeachAI initiative with tech industry partners to convince K-12 schools to embrace AI to increase the productivity of teachers and students not only in CS, but also in all other areas of education. Johnson's departure from Code.org — she was a founding Board member in 2013 — follows that of Microsoft President Brad Smith, Code.org's other founding Board member from industry, who has been focused on promoting Microsoft's AI efforts. Unlike Google, Microsoft is still represented on Code.org's Board by CTO Kevin Scott, who is credited with forging Microsoft's OpenAI partnership (with Smith and Microsoft CEO Satya Nadella) and whose assistant Dee Templeton joined OpenAI's Board as Microsoft's nonvoting observer in January following Sam Altman's reinstatement as OpenAI's CEO. Hey, it's a small K-12 CS and AI education world!

Submission + - Gates Foundation Gives $500K to Model Karlie Kloss's Free Coding Camp for Girls

theodp writes: In March, the Bill and Melinda Gates Foundation committed $500,000 to Kode With Klossy Incorporated for general operating support. Kode with Klossy, a charity founded by supermodel Karlie Kloss, operates free coding camps for girls aged 13-18.

Last September, Kloss spoke about coding at the Bill and Melinda Gates Foundation's Goalkeepers event in New York City during a panel moderated by Phoebe Gates (BillG's daughter) on The Power of the Women's Potential: "You know, it was almost a decade ago I started Kode with Klossy out of just my own curiosity of what this thing that I kept hearing code was. And I was meeting all these entrepreneurs, building technology, transforming the world, and I wanted to understand what they knew that I and many others didn't know. And this superpower of code. And coding, really what I realized, is a language and a language that can allow you to write your own destiny, write your own narrative and solve problems that you're most passionate about. And to me, that is the ultimate definition of agency."

Kloss has credited Codecademy and Flatiron School for teaching her to code. Both startups received early funding from Thrive Capital, the VC firm headed by Kloss's then-boyfriend and now-husband Joshua Kushner. Before it was sold to WeWork, Flatiron School also helped Kloss create Kode With Klossy.

Submission + - Things Programmers Say On 4/1

theodp writes: Developers, tell me it's April Fool's Day without telling me it's April Fool's Day: @ "Don't worry about missed deadlines — we're using Agile, Scrum, and Jira now!" @ "I can't believe how much money and time we've saved with the Cloud!" @ "Thanks to ChatGPT, you can cut all my coding and testing time estimates by half! @ "Why would I resent being micromanaged by a Product Manager who can't be bothered to seek my input or try to understand what I do?" @ "Don't worry about quality, we've tested every line of code with every possible combination of data values!" @ "Thank goodness our Sr. Management is being advised by Accenture, McKinsey, and Gartner!" @ "Of course, we can block every bad thing on the Internet with a minimal staff!" @ "Just because this has never worked anywhere else doesn't mean it can't work here!" @ "We can always cut back on coding and testing time to make sure we get these PowerPoint presentations right!" @ "Hey, why don't we get people who no longer program to dictate the programming/architecture standards that must be followed?" @ "Just because Amazon, Apple, and Microsoft PhDs make mistakes doesn't mean we should expect any less than a 0% error rate from our people!" @ "There's no excuse for testing with anything but fully synthetic data when the tools make it so simple for even the most complicated data!" @ "I thought it'd be impossible to identify everywhere that data comes from, how it's transformed, where it goes, and how it's used and by whom, but Data Governance and Audit did the hard part by giving us with Excel templates (with no filled-in examples) to fill out to document it all!" @ "Don't worry about security — each request and every person in our global organization as well as contractors is thoroughly vetted to ensure every element they're permitted to access is justified!" @ "How is HR so spot-on with making sure each person gets the fair review and compensation they deserve?" @ "Why should I be paid extra for being on call when it's nobody's fault but my own if anything goes wrong?" @ "It's incredible how refactoring code makes even the most complex code drop dead simple to understand!" @ "Thankfully office politics plays no role in how decisions are made here!" @ "Isn't it great that we have whole departments whose job is only to tell you what you're doing wrong, not what's right?" @ Just because the most brilliant minds at tech giants can't solve certain problems doesn't mean your team shouldn't be expected to do so during the next sprint!" @ "Even if your CI/CD requirements are impossible, it's my fault if my code and tests can't pass them!" @ "Why shouldn't we be able to come up with precise estimates for incomplete requirements for something that's never been done before without any idea of who's going to be working on it or what technology they'll be using!"

Submission + - Growing Up Ballmer 1

theodp writes: Business Insider and others are running tidbits from an interesting conversation with 29-year-old Pete Ballmer (apparently drawn from a Cash Cuties podcast), a standup comedian living in San Francisco and one of the sons of billionaire and former Microsoft CEO Steve Ballmer. "Up until I was in late elementary school," Ballmer notes, "my understanding was that my family was rich, but I didn't know that we were globally and historically rich. [...] For both my mom and my dad, having a lot of money was a relatively new experience, as was raising children. They raised us in line with how their parents raised them, and since they didn't grow up talking about wealth, they didn't talk about it with us, either."

Interestingly, when Pete finally did get a life-changing windfall, it didn't come from his parents. "After college," Ballmer explains, "I never considered not having a job, so I became a product manager at a game development company [he's a Stanford CS grad]. Then I inherited a sum of money from my grandfather when I turned 25. He had worked his way up to middle management at Ford and put the money he saved into Microsoft stock, which did pretty well and ended up being worth hundreds of thousands of dollars by the time I received it. When I first heard about it, I was a junior in college. My initial reaction was that I would decline it — I was still pretty uncomfortable with my family's wealth and figured I could get a pretty high-paying job in tech and wouldn't need their money. But then I turned 25, and I didn't decline the money; in retrospect, that would have been a very silly decision. I'd also started doing standup comedy in college and continued doing it on the side while I worked. After about four years of working as a product manager, I quit to pursue comedy full-time."

For now, Ballmer is leading a modest life compared to some billionaires' kids ("I live in a two-bed, one-bath apartment") but notes, "As we're all older now, our family has started talking more proactively and intentionally about money. We've talked about what our wills might consist of, what happens to the Clippers — which my dad owns — once my parents have passed, how having the money affects what we choose to do career-wise, how the money has or has not 'corrupted' us, and the wariness we all have around money's general ability to do that to people."

Submission + - Google To 6-Year-Olds: Don't Be a 'Goofus', Get Permission Before Using Content 1

theodp writes: Last month, a Google-funded special edition of Highlights for Children based on Google’s Be Internet Awesome curriculum was created and 1.25 million copies of the print magazine were distributed to children, schools, and other organizations as part of a new partnership between Google and Highlights, the children's publication that targets kids aged 6-12.

A Google.org blog post calls out the special issue's Goofus and Gallant cartoon feature in which always-does-the-wrong-thing "Goofus promised Kayden he wouldn't share the silly photo, but he shares it anyway", while always-does-the-right-thing "Gallant asks others if it's OK to share their photos." Also called out is a Don't Fall for Fake puzzle, which Google explains is provided so "kids can learn to discern between what's real and what's fake online."

Submission + - DuckDB Swims Through JSON Data Like a jq Duck, But With SQL

theodp writes: Among the amazing features of the in-process analytical database DuckDB, writes Paul Gross in DuckDB as the New jq, is that it has many data importers included without requiring extra dependencies. This means it can natively read and parse JSON as a database table, among many other formats. "Once I learned DuckDB could read JSON files directly into memory," Gross explains, "I realized that I could use it for many of the things where I’m currently using jq. In contrast to the complicated and custom jq syntax, I’m very familiar with SQL and use it almost daily."

The stark difference of the two programming approaches to the same problem — terse-but-cryptic jq vs. more-straightforward-to-most SQL — also raises some interesting questions: Will the use of Generative AI coding assistants more firmly entrench the status quo of the existing programming paradigms on whose codebases it's been trained? Or could it help bootstrap the acceptance of new, more approachable programming paradigms? Had something like ChatGPT been around back in the Programming Windows 95 days, might people have been content to use Copilot to generate reams of difficult-to-maintain-and-enhance Windows C code using models trained on the existing codebases instead of exploring easier approaches to Windows programming like Visual BASIC?

Slashdot Top Deals

We gave you an atomic bomb, what do you want, mermaids? -- I. I. Rabi to the Atomic Energy Commission

Working...