AI

Anthropic Launches Claude Pro, a Subscription AI That May Rival ChatGPT Plus (arstechnica.com) 9

An anonymous reader quotes a report from Ars Technica: On Thursday, AI-maker and OpenAI competitor Anthropic launched Claude Pro, a subscription-based version of its Claude.ai web-based AI assistant, which functions similarly to ChatGPT. It's available for $20/month in the US or 18 pounds/month in the UK, and it promises five-times-higher usage limits, priority access to Claude during high-traffic periods, and early access to new features as they emerge. Like ChatGPT, Claude Pro can compose text, summarize, do analysis, solve logic puzzles, and more.

Claude.ai is what Anthropic offers as its conversational interface for its Claude 2 AI language model, similar to how ChatGPT provides an application wrapper for the underlying models GPT-3.5 and GPT-4. In February, OpenAI chose a subscription route for ChatGPT Plus, which for $20 a month also gives early access to new features, but it also unlocks access to GPT-4, which is OpenAI's most powerful language model. What does Claude have that ChatGPT doesn't? One big difference is a 100,000 token context window, which means it can process about 75,000 words at once. Tokens are fragments of words used while processing text. That means Claude can analyze longer documents or hold longer conversations without losing its memory of the subject at hand. ChatGPT can only process about 8,000 tokens in GPT-4 mode.

Anthropic's primary selling point for the Claude Pro subscription is "5x more usage," but the company doesn't clearly communicate what Claude's free-tier usage limits actually are. Dropping clues like cryptic breadcrumbs, the company has written a support document about the topic that says, "If your conversations are relatively short (approximately 200 English sentences, assuming your sentences are around 15-20 words), you can expect to send at least 100 messages every 8 hours, often more depending on Claude's current capacity. Over two thirds of all conversations on claude.ai (as of September 2023) have been within this length." In another somewhat cryptic statement, Anthropic writes, "If you upload a copy of The Great Gatsby, you may only be able to send 20 messages in that conversation within 8 hours." We're not attempting the math, but if you know the precise word count of F. Scott Fitzgerald's classic, it may be possible to glean Claude's actual limits. We reached out to Anthropic for clarification yesterday and have not received a response by press time.

Sony

Sony Sends Copyright Notices To TV Museum About Shows 40 To 60 Years Old (torrentfreak.com) 61

An anonymous reader quotes a report from TorrentFreak: Rick Klein and his team have been preserving TV adverts, forgotten tapes, and decades-old TV programming for years. Now operating as a 501(c)(3) non-profit, the Museum of Classic Chicago Television has called YouTube home since 2007. However, copyright notices sent on behalf of Sony, protecting TV shows between 40 and 60 years old, could shut down the project in 48 hours. "Our YouTube channel with 150k subscribers is in danger of being terminated by September 6th if I don't find a way to resolve these copyright claims that Markscan made," Klein told TorrentFreak on Friday. "At this point, I don't even care if they were issued under authorization by Sony or not -- I just need to reach a live human being to try to resolve this without copyright strikes. I am willing to remove the material manually to get the strikes reversed."

Over the weekend Klein shared details of the copyright complaints filed with YouTube. Two of the claims can be seen in the image below and on first view, appear straightforward enough. Two episodes of the TV series Bewitched dated 1964 aired on ABC Network and almost sixty years later, archive copies of those transmissions were removed from YouTube for violating Sony copyrights, with MCCTv receiving a strike. A claim targeting an upload titled Bewitched -- 'Twitch or Treat' -- WPWR Channel 60 (Complete Broadcast, 8/6/1984) follows the same pattern, but what isn't shown are the details added by MCCTv to place the episode (and the included commercials) in historical context. Another takedown target -- Bewitched -- 'Sam in the Moon' (Complete 16mm Network Print, 1/5/1967) is accompanied by even more detail, including references in the episode to then-current events.

Given that copyright law locks content down for decades, Klein understands that can sometimes cause issues, although 16 years on YouTube suggests that the overwhelming majority of rightsholders don't consider his channel a threat. If they did, the option to monetize the recordings can be an option. [...] Klein says MCCTv certainly doesn't set out to hurt copyright holders. However, there's always a balance between preserving "rare pieces of video ephemera" and the likelihood that nobody needs to enforce any rights, versus unusual circumstances like these where unexpected complaints need to be resolved with impossible-to-reach parties. Klein says the team is happy to comply with Sony's wishes and they hope that given a little leeway, the project won't be consigned to history. Perhaps Sony will recall the importance of time-shifting while understanding that time itself is running out for The Museum of Classic Chicago Television.

Crime

'Starfield' Fan Banned From Subreddit For Narcing On Leaker To Cops (kotaku.com) 127

Kotaku reports that last week 29-year old Darin Harris "allegedly stole dozens of copies of the game from a warehouse and started selling them online," prompting lots of pre-release leaks for the game.

"One Reddit user immediately reported the leaks to Bethesda and Memphis police," adds Kotaku. "And he's now been banned from the r/GamingLeaksAndRumours subreddit after posting about it." I know this because the commenter in question, Jasper Adkins, emailed Kotaku to inform us it had happened. "It seems to me that the subreddit is running on 'bread and circuses' mode mixed with bystander syndrome," he wrote in his initial email. "They're perfectly willing to ignore a crime that hurts a developer they claim to support, in exchange for a few minutes of shaky gameplay filmed from a phone...."

Despite the criminal charges against him, Harris has become something of a folk hero within the community of fans hungry for Starfield leaks. As the Commercial Appeal reported, memes hail him as "Lord Tyrone" (his middle name) and one player even vowed to name their Starfield ship "Memphian" in his honor...

[Adkins] was banned from r/GamingLeaksAndRumours on August 24 shortly after posting about how he tried to help get Harris arrested. "An officer at the station told me so himself when I called him about it," he wrote in the middle of a long comment thread. Adkins soon received a notification that he had violated the subreddit's rules. He protested, but the r/GamingLeaksAndRumours admins weren't having it. "Just not interested in having someone here who takes action against the community like that," they wrote back.

I reached out to one of the subreddit's admins to confirm what had happened and the thinking behind the ban. "If he just did it I wouldn't think badly of him but to come on the sub and brag about calling the cops on the dude just rubbed me the wrong way," one of them told Kotaku in a DM. "Might unban him at some point but for now he's behind the bars of the internet."

Government

IBM Returns To the Facial Recognition Market 17

During the Black Lives Matter protests in 2020, IBM announced that it would no longer offer "general purpose" facial recognition technology due to concerns about racial profiling, mass surveillance, and other human rights violations. Now, according to The Verge and Liberty Investigates, "IBM signed a $69.8 million contract with the British government to develop a national biometrics platform that will offer a facial recognition function to immigration and law enforcement officials." From the report: A contract notice for the Home Office Biometrics Matcher Platform outlines how the project initially involves developing a fingerprint matching capability, while later stages introduce facial recognition for immigration purposes -- described as "an enabler for strategic facial matching for law enforcement." The final stage of the project is described as delivery of a "facial matching for law enforcement use-case." The platform will allow photos of individuals to be matched against images stored on a database -- what is sometimes known as a "one-to-many" matching system. In September 2020, IBM described such "one-to-many" matching systems as "the type of facial recognition technology most likely to be used for mass surveillance, racial profiling, or other violations of human rights."

IBM spokesman Imtiaz Mufti denied that its work on the contract was in conflict with its 2020 commitments. "IBM no longer offers general-purpose facial recognition and, consistent with our 2020 commitment, does not support the use of facial recognition for mass surveillance, racial profiling, or other human rights violations," he said. "The Home Office Biometrics Matcher Platform and associated Services contract is not used in mass surveillance. It supports police and immigration services in identifying suspects against a database of fingerprint and photo data. It is not capable of video ingest, which would typically be needed to support face-in-a-crowd biometric usage."

Human rights campaigners, however, said IBM's work on the project is incompatible with its 2020 commitments. Kojo Kyerewaa of Black Lives Matter UK said: "IBM has shown itself willing to step over the body and memory of George Floyd to chase a Home Office contract. This won't be forgotten." Matt Mahmoudi, PhD, tech researcher at Amnesty International, said: "The research across the globe is clear; there is no application of one-to-many facial recognition that is compatible with human rights law, and companies -- including IBM -- must therefore cease its sale, and honor their earlier statements to sunset these tools, even and especially in the context of law and immigration enforcement where the rights implications are compounding."
AI

Call of Duty Will Use AI To Moderate Voice Chats 48

Activision has partnered with a company called Modulate to moderate voice chats using an AI technology called ToxMod. According to The Verge, the tool "will work to identify behaviors like hate speech, discrimination, and harassment in real time." From the report: ToxMod's initial beta rollout in North America begins today. It's active within Call of Duty: Modern Warfare II and Call of Duty: Warzone. A "full worldwide release" (it does not include Asia, the press release notes) will follow on November 10th with the release of Call of Duty: Modern Warfare III, this year's new entry in the franchise. Modulate's press release doesn't include too many details about how exactly ToxMod works. Its website notes that the tool "triages voice chat to flag bad behavior, analyzes the nuances of each conversation to determine toxicity, and enables moderators to quickly respond to each incident by supplying relevant and accurate context."

The company's CEO said in a recent interview that the tool aims to go beyond mere transcription; it takes factors like a player's emotions and volume into context as well in order to differentiate harmful statements from playful ones. It is noteworthy that the tool (for now, at least) will not actually take action against players based on its data but will merely submit reports to Activision's moderators.
AI

OpenAI Launches a ChatGPT Plan For Enterprise Customers 16

An anonymous reader quotes a report from TechCrunch: Seeking to capitalize on ChatGPT's viral success, OpenAI today announced the launch of ChatGPT Enterprise, a business-focused edition of the company's AI-powered chatbot app. ChatGPT Enterprise, which OpenAI first teased in a blog post earlier this year, can perform the same tasks as ChatGPT, such as writing emails, drafting essays and debugging computer code. But the new offering also adds "enterprise-grade" privacy and data analysis capabilities on top of the vanilla ChatGPT, as well as enhanced performance and customization options. That puts ChatGPT Enterprise on par, feature-wise, with Bing Chat Enterprise, Microsoft's recently launched take on an enterprise-oriented chatbot service.

ChatGPT Enterprise provides a new admin console with tools to manage how employees within an organization use ChatGPT, including integrations for single sign-on, domain verification and a dashboard with usage statistics. Shareable conversation templates allow employees to build internal workflows leveraging ChatGPT, while credits to OpenAI's API platform let companies create fully custom ChatGPT-powered solutions if they choose. ChatGPT Enterprise, in addition, comes with unlimited access to Advanced Data Analysis, the ChatGPT feature formerly known as Code Interpreter, which allows ChatGPT to analyze data, create charts, solve math problems and more, including from uploaded files. For example, given a prompt like "Tell me what's interesting about this data," ChatGPT's Advanced Data Analysis capability can look through the data -- financial, health or location information, for example -- to generate insights.

Advanced Data Analysis was previously available only to subscribers to ChatGPT Plus, the $20-per-month premium tier of the consumer ChatGPT web and mobile apps. To be clear, ChatGPT Plus is sticking around -- OpenAI sees ChatGPT Enterprise as complementary to it, the company says. ChatGPT Enterprise is powered by GPT-4, OpenAI's flagship AI model, as is ChatGPT Plus. But ChatGPT Enterprise customers get priority access to GPT-4, delivering performance that's twice as fast as the standard GPT-4 and with an expanded 32,000-token (~25,000-word) context window. Context window refers to the text the model considers before generating additional text, while tokens represent raw text (e.g. the word "fantastic" would be split into the tokens "fan," "tas" and "tic"). Generally speaking, models with large context windows are less likely to "forget" the content of recent conversations.
Crucially, OpenAI said that it "won't train models on business data sent to ChatGPT Enterprise or any usage data and that all conversations with ChatGPT Enterprise are encrypted in transit and at rest," notes TechCrunch.

"OpenAI says that its future plans for ChatGPT Enterprise include a ChatGPT Business offering for smaller teams, allowing companies to connect apps to ChatGPT Enterprise, 'more powerful' and 'enterprise-grade' versions of Advanced Data Analysis and web browsing, and tools designed for data analysts, marketers and customer support."

A blog post introducing ChatGPT Enterprise can be found here.
Youtube

YouTube TV Urged To Drop '$600 Less Than Cable' Ad Claim (lightreading.com) 22

An advertising watchdog has recommended that YouTube TV, Google's growing pay-TV streaming service, drops an ad claim that the service is "$600 less than cable." The recommendation from the National Advertising Division (NAD) stems from a complaint lodged by Charter Communications. From a report: NAD, which used an expedited process for single-issue advertising cases in making this decision, found that YouTube TV's pricing claim, which identifies "comparable standalone cable" as the basis of comparison, doesn't hold up. NAD noted that the price calculation underlying the challenged claim includes the cost of two set-top boxes per household for "standalone cable" services," but argued that such a comparison isn't a good fit because operators such as Charter offer pay-TV streaming options that may not require a set-top box. In Charter's case, its Spectrum TV app, billed as a platform that can "stream outside the cable box," is compatible with iOS and Android mobile devices along with several retail streaming devices and/or integrated connected TVs from companies such as Apple, Roku, Google and Samsung. "In the context of the 'cable' comparison, NAD found the claim reasonably conveys the cost of YouTube TV is compared to all cable services," the organization explained.
Programming

Creators of Python, Java, TypeScript, and SmallTalk Will Make a Joint Appearance for Charity (pydata.org) 45

The creators of four programming languages will appear together onstage for a historic conversation on September 19th.

- Adele Goldberg — Smalltalk
- Guido Van Rossum — Python
- Anders Hejlsberg — Turbo Pascal, C#, TypeScript
- James Gosling — Java

The announcement describes it as "a conversation about programming language design." The charity event brings together this unique group of computer science pioneers, unlike any event held before. These great minds come together for what will surely be a fantastic night of discussion as the panel delves into the past and future of programming language creation.
It's a fundraiser for two groups. NumFOCUS is a nonprofit charity sponsoring nearly all the major tools in the Python data science stack (including jupyter, numpy, pandas, and matplotlib), and it's also the group behind PyData conferences on open source data tools. And the Last Mile Education Fund offers financial support for low-income underrepresented students. It's being billed as the "inaugural charity event" of PyData Seattle.

This happened once before in 2019, when Puget Sound Programming Python arranged a four-way discussion with Python creator Guido van Rossum, Java creator James Gosling, Perl creator Larry Wall, and Anders Hejlsberg (Turbo Pascal, C#, TypeScript). They held a 90-minute discussion about "language design, the universe, and everything" as a benefit for CSforALL (a group promoting computer science classes at every grade level). During that discussion Gosling shared how Java "started out as kind of 'Do a better C', and it got out of control. The rest of the project really ended up just providing the context." And Anders Hejlsberg told the audience that TypeScript was inspired by massive "write-only" JavaScript code bases.

In their discussion on variable typing and its use in IDEs, Gosling mocked what he called the "real men use vi" mentality, leading to a lively back and forth. Perl's Larry Wall later acknowledged the importance of types and the careful consideration that went into implementing them for Perl 6, but also shared his unique perspective as a long-time designer of programming languages. "I think IDEs make language developers lazy."

At the end of the event, they all agreed that the most rewarding part of language design was the people — the excitement, the gratitude, and to see that community helping others in its community.
AI

Does 'Coning' Self-Driving Cars Protest Tech Industry Impacts? (npr.org) 145

In July "Safe Street Rebels" launched the "Week of Cone" pranks (which went viral on TikTok and Twitter). TechCrunch called it "a bid to raise awareness and invite more pissed-off San Franciscans to submit public comments" to regulatory agencies.

But NPR sees a larger context: Coning driverless cars fits in line with a long history of protests against the impact of the tech industry on San Francisco. Throughout the years, activists have blockaded Google's private commuter buses from picking up employees in the city. And when scooter companies flooded the sidewalks with electric scooters, people threw them into San Francisco Bay. "Then there was the burning of Lime scooters in front of a Google bus," says Manissa Maharawal, an assistant professor at American University who has studied these protests.

She points out that when tech companies test their products in the city, residents don't have much say in those decisions: "There's been various iterations of this where it's like, 'Oh, yep, let's try that out in San Francisco again,' with very little input from anyone who lives here...." Waymo is already giving rides in Phoenix and is testing with human safety drivers in Los Angeles and Austin. And Cruise is offering rides in Phoenix and Austin and testing in Dallas, Houston, Miami, Nashville and Charlotte.

Meanwhile, in San Francisco, members of Safe Street Rebel continue to go out at night and stalk the vehicles one cone at a time.

They're apparently bicycling activists, judging by their web site, advocating "for car-free spaces, transit equity, and the end of car dominance." ("We regularly protest the city's thoughtless reopening of the Upper Great Highway to cars by slowing traffic to show just how unnecessary of a route this road is.") Their long-term goal is to expand the group "to the point where we can make a city for people to safely walk, bike and take public transit, not a city dominated by cars..." The last half-century has been a failed experiment with car dominance. They bankrupt our cities, ruin our environment, and force working people to sacrifice an unacceptable amount of their income to pay for basic transpiration. It is time to end car dependence and rethink our streets around public transit, walking and bikes.
Their demands include unredacted data from self-driving car companies about safety incidents (and a better reporting system) — plus a mechanism for actually citing robotaxis for traffic violations. But they also raise concerns about surveillance, noting the possibility of "a city-wide, moving network observing and analyzing everything."

Their web page says they also want to see studies on the pollution impact of self-driving cars — and whether or not AVs will increase car usage. They support the concerns of San Francisco's Taxi Workers Alliance about the possibility of lost jobs and increased traffic congestion.

And they raise one more concern: Their cars are not wheelchair accessible and do not pull up to the curb. Profit-driven robotaxi companies see accessibility as an afterthought. Without enforcement, their promises for the future will likely never materialize. Paratransit and transit are accountable to the public, but Cruise and Waymo are only accountable to shareholders.
But their list of concerns is followed by an exhaustive list of 266 robotaxi incidents documented with links to news articles and social media reports. ("The cars have run red lights, rear-ended a bus and blocked crosswalks and bike paths," writes NPR. "In one incident, dozens of confused cars congregated in a residential cul-de-sac, clogging the street. In another, a Waymo ran over and killed a dog.")

NPR's article adds one final note. "Neither Cruise nor Waymo responded to questions about why the cars can be disabled by traffic cones."

Thanks to Slashdot reader Tony Isaac for sharing the news.
AI

Microsoft AI Suggests Food Bank As a 'Cannot Miss' Tourist Spot In Canada 50

An anonymous reader quotes a report from Ars Technica: Late last week, MSN.com's Microsoft Travel section posted an AI-generated article about the "cannot miss" attractions of Ottawa that includes the Ottawa Food Bank, a real charitable organization that feeds struggling families. In its recommendation text, Microsoft's AI model wrote, "Consider going into it on an empty stomach." Titled, "Headed to Ottawa? Here's what you shouldn't miss!," (archive here) the article extols the virtues of the Canadian city and recommends attending the Winterlude festival (which only takes place in February), visiting an Ottawa Senators game, and skating in "The World's Largest Naturallyfrozen Ice Rink" (sic).

As the No. 3 destination on the list, Microsoft Travel suggests visiting the Ottawa Food Bank, likely drawn from a summary found online but capped with an unfortunate turn of phrase: "The organization has been collecting, purchasing, producing, and delivering food to needy people and families in the Ottawa area since 1984. We observe how hunger impacts men, women, and children on a daily basis, and how it may be a barrier to achievement. People who come to us have jobs and families to support, as well as expenses to pay. Life is already difficult enough. Consider going into it on an empty stomach."

That last line is an example of the kind of empty platitude (or embarrassing mistaken summary) one can easily find in AI-generated writing, inserted thoughtlessly because the AI model behind the article cannot understand the context of what it is doing. The article is credited to "Microsoft Travel," and it is likely the product of a large language model (LLM), a type of AI model trained on a vast scrape of text found on the Internet.
Microsoft

Microsoft May Store Your Conversations With Bing If You're Not an Enterprise User (theregister.com) 13

An anonymous reader quotes a report from The Register: Microsoft prohibits users from reverse engineering or harvesting data from its AI software to train or improve other models, and will store inputs passed into its products as well as any output generated. The details emerged as companies face fresh challenges with the rise of generative AI. People want to know what corporations are doing with information provided by users. And users are likewise curious about what they can do with the content generated by AI. Microsoft addresses these issues in a new clause titled 'AI Services' in its terms of service.

The five new policies, which were introduced on 30 July and will come into effect on September 30, state that:

Reverse Engineering. You may not use the AI services to discover any underlying components of the models, algorithms, and systems. For example, you may not try to determine and remove the weights of models.
Extracting Data. Unless explicitly permitted, you may not use web scraping, web harvesting, or web data extraction methods to extract data from the AI services.
Limits on use of data from the AI Services. You may not use the AI services, or data from the AI services, to create, train, or improve (directly or indirectly) any other AI service.
Use of Your Content. As part of providing the AI services, Microsoft will process and store your inputs to the service as well as output from the service, for purposes of monitoring for and preventing abusive or harmful uses or outputs of the service.
Third party claims. You are solely responsible for responding to any third-party claims regarding Your use of the AI services in compliance with applicable laws (including, but not limited to, copyright infringement or other claims relating to content output during Your use of the AI services).
A spokesperson from Microsoft declined to comment on how long the company plans to store user inputs into its software. "We regularly update our terms of service to better reflect our products and services. Our most recent update to the Microsoft Services Agreement includes the addition of language to reflect artificial intelligence in our services and its appropriate use by customers," the representative told us in a statement.

Microsoft has previously said, however, that it doesn't save conversations or use that data to train its AI models for its Bing Enterprise Chat mode. The policies are a little murkier for its Microsoft 365 Copilot, although it doesn't appear to use customer data or prompts for training, it does store information. "[Copilot] can generate responses anchored in the customer's business content, such as user documents, emails, calendar, chats, meetings, contacts, and other business data. Copilot combines this content with the user's working context, such as the meeting a user is in now, the email exchanges the user has had on a topic, or the chat conversations the user had last week. Copilot uses this combination of content and context to help deliver accurate, relevant, contextual responses," it said.
Privacy

Iraq Lifts Ban On Telegram After Messaging App Complies With Authorities 11

An anonymous reader quotes a report from TechCrunch: Iraq's telecom ministry lifted the ban on Telegram over the weekend, days after the agency blocked the chat app over security concerns. The ministry said it lifted the ban because of the "response of the company that owns the application to the requirements of the security authorities," which required Telegram to reveal sources leaking data of officials and citizens, according to a translated statement. Telegram has shown commitment to communicating with authorities about security concerns, the ministry said, insisting that it "doesn't stand against freedom of expression."

Telegram told Reuters that the company forbids users from posting private data on the platform without consent. Telegram didn't share any private user data with Iraqi authorities, the messaging app operator told the publication. "We can confirm that our moderators took down several channels sharing personal data. However, we can also confirm that no private user data was requested from Telegram and that none has been shared," the company told the publication in a statement.
Context: "Last week, Iraq banned the chat app saying that many channels were publishing citizen's private data such as names, addresses, and family ties with other people," reports TechCrunch. "At that time, the ministry said that Telegram -- which has more than 800 million users globally -- didn't respond to its requests, and as a result, the country banned the app."
Cloud

In Generative AI Market, Amazon Chases Microsoft and Google with Custom AWS Chips (cnbc.com) 25

An anonymous reader shared this report from CNBC: In an unmarked office building in Austin, Texas, two small rooms contain a handful of Amazon employees designing two types of microchips for training and accelerating generative AI. These custom chips, Inferentia and Trainium, offer AWS customers an alternative to training their large language models on Nvidia GPUs, which have been getting difficult and expensive to procure. "The entire world would like more chips for doing generative AI, whether that's GPUs or whether that's Amazon's own chips that we're designing," Amazon Web Services CEO Adam Selipsky told CNBC in an finterview in June. "I think that we're in a better position than anybody else on Earth to supply the capacity that our customers collectively are going to want...."

In the long run, said Chirag Dekate, VP analyst at Gartner, Amazon's custom silicon could give it an edge in generative AI...

With millions of customers, Amazon's AWS cloud service "still accounted for 70% of Amazon's overall $7.7 billion operating profit in the second quarter," CNBC notes. But does that give them a competitive advantage?

A technology VP for the service tells them "It's a question of velocity. How quickly can these companies move to develop these generative AI applications is driven by starting first on the data they have in AWS and using compute and machine learning tools that we provide." In June, AWS announced a $100 million generative AI innovation "center."

"We have so many customers who are saying, 'I want to do generative AI,' but they don't necessarily know what that means for them in the context of their own businesses. And so we're going to bring in solutions architects and engineers and strategists and data scientists to work with them one on one," AWS CEO Selipsky said... For now, Amazon is only accelerating its push into generative AI, telling CNBC that "over 100,000" customers are using machine learning on AWS today. Although that's a small percentage of AWS's millions of customers, analysts say that could change.

"What we are not seeing is enterprises saying, 'Oh, wait a minute, Microsoft is so ahead in generative AI, let's just go out and let's switch our infrastructure strategies, migrate everything to Microsoft.' Dekate said. "If you're already an Amazon customer, chances are you're likely going to explore Amazon ecosystems quite extensively."

AI

Stack Overflow 'Evolves', Previewing AI-Powered Answers and Chat Followups (stackoverflow.blog) 64

"Stack Overflow is adding artificial intelligence to its offerings," reports ZDNet (which notes traffic to the Q&A site has dropped 5% in the last year).

So in a video, Stack Overflow's CEO Prashanth Chandrasekar says that search and question-asking "will evolve to provide you with instant summarized solutions with citations to sources, aggregated by generative AI — plus the option to ask follow-up questions in a chat-like format."

The New Stack provides some context: As computer scientist Santiago Valdarrama remarked in a tweet, "I don't remember the last time I visited Stack Overflow. Why would I when tools like Copilot and ChatGPT answer my questions faster without making me feel bad for asking?" It's a problem Stack Overflow CEO Prashanth Chandrasekar acknowledges because, well, he encountered it too.

"When I first started using Stack Overflow, I remember my first experience was quite harsh, because I basically asked a fairly simple question, but the standard on the website is pretty high," Chandrasekar told The New Stack. "When ChatGPT came out, it was a lot easier for people to go and ask ChatGPT without anybody watching...."

But what may be of more interest to developers is that Stack Overflow is now offering an IDE (integrated development environment) extension for Visual Studio Code that will be powered by OverflowAI. This means that coders will be able to ask a conversational interface a question and find solutions from within the IDE.

Stack Overflow also is launching a GenAI Stack Exchange, where the community can post and share knowledge on prompt engineering, getting the most out of AI and similar topics.

And they're integrating it into other workflows as well. "Of course, AI isn't replacing humans any time soon," CEO Chandrasekar says in the video. "But it can help you draft a question to pose to our community..."

Signups for the OverflowAI preview are available now. "With your help, we'll be putting AI to work," CEO Chandrasekar says in the video.
AI

Anthropic Launches Improved Version of Its Entry-Level LLM (techcrunch.com) 5

Anthropic, the AI startup co-founded by ex-OpenAI execs, has released an updated version of its faster, cheaper, text-generating model available through an API, Claude Instant. TechCrunch reports: The updated Claude Instant, Claude Instant 1.2, incorporates the strengths of Anthropic's recently announced flagship model, Claude 2, showing "significant" gains in areas such as math, coding, reasoning and safety, according to Anthropic. In internal testing, Claude Instant 1.2 scored 58.7% on a coding benchmark compared to Claude Instant 1.1, which scored 52.8%, and 86.7% on a set of math questions versus 80.9% for Claude Instant 1.1. "Claude Instant generates longer, more structured responses and follows formatting instructions better," Anthropic writes in a blog post. "Instant 1.2 also shows improvements in quote extraction, multilingual capabilities and question answering."

Claude Instant 1.2 is also less likely to hallucinate and more resistant to jailbreaking attempts, Anthropic claims. In the context of large language models like Claude, "hallucination" is where a model generates text that's incorrect or nonsensical, while jailbreaking is a technique that uses cleverly-written prompts to bypass the safety features placed on large language models by their creators. And Claude Instant 1.2 features a context window that's the same size of Claude 2's -- 100,000 tokens. Context window refers to the text the model considers before generating additional text, while tokens represent raw text (e.g. the word "fantastic" would be split into the tokens "fan," "tas" and "tic"). Claude Instant 1.2 and Claude 2 can analyze roughly 75,000 words, about the length of "The Great Gatsby." Generally speaking, models with large context windows are less likely to "forget" the content of recent conversations.

Google

Google's AI Search is Getting More Video and Better Links (theverge.com) 13

Google's AI-powered Search Generative Experience is getting a big new feature: images and video. From a report: If you've enabled the AI-based SGE feature in Search Labs, you'll now start to see more multimedia in the colorful summary box at the top of your search results. Google's also working on making that summary box appear faster and adding more context to the links it puts in the box. SGE may still be in the "experiment" phase, but it's very clearly the future of Google Search.

"It really gives us a chance to, now, not always be constrained in the way search was working before," CEO Sundar Pichai said on Alphabet's most recent earnings call. "It allows us to think outside the box." He then said that "over time, this will just be how search works." The SGE takeover raises huge, thorny questions about the very future of the web, but it's also just a tricky product to get right. Google is no longer simply trying to find good links for you every time you search -- it's trying to synthesize and generate relevant, true, helpful information. Video in particular could go a long way here: Google has integrated YouTube more and more into search results over the years, linking to a specific chapter or moment inside a video that might help you with that "why is my dryer making that noise" query.

Businesses

Amazon Achieved Its 'Fastest Prime Speeds Ever' Last Quarter (techcrunch.com) 45

An anonymous reader quotes a report from TechCrunch: Amazon achieved its "fastest Prime speeds ever" last quarter, the online retail giant announced on Monday. The company says it has delivered more than 1.8 billion units to U.S. Prime members so far this year, nearly four times what it delivered at those speeds by this point in 2019. Across the top 60 largest U.S. metro areas, more than half of Prime member orders arrived the same or next day. Same-day delivery is currently available on millions of items for customers across more than 90 U.S. metro areas, and Amazon plans to double the number of delivery sites in the coming years.

Amazon notes that the average time from picking a customer's items to positioning the customer's package on the outbound dock is 11 minutes in same-day facilities, more than an hour faster than its traditional fulfillment centers. For context, same-day facilities are stocked with what customers in those areas are purchasing, while traditional facilities are larger and include a more vast assortment of items. "As we make these changes, we are seeing that they have the added benefit of reducing costs, too," Doug Herrington, CEO of Worldwide Amazon Stores, wrote in a blog post. "Regionalizing our network reduces miles traveled and handoffs. Since the beginning of this year alone, the distance between our sites and the customer decreased by 15%, with 12% fewer touchpoints within our middle mile network. Improved product placement gets items even closer to customers, making our delivery system more efficient. And our Same-Day Delivery network is not only our fastest way to get products to customers, it is also one of our lowest cost ways."

Social Networks

Reddit Users Heckle Search for New Mods, as Some Mods Move to Lemmy and Discord (arstechnica.com) 73

"Over the past week, a Reddit employee has posted to subreddits with ousted mods, asking for new volunteers," reports Ars Technica.

But it's not always going smoothly... A Reddit employee going by ModCodeofConduct (Reddit has refused to disclose the real names of admins representing the company on the platform) has posted to numerous subreddits over recent days, including r/IRLEasterEggs, r/donthelpjustfilm, r/ActLikeYouBelong, r/malefashionadvice, and r/AccidentalRenaissance... Like most official Reddit posts since the API war began, the comments under the job ads display users' discontent.

"May I nominate a mod? I think u/ConspirOC would be a great mod, as he created this subreddit and has successfully run it for years, before you forcibly removed him," a user going by LittleManOnACan wrote on ModCodeofConduct's post seeking replacement r/IRLEasterEggs mods. "Additionally, fire Steve Huffman (Fuck u/Spez)."

There's also a desire among Reddit users for a return to not just how things were but an acknowledgment of the efforts made by many previous moderators, how things changed, and why things are different now. A Redditor going by QuicklyThisWay wrote on ModCodeofConduct's post for news mods for r/IRLEasterEggs:

"Just to be clear for anyone 'applying' to be a moderator. The user that created the subreddit and any other mods were removed by admins for making the community private. Even though the option to change to private is available to all subreddits at any time, the admins have not and will not respect any 'autonomy' moderators appear to have...

As Ars has previously detailed, user protests didn't prevent third-party Reddit apps from closing. However, they have disrupted the platform.

Reddit didn't answer questions Ars sent about its replacement mod criteria or how it'll help ensure new mods can properly handle their newfound volunteer duties...

"mods Ars has spoken with over the weeks have frequently pointed to the potential for burnout, death threats, long training sessions (from other volunteer mods), and rapid turnover for Reddit mods..." the article notes, adding "Without mods proven to be dedicated and experienced, it's unclear how fervently such efforts will continue in the future...

"Disgruntled mods and ex-mods continue seeking new platforms to continue community discussions, including Lemmy and Discord. And as of this writing, there are still 1,900 subreddits private, per the Reddark_247 tracker."

Meanwhile, the third annual edition of Reddit's annual pixel-placing event r/Place "turned into a battleground for dunking on the CEO," reports Polygon. A Reddit spokesperson declined to comment about this year's edition of r/Place, telling Polygon via email "redditors are going to reddit."
Gizmodo's article includes a timelapse video (from YouTube) that they say captures "the whimsy — and anger — of its users," including "plenty of protest art directed at CEO Steve Huffman, who goes by u/spez on the platform..." While there are plenty of examples of "Fuck Spez" to go around, the most creative moment occurred at the end of the project. As r/Place wound to a close, users were able to place a pixel once every thirty seconds, but the pixel had to be white — an effort to wipe the slate clean. However, in the final moments of the project, users collaborated to leave one massive "FUCK SPEZ" across the canvas.
Red Hat Software

AlmaLinux Discovers Working with Red Hat (and CentOS Stream) Isn't Easy (zdnet.com) 73

After Red Hat's decision to only share RHEL source code with subscribers, AlmaLinux asked their bug report submitters to "attempt to test and replicate the problem in CentOS Stream as well, so we can focus our energy on correcting it in the right place."

Red Hat told Ars Technica they are "eager to collaborate" on their CentOS Stream distro, "even if we ultimately compete in a business sense. Differentiated competition is a sign of a healthy ecosystem."

But Red Hat still managed to ruffled some feathers, reports ZDNet: AlmaLinux Infrastructure Team Leader Jonathan Wright recently posted a CentOS Stream fix for CVE-2023-38403, a memory overflow problem in iperf3. Iperf3 is a popular open-source network performance test. This security hole is an important one, but not a huge problem.

Still, it's better by far to fix it than let it linger and see it eventually used to crash a server. That's what I and others felt anyway. But, then, a senior Red Hat software engineer replied, "Thanks for the contribution. At this time, we don't plan to address this in RHEL, but we will keep it open for evaluation based on customer feedback."

That went over like a lead balloon.

The GitLab conversation proceeded:

AlmaLinux: "Is customer demand really necessary to fix CVEs?"

Red Hat: "We commit to addressing Red Hat defined Critical and Important security issues. Security vulnerabilities with Low or Moderate severity will be addressed on demand when [a] customer or other business requirements exist to do so."

AlmaLinux: "I can even understand that, but why reject the fix when the work is already done and just has to be merged?"

At this point, Mike McGrath, Red Hat's VP of Core Platforms, AKA RHEL, stepped in. He explained, "We should probably create a 'what to expect when you're submitting' doc. Getting the code written is only the first step in what Red Hat does with it. We'd have to make sure there aren't regressions, QA, etc. ... So thank you for the contribution, it looks like the Fedora side of it is going well, so it'll end up in RHEL at some point."

Things went downhill rapidly from there...

On Reddit, McGrath said, "I will admit that we did have a great opportunity for a good-faith gesture towards Alma here and fumbled."

Finally, though the Red Hat Product Security team rated the CVE as "'Important,' the patch was merged.

Coincidentally, last month AlmaLinux announced that its move away from 1:1 compatibility with RHEL meant "we can now accept bug fixes outside of Red Hat's release cycle."

This Thursday AlmaLinux also reiterated that they're "fully committed to delivering the best possible experience for the community, no matter where or what you run." And in an apparent move to beef up compatibility testing, they announced they'd be bringing openQA to the RHEL ecosystem. (They describe openQA as a tool using virtual machines that "simplifies automated testing of the whole installation process of an operating system in a wide combination of software and hardware configurations.")
Games

Ubisoft Will Suspend and Then Delete Long-Inactive Accounts (pcgamer.com) 51

Leaving a Ubisoft account inactive for too long "apparently puts it at risk of permanent deletion," writes PC Gamer, calling the policy "a customer-unfriendly practice." A piracy and anti-DRM focused Twitter account, PC_enjoyer, recently shared a screenshot of a Ubisoft support email telling the user that their Ubisoft account had been suspended for "inactivity," and would be "permanently closed" after 30 days. The email provided a link to cancel the move. Now, that sounds like a phishing scam, right? I and many commenters wondered that, looking at the original post, but less than a day later, Ubisoft's verified support account responded to the tweet, seemingly confirming the screenshotted email's legitimacy.

"You can avoid the account closure by logging into your account within the 30 days (since receiving the email pictured) and selecting the Cancel Account Closure link contained in the email," Ubisoft Support wrote. "We certainly do not want you to lose access to your games or account so if you have any difficulties logging in then please create a support case with us."

I was unable to find anything regarding account closure for inactivity in Ubisoft's US terms of use or its end user licence agreement, but the company does reserve the right to suspend or end services at any time. Ubisoft has a support page titled "Closure of inactive Ubisoft accounts." The page first describes instances where the service clashes with local data privacy laws, then reads: "We may also close long-term inactive accounts to maintain our database. You will be notified by email if we begin the process of closing your inactive account."

This page links to another dedicated to voluntarily closing one's Ubisoft account, and seems to operate by the same rules: a 30-day suspension before permanent deletion. "As we will be unable to recover the account once it has been closed, we strongly recommend only putting in the request if you are absolutely sure you would like to close your account."

"If you have a good spam filter or just reasonably assume it's a phishing attempt, then you might one day try your old games and find they're just gone," worries long-time Slashdot reader Baron_Yam. "If you're someone who still plays games from decades ago every so often, this is a scenario you might want to think about."

The site Eurogamer reports that when a Twitter user complained that "I lost my Ubisoft account, and all the Ubisoft Steam game[s] I've bought are now useless", Ubisoft Support "responded to say that players can raise a ticket if they would like to recover their account."

The original tweet now includes this "reader-added context" supplied by other Twitter users — along with three informative links: For added context, Ubisoft can be required under certain data protection laws, such as the GDPR, to close inactive accounts if they deem the data no longer necessary for collection.

Ubisoft has claimed they don't close accounts that are inactive for less than 4 years.

Slashdot Top Deals