Connect with us

Tech

All the AI news of the week: Hands-on with Metas AI app, ChatGPT and and leaderboard drama

Published

on

Just like AI models, AI news never sleeps.

Every week, we're inundated with new models, products, industry rumors, legal and ethical crises, and viral trends. If that's not enough, the rival AI hype/doom chatter online makes it hard to keep track of what's really important. But we've sifted through it all to recap the most notable AI news of the week from the heavyweights like OpenAI and Google, as well as the AI ecosystem at large. Read our last recap, and check back next week for a new edition.

Another week, another batch of AI news coming your way.

This week, Meta held its inaugural LlamaCon event for AI developers, OpenAI struggled with model behavior, and LM Arena was accused of helping AI companies game the system. Congress also passed new laws protecting victims of deepfakes, and new research examines AI's current and potential harms. Plus, Duolingo and Wikipedia have very different approaches to their new AI strategies.

What happened at Meta's first LlamaCon

mark zuckerberg in black t-shirt with gold chain


Credit: Chris Unger / Zuffa LLC / Getty Images

At LlamaCon, Meta's first conference for AI developers, the two big announcements were the launch of a standalone Meta AI app to compete more directly with ChatGPT and the Llama API, now in limited preview. Following reports that this was in the works, CEO Sam Altman once joked that maybe OpenAI should do its own social media app, but now that is reportedly happening for real.

We also went hands-on with the new Llama-powered Meta AI app. For more details about Meta AI's top features, read Mashable's breakdown.

During LlamaCon's closing keynote, Mark Zuckerberg interviewed Microsoft CEO Satya Nadella about a bunch of trends, ranging from agentic AI capabilities to how we should measure AI's advancements. Nadella also revealed that up to 30 percent of Microsoft's code is written by AI. Not to be outdone, Zuckerberg said he wants AI to write half of Meta's code by next year.

ChatGPT has safety issues, goes shopping

Meta AI and ChatGPT both got busted this week for sexting minors.

OpenAI said this was a bug and they're working to fix it. Another ChatGPT issue this week made the latest GPT-4o update too much of a suck-up. Altman described the model's behavior as "sycophant-y and annoying," but users were concerned about the dangers of releasing a model like this, highlighting problems with iterative deployment and reinforcement learning.

OpenAI was even accused of intentionally tuning the model to keep users more engaged. Joanne Jang, OpenAI's head of model behavior, jumped on a Reddit AMA to do damage control. "Personally, the most painful part of the latest sycophancy discussions has been people assuming that my colleagues are irresponsibly trying to maximize engagement for the sake of it," wrote Jang.

Earlier in the week, OpenAI announced new features to make products mentioned in ChatGPT responses more shoppable. The company said it isn't earning purchase commissions, but it smells an awful lot like the beginnings of a Google Shopping competitor. Did we mention OpenAI would buy Chrome if Google is forced to divest it? Because they totally would, FYI.

The ChatGPT maker has had a few more problems with its recent models. Last week, we reported that o3 and o4-mini hallucinate more than previous models, by OpenAI's own admission.

Anyone in the U.S. can now sign up for Google AI Mode

Meanwhile, Google is barreling ahead with AI-powered search features. On Thursday, the tech giant announced that it's removing the waitlist to test out AI Mode in Labs, so anyone over 18 in the U.S. can try it out. We spoke with Robby Stein, VP of product for Google Search, about how users have responded to its AI features, the future of search, and Google's responsibility to publishers.

Your browser does not support the mp4 video codec.

Via Giphy

Google also updated Gemini with image editing tools and expanded NotebookLM, its AI podcast generator, to over 50 languages. Bloomberg also reported that Google has been quietly testing ads inside third-party chatbot responses.

We're keeping a close eye on that final development, and we are very curious how Google plans to inject ads into AI search. Would you trust a chatbot that gave you sponsored answers?

Leaderboard drama

Researchers from AI company Cohere, Princeton, Stanford, MIT, and Ai2, published a paper this week calling out Chatbot Arena for essentially helping AI heavyweights rig their benchmarking results. The study said the popular crowdsourced benchmarking tool from UC Berkeley allowed Meta, Google, OpenAI, and Amazon "extensive private testing" and gave them more prompt data, which "significantly" improved their rankings.

In response, LM Arena, the group behind Chatbot Arena said "there are a number of factual errors and misleading statements in this writeup" and posted a pointy-by-point rebuttal to the paper's claims on X.

The issue of benchmarking AI models has become increasingly problematic. Benchmark results are largely self-reported by the companies that release them, and the AI community has called for more transparency and accountability by objective third parties. Chatbot Arena seemed to provide a solution by allowing users to choose the best responses in blind tests. But now LM Arena's practices have come into question, further fueling the conversation around objective evaluations.

A few weeks ago, Meta got in trouble for using an unreleased version of its Llama 4 Maverick model on LM Arena, which scored a high ranking. LM Arena updated its leaderboard policies, and the publicly available version of Llama 4 Maverick was added instead, ranking way lower than the unreleased version.

Lastly, LM Arena recently announced plans to form a company of its own.

Regulators and researchers tackle AI's real-world harms

Now that generative AI has been in the wild for a few years, the real-world implications have started to crystallize.

This week, U.S. Congress passed the "Take It Down" Act, which requires tech companies to remove nonconsensual intimate imagery within 48 hours of a request. The law also outlines strict punishment for deepfake creators. The legislation had bipartisan support and is expected to be signed by President Donald Trump.

The nonpartisan U.S. Government Accountability Office (GAO) published a report on generative AI's impact on humans and the environment. The conclusion is that the potential impacts are huge, but exactly how much is unknown because "private developers do not disclose some key technical information."

And in the realm of the frighteningly real and specific harms of AI, a study from Common Sense Media said AI companion apps like Character.AI and Replika are unequivocally unsafe for teens. The researchers say if you're too young to buy cigarettes, you're too young for your own AI companion.

Then there was the report that researchers from the University of Zurich secretly deployed AI bots in the r/changemyview subreddit to try and convince people to change their minds. Some of the bot identities included a statutory rape victim, "a trauma counselor specializing in abuse," and "a black man opposed to Black Lives Matter."

Other AI news…

In other news, Duolingo is taking an "AI-first" approach, which means replacing its contract workers with AI whenever possible. On the flip side, Wikipedia announced it's taking a "human-first" approach to its AI strategy. It won't replace its volunteers and editors with AI, but will instead "use AI to build features that remove technical barriers to allow the humans at the core of Wikipedia."

Yelp deployed a bunch of AI features this week, including an AI-powered answering service that takes calls for restaurants, and Governor Gavin Newsom wants to use genAI to solve California's legendary traffic jams.

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Tech

Multiple porn sites sued by Florida attorney general

Published

on

By

Florida Attorney General James Uthmeier is suing several porn companies, according to a press release on Tuesday.

Uthmeier states that these porn sites aren't complying with Florida's age-verification law, which went into effect on Jan. 1. The law, HB 3, requires sites that publish a "substantial portion" of material that is "harmful to minors" to use a method to prove that visitors are over 18. HB 3 requires this method to keep personal information anonymous and be conducted by a nongovernmental, independent third party.

Florida's version of age verification is similar to that in other states, but some are more specific in that they require scanning a face or a government ID. These laws started to sprout up in states in 2022, beginning with Louisiana, and since then, free speech advocates and adult industry workers have told Mashable that the laws won't work for their intended purpose. A preliminary study out of NYU also suggests that age-verification laws don't work.

One reason is that they can be circumvented with software like VPNs, so visitors can pretend to be elsewhere. Another is that not every single website will comply.

Now, Uthmeier is suing companies that operate out of the Czech Republic, including the parent companies of XVideos and XNXX:

  • Webgroup Czech Republic (which operates XVideos)

  • NKL Associates (XNXX)

  • Sonesta Technologies, Inc. (BangBros)

  • Traffic F (an advertising network)

The AG is also suing GGW Group and GTFlix TV, distributors of GirlsGoneWild. The latter apparently also operates out of the Czech Republic.

The press release states that Uthmeier wrote two letters to two of the companies in April, demanding that they comply or face legal action.

"Multiple porn companies are flagrantly breaking Florida's age verification law by exposing children to harmful, explicit content. As a father of young children, and as Attorney General, this is completely unacceptable," Uthmeier stated in the press release. "We are taking legal action against these online pornographers who are willfully preying on the innocence of children for their financial gain."

When SCOTUS upheld Texas's age-verification law in June, experts told Mashable that it was a blow to free speech, as such laws quell adults' free speech, while also not actually stopping minors from accessing porn. Yet, these laws have also extended outside the U.S., as the UK has enacted age verification just last month. Already, internet users have found a way to bypass the law: using a photo of a video game character.

Continue Reading

Tech

Leaks may have revealed the iPhone 17 lineup release date

Published

on

By

According to leaked documents, Apple may be gearing up to unveil its iPhone 17 lineup — including the iPhone 17, 17 Air, and 17 Pro — on Sept. 9.

The rumor originates from iPhone-Ticker, a German blog, and was picked up by 9to5Mac, which reports that a local wireless carrier leaked internal documents pointing to an early September reveal.

While still unconfirmed, the date tracks with Apple’s usual playbook. The tech giant typically holds its iPhone launch events in the second week of September, excluding 2020, which was disrupted by COVID. The company also favors Tuesday announcements, though last year’s reveal was pushed due to the presidential debate.

If the leak holds true, we could be just weeks away from Apple’s next big drop.

This year, the spotlight is on the iPhone 17 Air, Apple’s rumored ultra-thin flagship measuring just 5.65mm thick. As Mashable’s Alex Perry put it, "that’s even thinner than a pencil."

Meanwhile, if you’ve been paying even casual attention to Apple leaks, most of the iPhone 17 and 17 Pro details are already out in the wild. One of the most eye-catching leaks is the new orange finish for the Pro models, which, to some (mostly me), is similar to the color scheme for the Charlotte Bobcats.

Aside from that, 9to5Mac notes that if the rumored Sept. 9 reveal date holds, Apple will likely stick to its usual rollout pattern—meaning pre-orders could open that Friday, Sept. 12, with the official launch landing a week later on Sept. 19.

Continue Reading

Tech

Delta and other airlines are working with an AI startup that personalizes prices

Published

on

By

Artificial intelligence may soon play a bigger role in your air travel fares.

Airlines are reportedly working with AI companies to deliver "personalized" prices to customers by using AI tools to analyze their personal information and data.

Delta Air Lines is currently using AI technology from the Israeli startup Fetcherr for some domestic flights, said President Glen Hauenstein in an earnings call last month. Hauenstein said the technology is still being tested, but told shareholders that Delta intends to expand its use of AI by the end of this year. As of now, the airline uses AI for only 3 percent of its domestic flight fares, but wants to increase this to 20 percent, according to ABC News.

However, in a recent letter to members of Congress, the company denied using AI tools to price-gouge customers, as Reuters reported last week.

Fetcherr is one of the prominent suppliers of AI-powered dynamic pricing, and it already works with several airlines, including Delta, Azul, Virgin Atlantic, WestJet, and Royal Air Maroc, according to Aviation Week. Delta has said it doesn't share personal customer data with Fetcherr.

But the airline has come under scrutiny for its rhetoric around using AI to optimize some fare prices. US lawmakers, including Democratic Arizona Senator Ruben Gallego, have accused Delta of "telling their investors one thing, and then turning around and telling the public another," said Gallego, who also said he believes Delta is engaging in "predatory pricing."

In a letter to Delta CEO Ed Bastian, Senators Gallego, Mark Warner, and Richard Blumenthal cited a comment made during an investor conference last December by Hauenstein, who said the company's AI price-setting technology sets fares by predicting "the amount people are willing to pay for the premium products related to the base fares."

"Consumers have no way of knowing what data and personal information your company and Fetcherr plan to collect or how the AI algorithm will be trained," reads the lawmakers' letter. The senators asked Delta to explain what data it collects and uses for its fares. Delta hasn't specified what data it relies on to set these individualized prices.

In response, the airline assured US Democratic senators that their ticket pricing "never takes into account personal data" but also spoke of the merits of using AI to set prices.

"Given the tens of millions of fares and hundreds of thousands of routes for sale at any given time, the use of new technology like AI promises to streamline the process by which we analyze existing data and the speed and scale at which we can respond to changing market dynamics," read Delta's letter to lawmakers.

While Delta insisted to US lawmakers that it’s not fixing prices with AI, recent revelations about Fetcherr raise serious questions about its technology.

Bloomberg reported this week on an alarming white paper by Fetcherr co-founder and chief AI officer Uri Yerushalmi. In the paper, Yerushalmi describes working with an unnamed airline to use artificial intelligence to create a pricing structure so complicated that it would “go beyond human cognitive limits,” according to Bloomberg.

So, even if AI isn’t used to “fix prices” in the traditional sense, it could still be used to make fare pricing so complex that consumers inadvertently end up paying more.

Rival airlines have also expressed concern. American Airlines CEO Robert Isom said using AI to set individualized fares could have an impact on consumer trust. He also said the strategy is not something AA would do.

Dynamic pricing has long been a part of the airline industry's strategy, but the use of AI has the potential to drastically change travel bookings. As airlines look to maximize revenue by harnessing AI, many policy experts fear consumers could face much higher prices, as expressed to The Lever. Another looming concern is that AI-powered pricing schemes can lead to price collusion between companies. Some, like Scott Keyes of Scott’s Cheap Flights, believe prices could actually be lowered, as he wrote in Time.

Last week, Democratic lawmakers Greg Casar and Rashida Tlaib introduced the Stop AI Price Gouging and Wage Fixing Act, a piece of legislation that would ban companies from using AI to fix prices or wages based on Americans' personal data. The lawmakers cited Delta's plans to increase their use of AI to set prices.

"Giant corporations should not be allowed to jack up your prices or lower your wages using data they got spying on you," said Congressman Casar in a statement. "Whether you know it or not, you may already be getting ripped off by corporations using your personal data to charge you more. This problem is only going to get worse, and Congress should act before this becomes a full blown crisis."

Continue Reading

Trending