OpenAI will reportedly pay $250 million to put News Corp’s journalism in ChatGPT

OpenAI and News Corp, the owner of The Wall Street Journal, MarketWatch, The Sun, and more than a dozen other publishing brands, have struck a multi-year deal to display news from these publications in ChatGPT, News Corp announced on Wednesday. OpenAI will be able to access both current and well as archived content from News Corp’s publications and use the data to further train its AI models. Neither company disclosed the terms of the deal, but a report in The Wall Street Journal estimated that News Corp would get $250 million over five years in cash and credits.

“The pact acknowledges that there is a premium for premium journalism,” News Corp Chief Executive Robert Thomson reportedly said in a memo to employees on Wednesday. “The digital age has been characterized by the dominance of distributors, often at the expense of creators, and many media companies have been swept away by a remorseless technological tide. The onus is now on us to make the most of this providential opportunity.”

Generative AI has exploded in popularity ever since OpenAI released ChatGPT at the end of 2022. But the quality of the responses provided by AI-powered chatbots is only as good as the data that is used to train the models that power it. So far, AI companies have trained their models by scraping publicly available data from the internet often without the consent of creators. But in recent times, they have been striking financial deals with the news industry to make sure that AI models can be trained on information that is current and authoritative. Over the last few months alone, OpenAI has announced partnerships with Reddit, the Financial Times, Dotdash Meredith, the Associated Press, German publisher Axel Springer, which owns Politico and Business Insider in the US and Bild and Die Welt in Germany, and Spain’s Prisa Media. Last month, News Corp also struck a deal reportedly between $5 and $6 million with Google to train its AI models, according to a report in The Information.

Google and OpenAI aren’t the only companies striking these deals to train their AI models. Hours before the News Corp announcement, Business Insider reported that Meta, which recently stuffed its own AI chatbot into Facebook, Messenger, WhatsApp, and Instagram, and also sells AI-powered sunglasses, was thinking about striking its own deals with news publishers to get access to training data.

Money from AI companies is increasingly a growing revenue source for a struggling news industry. But some publishers are still wary of striking these deals. The New York Times has sued OpenAI and Microsoft over using content for training AI systems. And the NYT, the BBC and The Verge have blocked OpenAI from scraping their websites.

This article originally appeared on Engadget at https://www.engadget.com/openai-will-reportedly-pay-250-million-to-put-news-corps-journalism-in-chatgpt-214615249.html?src=rss

Apple is battling a $2 billion EU fine over App Store practices

Apple has formally challenged a €1.8 billion ($1.95 billion) antitrust fine issued by the European Union, according to a report. The bloc handed down the penalty in March after determining that Apple had constrained competing music streaming services on the App Store following a 2019 complaint from Spotify.

At the time, Apple pledged to appeal the decision, arguing that the EU failed to "uncover any credible evidence of consumer harm." Bloomberg reports that Apple has now filed a suit in an attempt to overturn the ruling.

Along with the fine, the EU instructed Apple to stop blocking rival music-streaming platforms from telling users that they could sign up for their services at a lower cost away from the App Store. Spotify claimed it had to increase subscription prices to cover costs related to how Apple runs the App Store. That's despite Spotify not making it possible to upgrade to Premium directly through its iOS app — doing so would mean having to fork over a commission to Apple. For its part, Apple maintains that Spotify doesn't pay it anything, even though the latter taps into its APIs, beta testing tools and more.

Spotify's complaint predated the Digital Markets Act coming into force. That law stops defined gatekeepers — including Apple and Play Store operator Google — banning developers from telling users about cheaper ways to pay for their products outside of their app marketplaces. The EU is currently investigating both companies over their compliance with that aspect of the law.

This article originally appeared on Engadget at https://www.engadget.com/apple-is-battling-a-2-billion-eu-fine-over-app-store-practices-160032104.html?src=rss

Yuck: Slack has been scanning your messages to train its AI models

Slack trains machine-learning models on user messages, files and other content without explicit permission. The training is opt-out, meaning your private data will be leeched by default. Making matters worse, you’ll have to ask your organization’s Slack admin (human resources, IT, etc.) to email the company to ask it to stop. (You can’t do it yourself.) Welcome to the dark side of the new AI training data gold rush.

Corey Quinn, an executive at DuckBill Group, spotted the policy in a blurb in Slack’s Privacy Principles and posted about it on X (via PCMag). The section reads (emphasis ours), “To develop AI/ML models, our systems analyze Customer Data (e.g. messages, content, and files) submitted to Slack as well as Other Information (including usage information) as defined in our Privacy Policy and in your customer agreement.”

In response to concerns over the practice, Slack published a blog post on Friday evening to clarify how its customers’ data is used. According to the company, customer data is not used to train any of Slack’s generative AI products — which it relies on third-party LLMs for — but is fed to its machine learning models for products “like channel and emoji recommendations and search results.” For those applications, the post says, “Slack’s traditional ML models use de-identified, aggregate data and do not access message content in DMs, private channels, or public channels.” 

A Salesforce spokesperson reiterated this in a statement to Engadget, also saying that “we do not build or train these models in such a way that they could learn, memorize, or be able to reproduce customer data.”

The opt-out process requires you to do all the work to protect your data. According to the privacy notice, “To opt out, please have your Org or Workspace Owners or Primary Owner contact our Customer Experience team at feedback@slack.com with your Workspace/Org URL and the subject line ‘Slack Global model opt-out request.’ We will process your request and respond once the opt out has been completed.”

The company replied to Quinn’s message on X: “To clarify, Slack has platform-level machine-learning models for things like channel and emoji recommendations and search results. And yes, customers can exclude their data from helping train those (non-generative) ML models.”

How long ago the Salesforce-owned company snuck the tidbit into its terms is unclear. It’s misleading, at best, to say customers can opt out when “customers” doesn’t include employees working within an organization. They have to ask whoever handles Slack access at their business to do that — and I hope they will oblige.

Inconsistencies in Slack’s privacy policies add to the confusion. One section states, “When developing Al/ML models or otherwise analyzing Customer Data, Slack can’t access the underlying content. We have various technical measures preventing this from occurring.” However, the machine-learning model training policy seemingly contradicts this statement, leaving plenty of room for confusion. 

In addition, Slack’s webpage marketing its premium generative AI tools reads, “Work without worry. Your data is your data. We don’t use it to train Slack AI. Everything runs on Slack’s secure infrastructure, meeting the same compliance standards as Slack itself.”

In this case, the company is speaking of its premium generative AI tools, separate from the machine learning models it’s training on without explicit permission. However, as PCMag notes, implying that all of your data is safe from AI training is, at best, a highly misleading statement when the company apparently gets to pick and choose which AI models that statement covers.

Update, May 18 2024, 3:24 PM ET: This story has been updated to include additional information from Slack, which published a blog post explaining its practices in response to the community's concerns. 

This article originally appeared on Engadget at https://www.engadget.com/yuck-slack-has-been-scanning-your-messages-to-train-its-ai-models-181918245.html?src=rss

Slack has been using data from your chats to train its machine learning models

Slack trains machine-learning models on user messages, files and other content without explicit permission. The training is opt-out, meaning your private data will be leeched by default. Making matters worse, you’ll have to ask your organization’s Slack admin (human resources, IT, etc.) to email the company to ask it to stop. (You can’t do it yourself.) Welcome to the dark side of the new AI training data gold rush.

Corey Quinn, an executive at DuckBill Group, spotted the policy in a blurb in Slack’s Privacy Principles and posted about it on X (via PCMag). The section reads (emphasis ours), “To develop AI/ML models, our systems analyze Customer Data (e.g. messages, content, and files) submitted to Slack as well as Other Information (including usage information) as defined in our Privacy Policy and in your customer agreement.”

In response to concerns over the practice, Slack published a blog post on Friday evening to clarify how its customers’ data is used. According to the company, customer data is not used to train any of Slack’s generative AI products — which it relies on third-party LLMs for — but is fed to its machine learning models for products “like channel and emoji recommendations and search results.” For those applications, the post says, “Slack’s traditional ML models use de-identified, aggregate data and do not access message content in DMs, private channels, or public channels.” That data may include things like message timestamps and the number of interactions between users. 

A Salesforce spokesperson reiterated this in a statement to Engadget, also saying that “we do not build or train these models in such a way that they could learn, memorize, or be able to reproduce customer data.”

The opt-out process requires you to do all the work to protect your data. According to the privacy notice, “To opt out, please have your Org or Workspace Owners or Primary Owner contact our Customer Experience team at feedback@slack.com with your Workspace/Org URL and the subject line ‘Slack Global model opt-out request.’ We will process your request and respond once the opt out has been completed.”

The company replied to Quinn’s message on X: “To clarify, Slack has platform-level machine-learning models for things like channel and emoji recommendations and search results. And yes, customers can exclude their data from helping train those (non-generative) ML models.”

How long ago the Salesforce-owned company snuck the tidbit into its terms is unclear. It’s misleading, at best, to say customers can opt out when “customers” doesn’t include employees working within an organization. They have to ask whoever handles Slack access at their business to do that — and I hope they will oblige.

Inconsistencies in Slack’s privacy policies add to the confusion. One section states, “When developing Al/ML models or otherwise analyzing Customer Data, Slack can’t access the underlying content. We have various technical measures preventing this from occurring.” However, the machine-learning model training policy seemingly contradicts this statement, leaving plenty of room for confusion. 

In addition, Slack’s webpage marketing its premium generative AI tools reads, “Work without worry. Your data is your data. We don’t use it to train Slack AI. Everything runs on Slack’s secure infrastructure, meeting the same compliance standards as Slack itself.”

In this case, the company is speaking of its premium generative AI tools, separate from the machine learning models it’s training on without explicit permission. However, as PCMag notes, implying that all of your data is safe from AI training is, at best, a highly misleading statement when the company apparently gets to pick and choose which AI models that statement covers.

Update, May 18 2024, 3:24 PM ET: This story has been updated to include new information from Slack, which published a blog post explaining its practices in response to the community's concerns. 

Update, May 19 2024, 12:41 PM ET: This story and headline have been updated to reflect additional context provided by Slack about how it uses customer data.

This article originally appeared on Engadget at https://www.engadget.com/yuck-slack-has-been-scanning-your-messages-to-train-its-ai-models-181918245.html?src=rss

Netflix is becoming an ad-tech company

There was a time when streamers wooed potential customers with the promise of an ad-free experience. In recent years, however, companies such as Netflix, Amazon, Disney and more have hiked up their prices and made an ad-supported tier the most affordable option. Now, Netflix is taking the next step towards becoming a de-facto ad tech company by moving its development in-house, according to The Hollywood Reporter

Netflix announced the shift during its upfront preview, in which the company also shared that its $7 per month ad-supported tier has 40 million monthly active users. The ad-supported plan is reportedly getting 40 percent of new signups, with it having 15 million users just six months ago, in November. 

The streaming company has relied heavily on Microsoft to reach this success, partnering with the tech giant in 2022 on advertising and sales. But, the training wheels are coming off with Netflix's choice to move things in house, a choice that "will allow us to power the ads plan with the same level of excellence that’s made Netflix the leader in streaming technology today," Netflix ads chief Amy Reinhard said. Microsoft will also no longer be Netflix's sole ad tech partner, as the streamer will start working with companies like Google’s Display & Video 360 and The Trade Desk later this summer. 

This article originally appeared on Engadget at https://www.engadget.com/netflix-is-becoming-an-ad-tech-company-130004240.html?src=rss

Netflix is becoming an ad-tech company

There was a time when streamers wooed potential customers with the promise of an ad-free experience. In recent years, however, companies such as Netflix, Amazon, Disney and more have hiked up their prices and made an ad-supported tier the most affordable option. Now, Netflix is taking the next step towards becoming a de-facto ad tech company by moving its development in-house, according to The Hollywood Reporter

Netflix announced the shift during its upfront preview, in which the company also shared that its $7 per month ad-supported tier has 40 million monthly active users. The ad-supported plan is reportedly getting 40 percent of new signups, with it having 15 million users just six months ago, in November. 

The streaming company has relied heavily on Microsoft to reach this success, partnering with the tech giant in 2022 on advertising and sales. But, the training wheels are coming off with Netflix's choice to move things in house, a choice that "will allow us to power the ads plan with the same level of excellence that’s made Netflix the leader in streaming technology today," Netflix ads chief Amy Reinhard said. Microsoft will also no longer be Netflix's sole ad tech partner, as the streamer will start working with companies like Google’s Display & Video 360 and The Trade Desk later this summer. 

This article originally appeared on Engadget at https://www.engadget.com/netflix-is-becoming-an-ad-tech-company-130004240.html?src=rss

Apple’s big AI rollout at WWDC will reportedly focus on making Siri suck less

Apple will reportedly focus its first round of generative AI enhancements on beefing up Siri’s conversational chops. Sources speaking with The New York Times say company executives realized early last year that ChatGPT made Siri look antiquated. The company allegedly decided that the large language model (LLM) principles behind OpenAI’s chatbot could give the iPhone’s virtual assistant a much-needed shot in the arm. So Apple will reportedly roll out a new version of Siri powered by generative AI at its WWDC keynote on June 10.

Apple Senior Vice Presidents Craig Federighi and John Giannandrea reportedly tested ChatGPT for weeks before the company realized that Siri looked outdated. (I would argue that the epiphany came about a decade late.) What followed was what The NYT describes as Apple’s “most significant reorganization in more than a decade.”

The company sees generative AI as a once-in-a-decade tentpole area worth shifting heaps of resources to address. You may recall the company canceled its $10 billion “Apple Car” project earlier this year. Apple reportedly reassigned many of those engineers to work on generative AI.

Apple executives allegedly fear AI models could eventually replace established software like iOS, turning the iPhone into “a dumb brick” by comparison. The clunky, awkward and overall unconvincing first wave of dedicated AI gadgets we’ve reviewed, like the Human AI Pin and Rabbit R1, aren’t good enough to pose a threat. But that could change as software evolves, other smartphone makers incorporate more AI into their operating systems and other hardware makers have a chance to innovate.

So, at least for now, it appears Apple isn’t launching direct competitors to generative AI stalwarts like ChatGPT (words), Midjourney (images) or ElevenLabs (voices). Instead, it will start with a new Siri and updated iPhone models with expanded memory to better handle local processing. In addition, the company will reportedly add a text-summarizing feature to the Messages app.

Apple’s John Ternus standing in front of a digital slide of the M4 chip.
Apple’s M4 chip (shown next to VP John Ternus) could help process local Siri requests.
Apple

Apple’s first foray into generative AI, if The NYT’s sources are correct, sounds like less of an immediate threat to creators than some had imagined. At its May iPad event, the company ran a video plugging the new iPad Pro that showed various creative tools crushed by a hydraulic press. The clip accidentally served as the perfect metaphor for the (legitimate) fears of artists, musicians and other creators, whose work AI models have trained on — and who stand to be replaced by those same tools as they become more normalized for content creation.

On Thursday, Apple apologized for the ad and said it canceled plans to run it on TV. 

Samsung and Google have already loaded their flagship phones with various generative AI features that go far beyond improving their virtual assistants. These include tools for editing photos, generating text and enhancing transcription (among other things). These features typically rely on cloud-based servers for processing, whereas Apple’s approach will allegedly prioritize privacy and handle requests locally. So Apple will apparently start with a more streamlined approach that sticks to improving what’s already there, as well as keeping most or all processing on-device.

The New York Times’ sources add that Apple’s culture of internal secrecy and privacy-focused marketing have stunted its AI progress. Former Siri engineer John Burkey told the paper that the company’s tendency to silo off the information various divisions share with each other has been another primary culprit in Siri’s inability to evolve far past where the assistant was when it launched a day before Steve Jobs died in 2011.

This article originally appeared on Engadget at https://www.engadget.com/apples-big-ai-rollout-at-wwdc-will-reportedly-focus-on-making-siri-suck-less-203035673.html?src=rss

OpenAI partners with People publisher Dotdash Meredith

OpenAI is partnering with another publisher as it moves towards a licensed approach to training materials. Dotdash Meredith, the owner of brands like People and Better Homes & Gardens, will license its content for OpenAI to train ChatGPT while the publisher will use the AI company’s models to boost its in-house ad-targeting tool.

As part of the arrangement, ChatGPT will display content and links attributed to Dotdash Meredith’s publications. It also provides OpenAI with fully licensed training material from trusted publications.

That’s a welcome change after the company got in hot water for allegedly using content for training purposes without permission. The New York Times and Alden Capital Group (owner of The Chicago Tribune, New York Daily News and the Orlando Sentinel) have sued the ChatGPT maker, accusing it of using its content without permission. Comedian Sarah Silverman and a conspiracy-mongering car salesman (the latter for different reasons) have, too.

“We have not been shy about the fact that AI platforms should pay publishers for their content and that content must be appropriately attributed,” Neil Vogel, Dotdash Meredith CEO, wrote in a press release. “This deal is a testament to the great work OpenAI is doing on both fronts to partner with creators and publishers and ensure a healthy Internet for the future.”

Before the Dotdash Meredith deal, OpenAI struck an agreement with The Financial Times. “It is right, of course, that AI platforms pay publishers for the use of their material,” the paper’s CEO, John Ridding, said in a statement last month.

Dotdash Meredith, which also owns Investopedia, Food & Wine, InStyle and Verywell, will use OpenAI’s models to supercharge its D/Cipher ad-targeting tool. The publisher says its advertising system “connects advertisers directly to consumers based on the context of content being consumed, without using personal identifiers like cookies.” That’s an industry-wide shift on the horizon, as Google is moving to a cookie-less future — albeit later than initially advertised.

This article originally appeared on Engadget at https://www.engadget.com/openai-partners-with-people-publisher-dotdash-meredith-212832821.html?src=rss

Google prohibits ads promoting websites and apps that generate deepfake porn

Google has updated its Inappropriate Content Policy to include language that expressly prohibits advertisers from promoting websites and services that generate deepfake pornography. While the company already has strong restrictions in place for ads that feature certain types of sexual content, this update leaves no doubt that promoting "synthetic content that has been altered or generated to be sexually explicit or contain nudity" is in violation of its rules. 

Any advertiser promoting sites or apps that generate deepfake porn, that show instructions on how to create deepfake porn and that endorse or compare various deepfake porn services will be suspended without warning. They will no longer be able to publish their ads on Google, as well. The company will start implementing this rule on May 30 and is giving advertisers the chance to remove any ad in violation of the new policy. As 404 Media notes, the rise of deepfake technologies has led to an increasing number of ads promoting tools that specifically target users wanting to create sexually explicit materials. Some of those tools reportedly even pretend to be wholesome services to be able to get listed on the Apple App Store and Google Play Store, but it's masks off on social media where they promote their ability to generate manipulated porn. 

Google has, however, already started prohibiting services that create sexually explicit deepfakes in Shopping ads. Similar to its upcoming wider policy, the company has banned Shopping ads for services that "generate, distribute, or store synthetic sexually explicit content or synthetic content containing nudity. " Those include deepfake porn tutorials and pages that advertise deepfake porn generators. 

This article originally appeared on Engadget at https://www.engadget.com/google-prohibits-ads-promoting-websites-and-apps-that-generate-deepfake-porn-130059324.html?src=rss

Google prohibits ads promoting websites and apps that generate deepfake porn

Google has updated its Inappropriate Content Policy to include language that expressly prohibits advertisers from promoting websites and services that generate deepfake pornography. While the company already has strong restrictions in place for ads that feature certain types of sexual content, this update leaves no doubt that promoting "synthetic content that has been altered or generated to be sexually explicit or contain nudity" is in violation of its rules. 

Any advertiser promoting sites or apps that generate deepfake porn, that show instructions on how to create deepfake porn and that endorse or compare various deepfake porn services will be suspended without warning. They will no longer be able to publish their ads on Google, as well. The company will start implementing this rule on May 30 and is giving advertisers the chance to remove any ad in violation of the new policy. As 404 Media notes, the rise of deepfake technologies has led to an increasing number of ads promoting tools that specifically target users wanting to create sexually explicit materials. Some of those tools reportedly even pretend to be wholesome services to be able to get listed on the Apple App Store and Google Play Store, but it's masks off on social media where they promote their ability to generate manipulated porn. 

Google has, however, already started prohibiting services that create sexually explicit deepfakes in Shopping ads. Similar to its upcoming wider policy, the company has banned Shopping ads for services that "generate, distribute, or store synthetic sexually explicit content or synthetic content containing nudity. " Those include deepfake porn tutorials and pages that advertise deepfake porn generators. 

This article originally appeared on Engadget at https://www.engadget.com/google-prohibits-ads-promoting-websites-and-apps-that-generate-deepfake-porn-130059324.html?src=rss