Instagram’s status update feature is coming to user profiles

Instagram’s status update feature, Notes, will soon be more prominent in the app. Up until now, Notes have only been visible from Instagram’s inbox, but the brief updates will soon also be visible directly on users’ profiles.

The change should increase the visibility of the feature and give people a new place to interact with their friends’ updates. (Instagram added reply functionality to Notes back in December.) The app is also experimenting with “prompts” for Notes, which will allow users to share questions for their friends to answer in their updates, much like the collaborative “add yours” templates for Stories.

Notes are similar to Stories in that the updates only stick around for 24 hours, though they are only visible to mutual followers, so they aren’t meant to be as widely shared as a typical grid or Stories post. The latest updates are another sign of how Meta has used the feature, first introduced in 2022, to encourage users to post more often for smaller, more curated groups of friends.

Separately, the app is also adding a new “cutouts” feature, which allows users to make stickers out of objects in their photos, much like the iOS sticker feature. On Instagram, these stickers can be shared in Stories or in a Reel. Cutouts can also be made from other users’ public posts, effectively giving people a new way to remix content from others (Instagram’s help page notes that users can disable this feature if they prefer for their content to not be reused.)

This article originally appeared on Engadget at https://www.engadget.com/instagrams-status-update-feature-is-coming-to-user-profiles-182621692.html?src=rss

Meta’s AI image generator struggles to create images of couples of different races

Meta AI is consistently unable to generate accurate images for seemingly simple prompts like “Asian man and Caucasian friend,” or “Asian man and white wife,” The Verge reports. Instead, the company’s image generator seems to be biased toward creating images of people of the same race, even when explicitly prompted otherwise.

Engadget confirmed these results in our own testing of Meta’s web-based image generator. Prompts for “an Asian man with a white woman friend” or “an Asian man with a white wife” generated images of Asian couples. When asked for “a diverse group of people,” Meta AI generated a grid of nine white faces and one person of color. There were a couple occasions when it created a single result that reflected the prompt, but in most cases it failed to accurately depict the prompt.

As The Verge points out, there are other more “subtle” signs of bias in Meta AI, like a tendency to make Asian men appear older while Asian women appeared younger. The image generator also sometimes added “culturally specific attire” even when that wasn’t part of the prompt.

It’s not clear why Meta AI is struggling with these types of prompts, though it’s not the first generative AI platform to come under scrutiny for its depiction of race. Google’s Gemini image generator paused its ability to create images of people after it overcorrected for diversity with bizarre results in response prompts about historical figures. Google later explained that its internal safeguards failed to account for situations when diverse results were inappropriate.

Meta didn’t immediately respond to a request for comment. The company has previously described Meta AI as being in “beta” and thus prone to making mistakes. Meta AI has also struggled to accurately answer simple questions about current events and public figures.

This article originally appeared on Engadget at https://www.engadget.com/metas-ai-image-generator-struggles-to-create-images-of-couples-of-different-races-231424476.html?src=rss

Meta’s AI image generator struggles to create images of couples of different races

Meta AI is consistently unable to generate accurate images for seemingly simple prompts like “Asian man and Caucasian friend,” or “Asian man and white wife,” The Verge reports. Instead, the company’s image generator seems to be biased toward creating images of people of the same race, even when explicitly prompted otherwise.

Engadget confirmed these results in our own testing of Meta’s web-based image generator. Prompts for “an Asian man with a white woman friend” or “an Asian man with a white wife” generated images of Asian couples. When asked for “a diverse group of people,” Meta AI generated a grid of nine white faces and one person of color. There were a couple occasions when it created a single result that reflected the prompt, but in most cases it failed to accurately depict the prompt.

As The Verge points out, there are other more “subtle” signs of bias in Meta AI, like a tendency to make Asian men appear older while Asian women appeared younger. The image generator also sometimes added “culturally specific attire” even when that wasn’t part of the prompt.

It’s not clear why Meta AI is struggling with these types of prompts, though it’s not the first generative AI platform to come under scrutiny for its depiction of race. Google’s Gemini image generator paused its ability to create images of people after it overcorrected for diversity with bizarre results in response prompts about historical figures. Google later explained that its internal safeguards failed to account for situations when diverse results were inappropriate.

Meta didn’t immediately respond to a request for comment. The company has previously described Meta AI as being in “beta” and thus prone to making mistakes. Meta AI has also struggled to accurately answer simple questions about current events and public figures.

This article originally appeared on Engadget at https://www.engadget.com/metas-ai-image-generator-struggles-to-create-images-of-couples-of-different-races-231424476.html?src=rss

X names its third head of safety in less than two years

X has named a new head of safety nearly a year after the last executive in the position resigned. The company said Tuesday that it had promoted Kylie McRoberts to Head of Safety and hired Yale Cohen as Head of Brand Safety and Advertiser Solutions.

The two will have the unenviable task of leading X’s safety efforts, including its attempts to reassure advertisers that the platform doesn’t monetize hate speech or terrorist content. The company said earlier this year it planned to hire 100 new safety employees after previously cutting much of its safety staff.

Head of safety has been a particularly fraught position since Elon Musk took over the company previously known as Twitter. Musk has previously clashed with his safety leads and McRoberts is the third person to hold the title in less than two years. Previously, Yoel Roth resigned shortly after the disastrous rollout of Twitter Blue in 2022. Roth was replaced by Ella Irwin, who resigned last year after Musk publicly criticized employees for enforcing policies around misgendering.

Not much is known about McRoberts, but she is apparently an existing member of X’s safety team (her X account is currently private and a LinkedIn profile appears to have been recently deleted). “During her time at X, she has led initiatives to increase transparency in our moderation practices through labels, improve security with passkeys, as well as building out our new Safety Center of Excellence in Austin,” X said in a statement.

This article originally appeared on Engadget at https://www.engadget.com/x-names-its-third-head-of-safety-in-less-than-two-years-213004771.html?src=rss

X names its third head of safety in less than two years

X has named a new head of safety nearly a year after the last executive in the position resigned. The company said Tuesday that it had promoted Kylie McRoberts to Head of Safety and hired Yale Cohen as Head of Brand Safety and Advertiser Solutions.

The two will have the unenviable task of leading X’s safety efforts, including its attempts to reassure advertisers that the platform doesn’t monetize hate speech or terrorist content. The company said earlier this year it planned to hire 100 new safety employees after previously cutting much of its safety staff.

Head of safety has been a particularly fraught position since Elon Musk took over the company previously known as Twitter. Musk has previously clashed with his safety leads and McRoberts is the third person to hold the title in less than two years. Previously, Yoel Roth resigned shortly after the disastrous rollout of Twitter Blue in 2022. Roth was replaced by Ella Irwin, who resigned last year after Musk publicly criticized employees for enforcing policies around misgendering.

Not much is known about McRoberts, but she is apparently an existing member of X’s safety team (her X account is currently private and a LinkedIn profile appears to have been recently deleted). “During her time at X, she has led initiatives to increase transparency in our moderation practices through labels, improve security with passkeys, as well as building out our new Safety Center of Excellence in Austin,” X said in a statement.

This article originally appeared on Engadget at https://www.engadget.com/x-names-its-third-head-of-safety-in-less-than-two-years-213004771.html?src=rss

LinkedIn is testing a TikTok-like feed for vertical video

LinkedIn is testing a new feed of TikTok-like vertical videos. The feature hasn’t been publicly announced but it’s been spotted by users in recent days and the company confirmed the tests to TechCrunch.

According to a screenshot shared by Instagram employee Jenny Eishingdrelo and a video posted to LinkedIn by influencer marketing exec Austin Null, the new feed will appear in a separate “video” tab in the LinkedIn app. Users will be able to scroll vertically to move between clips, much like TikTok or Instagram Reels.

It’s not the first time the company has hopped on a trendy format. LinkedIn previously experimented with a Stories feature for disappearing posts. That feature lasted less than a year, though the professional network hinted at the time that it wasn’t done with its video experiments, saying it was working “to evolve the Stories format into a reimagined video experience across LinkedIn.”

Presumably, LinkedIn is hoping the feed will showcase content from its ranks of professional creators and thought leaders, many of whom are already posting video to their feeds. However, it’s not clear how many of the site’s users are interested in a dedicated video feed for workplace-related content.

This article originally appeared on Engadget at https://www.engadget.com/linkedin-is-testing-a-tiktok-like-feed-for-vertical-video-233454044.html?src=rss

LinkedIn is testing a TikTok-like feed for vertical video

LinkedIn is testing a new feed of TikTok-like vertical videos. The feature hasn’t been publicly announced but it’s been spotted by users in recent days and the company confirmed the tests to TechCrunch.

According to a screenshot shared by Instagram employee Jenny Eishingdrelo and a video posted to LinkedIn by influencer marketing exec Austin Null, the new feed will appear in a separate “video” tab in the LinkedIn app. Users will be able to scroll vertically to move between clips, much like TikTok or Instagram Reels.

It’s not the first time the company has hopped on a trendy format. LinkedIn previously experimented with a Stories feature for disappearing posts. That feature lasted less than a year, though the professional network hinted at the time that it wasn’t done with its video experiments, saying it was working “to evolve the Stories format into a reimagined video experience across LinkedIn.”

Presumably, LinkedIn is hoping the feed will showcase content from its ranks of professional creators and thought leaders, many of whom are already posting video to their feeds. However, it’s not clear how many of the site’s users are interested in a dedicated video feed for workplace-related content.

This article originally appeared on Engadget at https://www.engadget.com/linkedin-is-testing-a-tiktok-like-feed-for-vertical-video-233454044.html?src=rss

X is working on NSFW Communities for adult content

X is working on features that will allow admins of “Communities,” the platform’s tool for subreddit-like groups, to designate the spaces as containing “adult content.” The change was confirmed by an engineer at X amid reports that the Elon Musk-owned company was working on enabling NSFW groups.

In a post on X, engineer Dong Wook Chung noted that “soon” NSFW content would be automatically filtered in the app’s Communities feature. “Admins can now set 'Adult content' in Settings to avoid auto-filtering of the content,” Chung said.

As Bloomberg reported, researchers had previously spotted clues that X planned to enable settings for “adult-sensitive” content. X permits users to share nudity and other “graphic” content, but doesn’t allow it to appear in certain parts of the app, like profile photos and cover images for Communities.

X’s Communities feature predates Musk’s takeover of the company. Twitter began experimenting with the idea in 2021, saying it would provide “a more intimate space for conversations” on the platform. Though Twitter never publicly discussed enabling NSFW features for Communities, the app allowed adult content, unlike most of its social media peers. The company reportedly looked into creating an OnlyFans competitor with its creator subscription product in 2022. The plan was eventually scrapped, according to the Platformer newsletter, due to concerns it would “worsen” the company’s problems with illegal child exploitation content.

It’s not clear if X’s current leadership has addressed those concerns. In a separate post, Chung, the X engineer, stated that the new filtering settings “is about making Communities safer for everyone by automatically filtering out” adult content. “Only users who have specified their age will be able to search Communities with NSFW content.” 

X didn’t immediately respond to a request for comment.

This article originally appeared on Engadget at https://www.engadget.com/x-is-working-on-nsfw-communities-for-adult-content-184629839.html?src=rss

X is working on NSFW Communities for adult content

X is working on features that will allow admins of “Communities,” the platform’s tool for subreddit-like groups, to designate the spaces as containing “adult content.” The change was confirmed by an engineer at X amid reports that the Elon Musk-owned company was working on enabling NSFW groups.

In a post on X, engineer Dong Wook Chung noted that “soon” NSFW content would be automatically filtered in the app’s Communities feature. “Admins can now set 'Adult content' in Settings to avoid auto-filtering of the content,” Chung said.

As Bloomberg reported, researchers had previously spotted clues that X planned to enable settings for “adult-sensitive” content. X permits users to share nudity and other “graphic” content, but doesn’t allow it to appear in certain parts of the app, like profile photos and cover images for Communities.

X’s Communities feature predates Musk’s takeover of the company. Twitter began experimenting with the idea in 2021, saying it would provide “a more intimate space for conversations” on the platform. Though Twitter never publicly discussed enabling NSFW features for Communities, the app allowed adult content, unlike most of its social media peers. The company reportedly looked into creating an OnlyFans competitor with its creator subscription product in 2022. The plan was eventually scrapped, according to the Platformer newsletter, due to concerns it would “worsen” the company’s problems with illegal child exploitation content.

It’s not clear if X’s current leadership has addressed those concerns. In a separate post, Chung, the X engineer, stated that the new filtering settings “is about making Communities safer for everyone by automatically filtering out” adult content. “Only users who have specified their age will be able to search Communities with NSFW content.” 

X didn’t immediately respond to a request for comment.

This article originally appeared on Engadget at https://www.engadget.com/x-is-working-on-nsfw-communities-for-adult-content-184629839.html?src=rss

Snapchat’s latest paid perk is an AI Bitmoji of your pet

Snapchat has a new AI-powered perk for subscribers: Bitmoji versions of your pet. The feature, which is unfortunately not called “petmoji,” allows users to snap a photo of their four-legged friend to create a cartoon-like avatar to accompany their Bitmoji in the Snap Map.

Based on screenshots shared by the company, it seems users will be able to choose from a few different variations of the AI-generated images after sharing a photo of their pet. That’s considerably less customization than what you can do with your own human-inspired Bitmoji,though it should allow users to create something that looks similar to their IRL pet. (No word on if Snap could one day introduce branded pet accessories for animal avatars like they do for human Bitmoji.)

The addition is also the latest example of how Snap has embraced AI features in its subscription offering. Since debuting Snapchat+ in 2022, the company has used the premium service to experiment with generative AI features, including its MyAI assistant as well as camera-powered features like Dreams and AI-generated snaps. Snapchat+ has more than 7 million subscribers, the company announced in December.

Elsewhere, Snap added some updates for non-subscribers, too. The app is adding a new template feature to make it easier to edit clips, and new swipe-based gestures to send and edit snaps more quickly. Snapchat will also support longer video uploads for Stories and Spotlight. In-app captures can now be three minutes long, while the app will support uploads of up to five minutes.

This article originally appeared on Engadget at https://www.engadget.com/snapchats-latest-paid-perk-is-an-ai-bitmoji-of-your-pet-235027028.html?src=rss