Connect with us

SOCIAL

Instagram Expands Sensitive Content Control Options

Published

on

Instagram Expands Sensitive Content Control Options

Instagram has announced an expansion of its Sensitive Content Control options, which it first launched last July, providing more ways for users to manage their in-app experience.

Originally only covering the app’s ‘Explore’ element, Instagram will now enable users to manage the content that they see in all sections of the app.

As explained by Instagram:

Starting today, the Sensitive Content Control will cover all surfaces where we make recommendations. In addition to Explore, you will now be able to control the amount of sensitive content and accounts you see in Search, Reels, Accounts You Might Follow, Hashtag Pages and In-Feed Recommendations. With this update, we’re also applying the technology we use to enforce our Recommendation Guidelines to Instagram’s recommendations on Search and Hashtag pages. This update will be available to everyone in the coming weeks.”

As you can see in the screenshots above, Instagram’s sensitive content controls provide three options: “More”, “Standard” and “Less”.

“Standard” is the default state, and will prevent people from seeing some sensitive content and accounts. “More” enables people to see more sensitive content and accounts, whereas “Less” means they see less of this content than the default state. For people under the age of 18, the “More” option is unavailable.

That’ll help to reduce exposure to potentially harmful content in the app, which is a significant concern for parents, in particular, especially given the amount of time their kids are now spending in online environments.

That’s been further amplified by the pandemic, with youngsters forced online to maintain social connection and engagement. It’s impossible for parents to monitor what their kids are seeing at all times, but any measure to better protect them from the worst of the web is a bonus, as they seek to minimize confusion and harm caused by such.

Parental controls also provide additional assurance on this front, and Instagram also added new parental control options in December last year.

Instagram Parental Controls

Sensitive content controls are another element, which are obviously not solely focused on youngsters, as such, but could help in adding extra assurance for parents.

As well as, of course, for regular users. One misguided search could lead you down the wrong path very easily, and it can be handy to have these extra measures to avoid unexpected, and unwanted visuals popping up on screen.

It’s another important step for IG, which continues to develop new tools to protect users from negative experiences.

To change your Sensitive Content Controls on Instagram:

  • Go to your profile
  • Tap the ‘Settings’ menu in the upper right corner
  • Tap ‘Account’
  • Tap ‘Sensitive Content Control’

Source link

Keep an eye on what we are doing
Be the first to get latest updates and exclusive content straight to your email inbox.
We promise not to spam you. You can unsubscribe at any time.
Invalid email address

SOCIAL

Walmart says it has stopped advertising on Elon Musk’s X platform

Published

on

Walmart says it has stopped advertising on Elon Musk's X platform

Walmart said Friday that it is scaling back its advertising on X, the social media company formerly known as Twitter, because “we’ve found some other platforms better for reaching our customers.”

Walmart’s decision has been in the works for a while, according to a person familiar with the move. Yet it comes as X faces an advertiser exodus following billionaire owner Elon Musk’s support for an antisemitic post on the platform. 

The retailer spends about $2.7 billion on advertising each year, according to MarketingDive. In an email to CBS MoneyWatch, X’s head of operations, Joe Benarroch, said Walmart still has a large presence on X. He added that the company stopped advertising on X in October, “so this is not a recent pausing.”

“Walmart has a wonderful community of more than a million people on X, and with a half a billion people on X, every year the platform experiences 15 billion impressions about the holidays alone with more than 50% of X users doing most or all of their shopping online,” Benarroch said.

Musk struck a defiant pose earlier this week at the New York Times’ Dealbook Summit, where he cursed out advertisers that had distanced themselves from X, telling them to “go f— yourself.” He also complained that companies are trying to “blackmail me with advertising” by cutting off their spending with the platform, and cautioned that the loss of big advertisers could “kill” X.

“And the whole world will know that those advertisers killed the company,” Musk added.


Elon Musk faces backlash from lawmakers, companies over endorsement of antisemitic X post

02:23

Dozens of advertisers — including players such as Apple, Coca Cola and Disney — have bailed on X since Musk tweeted that a post on the platform that claimed Jews fomented hatred against White people, echoing antisemitic stereotypes, was “the actual truth.”

Advertisers generally shy away from placing their brands and marketing messages next to controversial material, for fear that their image with consumers could get tarnished by incendiary content. 

The loss of major advertisers could deprive X of up to $75 million in revenue, according to a New York Times report

Musk said Wednesday his support of the antisemitic post was “one of the most foolish” he’d ever posted on X. 

“I am quite sorry,” he said, adding “I should in retrospect not have replied to that particular post.”

Source link

Keep an eye on what we are doing
Be the first to get latest updates and exclusive content straight to your email inbox.
We promise not to spam you. You can unsubscribe at any time.
Invalid email address
Continue Reading

SOCIAL

US Judge Blocks Montana’s Effort to Ban TikTok

Published

on

U.S. Judge Blocks Montana’s Effort to Ban TikTok in the State

TikTok has won another reprieve in the U.S., with a district judge blocking Montana’s effort to ban the app for all users in the state.

Back in May, Montana Governor Greg Gianforte signed legislation to ban TikTok outright from operating in the state, in order to protect residents from alleged intelligence gathering by China. There’s no definitive evidence that TikTok is, or has participated in such, but Gianforte opted to move to a full ban, going further than the government device bans issued in other regions.

As explained by Gianforte at the time:

The Chinese Communist Party using TikTok to spy on Americans, violate their privacy, and collect their personal, private, and sensitive information is well-documented. Today, Montana takes the most decisive action of any state to protect Montanans’ private data and sensitive personal information from being harvested by the Chinese Communist Party.”

In response, a collection of TikTok users challenged the proposed ban, arguing that it violated their first amendment rights, which led to this latest court challenge, and District Court Judge Donald Molloy’s decision to stop Montana’s ban effort.

Montana’s TikTok ban had been set to go into effect on Jan. 1, 2024.

In issuing a preliminary injunction to stop Montana from imposing a full ban on the app, Molloy said that Montana’s legislation does indeed violate the Constitution and “oversteps state power.”

Molloy’s judgment is primarily centered on the fact that Montana has essentially sought to exercise foreign policy authority in enacting a TikTok ban, which is only enforceable by federal authorities. Molloy also noted that there was apervasive undertone of anti-Chinese sentiment” within Montana’s proposed legislation.

TikTok has welcomed the ruling, issuing a brief statement in response:

Montana attorney general, meanwhile, has said that it’s considering next steps to advance its proposed TikTok ban.

The news is a win for TikTok, though the Biden Administration is still weighing a full TikTok ban in the U.S., which may still happen, even though the process has been delayed by legal and legislative challenges.

As I’ve noted previously, my sense here would be that TikTok won’t be banned in the U.S. unless there’s a significant shift in U.S.-China relations, and that relationship is always somewhat tense, and volatile to a degree.

If the U.S. government has new reason to be concerned, it may well move to ban the app. But doing so would be a significant step, and would prompt further response from the C.C.P.

Which is why I suspect that the U.S. government won’t act, unless it feels that it has to. And right now, there’s no clear impetus to implement a ban, and stop a Chinese-owned company from operating in the region, purely because of its origin.

Which is the real crux of the issue here. A TikTok ban is not just banning a social media company, it’s blocking cross-border commerce, because the company is owned by China, which will remain the logic unless clear evidence arises that TikTok has been used as a vector for gathering information on U.S. citizens.

Banning a Chinese-owned app because it is Chinese-owned is a statement, beyond concerns about a social app, and the U.S. is right to tread carefully in considering how such a move might impact other industries.

So right now, TikTok is not going to be banned, in Montana, or anywhere else in the U.S. But that could still change, very quickly.



Source link

Keep an eye on what we are doing
Be the first to get latest updates and exclusive content straight to your email inbox.
We promise not to spam you. You can unsubscribe at any time.
Invalid email address
Continue Reading

SOCIAL

EU wants to know how Meta tackles child sex abuse

Published

on

The investigation is the first step in procedures launched under the EU's new online content law known as the Digital Services Act

The investigation is the first step in procedures launched under the EU’s new online content law known as the Digital Services Act – Copyright AFP Kirill KUDRYAVTSEV

The EU on Friday demanded Instagram-owner Meta provide more information about measures taken by the company to address child sexual abuse online.

The request for information focuses on Meta’s risk assessment and mitigation measures “linked to the protection of minors, including regarding the circulation of self-generated child sexual abuse material (SG-CSAM) on Instagram”, the European Commission said.

Meta must also give information about “Instagram’s recommender system and amplification of potentially harmful content”, it added.

The investigation is the first step in procedures launched under the EU’s Digital Services Act (DSA), but does not itself constitute an indication of legal violations or a move towards punishment.

Meta must respond by December 22.

A report by Stanford University and the Wall Street Journal in June this year said Instagram is the main platform used by paedophile networks to promote and sell content showing child sexual abuse.

Meta at the time said it worked “aggressively” to fight child exploitation.

The commission has already started a series of investigations against large digital platforms seeking information about how they are complying with the DSA.

It has sought more information from Meta in October about the spread of disinformation as well as a request for information last month about how the company protects children online.

The DSA is part of the European Union’s powerful regulatory armoury to bring big tech to heel, and requires digital giants take more aggressive action to counter the spread of illegal and harmful content as well as disinformation.

Platforms face fines that can go up to six percent of global turnover for violations.

Source link

Keep an eye on what we are doing
Be the first to get latest updates and exclusive content straight to your email inbox.
We promise not to spam you. You can unsubscribe at any time.
Invalid email address
Continue Reading

Trending