Connect with us


Meta Releases New Insights into its Evolving Efforts to Detect Coordinated Manipulation Programs



Meta Releases New Insights into its Evolving Efforts to Detect Coordinated Manipulation Programs

Meta has shared some new insights into its ongoing efforts to combat coordinated misinformation networks operating across its platforms, which became a major focus for the company following the 2016 US Election, and the revelations that Russian-backed teams had sought to sway the opinions of American voters.

As explained by Meta:

Since 2017, we’ve reported on over 150 influence operations with details on each network takedown so that people know about the threats we see – whether they come from nation states, commercial firms or unattributed groups. Information sharing enabled our teams, investigative journalists, government officials and industry peers to better understand and expose internet-wide security risks, including ahead of critical elections.

Meta publishes a monthly round-up of the networks that it’s detected and removed, via automated, user-reported, and other collaborative means, which has broadened its net in working to catch out these groups.

And some interesting trends have emerged in Meta’s enforcement data over time – first off, Meta has provided this overview of where the groups that it has detected and taken action on have originated from.

As you can see, while there have been various groups detected within Russia’s borders, there’s also been a cluster of activity originating from Iran and the surrounding regions, while more recently, Meta has taken action against several groups operating in Mexico.

But even more interesting is Meta’s data on the regions that these groups have been targeting, with a clear shift away from foreign interference, and towards domestic misinformation initiatives.

Meta CIB enforcement over time

As shown in these charts, there’s been a significant move away from international pushes, with localized operations becoming more prevalent, at least in terms of what Meta’s teams have been able to detect.

See also  Twitter Shares New Insights into the Rising Shopping Conversation via Tweet [Infographic]

Which is the other side of the research – those looking to utilize Meta’s platforms for such purpose are always evolving their tactics, in order to avoid detection, and it could be that more groups are still operating outside of Meta’s scope, so this may not be a complete view of misinformation campaign trends, as such.

But Meta has been upping its game, and it does appear to be paying off, with more coordinated misinformation pushes being caught out, and more action being taken to hold perpetrators accountable, in an effort to disincentivize similar programs in future.

But really, it’s going to keep happening. Facebook has reach to almost 3 billion people, while Instagram has over a billion users (reportedly now over 2 billion, though Meta has not confirmed this), and that’s before you consider WhatsApp, which has more than 2 billion users in its own right. At such scale, each of these platforms offers a massive opportunity for amplification of politically-motivated messaging, and while bad actors are able to tap into the amplification potential that each app provides, they will continue to seek ways to do so.

Which is a side effect of operating such popular networks, and one that Meta, for a long time, had either overlooked or refused to see. Most social networks were founded on the principle of connecting the world, and bringing people together, and that core ethos is what motivates all of their innovations and processes, with a view to a better society through increased community understanding, in global terms.

That’s an admirable goal, but the flip side of that is that social platforms also enable those with bad motivations to also connect and establish their own networks, and expand their potentially dangerous messaging throughout the same networks.

See also  Australia sues Facebook owner Meta over scam ads

The clash of idealism and reality has often seemed to flummox social platform CEOs, who, again, would prefer to see the potential good over all else. Crypto networks are now in a similar boat, with massive potential to connect the world, and bring people together, but equally, the opportunity to facilitate money laundering, large-scale scams, tax evasion and potentially worse.

Getting the balance right is difficult, but as we now know, through experience, the impacts of failing to see these gaps can be significant.

Which is why these efforts are so important, and it’s interesting to note both the increasing push from Meta’s teams, and the evolution in tactics from bad actors.


My view? Localized groups, after learning how Russian groups sought to influence the US election, have sought to utilize the same tactics on a local level, meaning that past enforcement has also inadvertently highlighted how Meta’s platforms can be used for such purpose.

That’s likely to continue to be the case moving forward, and hopefully, Meta’s evolving actions will ensure better detection and removal of these initiatives before they can take effect.

You can read Meta’s Coordinated Misinformation Report for December 2021 here.

Source link


New Screenshots Highlight How Snapchat’s Coming ‘Family Center’ Will Work



New Screenshots Highlight How Snapchat's Coming 'Family Center' Will Work

Snapchat’s parental control options look close to launch, with new screenshots based on back-end code showing how Snap’s coming ‘Family Center’ will look in the app.

As you can see in these images, shared by app intelligence company Watchful (via TechCrunch), the Family Center will enable parents to see who their child is engaging with in the app, along with who they’ve added, who they’re following, etc.

That could provide a new level of assurance for parents – though it could also be problematic for Snap, which has become a key resource for more private, intimate connection, with its anti-public posting ethos, and disappearing messages, helping to cement its place as an alternative to other social apps.

That’s really how Snap has embedded its niche. While other apps are about broadcasting your life to the wider world, Snap is about connecting with a small group of friends, where you can share your more private, secret thoughts, without concern of them living on forever, and coming back to bite you at a later stage.

That also, of course, means that more questionable, dangerous communications are happening in the app. Various reports have investigated how Snap is used for sending lewd messages, and arranging hook-ups, while drug dealers reportedly now use Snap to organize meet-ups and sales.

Which, of course, is why parents will be keen to get more insight into such, but I can’t imagine Snap users will be so welcoming of an intrusive tool in this respect.

But if parents know that it exists, they may have to, and that could be problematic for Snap. Teen users will need to accept their parents’ invitation to enable Family Center monitoring, but you can see how this could become an issue for many younger users in the app.


Still, the protective benefits may well be worth it, with random hook-ups and other engagements posing significant risks. And with kids as young as 13 able to create a Snapchat account, there are many vulnerable youngsters engaging in the app.

See also  Your Simple Guide to Twitter #Hashtags

But it could reduce Snap’s appeal, as more parents become aware of the tool.

Snapchat hasn’t provided any further insight into the new Family Center, or when it will be released, but it looks close to launch based on these images.  

Source link

Continue Reading

Subscribe To our Newsletter
We promise not to spam you. Unsubscribe at any time.
Invalid email address