Tech & IT
Meta Accused of “Blindfolding” Itself to Child Sex Abuse with Expanded Encryption
Law enforcement agencies worldwide, including the FBI and Interpol, are expressing concern over parent company Meta’s plans to roll out expanded end-to-end encryption, warning it could effectively “blindfold” the firm from detecting incidents of child sex abuse.
The Virtual Global Taskforce, a coalition of 15 law enforcement organizations tasked with protecting children from such crimes, pointed out Meta in a joint statement urging tech companies to consider secure protocols when instituting end-to-end encryption.
“The announced implementation of [end-to-end encryption] on Meta platforms Instagram and Facebook is an example of a purposeful design choice that degrades safety systems and weakens the ability to safeguard child users,” the Virtual Global Taskforce said in a policy statement. The officials argue end-to-end encryption, while a sought-after privacy feature for secure communications, could make it more difficult for companies like Meta to identify criminal behavior occurring on their platforms.
“The VGT calls for all industry partners to fully understand the impact of implementing system design decisions that result in blinding themselves to [child sex abuse] occurring on their platforms, or reduce their capacity to detect CSA and keep kids secure,” the agencies added.
“The abuse will not stop just because companies decide to cease looking,” the agencies added.
Meta has indicated plans to roll out end-to-end encryption for messages on all of its platforms – with one company executive once stating the feature would be enabled by default “sometime in 2023.” Meta-controlled WhatsApp already offers the feature by default.
Earlier this year, Meta published a blog post detailing expanded end-to-end encryption on its Messenger platform.
The Financial Times was first to report on the Virtual Global Taskforce’s statement.
The outlet noted that UK lawmakers are currently working on an online safety bill that has drawn criticism from tech giants who allege it will hurt user privacy.
The proposed legislation would empower the UK’s telecom regulator, the Office of Communications, to require companies to monitor some messages for instances of child abuse.
An open letter signed by various tech bosses, including WhatsApp chief Will Cathcart, argued the bill would “give an unelected official the power to weaken the privacy of billions of people around the world” by scrutinizing encrypted messages.
Meta defended its safety practices in a statement obtained by the FT.
“The vast majority of Brits already rely on apps that use encryption. We don’t think people want us reading their private messages, so we have developed safety measures that prevent, detect and allow us to take action against this appalling abuse, while preserving online privacy and security,” a Meta spokesperson said in the statement.
“As we continue to roll out our end-to-end encryption plans, we remain dedicated to working with law enforcement and child safety experts to ensure that our platforms are safe for young people.” The Post has reached out to Meta for further comment.
Meta has faced intense criticism from US legislators over its safety practices, with detractors arguing the tech giant hasn’t gone far enough to protect its underage users from harmful content and abuse.
As The Post reported earlier this month, online safety experts penned an open letter urging Meta CEO Mark Zuckerberg to abandon the company’s plans to let children and teen users access its new metaverse service “Horizon Worlds” due to concern about potential abuse.
The Virtual Global Taskforce is an alliance of 15 law enforcement agencies from around the world, including the FBI, US Immigration and Customs Enforcement (ICE), Interpol, Europol, and the United Kingdom’s National Crime Agency, with the latter serving as the group’s chair.
The task force’s website describes the group as “an international coalition of 15 committed law enforcement agencies collaborating to address the global threat from child sexual abuse.”
Source: The Post, FT.com
You may like
Meta Seeks Dismissal of User Safety Lawsuit
Twitter Contemplates Legal Action against Meta Over the Rapidly Growing Threads App
Court Gives Green Light to Lawsuit Claiming Meta’s Ads Discriminated
Meta: Investigating Network of Pedophiles on Instagram, Linked by Algorithms
The Meta Quest 3 VR and AR Headset: Impressions of a Pre-Release Trial
Instagram disabled by “mistake” Suzanne Hillinger’s account
Facebook and Instagram Tighten Advertising Rules for Teens
Tech & IT
The Truth About Free Speech, Big Tech, and Protecting Our Children
In today’s world, there is a lot of confusion about what “free speech” truly means, especially regarding the influence of Big Tech. As Americans continue to idolize tech billionaires, it’s essential to understand the legal boundaries of free speech and how these platforms operate, especially when children’s safety is at stake.
What Is Free Speech?
Free speech, as protected under the First Amendment of the U.S. Constitution, is often misunderstood. The First Amendment restricts the government’s ability to limit speech, but it doesn’t grant individuals the right to say whatever they want on private platforms. Whether it’s a social media site, a restaurant, or a business, private companies have the right to moderate or restrict speech on their terms. The idea that users are entitled to free speech on platforms like Facebook, Twitter, or Telegram is a misconception. These platforms are private businesses, not public forums.
However, these tech companies promote themselves as champions of free speech while still exercising significant control over the content they allow. This creates an illusion of free speech where, in reality, users must follow the rules set by these billionaires.
The Cost of Unchecked Platforms on Children’s Safety
One of the most alarming issues today is the way tech platforms are being exploited for child abuse and sex trafficking. While some Big Tech companies claim they are creating safe spaces, many have been slow or reluctant to address the growing epidemic of child sexual exploitation on their platforms. Reports have shown that Facebook, Instagram, Twitter, and even Telegram have become hotbeds for child trafficking and the distribution of abusive material.
For example, Telegram’s founder, Pavel Durov, has been praised for allowing free speech on his platform. However, recent investigations reveal that Telegram has been slow to cooperate with law enforcement, particularly in cases involving child abuse. French authorities recently arrested Durov for allegedly failing to provide information in child exploitation cases. This arrest raises serious concerns about the safety of children online and how tech platforms, even those claiming to defend free speech, might be complicit in illegal activities.
These companies prioritize profit over safety. They know tightening security would cost them time and money, so they continue allowing unsafe environments to thrive. Children are the ones paying the price as these platforms enable predators to find and exploit them.
The Greed Behind Big Tech
At the heart of the problem is greed. Tech billionaires like Durov, Mark Zuckerberg (Facebook), and Elon Musk (Twitter) have made fortunes by creating platforms that allow anyone to voice their opinions. However, these platforms have also created opportunities for criminals, including child traffickers. Instead of focusing on safety, these companies prioritize user engagement, which increases ad revenue, data collection, and, ultimately, their bottom line.
Despite the ongoing abuse, companies like Twitter have cut teams responsible for monitoring child exploitation. Under Elon Musk’s leadership, Twitter reduced its child safety monitoring staff, even though Musk publicly stated that protecting children would be a top priority. The result? An increase in dangerous and illegal content that harms vulnerable young users.
Similarly, Facebook and Instagram have failed to take meaningful steps to combat child trafficking on their platforms. Lawsuits have even been filed against these tech giants, accusing them of promoting child trafficking. Instead of acting decisively to protect children, these billionaires protect their business models and profits.
Protecting Free Speech While Safeguarding Children
There is a clear need to balance free speech with the responsibility to protect children. While people have the right to express their opinions, this does not mean tech platforms should turn a blind eye to illegal and harmful activities on their sites. Big Tech’s refusal to adopt stronger protections is not about defending free speech—it’s about greed and profit.
It is crucial to demand more accountability from these platforms. The public must understand that free speech doesn’t give anyone the right to endanger others, particularly children. If platforms are not ensuring safety, they should be held accountable for their negligence.
The Solution
To protect free speech and ensure the safety of our children, tech companies need to take a stand against illegal activities. This means investing in moderation, cooperating with law enforcement, and putting ethics before profit. While Big Tech platforms offer valuable services, they cannot continue to put children at risk to grow their empires.
Parents, governments, and communities must stay vigilant and pressure these platforms to enforce stronger safety measures while protecting free speech. Free speech should never come at the cost of our children’s safety.
In conclusion, the battle for free speech must not ignore the importance of protecting society’s most vulnerable. As long as greed drives tech companies’ decision-making processes, our children will remain in danger. It’s time to demand better.
Source: healthimpactnews.com
Tech & IT
Low Users, Sex Scandals Sink Korean Metaverses; 3AC Sues Terra
South Korea’s tech-savvy population, high-speed internet, and deep gaming culture once made it the perfect setting for metaverse platforms. Despite these strengths, several leading South Korean metaverses are now shutting down due to low user engagement and controversies.
The most recent closure is 2nd Block, a metaverse created by Dunamu, the operator of South Korea’s largest cryptocurrency exchange, Upbit. Dunamu announced it will close on September 9 due to declining user activity since the pandemic began to ease.
This follows the closure of Seoul’s $4.5 million metaverse project, which will shut down on October 16 after less than two years. The platform was designed to offer a digital space for residents and tourists, but it failed to maintain interest.
Even metaverses still in operation are struggling. Zepeto, developed by South Korean internet giant Naver, has shifted focus to the global market, boasting 20 million monthly overseas users, compared to just 1.2 million in South Korea. However, Zepeto has faced domestic criticism after reports emerged of children’s avatars being sexually harassed, with predators coercing minors into sharing explicit images in exchange for in-game items.
In response, South Korean lawmakers proposed stricter penalties for virtual sex offenders, but the bill failed to pass, leaving the issue unresolved. The challenge of tracking predators across international borders complicates enforcement.
In a related legal battle, liquidators of the defunct Singapore crypto hedge fund Three Arrows Capital (3AC) have filed a $1.3 billion lawsuit against Terraform Labs. The lawsuit alleges that Terraform misled 3AC about the stability of its stablecoin TerraUSD (UST) and its sister token LUNA. Terra’s collapse in May 2022 led to a $40 billion loss in the Terra ecosystem and forced 3AC into bankruptcy. Terraform Labs settled a $4.47 billion SEC lawsuit in June. Meanwhile, Terraform’s co-founder, Do Kwon, remains in Montenegro awaiting extradition.
Source: cointelegraph.com
Tech & IT
Apple Faces Class Action Over Child Abuse Content
Apple Inc. is facing a class action lawsuit accusing the company of failing to prevent the upload and distribution of child sexual abuse material (CSAM) on its iCloud service. The lawsuit, filed by a minor identified as Jane Doe, claims that Apple has engaged in “privacy washing”—promoting a public image of strong privacy protection while allegedly neglecting to implement effective safeguards against the transmission of harmful content.
According to the lawsuit, Apple is fully aware that iCloud has become a significant platform for the distribution of child pornography. The plaintiff alleges that despite this knowledge, Apple has chosen not to adopt industry-standard practices for detecting and removing CSAM from its services. Instead, the company is accused of shifting the responsibility and costs of ensuring a safe online environment onto children and their families.
The lawsuit criticizes Apple for failing to live up to its stated privacy policies. “Apple’s privacy policy claims that it collects users’ private information to prevent the spread of CSAM, but it fails to do so in practice—a failure that Apple was already aware of,” the lawsuit states. This alleged negligence has led to Apple being accused of misrepresentation, unjust enrichment, and violations of federal sex trafficking laws, California business and professional codes, and North Carolina consumer protection laws.
Jane Doe, through her guardian, seeks to represent a nationwide class of individuals who have been victims of child sexual abuse due to the transmission of CSAM on iCloud over the past three years. The lawsuit demands a jury trial and calls for declaratory and injunctive relief, as well as compensatory and punitive damages for all class members.
This legal action against Apple adds to a growing list of lawsuits the company is facing. Earlier this year, two antitrust class actions were filed against Apple, accusing it of monopolizing the smartphone market with its iPhone and engaging in anticompetitive practices.
As the case progresses, it raises significant questions about the responsibility of tech giants like Apple to protect their users, particularly vulnerable groups such as children, from harmful online content. The outcome of this lawsuit could have far-reaching implications for how companies handle privacy and security in the digital age.
The plaintiff in this case is represented by Juyoun Han and Eric Baum of Eisenberg & Baum LLP, and John K. Buche and Byron Ma of The Buche Law Firm PC. The case, Doe, et al. v. Apple Inc., is currently being heard in the U.S. District Court for the Northern District of California under Case No. 5:24-cv-05107.
If you have been a victim of child sexual abuse as a result of CSAM being uploaded to iCloud, you are encouraged to share your experience in the comments.
Source: topclassactions.com
Trending
- Selfcare & Sexual Wellness8 months ago
The Dark Side of the Playboy Mansion
- Selfcare & Sexual Wellness8 months ago
Adult Star’s Choice to Avoid Pregnancy at Work
- Cam Models3 years ago
EmilyJonesChat
- Finance & Business2 months ago
BCAMS Magazine, the 22th issue!
- Finance & Business6 months ago
BCAMS Magazine, the 21th issue!
- Cam Models2 years ago
Demora Avarice
Warning: Undefined variable $user_ID in /home/bcamsmagazine/public_html/wp-content/themes/zox-news/comments.php on line 49
You must be logged in to post a comment Login
This site uses User Verification plugin to reduce spam. See how your comment data is processed.