How Safe Is Signal Messaging App Used by Trump Aides to Share War Plans?

The Signal messaging app logo is seen on a smartphone, in front of the same displayed same logo, in this illustration taken, January 13, 2021. (Reuters)
The Signal messaging app logo is seen on a smartphone, in front of the same displayed same logo, in this illustration taken, January 13, 2021. (Reuters)
TT
20

How Safe Is Signal Messaging App Used by Trump Aides to Share War Plans?

The Signal messaging app logo is seen on a smartphone, in front of the same displayed same logo, in this illustration taken, January 13, 2021. (Reuters)
The Signal messaging app logo is seen on a smartphone, in front of the same displayed same logo, in this illustration taken, January 13, 2021. (Reuters)

Top Trump administration officials used messaging app Signal to share war plans and mistakenly included a journalist in the encrypted chat, spurring calls by Democratic lawmakers for a congressional investigation into the security breach.

Under US law, it can be a crime to mishandle, misuse or abuse classified information, though it is unclear whether those provisions might have been violated in this case.

Below are some of the main facts about Signal:

HOW SAFE IS IT?

Signal is an open-source and fully encrypted messaging service that runs on centralized servers maintained by Signal Messenger.

The only user data it stores on its servers are phone numbers, the date a user joined the service, and the last login information.

Users' contacts, chats and other communications are instead stored on the user's phone, with the possibility of setting the option to automatically delete conversations after a certain amount of time.

The company uses no ads or affiliate marketers, and doesn't track users' data, as stated on its website.

Signal also gives users the possibility to hide their phone number from others and use an additional safety number to verify the safety of their messages, it adds.

Signal does not use US government encryption or that of any other governments, and is not hosted on government servers.

The messaging app has a "stellar reputation and is widely used and trusted in the security community", said Rocky Cole, whose cybersecurity firm iVerify helps protect smartphone users from hackers.

"The risk of discussing highly sensitive national security information on Signal isn't so much that Signal itself is insecure," Cole added.

Actors who pose threats to nation states, he said, "have a demonstrated ability to remotely compromise the entire mobile phone itself. If the phone itself isn't secure, all the Signal messages on that device can be read."

HOW DOES SIGNAL WORK?

Signal is a secure messaging service that uses end-to-end encryption, meaning the service provider cannot access and read private conversations and calls from users on its app, therefore guaranteeing its users' privacy.

Signal's software is available across platforms, both on smartphones and computers, and enables messaging, voice and video calls. A telephone number is necessary to register and create an account.

Unlike other messaging apps, Signal does not track or store user data, and its code is publicly available, so security experts can verify how it works and ensure it remains safe.

Signal President Meredith Whittaker on Tuesday defended the app's security: "Signal is the gold standard in private comms."

She added in a post on X: "WhatsApp licenses Signal’s cryptography to protect message contents for consumer WhatsApp."

WHO FOUNDED SIGNAL?

Signal was founded in 2012 by entrepreneur Moxie Marlinspike and Whittaker, according to the company's website.

In February 2018, Marlinspike alongside WhatsApp co-founder Brian Acton started the non-profit Signal Foundation, which currently oversees the app.

Acton provided an initial funding of $50 million. Acton left WhatsApp in 2017 due to differences around the use of customer data and targeted advertising.

Signal is not tied to any major tech companies and will never be acquired by one, it says on its website.

WHO USES SIGNAL?

Widely used by privacy advocates and political activists, Signal has gone from an exotic messaging app used by dissidents to a whisper network for journalists and media, to a messaging tool for government agencies and organizations.

Signal saw "unprecedented" growth in 2021 after a disputed change in rival WhatsApp's privacy terms, as privacy advocates jumped off WhatsApp on fears users would have to share their data with both Facebook and Instagram.

Reuters lists Signal as one of the tools tipsters can use to share confidential news tips with its journalists, while noting that "no system is 100 percent secure".

Signal's community forum, an unofficial group which states that its administration is composed of Signal employees, also lists the European Commission as a user of the tool. In 2017, the US Senate Sergeant at Arms approved the use of Signal for Senate staff.

"Although Signal is widely regarded as offering very secure communications for consumers due to its end-to-end encryption and because it collects very little user data, it is hard to believe it is suitable for exchanging messages related to national security," said Ben Wood, chief analyst at CCS Insight - alluding to the breach involving top Trump aides discussing plans for military strikes on Yemeni Houthi militants.

Google's message services Google Messages and Google Allo, as well as Meta's Facebook Messenger and WhatsApp, use the Signal Protocol, according to Signal's website.



Reddit Sues AI Giant Anthropic Over Content Use

Dario Amodei, co-founder and CEO of Anthropic. JULIEN DE ROSA / AFP
Dario Amodei, co-founder and CEO of Anthropic. JULIEN DE ROSA / AFP
TT
20

Reddit Sues AI Giant Anthropic Over Content Use

Dario Amodei, co-founder and CEO of Anthropic. JULIEN DE ROSA / AFP
Dario Amodei, co-founder and CEO of Anthropic. JULIEN DE ROSA / AFP

Social media outlet Reddit filed a lawsuit Wednesday against artificial intelligence company Anthropic, accusing the startup of illegally scraping millions of user comments to train its Claude chatbot without permission or compensation.

The lawsuit in a California state court represents the latest front in the growing battle between content providers and AI companies over the use of data to train increasingly sophisticated language models that power the generative AI revolution.

Anthropic, valued at $61.5 billion and heavily backed by Amazon, was founded in 2021 by former executives from OpenAI, the creator of ChatGPT.

The company, known for its Claude chatbot and AI models, positions itself as focused on AI safety and responsible development.

"This case is about the two faces of Anthropic: the public face that attempts to ingratiate itself into the consumer's consciousness with claims of righteousness and respect for boundaries and the law, and the private face that ignores any rules that interfere with its attempts to further line its pockets," the suit said.

According to the complaint, Anthropic has been training its models on Reddit content since at least December 2021, with CEO Dario Amodei co-authoring research papers that specifically identified high-quality content for data training.

The lawsuit alleges that despite Anthropic's public claims that it had blocked its bots from accessing Reddit, the company's automated systems continued to harvest Reddit's servers more than 100,000 times in subsequent months.

Reddit is seeking monetary damages and a court injunction to force Anthropic to comply with its user agreement terms. The company has requested a jury trial.

In an email to AFP, Anthropic said "We disagree with Reddit's claims and will defend ourselves vigorously."

Reddit has entered into licensing agreements with other AI giants including Google and OpenAI, which allow those companies to use Reddit content under terms that protect user privacy and provide compensation to the platform.

Those deals have helped lift Reddit's share price since it went public in 2024.

Reddit shares closed up more than six percent on Wednesday following news of the lawsuit.

Musicians, book authors, visual artists and news publications have sued the various AI companies that used their data without permission or payment.

AI companies generally defend their practices by claiming fair use, arguing that training AI on large datasets fundamentally changes the original content and is necessary for innovation.

Though most of these lawsuits are still in early stages, their outcomes could have a profound effect on the shape of the AI industry.