AI Tool Uses Selfies to Predict Biological Age and Cancer Survival

Three pedestrians take a selfie on the picturesque alleyway at the end of Rue de l'Universite, Paris. Ian LANGSDON / AFP/File
Three pedestrians take a selfie on the picturesque alleyway at the end of Rue de l'Universite, Paris. Ian LANGSDON / AFP/File
TT
20

AI Tool Uses Selfies to Predict Biological Age and Cancer Survival

Three pedestrians take a selfie on the picturesque alleyway at the end of Rue de l'Universite, Paris. Ian LANGSDON / AFP/File
Three pedestrians take a selfie on the picturesque alleyway at the end of Rue de l'Universite, Paris. Ian LANGSDON / AFP/File

Doctors often start exams with the so-called "eyeball test" -- a snap judgment about whether the patient appears older or younger than their age, which can influence key medical decisions.

That intuitive assessment may soon get an AI upgrade.

FaceAge, a deep learning algorithm described Thursday in The  Lancet Digital Health, converts a simple headshot into a number that more accurately reflects a person's biological age rather than the birthday on their chart.

Trained on tens of thousands of photographs, it pegged cancer patients on average as biologically five years older than healthy peers. The study's authors say it could help doctors decide who can safely tolerate punishing treatments, and who might fare better with a gentler approach.

"We hypothesize that FaceAge could be used as a biomarker in cancer care to quantify a patient's biological age and help a doctor make these tough decisions," said co-senior author Raymond Mak, an oncologist at Mass Brigham Health, a Harvard-affiliated health system in Boston.

Consider two hypothetical patients: a spry 75‑year‑old whose biological age clocks in at 65, and a frail 60‑year‑old whose biology reads 70. Aggressive radiation might be appropriate for the former but risky for the latter.

The same logic could help guide decisions about heart surgery, hip replacements or end-of-life care.

Sharper lens on frailty

Growing evidence shows humans age at different rates, shaped by genes, stress, exercise, and habits like smoking or drinking. While pricey genetic tests can reveal how DNA wears over time, FaceAge promises insight using only a selfie.

The model was trained on 58,851 portraits of presumed-healthy adults over 60, culled from public datasets.

It was then tested on 6,196 cancer patients treated in the United States and the Netherlands, using photos snapped just before radiotherapy. Patients with malignancies looked on average 4.79 years older biologically than their chronological age.

Among cancer patients, a higher FaceAge score strongly predicted worse survival -- even after accounting for actual age, sex, and tumor type -- and the hazard rose steeply for anyone whose biological reading tipped past 85.

Intriguingly, FaceAge appears to weigh the signs of aging differently than humans do. For example, being gray-haired or balding matters less than subtle changes in facial muscle tone.

FaceAge boosted doctors' accuracy, too. Eight physicians were asked to examine headshots of terminal cancer patients and guess who would die within six months. Their success rate barely beat chance; with FaceAge data in hand, predictions improved sharply.

The model even affirmed a favorite internet meme, estimating actor Paul Rudd's biological age as 43 in a photo taken when he was 50.

Bias and ethics guardrails

AI tools have faced scrutiny for under‑serving non-white people. Mak said preliminary checks revealed no significant racial bias in FaceAge's predictions, but the group is training a second‑generation model on 20,000 patients.

They're also probing how factors like makeup, cosmetic surgery or room lighting variations could fool the system.

Ethics debates loom large. An AI that can read biological age from a selfie could prove a boon for clinicians, but also tempting for life insurers or employers seeking to gauge risk.

"It is for sure something that needs attention, to assure that these technologies are used only in the benefit for the patient," said Hugo Aerts, the study's co-lead who directs MGB's AI in medicine program.

Another dilemma: What happens when the mirror talks back? Learning that your body is biologically older than you thought may spur healthy changes -- or sow anxiety.

The researchers are planning to open a public-facing FaceAge portal where people can upload their own pictures to enroll in a research study to further validate the algorithm. Commercial versions aimed at clinicians may follow, but only after more validation.



Reddit Sues AI Giant Anthropic Over Content Use

Dario Amodei, co-founder and CEO of Anthropic. JULIEN DE ROSA / AFP
Dario Amodei, co-founder and CEO of Anthropic. JULIEN DE ROSA / AFP
TT
20

Reddit Sues AI Giant Anthropic Over Content Use

Dario Amodei, co-founder and CEO of Anthropic. JULIEN DE ROSA / AFP
Dario Amodei, co-founder and CEO of Anthropic. JULIEN DE ROSA / AFP

Social media outlet Reddit filed a lawsuit Wednesday against artificial intelligence company Anthropic, accusing the startup of illegally scraping millions of user comments to train its Claude chatbot without permission or compensation.

The lawsuit in a California state court represents the latest front in the growing battle between content providers and AI companies over the use of data to train increasingly sophisticated language models that power the generative AI revolution.

Anthropic, valued at $61.5 billion and heavily backed by Amazon, was founded in 2021 by former executives from OpenAI, the creator of ChatGPT.

The company, known for its Claude chatbot and AI models, positions itself as focused on AI safety and responsible development.

"This case is about the two faces of Anthropic: the public face that attempts to ingratiate itself into the consumer's consciousness with claims of righteousness and respect for boundaries and the law, and the private face that ignores any rules that interfere with its attempts to further line its pockets," the suit said.

According to the complaint, Anthropic has been training its models on Reddit content since at least December 2021, with CEO Dario Amodei co-authoring research papers that specifically identified high-quality content for data training.

The lawsuit alleges that despite Anthropic's public claims that it had blocked its bots from accessing Reddit, the company's automated systems continued to harvest Reddit's servers more than 100,000 times in subsequent months.

Reddit is seeking monetary damages and a court injunction to force Anthropic to comply with its user agreement terms. The company has requested a jury trial.

In an email to AFP, Anthropic said "We disagree with Reddit's claims and will defend ourselves vigorously."

Reddit has entered into licensing agreements with other AI giants including Google and OpenAI, which allow those companies to use Reddit content under terms that protect user privacy and provide compensation to the platform.

Those deals have helped lift Reddit's share price since it went public in 2024.

Reddit shares closed up more than six percent on Wednesday following news of the lawsuit.

Musicians, book authors, visual artists and news publications have sued the various AI companies that used their data without permission or payment.

AI companies generally defend their practices by claiming fair use, arguing that training AI on large datasets fundamentally changes the original content and is necessary for innovation.

Though most of these lawsuits are still in early stages, their outcomes could have a profound effect on the shape of the AI industry.