WeTransfer reignites fears over training AI on user data

When WeTransfer quietly altered its terms of service to include language suggesting it could use user uploads for training AI, it triggered a storm of concern. For many creatives artists, writers, filmmakers this vague wording felt like a serious incursion on intellectual property. The episode is a timely reminder: training AI on user data isn’t just a technical possibility; it’s a trust issue.

1. The Trigger: A Vague Clause That Shook Trust

In mid‑2025, WeTransfer updated its terms (effective August 8), granting itself a royalty‑free license to use user‑uploaded content for improving machine learning models, notably tied to content moderation. The clause was broad, unclear, and alarmed users who feared their files might be used or even sold for training AI without permission.

The backlash spread fast across social media, with creatives like Sarah McIntyre and Matt Lieb voicing outrage. One reaction captured the sentiment: “I pay you to shift my big artwork files. I DON’T pay you to have the right to use them to train AI…”

2. Why Training AI on User Data Sparks Anxiety

The fallout wasn’t just about legal wording it was about trust, consent, and ownership. Fear that generative tools might be built on user uploads without clarity or compensation ignited concerns from creative communities.

This incident echoed similar controversies at Zoom, Adobe, Slack, Dropbox, and others where ambiguous AI related terms raised alarms before being retracted or clarified.

3. Backpedal and Cleanup

In response, WeTransfer quickly clarified that it neither used nor ever intended to use user content to train AI. They emphasized that the clause was meant for potential internal content moderation not for model development or commercialization and that no such systems had been deployed.

They updated the terms: users now grant a royalty‑free license strictly for operating, developing, and improving the service, aligned with the Privacy & Cookie Policy explicitly removing references to machine learning.

4. The Broader Implication: A Need for Clearer Communication

This episode underscores how even well‑intentioned language can erode trust if not communicated clearly. Brands like WeTransfer are celebrated for being privacy-conscious so any hint of training AI on user data feels like a betrayal to the creative base.

The backlash reflects wider frustration with tech firms that employ overly broad, legal‑ese terms that blur boundaries. Users now expect clarity: “What are you doing with my stuff?”.

5. Lessons for Platforms and Users

  • For platforms: Language must be precise and transparent. Mentions of AI should have clear definitions especially when they touch on creative or sensitive content.
  • For users: It’s more important than ever to read updates to terms of service, even if long and dense. These changes can have real privacy implications.
  • Industry takeaway: This isn’t the first nor will it likely be the last such incident. Until regulation and awareness catch up, training AI on user data remains a flashpoint for trust.

6. Final Take

WeTransfer’s misstep shows how a single phrase loosely tied to training AI can ignite serious backlash among users who expect control over their creative work. Transparent corrections can restore some confidence, but the underlying tension between innovation and privacy persists.
Explore more insights on digital privacy and AI trends at Infoproweekly your gateway to staying informed and empowered.