AI ‘undressing apps’ are a warning shot for AI in construction
AOL.com • 4/1/2026, 12:01:10 PM
By WorksRecorded Field Desk — practical notes on AI tools and AI in construction.

The short version
TikTok has reportedly carried ads for AI apps that can digitally undress people in photos. It’s a grim little corner of the internet, but it should ring alarms far beyond social media. If consumer-grade AI tools can strip clothes off a selfie, they can also fake IDs, alter safety footage, or manipulate project documentation.
For anyone betting big on AI in construction, this story is less about TikTok and more about what happens when powerful automation leaves the lab and lands in the wild. The same construction technology we celebrate for streamlining submittals and clash detection can, in the wrong hands, quietly distort reality.
When AI can cheaply rewrite what the camera saw, every photo, plan, and report on a jobsite becomes both an asset and a liability.
Why this matters on real projects
The reported TikTok ads highlight a class of AI tools that do one thing very well: they generate convincing fakes from ordinary images. In social media, that means deepfake-style harassment. On a jobsite, the same underlying capability could target people, records, and even legal evidence.
Consider a few uncomfortable but plausible extensions into AI in construction:
- **Faked safety incidents or the absence of them.** If an app can undress someone, it can just as easily "dress" someone in PPE that wasn’t there. Imagine AI quietly adding hard hats and harnesses to site photos to pass an audit—or, conversely, stripping them off to fabricate a safety violation in a dispute.
- **Manipulated progress photos.** Owners and lenders increasingly rely on photo documentation to verify schedule and payment milestones. An AI model trained to erase defects, hide incomplete scopes, or touch up cracked concrete is already within reach of off-the-shelf automation.
- **Identity and credential abuse.** Deepfake-style tools can alter faces, badges, and documents. That’s not science fiction; it’s the same class of technology behind these undressing apps. For construction technology platforms that use selfies for access control or timekeeping, this is a real attack surface.
The TikTok case underscores two key truths about AI tools:
1. **They’re becoming cheap, fast, and casual.** You don’t need a research lab to run this kind of model anymore. A phone, a credit card, and an ad network are enough. That’s the same ecosystem construction is plugging into when it adopts plug-and-play AI features. 2. **They don’t care what industry they hit.** An AI model that can redraw clothing can also redraw scaffolding, guardrails, or rebar. Once the capability exists, it will be repurposed.
For construction companies racing to digitize, this is a reminder that **governance has to keep pace with innovation**. It’s not enough to roll out AI tools that label RFIs or summarize daily reports; you also need policies for:
- What counts as an original record.
- How photos and videos are stored, watermarked, and audited.
- When AI-generated content is allowed—and how it must be labeled.
The uncomfortable lesson from TikTok’s reported ads is that the market will happily monetize harmful AI uses until someone draws a line. On a jobsite, that "someone" has to be you: the GC, the owner, the tech vendor.
What to watch next
- **Stronger content controls in construction platforms.** Expect project management and reality-capture tools to add tamper-evident logs, cryptographic watermarks, and AI-detection layers so teams can prove when an image or report has been altered.
- **Contract language about AI manipulation.** We’re likely to see contracts start naming AI in construction explicitly—defining what counts as acceptable automation and what constitutes falsification of records.
- **Insurer and regulator scrutiny.** Insurers, OSHA investigators, and building officials may become more skeptical of digital evidence, pushing for independent verification or certified capture tools.
- **Workforce and HR exposure.** The same tech behind undressing apps can target employees and subcontractors. Expect new policies around image use, harassment, and personal-device apps on jobsites.
- **Ethical AI standards from vendors.** Construction technology providers will face pressure to prove that their AI tools are designed with guardrails, especially when they touch imagery, identity, or safety data.
Field note from the editor
When I read about TikTok running ads for AI apps that digitally undress people, my first reaction wasn’t "that’s a social media problem"—it was, "this is exactly the kind of AI that will wander onto jobsites if we’re not careful."
We’re in a moment where AI in construction is finally doing useful, almost boring things: auto-filling forms, sifting RFIs, flagging clashes, nudging schedules. That’s the good side of automation. But the same wave is bringing in tools that can casually rewrite reality—photos, faces, documents—with a swipe.
If there’s a takeaway for builders, it’s this: **don’t separate "AI ethics" from "field operations."** They’re now the same conversation. The sooner project teams, IT, and legal sit down together to set boundaries, the more we can enjoy the upside of AI tools without getting blindsided by their darker, TikTok-style cousins.