Introduction

In our first post in this series, “Your Data Is Listening — And Watching”, we explained how your daily digital footprint—every click, swipe, or voice command—gets logged and stitched into an invisible profile that algorithms use to influence everything you see.

In our second post, “Cambridge Analytica Rebranded”, we peeled back the curtain on how those same psychological targeting tools didn't vanish—they just reappeared under new names/new coat of paint, quietly shaping opinions behind a friendlier company veneer.

Now, in this, our third post in the series, we’re fast-forwarding to 2025—where profiling isn’t reactive or static. It’s real-time, relentless, and happening whether you notice it or not.

Forget surveys or “take this quiz”—today’s AI watches how you speak, move, and even breathe, then clusters that info into a profile before you can blink.

In this post we'll explain how your devices—your phone, smartwatch, smart car, and even house IOT sensors—are all “phoning home,” feeding AI systems that build emotional and behavioural snapshots of you.

We’ll map the journey from old-school psychometrics to always-on surveillance, walk through how your data becomes someone’s ad strategy or mental health nudges, and dig into why that should freak you out—or at least make you ask for your rights back.

Annie are you ok?
🙈 AI companies love to look smooth, but time will tell if it’s just a polished cover for something more despicable.

How AI Builds You: From Manual Profiling to Real-Time AI Insights

Imagine the difference between sitting in a quiet room, filling out a personality test, and going about your day while invisible AI quietly pieces together your emotional landscape. That’s the gulf between old-school CA profiling and today’s real-time AI profiling.

Manual CA Profiling: Solid, but Old-School

Back in the day, profiling meant answering questionnaires or standardised tests. You consciously report how you feel or what you prefer. It’s procedural and explicit, but slow and stuck in static categories—think MBTI (Myers Briggs Type Indicator), demographics, or psychometric buckets.

Real-Time AI Profiling: Always On, Always Learning

Fast forward to 2025. AI isn’t waiting for your answers anymore—why would it, it doesn't need to ask for permission, it’s observing you on device. Through smartphone apps, wearables, connected cars, smart home gizmos, and even the software you interact with, AI is collecting a continuous stream of your behavioural signals:

  • Wearables & physiological cues: Smartwatches and wristbands don’t just track steps or heartbeat—they can infer stress or anxiety in real time by monitoring skin conductance or heart rate variability.
  • Multimodal emotion sensing: AI frameworks like IoT‑ConvNet process sensors, expressions, speech, and gestures through deep learning to detect emotional states across environments like smart homes or healthcare.
  • App and device telemetry: Ever noticed how your phone or navigation system seems to “phone home”? Telemetry collects data on your habits—from when you open an app to how long you stare at a screen. This data fuels AI models that spot patterns at scale.
  • IoT systems everywhere: Your connected fridge, home sensors, or even the car's audio system, are constantly generating data. AI analyzes all these signals—sometimes inferencing mood or behavior without you even realising it.

With so much data streaming—from devices, apps, infrastructure AI can:

  • Detect emotion in real time (e.g., joy, stress, anger) without self-reporting
  • Cluster users dynamically based on behaviour, time of use, mood patterns, and usage contexts—not fixed typologies.
  • Scale massively and instantly—profiling groups of people on the fly as data pours in.

From Data Collection to Commercial Targeting

So what happens once AI has grouped you into a cluster? Simple: businesses—and plenty of other players—are lining up to buy and use that information. Suddenly, the ads you see aren’t just based on what you searched or bought, but on when you’re distracted, relaxed, anxious, or primed to spend.

How it works (generally speaking):

  1. Your phone, smartwatch, or even your smart car picks up micro behaviours.
  2. AI clusters you with similar patterns—“late-night browsers,” “stressed commuters,” “enthusiastic weekend exercisers.”
  3. Those clusters are sold or leveraged across marketing systems to tailor offers, ads, or even influence you.

All of this happens quietly in the background—no forms, no consent, just invisible AI constantly building, tweaking, and monetising your profile. And you don’t earn a cent. In fact, you probably paid for the privilege—only to have your own data packaged up and sold back to you.

Ethical & Privacy Risks

It’s hard to ignore the darker side of invisible profiling. As AI gets smarter and more independent, the privacy worries keep stacking up. Take agentic AI—the kind that runs with little human oversight. Sounds efficient, sure—but when it gets it wrong, who’s actually accountable?

Want to know who’s to blame when AI goes off the rails? Keep an eye on this case (link below)—it’ll show just how much (or how little) Big Tech will actually be held to account.

Sadly, I’m already calling it: a “thoughts and prayers” outcome at best. Once the T&Cs get waved around by high-priced lawyers, expect the blame to shift—service providers painted as victims, and parents demonised as “bad parents.”

Parents sue OpenAI over ChatGPT’s role in son’s suicide | TechCrunch
Before sixteen-year-old Adam Raine died by suicide, he had spent months telling ChatGPT about his plans to end his life.

August 26th 2025

The Stanford 2025 AI Index Report (below) found a 56% jump in AI-related privacy and security incidents in a single year.

Most of those were data breaches or flat-out misuse. In other words, the very systems profiling us in real time aren’t just busy watching—they’re also pretty exposed to bad actors.

Regulation & Governance Landscape

guardrails are finally being bolted on. In Europe, the EU’s AI Act rolled out in August 2024. It bans the worst stuff—like real-time biometric ID and social scoring—and forces extra transparency when it comes to high-risk AI systems.

The UK followed with its Data (Use and Access) Act 2025, basically saying: “Okay, AI can make some automated decisions, but only if humans can step in, you’ve got the right to appeal, and the process isn’t a total black box.”

And closer to home, Australia’s government has been pushing back against the likes of Meta, signalling that tougher privacy reforms are on the way—this time written with the AI era in mind.

🐌
But let’s be real: laws move at a snail’s pace, while tech sprints. So don’t hold your breath waiting for regulation to catch up any time soon.

Conclusion

The thing about AI companies? They’re in an all-out sprint—hoovering up data and scrambling to beat the competition. But in that mad dash, something always gets trashed: your rights, your privacy, and, bit by bit, the social fabric we all rely on.

But with that power comes a responsibility that’s too often missing: where’s the consent?

The oversight? The fairness? This isn’t just about slicker marketing or endless consumerism—it’s about protecting autonomy in a world where the invisible commercial gaze never blinks.

power and responsibility
Everyone agrees 🫳🎤

Don’t just take my word for it—look at the real toll social media has taken on young people. Studies and internal TikTok documents reveal (READ MORE HERE) that the app can become addictive in as little as 35 minutes, and that excessive usage is tied to cognitive issues like anxiety, memory loss, and diminished empathy.

Even more chilling, unsealed internal videos show TikTok staffers raising concerns about the algorithm pushing disordered-eating content to teens, saying some users feel they “never want to leave.” And in France, seven grieving families have filed a lawsuit claiming TikTok’s algorithm exposed their children to harmful, self-harm content—contributing to devastating outcomes.

These stories aren’t isolated. They're a warning: the race for growth and to embrace the new often tramples real human costs.

So here’s the question facing us now: Are we going to learn from this? Or will we keep running hard toward the future—only to end up spending the next era cleaning up the wreckage caused by unchecked AI profiling and corporate ambition?

🗣️
If this sounds concerning, you're not alone. Speak up—demand clear privacy policies from your apps and devices. Support legislation that protects against opaque surveillance. And, most importantly, remain conscious of what you share—because in a world where your profile is being built behind the scenes, knowledge is your most powerful tool.
trust and privacy not at the cost of AI implementation
👉️ Trust’s already been gutted by corporate greed and the obsession with “shareholder value.” Will AI just be the next trick to distract us while companies dig deeper into our pockets? Let’s hope not.

Posts In This Series

CA’s Legacy: Emerdata, Auspex & Data Propria
Cambridge Analytica collapsed in 2018, but its methods never died. Former staff and investors re-emerged in new firms—Emerdata, Auspex, and Data Propria—still using psychological profiling and micro-targeting to influence politics. Same game, same players, new names - your the target!

POST 2

Cambridge Analytica: How Your Data Became a Weapon
Cambridge Analytica wasn’t just a scandal, it was a preview of how data could be weaponised. In this first post of our six-part series, we break down how they pulled it off—and why the same tactics still shape what you think, how you vote, and what you see online.

POST 1

Further Reading + Sources + References

The Impact of Social Media on the Mental Health of Adolescents and Young Adults: A Systematic Review - SOURCE

Khalaf AM, Alubied AA, Khalaf AM, Rifaey AA. The Impact of Social Media on the Mental Health of Adolescents and Young Adults: A Systematic Review. Cureus. 2023 Aug 5;15(8):e42990. doi: 10.7759/cureus.42990. PMID: 37671234; PMCID: PMC10476631.

TikTok knew depth of app’s risks to children, court document alleges - SOURCE

📱 Your Phone Is Listening — And We Can Prove It
iPhone or Android, Privacy Is Dead Unless You Fight for It Let’s stop pretending this is a conspiracy theory. You say something—like “I need a massa
Facebook–Cambridge Analytica data scandal - Wikipedia
USA: Families sue Character.AI over alleged role in encouraging violence, self-harm and sexual content - Business & Human Rights Resource Centre
Check out this page via the Business and Human Rights Resource Centre
Putting the ‘Capitalism’ in ‘Surveillance Capitalism’
In the 21st century, every life experience is a monetizable data point.
How big five personality traits influence information sharing on social media: A meta analysis - PMC
Research interest in information sharing behavior on social media has significantly increased over the past decade. However, empirical studies on the relationship between Big Five personality traits and information sharing behavior have yielded…

Ditch Microsoft & Apple: Come to the Dark Side - Try MX Linux + Save A Packet
Fed up with Microsoft and Apple subscription fees? Embrace MX Linux: subscription-free, lightning-fast, and refreshingly easy. Revive your previous-generation computer with zero hassle. The dark side is peaceful here—no sharks allowed.
GrapheneOS: Secure Your Data and Privacy on Your Smartphone
Tired of big tech companies exploiting your personal data? GrapheneOS offers a secure, privacy-focused alternative to mainstream operating systems. In this guide, we explore what GrapheneOS is, why you should consider using it, and how to install it. Take control of your digital life today.
Secure Your Data with Cryptomator: Easy NSA-Grade Protection
Discover how Cryptomator provides NSA-grade security for your personal data. This user-friendly encryption tool ensures your files are protected from unauthorized access. Learn the importance of data security and how to easily install and use Cryptomator on various devices.
BorgBase & Vorta: Secure, Budget-Friendly Data Backup
Learn why you need solid data prevention plans—and how BorgBase + Vorta deliver encrypted, cost-effective backups with easy setup across Mac, Linux and more for under about $120 AUD/year (non sponsored post)

✅ Want to learn how to update your privacy settings?

Naomi Brockwell’s NBTV YouTube channel features great explainers that are clear and easy to understand.

Naomi Brockwell TV

Want to dig a little deeper into the dark side of tech? Don’t worry—not tinfoil-hat deep, just enough to actually understand what’s going on.

Vanessa Wingårdh
Developer exploring the dark side of tech