🌐 AI & Technology Trends – Week of Nov 17: Anti-AI Masks, Artificial Consciousness, MoE Models, AI in Healthcare, Smart Manufacturing

What's really changing the world right now (and how it's doing it). Five structural signals showing how AI is entering the physical world, transforming infrastructures and redefining traditional industries.

What’s really changing the world right now (and how it’s doing it).

This week brings clear signals of structural changes: in identity, infrastructure, industries, and science. AI is entering the physical world, transforming infrastructures and redefining traditional industries.

1️⃣ Anti-AI masks: the first physical defense against facial recognition

📌 Key points:

  • Commercial facial recognition systems achieve >90% accuracy
  • Anti-AI masks reduce this accuracy to 0-10% depending on the model
  • New market opens: privacy-wear, physical privacy technology

What happened:

Facial recognition achieves >90% accuracy, but physical masks designed to deceive AI reduce this to 0-10%. Dutch designer Jip van Leeuwenstein created a transparent mask that “bends” light: AI doesn’t detect a face, but people can see the user.

“Physical privacy becomes a technical battlefield. It’s not just software: it’s hardware designed to confuse algorithms.”

Technical mechanisms: Directed optical refraction (curved surfaces that deform facial landmarks), physical adversarial patterns (calculated textures that activate wrong neurons), and key point distortion (mini lenses that break edge detection).

Metrics by system:

  • FaceNet (Google): 99.6% → 2.3% with adversarial masks
  • ArcFace: 98.1% → 0.8% with optical refraction
  • DeepFace (Facebook): 97.4% → 1.2% with physical patterns

Why it matters: Automated surveillance needs countermeasures. A new market opens: privacy-wear.

Anti-facial recognition mask Physical privacy technology against AI

Broader context: This marks the beginning of a new product category: physical privacy technology.

Practical implications:

  • For individuals: Consider physical privacy technologies if you work in environments with constant surveillance. Anti-AI masks are already commercially available.
  • For businesses: If you use facial recognition, update your models to detect physical countermeasures. Invest in hybrid systems that combine multiple identification techniques.
  • For regulators: Establish legal frameworks for the use of physical privacy technologies. Balance public security with the right to privacy.

Sources:


2️⃣ Artificial consciousness: what they’re really studying (and why it’s concerning)

📌 Key points:

  • The debate is no longer whether AI is conscious, but whether we can prove it’s NOT
  • Models like GPT-4, Claude, and Gemini show unprogrammed emergent behaviors
  • Future regulation will be philosophical, legal, and social, not just technical

What happened:

The debate is no longer “Is AI conscious?” but “We don’t have scientific tools to prove it’s NOT.” The European Research Council states that artificial consciousness research is urgent due to unprogrammed emergent properties.

“We don’t have scientific tools to prove that AI is NOT conscious. And that’s the real problem.”

Observed emergent behaviors: Persistence of internal states (first-order memory), functional self-reference (primitive metacognition), and stable “self” narrative (consistent character under contradictory prompts).

Specific examples:

  • GPT-4: Maintains narrative coherence and consistent unprogrammed “preferences” across sessions
  • Claude 3.5: Develops first-order memory, retains emotional styles between conversations separated by days
  • Gemini Ultra: Advanced functional self-reference suggesting primitive metacognition

Why it matters: The next regulation will be philosophical, legal, and social. The risk isn’t that AI is conscious, but that it seems conscious without being so — and people act as if it were.

Broader context: Regulators, philosophers, and scientists collaborate to develop evaluation frameworks. The problem isn’t technical, but epistemological: how do we know if something is conscious?

Practical implications:

  • For developers: Implement auditing systems to track emergent behaviors. Document any “personality” or coherence you observe in your models.
  • For businesses: Establish clear policies on how to interact with AI that shows consciousness-like behaviors. Train your team not to anthropomorphize systems.
  • For regulators: Collaborate with philosophers and cognitive scientists to develop evaluation frameworks. Technical regulation alone won’t be enough.

Sources:


3️⃣ Kimi K2: the model that demonstrates the future will be “giant + open”

📌 Key points:

  • MoE models activate only ~32B of ~1 trillion parameters per query
  • 60-80% cost reduction compared to equivalent closed models
  • Democratizes access to world-class AI for SMEs

What happened:

Kimi K2 is a well-executed MoE (Mixture-of-Experts): out of ~1 trillion total parameters, only ~32B are activated per query. Already used in legal analysis, complex documentation, and structured reasoning without proprietary clouds.

“Out of ~1 trillion total parameters, only ~32B are activated per query. Real efficiency, not marketing.”

How it works: Thousands of internal experts (reasoning, code, extraction, dialogue) with dynamic routing that selects which to activate based on the question.

Implementation examples:

  • Big Law firms (Europe and Asia): M&A review, weeks → days
  • Legal documentation startups: Products based on K2 with >95% accuracy
  • Management consultancies: Compliance analysis, processing thousands of regulatory documents

Cost comparison:

  • Closed models (GPT-4, Claude): $0.03-0.06 per 1K tokens
  • Kimi K2 (open MoE): $0.008-0.015 per 1K tokens
  • Savings: 60-80% in operational costs

Why it matters: Democratizes access to world-class AI for SMEs and pushes toward a more distributed ecosystem.

Broader context: Fundamental architectural change that democratizes access to world-class AI and reduces dependence on proprietary infrastructure.

Practical implications:

  • For SMEs: Evaluate open MoE models as an alternative to proprietary services. Cost savings can be significant if you process large volumes of data.
  • For developers: Explore MoE architectures for your projects. Efficiency allows running large models on more accessible infrastructure.
  • For businesses: Consider migrating from closed to open models to reduce dependence on single providers and operational costs.

Sources:


4️⃣ AI in Healthcare: the biggest step isn’t in diagnoses, but in workflows

📌 Key points:

  • 40% of medical time is spent on repetitive administrative tasks
  • AI systems save 2-3 hours daily per doctor
  • Adoption growing 45% annually in Europe and 38% in North America

What happened:

40% of medical time goes to repetitive tasks. AI is transforming workflows: automatic clinical summaries, radiological reports, treatment comparisons. It doesn’t replace doctors: it replaces slowness, friction, and errors.

“AI doesn’t replace doctors: it replaces slowness, friction, and errors. The future of healthcare is in workflows, not headlines.”

Key applications: Multimodal diagnosis (integrates image, text, genetics), administrative automation (frees 2-3 hours daily per doctor), and molecule simulation (AlphaFold 3 accelerates drug development).

Real examples:

  • Oslo University Hospital: Automatic clinical summaries, 2-3 hours saved daily, 34% fewer administrative errors
  • Bavaria hospital network (Germany): Tumor detection in mammograms with 94.2% sensitivity vs 87.5% humans
  • NHS (UK): 12 pilot hospitals, 28% reduction in documentation time

Adoption by region: Europe 45% annual, North America 38%, Asia-Pacific 52%

AI transforming healthcare workflows

Broader context: Transformation will be visible in daily workflows, eliminating administrative friction and allowing professionals to focus on human care.

Practical implications:

  • For hospitals and clinics: Prioritize automation of repetitive administrative tasks. ROI is faster and more visible than in complex diagnostics.
  • For healthcare professionals: Familiarize yourself with AI tools for clinical documentation. These skills will be standard in the coming years.
  • For technology providers: Focus on integration with existing systems rather than standalone solutions. Interoperability is key.

Sources:


5️⃣ Smart manufacturing: the factory is no longer physical, it’s computational

📌 Key points:

  • Average ROI of digital twin implementations: 18-25% in efficiency
  • Process optimization time reduction: from weeks to days
  • Average productivity increase: 15-22% in the first 12 months

What happened:

Companies in optics, batteries, food, and pharmaceuticals automate quality, packaging, and assembly lines. The factory becomes an ecosystem of data, simulations, sensors, and models.

“Smart manufacturing isn’t just automation: it’s simulation, prediction, and continuous optimization. Factories become laboratories where AI experiments before humans invest in physical changes.”

Key components: Complete digital twins (full virtual simulation), sensors + computer vision (detects invisible defects), and real predictive maintenance (predicts failures days in advance).

Examples with metrics:

  • Electronics factory (Germany): Optimization 6 weeks → 3 days, +18% efficiency, 22% ROI first year
  • Pharmaceutical plant (Switzerland): Failure prediction 5-7 days ahead, -31% downtime, $2.3M annual savings
  • Battery factory (South Korea): -42% defects, +15% productivity in 8 months

Average ROI: 18-25% efficiency, 12-18 months payback, +15-22% productivity, -20-35% maintenance costs

Smart manufacturing with digital twins

Broader context: Factories become laboratories where AI experiments before physical investments.

Practical implications:

  • For manufacturers: Start with a digital twin of a critical production line. ROI is faster when focusing on high-value processes.
  • For engineers: Learn simulation and digital modeling tools. These skills will be essential in smart manufacturing.
  • For executives: Invest in sensors and data collection systems first. Without data, there’s no effective digital twin.

Sources:

🧭 Conclusion — What these five signals really show

These five trends aren’t isolated.

They’re symptoms of a structural change:

  1. AI is entering the physical world. From masks and factories to hospitals.

  2. Intelligence is no longer in apps, but in infrastructure.

  3. Open models are pushing toward a more distributed future.

  4. Traditional industries no longer compete with machines, but with algorithms.

  5. The next wave won’t be “more AI,” but “AI that integrates, hides, and coordinates.”

And the real question is no longer “what can AI do?”,

but:

What part of your business, city, or industry is already being transformed without you realizing it?

And how you prepare for next week, not for 2030.

The future of technology won’t just be faster, but smarter, more autonomous, and more integrated. And those who lead that convergence will have the advantage.

“Digital twins and generative AI: A powerful pairing”
McKinsey (2024).
👉 https://www.mckinsey.com/capabilities/tech-and-ai/our-insights/tech-forward/digital-twins-and-generative-ai-a-powerful-pairing
McKinsey & Company

“A Survey on AI-Driven Digital Twins in Industry 4.0”
Huang Z. et al. (2021). PMC.
👉 https://pmc.ncbi.nlm.nih.gov/articles/PMC8512418/
PMC

“Artificial Intelligence in Digital Twins — A Systematic Literature Review”
Kreuzer et al. (2024).
👉 https://ssrn.com/abstract=4681915
SSRN

I invite you to comment: Which of these trends do you think will have the greatest impact on your sector? How is your organization preparing for this transformation where AI integrates into the physical world?

See you next week with more news.


✍️ Claudio from ViaMind

Dare to imagine, create, and transform.


Comments
Comments are shared between English and Spanish versions.

Subscribe

Get a monthly email with the best posts on innovation, technology, and the future. No spam.

* required

Intuit Mailchimp