Samuel Thompson
-
Enhancing Enterprise Governance for Emerging Edge AI Workloads
The rise of powerful, locally executable AI models like Google’s Gemma 4 challenges traditional enterprise security. These models can run on edge devices, bypassing traditional perimeter defenses and making network traffic monitoring ineffective for offline AI processing. This creates blind spots for security, especially in regulated industries requiring auditability. CISOs must shift focus from blocking models to controlling intent and system access, as identity platforms become the new firewall. Enterprises need new endpoint detection tools to monitor local AI inference and adapt security policies to acknowledge that compute execution is no longer solely cloud-based.
-
Companies Enhance AI Adoption with Controlled Integration
Businesses are adopting a cautious approach to AI, favoring tools that augment human decision-making over full automation. Sectors with high financial or legal risks prioritize AI that is manageable, verifiable, and trustworthy. S&P Global Market Intelligence exemplifies this by integrating AI into its Capital IQ Pro platform to assist financial analysts, ensuring AI outputs are tied to verified sources. While AI adoption is widespread, many organizations use it for tasks like summarization, not independent action. Trust in AI systems relies on their ability to explain reasoning, cite sources, and operate within defined parameters.
-
Robust AI Governance: Safeguarding Enterprise Margins
To protect margins and foster innovation, businesses must prioritize robust AI governance, moving from opaque proprietary systems to open infrastructure. As AI becomes foundational, closed development becomes untenable due to complexity, security risks, and integration challenges. Open-source AI enhances operational resilience through broad scrutiny and collaborative improvement. Embracing transparency and open foundations is crucial for enterprise AI’s future, enabling adaptability, innovation, and public legitimacy.
-
Why Apple and Others Are Building AI Agents with Limits
Next-generation AI assistants, from Apple and Qualcomm, will offer advanced capabilities for task management and app navigation. However, development prioritizes user control and security, incorporating explicit approval checkpoints for sensitive actions like financial transactions. This “human-in-the-loop” model limits AI autonomy, ensuring users retain final decision-making authority and data privacy through granular access controls and integration with secure partner services. This controlled approach aims to mitigate risks and foster trust in agentic AI.
-
Meta’s AI Ambition: Open Source Ethics vs. Competitive Edge
Meta’s Muse Spark marks a significant shift from its open-source Llama initiative. This proprietary, multimodal AI model, developed after a $14.3 billion investment, surpasses Llama 4 in benchmarks, particularly in healthcare. Unlike its predecessors, Muse Spark’s advanced capabilities are not freely available. This move, while driving Meta’s stock up and targeting billions of users directly, has generated skepticism within the developer community awaiting future open-source releases.
-
Navigating Agentic AI Governance: The EU AI Act’s Impact in 2026
To mitigate risks in AI agentic systems, organizations must prioritize robust identity protocols, auditable logs, policy enforcement, human oversight, rapid revocation, and thorough documentation. Employing tools like SDKs that cryptographically sign actions, similar to blockchain, creates an immutable audit trail. A centralized, verbose system of record for all agentic AI activities is crucial for governance, surpassing scattered text logs. Maintaining an “agentic asset list” with unique identifiers, capabilities, and permissions is essential, aligning with regulations like the EU AI Act’s mandates for ongoing, evidence-based risk management and interpretable AI systems.
-
Anthropic’s New AI Model Kept Private After Discovering Thousands of External Vulnerabilities
Anthropic has kept its advanced AI model private due to discovering thousands of external vulnerabilities during testing. This decision emphasizes responsible AI development, prioritizing security over rapid release. The discovery highlights the complexity and potential risks of next-generation AI, underscoring the need for rigorous, continuous security measures to ensure AI trustworthiness and safety.
-
The AI Cybersecurity Crisis Driving Anthropic’s Project Glasswing
Anthropic’s Claude Mythos Preview, an advanced AI, excels at identifying cybersecurity vulnerabilities. Instead of public release, Anthropic uses “Project Glasswing” for controlled distribution to industry leaders and critical organizations. The initiative includes significant funding for open-source security. This strategy aims to bolster global digital defenses while managing the AI’s dual-use potential, acknowledging the severe risks of widespread, uncontrolled access to such powerful capabilities.
-
AI Workflows for Software Developers: The Imperative of Oversight
Enterprises are increasingly trusting autonomous AI agents, with 73% expressing high or moderate confidence, up from the previous year. Reliance on AI-generated code has also surged to 67%. However, robust governance lags, with only 36% of organizations having a centralized strategy. Technical hurdles in implementing human-in-the-loop oversight and concerns about “AI sprawl” (94% of leaders worried) pose challenges, potentially outpacing accountability mechanisms. For regulated sectors, auditability and orchestration are critical.
-
Asylon and Thrive Logic Partner for Physical AI in Enterprise Perimeter Security
Thrive Logic and Asylon partner to introduce “Physical AI” for network edge security. This integration combines Asylon’s robotic patrols with Thrive Logic’s AI agent analytics for proactive, autonomous incident detection and response. The goal is to minimize response times, enhance operational resilience, and provide security leaders with reliable, auditable coverage in exterior security zones. This human-AI collaboration shifts security from reactive to strategic oversight.