Demo Request
Take a personalized product tour with a member of our team to see how we can help make your existing security teams and tools more effective within minutes.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
Home
Blog

Navigating the New Frontier of AI Governance: Insights from Digital World Conference Summit

Tal Shapira
Updated
April 18, 2024
November 29, 2024
5 min read
Ready to Close the SaaS Security Gap?
Chat with us

Introduction

As we are at the convergence of unprecedented technology advancements and global challenges, the significance of AI safety governance has never been more important. On April 14, 2024, I had the distinct honor of participating as a panelist in the Generative AI Governance discussion at the recent Digital World Conference (DWC) Summit “Building Trust in the Digital World” in Geneva, Switzerland. This event was organized by the World Digital Technology Academy (WDTA).

Participating in the "Building Trust in the Digital World" event

Insights from the DWC Summit

The DWC Summit was a melting pot of ideas and initiatives supported by WDTA’s commitment to advancing digital technology for all. The panel discussion I participated in brought together experts from the United Nations International Computing Centre, the Norwegian Academy of Engineering, Georgetown’s CSET, OpenAI, and of course, Reco AI.

During the session, we discussed the challenges of governing GenAI. This technology’s rapid evolution has brought significant governance issues that require frameworks to ensure that it contributes positively to society. At Reco, we've observed a sharp increase in the deployment of GenAI applications, which means that organizations require a new approach to SaaS security.

Our discussion extended to the rise in usage of GenAI apps and the subsequent rise in 3rd-party app connections, authentication permissions, and tokens. This trend shows the urgent need for comprehensive AI-driven security measures that not only protect but also anticipate potential threats.

During the panel, I was approached by the moderator, Ken Huang, Chair of the WDTA AI STR Working Group, with the following question:

"Given Reco's recent research on the potential risks of Microsoft Copilot, can you discuss the risks of generative AI usage and how to defend against them?"

On March 14, Microsoft introduced Copilot to their 365 environment. Copilot is designed to enhance productivity through a conversational AI interface that helps users conduct research and create content. However, our research at Reco AI reveals significant risks associated with the broad permissions often granted to such tools. Copilot, like other generative AI tools, has access to all files a user can access, including those shared widely across the organization. This expansive access poses a notable risk of data leakage that can occur by a single prompt.

In our deep-dive analysis, “Are you ready for Microsoft Copilot?”, we discovered that generative AI tools could potentially access and leak sensitive organizational data. For example, in situations where an account is compromised, a malicious actor can utilize Copilot to quickly locate and exploit sensitive information, bypassing the traditional time-consuming process of searching through files manually. Shockingly, these interactions often leave minimal audit trails, unlike regular user access, which is logged and can be audited. Our team managed to retrieve names of restricted files simply by querying Copilot about their authors, highlighting a severe oversight in data access logging and control.


To defend against these risks, we recommend educating employees about risks involved when using GenAI tools, meticulously configuring AI applications, monitoring for configuration drifts, reducing data exposure, and implementing strict access controls. Additionally, deploying advanced AI-driven monitoring and detection systems is critical for identifying and mitigating potential breaches or misuse of AI applications.

Conclusion: How GenAI Tools Can Be Used for AI Governance

As is said in the cartoon show Jackie Chan, “You can only defeat magic with magic." This is the approach needed in AI governance—essentially, the only way to fight AI is with GenAI. Depending on the use case, generative AI tools can be adapted to govern themselves and other digital tools effectively. For instance, when training or fine-tuning foundational models, we can use large language models (LLMs) to scan for personally identifiable information (PII) or payment card information (PCI). This would prevent data privacy breaches before they occur. In production environments, implement AI guardrails that scan user inputs and outputs. These models can block attacks like prompt injection, which manipulates systems to grant unauthorized access, and can also filter out harmful content, maintaining the integrity of interactions.

At Reco, we use AI to detect shadow GenAI apps or any unsanctioned applications utilizing generative AI, managing unauthorized data access risks. Our AI tools can identify abnormal user behaviors, such as unauthorized data access or unusual app-to-app permissions, which could indicate sophisticated attacks like SaaS session hijacking. We also leverage generative AI to automate creation of security policies and compliance rules, helping maintain agile, relevant governance processes that are aligned with the latest cybersecurity best practices. This significantly reduces the administrative burden on security teams. Reco also uses generative AI to map compliance requirements to specific controls, policies, and posture checks within SaaS. By embracing generative AI, we ensure that our governance tools are not only effective but are also continually evolving to meet the challenges posed by new technologies, thereby securing our digital future.

No items found.

Dr. Tal Shapira

ABOUT THE AUTHOR

Tal is the Cofounder & CTO of Reco. Tal has a Ph.D. from the school of Electrical Engineering at Tel Aviv University, where his research focused on deep learning, computer networks, and cybersecurity. Tal is a graduate of the Talpiot Excellence Program, and a former head of a cybersecurity R&D group within the Israeli Prime Minister's Office. In addition to serving as the CTO, Tal is a member of the AI Controls Security Working Group with the Cloud Security Alliance.

Technical Review by:
Gal Nakash
Technical Review by:
Dr. Tal Shapira

Tal is the Cofounder & CTO of Reco. Tal has a Ph.D. from the school of Electrical Engineering at Tel Aviv University, where his research focused on deep learning, computer networks, and cybersecurity. Tal is a graduate of the Talpiot Excellence Program, and a former head of a cybersecurity R&D group within the Israeli Prime Minister's Office. In addition to serving as the CTO, Tal is a member of the AI Controls Security Working Group with the Cloud Security Alliance.

Ready to Close the SaaS Security Gap?
Chat with us
Table of Contents
Get the Latest SaaS Security Insights
Subscribe to receive updates on the latest cyber security attacks and trends in SaaS Security.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

Explore Related Posts

Malicious Extensions That Lock You Out While They Steal Your Session
Dr. Tal Shapira
Five malicious Chrome extensions disguised as enterprise productivity tools stole session tokens from Workday, NetSuite, and SuccessFactors while simultaneously blocking admins from revoking access or resetting credentials. The attack exposed a blind spot in SaaS security: the browser, where stolen session cookies render SSO and MFA irrelevant.
Why the Hidden Cost of AI Sprawl Is Rising in Modern Enterprises
Gal Nakash
AI adoption is accelerating across modern enterprises, but the rapid growth of AI tools and agents often introduces hidden operational and security risks. This article explores the hidden cost of AI sprawl, including duplicate tools, fragmented workflows, and expanding SaaS integrations. It also outlines practical frameworks and best practices that help organizations detect uncontrolled AI adoption and maintain visibility, governance, and security across enterprise environments.
Model Context Protocol (MCP) Is Rewiring SaaS Trust, One Agent Action at a Time
Gal Nakash
The Model Context Protocol (MCP) is an emerging standard that enables AI agents to seamlessly connect with SaaS tools and APIs, allowing them to perform actions like fetching files, updating records, and sending messages autonomously. However, this power introduces significant security risks, including identity drift, weak authentication, data leakage, and invisible access that bypasses traditional monitoring. Organizations can mitigate these risks by enforcing least-privilege OAuth scopes, using short-lived tokens, binding agents to human owners, and adopting platforms that provide continuous visibility into MCP-based trust paths.
See more featured resources

Ready for SaaS Security that can keep up?

Request a demo