Microsoft Ignite 2023 was an eventful one, with many announcements across Microsoft’s AI Copilot capabilities. The biggest announcement, in our opinion, is that of Microsoft Copilot Studio, a low-code tool that allows professional and citizen developers to build standalone AI Copilots, as well as customize Microsoft Copilot for Microsoft 365. These exciting technological advancements allow business users (of all technical backgrounds, mind you) to customize Copilot for Microsoft 365, publish standalone copilots and custom GPTs, and be left to their own devices to manage and secure these customizations and copilots.
While this may seem like the first step of a new Terminator movie with “Skynet” coming for us all, it really is an exciting breakthrough that Microsoft has delivered for its customers. Generative AI Copilots enable more business users to be more productive than ever before and do a lot in the way of bringing people closer to technology. HOWEVER. There are also massive security ramifications that must be accounted for, which is why we are so pleased to announce Zenity as the first security vendor to protect Microsoft Copilot Studio, as well as introduce new capabilities within the Zenity platform that enable security teams to maintain visibility and control over how Generative AI is used throughout the organization.
For starters, Copilot Studio makes it possible for any user to build new AI agents and bots from scratch; adding rocket fuel to both the citizen development movement as well as adoption of Gen AI capabilities. This is also increasingly favored by professional developers as they can now build smarter and more integrated bots with the power to build their own AI Copilots. What this means in practice is that any user can now build, test, and customize their own copilots and GPTs, rather than relying on the centralized GPT platform itself; with no developer knowledge required. Every user gets complete control over their standalone copilots and can do whatever they please.
Anyone can now connect Copilot to enterprise data sources, including pre-built or custom plugins, to tap into any system (think: M365, Azure, SaaS apps, databases, SQL servers, etc.), and build customized solutions that can generate real-time answers. Rather than just creating a chatbot that has pre-canned answers that eventually leave you with “for more information, reach out to the Help Desk,” business users can create adaptive apps and bots that provide answers in real-time based on real and evolving data.
There are massive risks with this type of development democratization, which can be boiled down to three main groups; data leakage, over-sharing, and a lack of visibility. To illustrate these risks, let’s walk through an example of how a business user might use Copilot Studio.
Say you’re in accounting and you want to build a bot that can help you stay on top of expense reports that employees are filing. Now, with Copilot Studio, you can build a bot in Power Agents and insert an AI Copilot to be able to accurately give end users up-to-date information when prompted. You, and members of your team can simply open up the bot and ask it “Tell me how much money we have left to spend on office equipment this quarter,” and it will query a database, look through reports, and more to give you an accurate number.
However, in order for the bot to be able to access this information in real-time, it must have access to a SQL database, therefore, it needs your SQL credential to authorize itself. In setting up this bot, the builder has two choices to establish a connection. First, they can configure the connection to force each individual user to bring their own credential to authenticate, or, more frequently, the bot will be built so that the bot builder/owner brings their credentials, and then those are used and shared with every other user of the bot.
The reason the latter is so much more frequent is for a couple reasons. For one, some cases, like SQL servers and Azure connections, will do this by default and citizen developers (and even professional developers) will not know, or be technically capable enough to change it, but also because citizen developers may not know the difference and just take the path of least resistance by entering in their own credentials.
Unfortunately, this results in data leaking, implicitly and overly sharing applications, as well as an overall lack of visibility for security teams.
The volume and speed of development continues to accelerate at breakneck speeds, and staying on top of each individual app and connection that is made is harder than ever before. At Zenity, we are proud to announce the first solution that helps organizations gain visibility and control over custom-built GPTs and bots that are used within the enterprise. By extending our security and governance platform to that of Copilot Studio we empower our customers to:
The recent closure of our Series A round was done with one eye on how to secure Generative AI, as this is a natural extension and added capability for both professional and citizen developers. We are excited to see how this creates a ripple effect, with more and more software vendors likely following suit to put more Generative AI capabilities into the hands of business users; which requires stringent security that does not block productivity. If you’d like to see this in action, or are experiencing challenges already, reach out to us to see a demo or schedule a call directly with one of our representatives!
All ArticlesIntroducing Zenity for ChatGPT Enterprise More than 80% of Fortune 500 companies rely on ChatGPT Enterprise, not...
In the realm of modern enterprise productivity suites, Copilot for Microsoft 365 stands as a huge driver for efficiency,...
Businesses of all shapes and sizes are leveraging Microsoft Power BI to find insights within their own data. This...
We’d love to chat with you about how your team can secure and govern AI Agents everywhere.
Book Demo