Generative AI has quickly changed what the world thought was possible with artificial intelligence, and its mainstream adoption may seem shocking to many who don\u2019t work in tech. It inspires awe and unease \u2014 and often both at the same time.\n\nSo, what are its implications for the enterprise and cybersecurity?\n\nA technology inflection point\n\nGenerative AI operates on neural networks powered by deep learning systems, just like the brain works. These systems are like the processes of human learning. But unlike human learning, the power of crowd-source data combined with the right information in Generative AI means that processing answers will be light years faster. What might take 30 years for an individual to process could take just an eyeblink. That is a benefit that can be derived depending on the quality as well as massive amounts of data that can be fed into it.\n\nIt is a scientific and engineering game-changer for the enterprise. A technology that can greatly improve the efficiency of organizations \u2013 allowing them to be significantly more productive with the same number of human resources. But the shock of how fast Generative AI applications such as ChatGPT, Bard, and GitHub Pilot emerged seemingly overnight has understandably taken enterprise IT leaders by surprise. So fast that in just six months, the popularization of Generative AI tools is already reaching a technology inflection point.\n\nThe cybersecurity challenges\n\nGenerative AI, including ChatGPT, is primarily delivered through a software as a service (SaaS) model by third parties. One of the challenges this poses is that interacting with Generative AI requires providing data to this third party. Large learning models (LLMs) that back these AI tools require storage of that data to intelligently respond to subsequent prompts.\n\nThe use of AI presents significant issues around sensitive data loss, and compliance. Providing sensitive information to Generative AI programs such as personally identifiable data (PII), protected health information (PHI), or intellectual property (IP) needs to be viewed in the same lens as other data processor and data controller relationships. As such, proper controls must be in place.\n\nInformation fed into AI tools like ChatGPT becomes part of its pool of knowledge. Any subscriber to ChatGPT has access to that common dataset. This means any data uploaded or asked about can then be replayed back within certain app guardrails to other third parties who ask similar questions. It\u2019s worth noting that this is very similar to software-as-a-service (SaaS) application problems as it can impact the response of future queries when used as a training set. As it stands today, most Generative AI tools do not have concrete data security policies for user-provided data.\n\nThe insider threat also becomes significant with AI. Insiders with intimate knowledge of their enterprise can use ChatGPT to create very realistic email. They can duplicate another\u2019s style, typos, everything. Moreover, attackers can also duplicate websites exactly.\n\nWhat enterprises need for security\n\nFortunately, there are Generative AI Protection solutions, such as Symantec DLP Cloud, Adaptive Protection on Symantec Endpoint Security Complete (SESC), and real time link in email security that address these emerging challenges and block attacks in different, targeted ways.\n\nSymantec DLP Cloud extends Generative AI Protection for enterprises, with the capabilities they need to discover, and subsequently monitor and control, interaction with generative AI tools within their organizations. Among other benefits, DLP can use AI to speed incident prioritization, helping senior analysts to triage the most significant and recognize those that are not a critical threat to the enterprise. \n\nThe benefits include:\n\nThe bottom line: Symantec Generative AI Protection allows enterprises to \u201csay yes\u201d to generative AI\u2019s productivity enhancing innovations without compromising data security and compliance.\n\nLearn more about the implications of Generative AI to the enterprise here. \n\nAbout Alex Au Yeung\n\nAlex Au Yeung is the Chief Product Officer of the Symantec Enterprise Division at Broadcom. A 25+ year software veteran, Alex is responsible for product strategy, product management and marketing for all of Symantec.