Writing Code with AI - Good Idea?

In recent years, there has been a surge in the development of chatbots powered by artificial intelligence (AI). ...

In recent years, there has been a surge in the development of chatbots powered by artificial intelligence (AI). Chatbots are used for various purposes, including customer service, marketing, and even code writing. ChatGPT is one such AI-powered chatbot that is used for code writing. However, as with any technology, there are potential security concerns that must be addressed.

The Seemingly Obvious

As stated in a recent article on Security Boulevard, using ChatGPT for client-side code writing can pose significant risks. A rather obvious and up-front concern would be that the chatbot may introduce vulnerabilities or malicious code into the codebase. This is because ChatGPT is not designed to check for security vulnerabilities or identify malicious code. As a result, it is possible for an attacker to exploit these vulnerabilities and gain access to sensitive information. In addition to the risks associated with using ChatGPT for code writing, there are also concerns regarding the security of the chatbot itself. A recent blog post by GitGuardian and even in the ModernCyber Humans vs Bots Blog highlights the fact that ChatGPT has been used in phishing attacks. Attackers have been known to use the chatbot to create convincing fake login pages and steal credentials from unsuspecting victims. This is a major concern as it exposes the potential for hackers to use the chatbot to infiltrate sensitive systems and gain unauthorized access to valuable data. With the rise of cyber threats, it is imperative that organizations take a proactive approach to security when leveraging ChatGPT for code writing.

Regulatory Concerns

Regulations Puzzle Block

Another concern with using ChatGPT for code writing is potential regulatory issues. Depending on the industry and jurisdiction, there may be regulations in place that require certain security measures to be taken when developing and deploying software. If ChatGPT is used for code writing in these contexts, it is important to ensure that the generated code meets these requirements and that the chatbot platform itself is compliant with relevant regulations and standards. Failure to do so could result in legal and financial consequences for the organization.

Privacy Concerns

Privacy Meme

The next major item for consideration is the privacy concerns with using ChatGPT for code development. There is a potential for sensitive information to be leaked. ChatGPT requires access to the codebase in order to generate code, which means that it has access to sensitive information such as API keys, passwords, and other credentials. If this information is leaked, it could result in significant damage to the organization and its reputation. Additionally, there is a lack of control over the data that is processed by the chatbot. ChatGPT is powered by machine learning algorithms that require large amounts of data to function effectively. This means that the chatbot is processing and potentially storing large amounts of data, some of which may be sensitive. Organizations need to ensure that they have appropriate controls in place to protect this data and prevent unauthorized access or misuse. It is also important to consider the potential for bias in the code generated by ChatGPT. Machine learning algorithms are only as good as the data they are trained on, and if the data is biased or incomplete, this can lead to biased or incomplete code. This can have serious implications for organizations that rely on this code to make important decisions.

To mitigate these privacy concerns, it is important to carefully consider the privacy implications of using ChatGPT for code development. Organizations should ensure that they have appropriate controls in place to protect sensitive information and prevent unauthorized access or misuse. Additionally, they should carefully review the code generated by ChatGPT to ensure that it is free from bias and meets any relevant privacy regulations or standards.


Organizations need to be aware of the potential threats posed by ChatGPT and the security concerns around it. The lack of security measures in place in the chatbot itself makes it vulnerable to exploitation by cyber criminals, making it a threat to the security of organizations that use it.

In conclusion, while ChatGPT can be a useful tool for code writing, it is important to be aware of the potential security concerns associated with its use rather than not using it at all. By taking a proactive approach to security and using secure chatbot platforms, it is possible to mitigate these risks and continue to benefit from the advantages that AI-powered chatbots can provide. With the right security measures in place, organizations can safely leverage ChatGPT for code writing while protecting their sensitive data and systems from cyber threats.

ModernCyber’s Services

If you are looking for help with Cisco ISE or other consultative or implementation services, we provide multiple service options that can be customized to the outcomes and requirements you need to meet.

Schedule some time to speak with one of our cybersecurity experts.

Similar posts