The Ethics of Generative AI: Understanding the Principles and Risks
Senior Staff Writer, Neo4j
13 min read
opens in new tabGenerative artificial intelligence (GenAI) has taken the business world by storm. According to opens in new tabMcKinsey, 65% of organizations are regularly using it. Its advancements are improving use cases in opens in new tabfraud detection, opens in new tabdrug discovery, product development, opens in new tabsupply chain improvement, customer support, engineering, and more. With its ability to analyze large quantities of data, inform strategic decision-making, streamline processes, and enable innovation, GenAI is transformational for many companies.
Executives, data scientists, and developers all have high expectations for generative AI technology, but a lack of ethical and responsible AI practices can interfere with implementation. As industry practitioners recognize GenAI’s potential to enhance mission-critical systems, they also face the challenge of addressing and managing ethical implications and associated risks.
In this blog post, we look at the current state of generative AI ethics, including the latest guidelines and legislation, along with some concrete ways to incorporate ethical best practices.
- The GenAI Achilles’ Heel
- Generative AI: The Wild West
- Ethically Questionable Opportunism
- Addressing Ethical Concerns
- Principles of Ethical AI
- Help from NIST
- Applying the Principles
- Add Responsible AI to Your Tech Stack
The GenAI Achilles’ Heel
Chief among ethical considerations is GenAI’s tendency to opens in new tabhallucinate, providing an inaccurate or nonsensical answer derived from its large language model (LLM) content.
An example of ChatGPT’s hallucination.
Why does GenAI hallucinate? LLMs are trained to prioritize human-like answers, even if they aren’t correct. Hallucinations may also result from poor data quality, model limitations, and sole reliance on pattern recognition, which interferes with full understanding.
Hallucination is an obvious ethical problem for AI because it means generated information isn’t guaranteed to be correct. When a generative AI model hallucinates, and someone makes a real-world decision based on the information, the consequences can be far-reaching. For instance, if an AI-powered chatbot provides an erroneous answer to a consumer’s question, the incident could cause the consumer to accidentally spread misinformation and also opens in new tabdamage the company’s reputation. Researchers, data scientists, and developers are, of course, avidly looking for ways to reduce hallucination, but so far, even when using generative AI with high-quality LLM training data, this phenomenon can be challenging to eliminate.
With GenAI, businesses are also concerned about::
- The fact that answers may lack domain-specific knowledge, or “context”
- An inability to trace, verify, and explain how algorithmic responses are derived
- The possibility of biased output that could perpetuate discrimination
- The risk of leaking sensitive data
- Keeping up with changing legal ramifications of using the technology
While GenAI is still considered a breakthrough technology that holds immense promise across industries, some company leaders feel that they can’t yet rely on it, let alone incorporate ethical best practices.
Generative AI: The Wild West
Ushering in a bold new tech era, GenAI opens in new tabdraws parallels to the American Wild West. Its pioneers, ranging from startups to established enterprises, are pushing forward with innovative applications, staking claims, and developing swaths of territory in a relatively unchecked environment.
In this land of opportunity, many corporate trailblazers want to do the right thing, such as building in checks and balances when designing GenAI algorithms. CIOs and other stakeholders making assessments about GenAI are keenly aware of the need to mitigate any ethical risks. Company leaders believe they have a duty to guide development and application in ways that incorporate protections and engender public trust, while still fostering innovation and success. AI-focused companies such as opens in new tabGoogle and opens in new tabMicrosoft have drafted best practices for using AI responsibly, and they’re actively proposing and testing prospective solutions.
This abundance of good-faith responses represents a strong start toward creating ethical GenAI frameworks that can help companies succeed while also benefiting society as a whole, but there’s more to be done.
Ethically Questionable Opportunism
Early applications of GenAI have brought attention to some worrisome activity:
- AI creation of “original” writing, music, and artwork. GenAI applications have copied proprietary material found in LLM data sets without providing attributions or obtaining permissions, resulting in copyright opens in new tabinfringements of opens in new tabauthors’, opens in new tabmusicians’, and opens in new tabartists’ work. Intellectual property opens in new tabchallenges are progressing.
- Creation of deepfakes, doctored videos and audio clips that pass as authentic. opens in new tabDeepfakes are easy to make and opens in new tabtough to detect. They opens in new tabfacilitate identity theft, and, when released on social media, can do things like opens in new tabinfluence elections. “The ease of creating AI-powered tools might lead to large-scale automated attacks,” warns venture capitalist opens in new tabAlon Hillel-Tuch. Some good news: California is opens in new tabbanning opens in new tabdeepfakes, and OpenAI has built software to help opens in new tabdetect those created by DALL-E.
- Empowerment of opens in new tabscammers through creation of deepfakes and availability of various applications.
These “Wild West” manifestations of GenAI prowess bolster the case for formalizing ethical approaches and establishing regulations as needed.
Addressing Ethical Concerns
At the societal-well-being level, as well as the day-to-day corporate level, can we put guardrails on certain aspects of GenAI and resolve the ethical challenges? World leaders are focused on creating and releasing universal legal frameworks. Some ethics-oriented proposals to date are opens in new tabcontroversial, but there’s growing momentum to adopt rules.
Most notably, the European Union recently passed the opens in new tabArtificial Intelligence Act, the first opens in new tabcomprehensive global regulatory framework for AI. This legislation distinguishes AI from traditional software, defining it as “a machine-based system that is designed to operate with varying levels of autonomy and that may exhibit adaptiveness after deployment, and that, for explicit or implicit objectives, infers, from the input it receives, how to generate outputs such as predictions, content, recommendations, or decisions that can influence physical or virtual environments.”
In addition to Europe, this sweeping law applies to United States companies that have global reach, spanning AI development, use, importation, and distribution in the European market, as well as to American companies that produce output for use in the EU market.
Targeting High-Risk AI Systems
According to opens in new tabCPO, The AI Act is focused on “high-risk AI systems” related to either regulated products or associated safety components, including vehicles, medical devices, aircraft, toys, and machinery.
Applicable use cases include:
- Biometric remote identification, biometric categorization, or emotion recognition
- Critical infrastructure
- Education and vocational training
- Employment, workers management, and access to self-employment
- Essential private services, and essential public services and benefits
- Law enforcement
- Migration, asylum, and border control management
- Administration of justice and democratic processes (e.g., election)
In addition to the AI Act, American companies must comply with broad-level existing regulations that could have implications for using AI, such as opens in new tabequal employment opportunity laws. For example, the EEOC is opens in new tabscrutinizing generative AI tools that could create discrimination in the opens in new tabhiring process, such as résumé scanners that elevate particular keywords and applicant rating systems that assess factors such as personality, aptitude, cognitive abilities, and expected cultural fit.
How can a company that uses GenAI technology for mission-critical projects move forward in an ethically acceptable way? Here are some guiding ethical principles, followed by practical steps executives, data scientists, and developers can take to get on track.
Principles of Ethical AI
A first step for applying ethics to GenAI has been to illuminate human-centered principles of good behavior as they may apply to GenAI. Under that lens, ethics-oriented guidelines include:
Do No Harm
Consumers are reminded by spiritual leaders, psychologists, and businesses to opens in new tabdo no harm, be kind and good, go in peace, and live and let live. Company executives hear similar directives for steering their organizations.
When it comes to using GenAI, how can a company strive to be a model corporate citizen? For starters, teams working with GenAI technology can:
- Monitor AI output and avoid releasing misinformation
- Get feedback to assess whether output could be detrimental
- When opens in new tabmaking decisions that impact people, avoid physical, mental, or emotional harm
- Learn about GraphRAG, a technique that enhances the accuracy and reliability of GenAI models
- Seek ways to build credibility and sustainability of GenAI technology
Be Fair
Fairness is treating everyone equally regardless of race, gender, socioeconomic status, and other factors.
Being fair sounds fairly straightforward, but GenAI processes may inadvertently introduce or perpetuate data bias. For example:
- If a car safety test is conducted with only opens in new tabmen drivers — typically larger-framed people — then, regardless of any fairness instructions given to the AI, the resulting data will be biased against women and other small-framed individuals.
- When programming self-driving cars, it’s opens in new tabnot easy to make morally perfect decisions.
- An AI model trained on material generated by biased AI can opens in new tabexaggerate stereotypes.
In opens in new tabLLM design, eliminating data bias that favors or disfavors certain groups of people is often easier said than done. A developer can instruct AI to be unbiased, and then follow up to root out any bias that snuck through, but that may not be enough. With the best of intentions, say, by favoring one category of sensitive personal data, a company may still discriminate, as evidenced by opens in new tabGoogle Gemini. “For SMEs relying on ChatGPT and similar language models for customer interactions, there’s a risk of unintentional bias affecting their brand,” notes industry expert opens in new tabBrad Drysdale.
Companies can promote GenAI fairness by:
- Monitoring and testing AI outcomes for group disparities. The National Institute of Standards and Technology (opens in new tabNIST) supports a “socio-technical” approach to mitigating bias in AI, recognizing the importance of operating in a larger social context and that “purely technically based efforts to solve the problem of bias will come up short.”
- Using diverse, relevant training data to populate an LLM.
- Including a variety of team input during model development.
- Embedding data context in opens in new tabgraphs, which can ensure more fair treatment. For example, opens in new tabQualicorp, a Brazil-based healthcare benefits administrator, has created a graph-based AI application that helps prevent information omission for insurance applicants. In this ethical use case, context provided in a graph database makes it easier to treat people as individuals instead of as a single, homogenous group.
Ensure Data Privacy
With awareness of widespread corporate data breaches, consumers have become advocates for stronger cybersecurity. GenAI models are in the spotlight because they often opens in new tabcollect personal information. And consumers’ sensitive data isn’t the only type at risk: 15% of employees opens in new tabput company data in ChatGPT, where it becomes public.
Most company leaders understand that they must do more to reassure their customers that data is used for legitimate purposes. In response to the need for stronger data security, according to opens in new tabCisco, 63% of organizations have placed limits on which data can be entered digitally, while 61% are limiting the GenAI tools employees are allowed to use.
Executives and developers can opens in new tabprotect sensitive data by:
- Setting up opens in new tabstrong enterprise defenses
- Using robust encryption for data storage
- Using only zero- or first-party data for GenAI tasks
- Denying LLMs access to sensitive information
- Processing only necessary data (a opens in new tabGDPR principle)
- Anonymizing user data
- opens in new tabFine-tuning models for particular tasks
Honor Human Autonomy
Applied to AI, what does this lofty directive mean?
In certain contexts, notes opens in new tabFrontiers in Artificial Intelligence, AI technologies threaten human autonomy by “over-optimizing the workflow, hyper-personalization, or by not giving users sufficient choice, control, or decision-making opportunities.” Respecting autonomy alludes to maintaining the natural order of what humans do, such as making choices. For example, with critical decision-making about opens in new tabhealthcare, the buck should stop with medical professionals.
When using GenAI, companies can strive to respect human autonomy by:
- Being ethically minded when nurturing talent for developing and using GenAI
Be Accurate
opens in new tabInaccurate information from LLMs that finds its way into the world is a pressing concern. Inaccurate “facts” end up in everything from search results detailing opens in new tabthe history of Mars to opens in new tablegal briefs.
Aside from fact-checking, how can a company ensure the accuracy of its LLM-sourced information? To help catch GenAI opens in new tabhallucinations and improve your LLM responses, you can use a technique called opens in new tabretrieval augmented generation (RAG). Informed by opens in new tabvector search, RAG gathers relevant information from external data sources to provide your GenAI with an authoritative knowledge base. By adding a final key ingredient, opens in new tabknowledge graphs, which “codify” an LLM’s facts, you can provide more context and structure than you could with metadata alone. RAG techniques that can reduce hallucinations are also continuing to evolve. One application is opens in new tabGraphRAG.
“By using a combination of vector search, RAG, and knowledge graph interfaces, we can synthesize the human, rich, contextual understanding of a concept with the more foundational ‘understanding’ a computer (LLM) can achieve,” explains opens in new tabJim Webber, chief scientist at opens in new tabNeo4j, the preferred data store for RAG over vector databases and other alternatives.
Be Transparent
When a consumer applies for a loan from a bank and is turned down as the result of a GenAI process, does management understand how the decision was made? If a job applicant is rejected based on their resume, is the reason easily identifiable? GenAI models can be “black boxes” in which the algorithmic processing is hidden or obscured.
Being unable to pinpoint what’s behind algorithmic decisions can lead to public distrust of AI, which is why ethical AI proponents are pushing for companies to opens in new tabprovide transparency. To identify what data was used in GenAI decision processing, a system can provide opens in new tabcontext — the peripheral information — which facilitates understanding of the pathways of logic processing. Explicitly incorporating context ensures that the technology doesn’t violate ethical principles. One way for a company to enhance transparency is to incorporate context by using a graph database such as opens in new tabNeo4j.
Be Explainable
With a bank denying a loan based on an algorithm and no way to trace the decision to an origination point, what does the loan officer tell the applicant? Saying “We have no idea why our AI did that, but we’re standing by it” isn’t going to go over well. In addition to damaging trust, not being able to explain how the decision is justified can hinder further application adoption.
In light of this dilemma, the subfield of opens in new tabexplainable AI (XAI), the ability to verify, trace, and explain how responses are derived, has emerged.
opens in new tabExplainability has four components:
- Being able to cite sources and provide links in a response to a user prompt
- Understanding the reasoning for using certain information
- Understanding patterns in the “grounding” source data
- Explaining the retrieval logic: how the system selected its source information
How can companies adapt their GenAI processes for this aspect of ethical operation? To build in explainability, tech leaders must identify training provenance. “You need a way to show the decision-owner information about individual inputs in the most detail possible,” says Philip Rathle, chief technology officer at Neo4j.
To that end, a company can use opens in new tabknowledge graphs and metadata tagging to allow for backward tracing to show how generated content was created. By storing connections between data points, linking data directly to sources, and including traceable evidence, opens in new tabknowledge graphs facilitate LLM data governance. For instance, if a company board member were to ask a GenAI chatbot for a summary of an HR policy for a specific geographic region, a model based on a knowledge graph could provide not just a response but the source content consulted.
Be Accountable
With ethical AI, the thinking goes, humans are typically held accountable for their decisions, so logically, generative AI systems should be, too. If there’s an error in an algorithm’s decision process, it’s imperative for the company to step up and be accountable. Acknowledging the issue, determining whether a change must be made, and, if so, making the change helps protect a company’s reputation. Companies can also stand behind their GenAI activity by opens in new tabbuilding in accountability.
Help from NIST
When pairing proprietary data with GenAI, applying these principles may seem relatively straightforward. However, because AI ethics guidelines and regulations are dynamically evolving, executives and developers who want to ensure ethical operation aren’t always sure of the best or most acceptable course of action.
To help with this challenge, NIST has drafted formal AI ethical guidelines. The organization first requested input from stakeholders across the country to create a plan for development of technical standards and related tools to support reliable, trustworthy AI systems.
Among the key contributors, graph database company opens in new tabNeo4j called for recognition of ethical principles as the foundation of AI risk management. Neo4j suggested that for AI to be situationally appropriate and “learn” in a way that leverages adjacency to understand and refine outputs, it must be underpinned by context. The company suggested that these principles be embedded in AI at the design stage, as well as defined in operational terms according to use case and implementation context. For example, fairness might be defined as group fairness (equal representation of groups) or procedural fairness (treating every individual the same way).
NIST followed its input stage by drafting an opens in new tabAI risk-management framework promoting fairness, transparency, and accountability. If you want to design and deploy ethically oriented AI systems, this document is a great place to start.
Applying the Principles
How can executives, data scientists, and developers start opens in new tabapplying ethical principles and building generative AI tools on a firm opens in new tabfoundation that can expediently adapt to a changing regulatory environment?
For starters, they can:
- Ask whether the use of generative AI truly makes sense for the application: Is it the most appropriate tool?
- When selecting a technology stack, keep your strategic requirements in focus
- Apply ethics to GenAI at the design stage to incorporate transparency
- Establish a risk-based system to review output for bias
- Set up a process to handle any discovered bias
- Teach staff to verify GenAI outputs and report suspect results
- Fact-check to make sure AI-generated information is accurate
- Ensure that private data stays safe
- Use graph databases, which incorporate data context to facilitate ethical outcomes
- Research new techniques that reduce hallucinations and factually ground data
Add Responsible AI to Your Tech Stack
Looking to build ethically oriented opens in new tabgenerative AI applications but wary of wading into an ethical AI sinkhole?
Neo4j opens in new tabgraph database technology is a proven, reliable way forward. Our opens in new tabknowledge graphs enable GenAI to understand complex relationships, enhancing the quality of your output. With our opens in new tabgraph retrieval augmented generation (GraphRAG) technology, you can confidently put AI ethical principles into practice.
With Neo4j, you can:
- Make your machine learning model outcomes more trustworthy and robust by giving your data meaningful opens in new tabcontext
- Provide a comprehensive, accurate view of data created by the connections in a opens in new tabknowledge graph
- Improve your LLM responses by using opens in new tabRAG (retrieval augmented generation) to retrieve relevant source information from external data stores
- Explain your retrieval logic: Trace your sources and understand the connections between knowledge and responses
- Improve data quality, transparency, and explainability as a result of generating more relevant information
- Integrate domain knowledge: Incorporate and connect your organizational data and facts for accurate, tailored responses
- Accelerate your GenAI development: build and deploy quickly with frameworks and flexible architecture
Get the details on how you can build responsible GenAI applications today!