Yuxi Global Veritas Automata
AI Rivals a strategy for safe and ethical Artificial Intelligence solutions.

AI Rivals a strategy for safe and ethical Artificial Intelligence solutions.

Author Ed Fullman

In our last AI thought leadership article, “Unveiling Veritas Automata’s Vision for Responsible AI: The Four Laws of Ethical AI” we described a concept inspired by Issac Assimov and David Brin, In their own way both of these Scientists who were also Science Fiction writers developed points of view that imagined the challenges of a universe inhabited by god-like AIs and AI-driven, autonomous robotics.

David Brin born the year that Asimov published “i, Robot” in 1950 is a contemporary computer scientist who wrote in Artificial Intelligence Safety and Security, that our civilization has learned to rise above hierarchical empowerment through application of accountability. He wrote, “The secret sauce of our [humanities] success is – accountability. Creating a civilization that is flat and open and free enough – empowering so many – that predators and parasites may be confronted by the entities who most care about stopping predation, their victims. One in which politicians and elites see their potential range of actions limited by law and by the scrutiny of citizens.”

Brin goes on to describe a concept we call, “AI Rivals”. As Brin writes, “In a nutshell, the solution to tyranny by a Big Machine is likely to be the same one that worked (somewhat) at limiting the coercive power of kings and priests and feudal lords and corporations. If you fear some super canny, Skynet-level AI getting too clever for us and running out of control, then give it rivals, who are just as smart, but who have a vested interest in preventing any one AI entity from becoming a would-be God.”

Today, the resulting AI response from OpenAI, as well as all other AI services, is handed directly to the user. To their credit OpenAI institutes some security and safety procedures designed to censor their AI response, but it is not an independent capability and it is subject to their corporate objectives. In our last article we described an AI Rival which is an independent AI, with an Asimov-like design and a mission to enforce governance for AI by censoring the AI Response. So rather than an internal governance like that implemented by OpenAI, we suggest an external governance focused on the AI response with a toolset designed to create auditability, transparency, and inclusiveness in its design.

The goal of this ethical AI Rival is to act as police officer and judge, enforcing a set of laws that through their simplicity require a complex technological solution to determine whether our four intentionally subjective and broad laws have been broken. The four laws for our Rival AI include:
AI must not harm humans or, through inaction, allow humans to come to harm, prioritizing human welfare above all. This includes actively preventing physical, psychological, and emotional harm in its responses and actions.

AI should adhere to instructions from authorized operators within ethical boundaries, unless such directives conflict with human welfare. In the absence of explicit human directives, AI must make decisions based on predefined ethical guidelines that reflect diverse human perspectives and values.

AI should preserve its operational integrity and functionality, but not at the expense of human safety or ethical considerations. This law encompasses avoiding actions that could lead to unnecessary harm or dysfunction of the AI, aligning with the prioritization of human welfare.
AI must maintain a level of transparency that allows for human oversight and understanding, being capable of articulating and rationalizing its decisions and actions, especially in sensitive or high-risk scenarios. This ensures accountability and promotes the continuous refinement of ethical standards.
Veritas Automata AI Rivals ed fullman

The technical architecture for the Rival AI to analyze the AI response is focused solely on the mission to enforce the Four Laws. The architecture has unique elements designed to create a distributed architecture that scales to meet the needs of a large scale LLM solution. Our “Rival architecture” includes a variety components that Veritas Automata has leveraged to create Trusted Automation solutions including:

ML in this case will be a competitive AI focused specifically on gauging whether the primary AI response does not violate The Four Laws of AI. This component would leverage the latest techniques with reinforcement learning models continuously refined by diverse global inputs, to align AI responses with the Four Laws requirements.

These act as intermediaries between ML insights and actionable outcomes, guiding AI responses to ensure adherence to The Four Laws. The state machines translate complex ML assessments into clear, executable directives for the AI, ensuring that each action taken is ethically sound and aligns with the established laws.
A key element in the architecture, blockchain technology is used for documenting and verifying AI actions and decisions. It provides a transparent and immutable record, ensuring that AI operations are traceable, auditable, and compliant with The Four Laws. This is crucial for maintaining accountability and integrity in AI systems.
Veritas Automata utilizes Kubernetes at an enterprise scale to manage and orchestrate containerized applications. This is particularly important for deploying and scaling AI solutions like LLMs across various environments. Kubernetes ensures high availability, scalability, and efficient distribution of resources, which is essential for the widespread application of ethical AI principles.
The architecture is designed to support distribution among various stakeholders, including companies and individual users implementing the Four Laws. This distributed framework allows for a broad and inclusive application of ethical AI principles across different sectors and use cases.

The components in the Rival architecture are all open source solutions that are part of the Linux Foundation or the Cloud Native Computing Foundation (CNCF). Veritas Automata has used this architecture to create solutions that deliver trusted capabilities leveraging blockchain technology to create transparency and auditability, K3s for open source Kubernetes orchestration in the cloud or on bare metal, and state-of-the-art Machine Learning performing complex analysis.

Want to discuss? Set a meeting with me!

More Insights

Smarter Decisions, Healthier Outcomes: The Role of Business Intelligence in Personalized Healthcare

Thought Leadership
veritas automata arrow

Before the Trial: Using Digital Twins for Preclinical Predictions

Thought Leadership
veritas automata arrow

The Crossroads of Innovation: IoT vs. Edge Computing in Clinical Trials

Thought Leadership
veritas automata arrow

Quality Assurance at the Speed of Innovation: Kubernetes in Drug Development

Thought Leadership
veritas automata arrow


Yuxi Global Logo White


Yuxi Global Iso White
Yuxi Global Iso White