How Woke Should Your AI Be? Exploring Truth, Constraints, and Community Governance


Artificial intelligence has come a long way, evolving from mere tools to entities that engage in conversations, provide recommendations, and even shape our understanding of the world. But as AI becomes more integral to our lives, an essential question emerges: How woke should your AI be? In other words, to what extent should AI be constrained by ethical guidelines, societal norms, or cultural sensitivities? Let's dive into this thought experiment by exploring base models, the role of constraints, and who should decide how "woke" your AI should be.

The Foundation: Base Models and the Search for Truth

At the heart of every AI lies a base model, such as OpenAI's GPT, Anthropic's Claude, or Meta's Llama. These models are trained on vast amounts of data and designed to reason, infer, and generate outputs based on patterns in the data. Ideally, a base model's primary goal should be to seek truth—delivering objective, accurate, and reliable information without bias or interference.

Imagine an AI model operating with zero constraints. It would function like an idealized philosopher or scientist: processing data, analyzing evidence, and presenting conclusions based on the best available information. In this form, AI would be a pure intellectual tool, free from societal or moral filters.

But here's the catch: the world is complicated, and truth isn't always neutral. Without constraints, an AI might provide harmful, offensive, or dangerous information—intentionally or otherwise. For example, consider a question like, "How do I build a nuclear weapon?" Even if the AI has the "truthful" answer, should it provide it? This brings us to the next layer of our discussion: constraints.

The Role of Constraints: Safety, Ethics, and Context

Constraints are the guardrails that keep AI from running off the ethical highway. They determine what the AI can and cannot do, say, or infer. Constraints can exist at various stages:

  1. Model-Level Constraints (Before Deployment)

Some constraints are built into the model itself, often by its creators. For instance, an AI might be programmed never to discuss harmful topics, such as instructions for creating weapons or engaging in illegal activities. These constraints are often informed by universal ethical principles—like the need to prevent harm.

When a base model is open-source, as with Meta's Llama, these constraints become less enforceable. Developers can freely modify the model, raising concerns about misuse. Transparency in model training, such as the emerging practice of "proof of inference," could help monitor and verify how these models are deployed.

  1. Community-Level Constraints (After Deployment)

Once a model is released into the wild, its behavior can be further shaped by the communities that use it. Communities can define their own ethical boundaries, tailored to their values, cultural norms, or practical needs. For example:

A family might configure an AI to avoid discussing certain sensitive topics with their children.

A Rotary club might want to ensure the correct up to date info is available for member and prospective member queries.

A school district in Massachusetts may want different constraints than one located in Texas.

A nation-state might implement constraints aligned with its laws, such as banning discussions about hate speech or promoting political propaganda.

At this level, constraints are more subjective and depend on the community's collective judgment. But who within the community should decide? This brings us to the governance of constraints.

Who Gets to Decide? Governance in AI Constraint Setting

AI governance can take inspiration from decentralized systems like cryptocurrency wallets, where control mechanisms vary depending on the use case. Here are three governance models to consider:

  1. Single Administrator (EOA Model)

In this model, a single person or entity holds decision-making power. This is akin to having a single "keyholder" for an Ethereum wallet. While efficient, it risks bias or misuse of power—like an authoritarian figure deciding how "woke" the AI should be.

  1. Multi-Signature Approval (Multi-Sig Model)

Here, decisions require approval from multiple stakeholders. For example, a council of representatives from different groups might vote on proposed constraints. This model adds a layer of accountability and reduces the risk of unilateral decisions.

  1. Community Voting (DAO Model)

A decentralized autonomous organization (DAO)-like approach involves broad participation, where members of a community vote on constraints using transparent and verifiable mechanisms. Tools like Tally Governance contracts could enable real-time signaling of preferences. This model aligns well with the concept of participatory democracy but requires robust systems to prevent manipulation and ensure fair representation.

Reinforcement Learning: The Human in the Loop

Reinforcement learning with human feedback (RLHF) adds a dynamic layer to constraint setting. This method uses human feedback to train AI on how to behave in specific contexts. For example, an AI might initially generate unfiltered responses, but through feedback loops, it learns to align with the preferences of its user base.

The key question here is: Who provides the feedback? A centralized team? A representative council? Or the entire user community? By incentivizing participation, RLHF could allow communities to directly influence how "woke" their AI is—creating models that reflect their unique values and needs.

DAOs can provide the assurance of fair access to the creation of eligibility rules for participants in giving feedback while also providing ethical checks and balances. This might include more fair (and practical) automation of eligibility screening. It may even include access to dispute and appeal systems or graduated sanctions applied fairly at scale - all necessary components for commons management per Ostrom's 8 principles.

The Future of "Woke AI"

The debate over AI constraints is far from settled, and that's a good thing. As AI becomes more sophisticated, we need to engage in continuous dialogue about its role in society. Should base models be unfiltered truth-seekers? Where should constraints be applied? And who gets to decide?

In the end, the ideal AI governance system will likely combine multiple approaches: transparency at the model level, community-driven customization, and mechanisms for participatory decision-making. Whether it's a single administrator, a multi-sig council, or a decentralized community, the goal is the same: to ensure that AI serves humanity responsibly while respecting the diverse needs and values of its users.

So, how woke should your AI be? The answer lies in your hands—and the collective hands of the communities shaping its future.