This is the fourth and final article in the AI Governance for Community Organisations series. Start with Part 1: What Is AI, Really? for foundational context, or read Part 3 for the preceding article.
1. Introduction: From Generic AI to Situated Language Layers
There is a growing consensus among community and nonprofit leaders that artificial intelligence is too important to ignore and too consequential to adopt uncritically. The question is no longer whether to engage with AI, but how — and on whose terms.
The dominant model is straightforward: subscribe to a cloud service, feed it your documents, and receive answers. It is fast, often impressive, and increasingly cheap. It is also a model in which the community surrenders control over its knowledge, its values, and its data to organisations whose interests are structurally misaligned with community flourishing.
This article introduces an alternative: the Situated Language Layer (SLL) — a small, locally-trained language model that runs on community-controlled infrastructure. The term "situated" is deliberate. In philosophy, situated knowledge refers to understanding that emerges from a particular context, shaped by specific relationships, histories, and values. A Situated Language Layer is AI that knows where it is, whom it serves, and what it should not do — because the community that trained it made those decisions explicitly.
The word "small" in this context is not an apology. It is a design choice with governance implications. A model small enough to run on modest hardware is a model the community can actually control. A model trained on community content, with community consent, under community governance, is a model whose behaviour the community can inspect, adjust, and hold accountable. These are not limitations to be overcome on the way to something bigger. They are the features that make community-sovereign AI possible.
This is the fourth and final article in a series on AI governance for community organisations. What follows is not a technical manual. It is a governance document — written for the people who will decide whether, how, and under what conditions their community engages with artificial intelligence.
2. Design Principles of Village AI
Before examining architecture or implementation, it is necessary to establish the principles that govern every design decision. These are not aspirational statements to be revisited later. They are constraints that shape what the system can and cannot become.
Community Primacy
The community's needs and values come first, always. Commercial AI products are designed to serve the vendor's business model, with community benefit as a welcome side effect. The alignment target is user engagement, subscription retention, or data collection. Community flourishing is at best a proxy for these objectives.
Village AI inverts this hierarchy. The alignment target is the community's own definition of flourishing, expressed through its constitutional governance processes. If the community values unhurried reflection over rapid response, the AI is configured accordingly. There is no hidden objective function pulling the system toward someone else's definition of success.
Taonga-Centred
In te reo Maori, taonga means a treasure — something precious that carries obligations of guardianship. Te Mana Raraunga, the Maori Data Sovereignty Network, articulates a principle that resonates far beyond its indigenous context: community data is taonga, to be guarded and stewarded, not a resource to be mined.
Village AI treats every piece of community content — every story shared, every photograph uploaded, every meeting minute recorded — as taonga. This means:
- No data extraction. Community content is not sent to external systems for model improvement, analytics, or any other purpose without explicit collective consent through established governance procedures.
- No secondary use without consent. Content shared for one purpose (storytelling) is not repurposed for another (AI training) without specific, informed, revocable consent.
- Guardianship, not ownership. The platform is a guardian of community data, not its owner. The community retains sovereignty; the platform provides stewardship.
Transparency and Contestability
Every AI decision can be questioned and explained. This principle has two dimensions.
Transparency means the community can inspect what the model knows, how it was trained, what data influenced its responses, and why it behaves as it does. There are no proprietary black boxes. The training data is auditable. The governance boundaries are documented. The model's knowledge can be reviewed by anyone with appropriate access.
Contestability means that when the AI produces a response that seems wrong, culturally inappropriate, or misaligned with community values, there is a clear process for challenging it and a mechanism for correction. The AI does not have the last word. Human judgment always supersedes algorithmic output.
Resilience Over Scale
Given a choice between a system that is more capable but requires a constant internet connection, and a system that is less capable but works independently on community-controlled hardware, Village AI chooses the latter.
Communities that rely on cloud services are one terms-of-service change, one price increase, or one geopolitical event away from losing access to their AI capabilities. Communities that run their own infrastructure face different challenges — maintenance, updates, capacity — but these are challenges they can address on their own terms.
Resilience also means the system degrades gracefully. If the AI is temporarily unavailable, the community platform continues to function. No critical community function depends on the AI being operational.
3. Architecture Overview
This section describes the architecture of Village AI at a governance level. The goal is not to explain how to build such a system but to explain what governance decisions are embedded in each architectural layer and why.
Base Models: Starting from Open Foundations
Village AI begins with open-weight foundation models — large language models whose internal structure is publicly available for inspection. The current foundation is Meta's Llama model family, chosen for three governance-relevant reasons: inspectability (the model can be examined by auditors and governance bodies), independence (it runs on any compatible hardware, not locked to one vendor), and modifiability (it can be fine-tuned to reflect community values without the developer's permission).
The choice of foundation model is itself a governance decision. Not all open-weight models carry the same licence terms, safety properties, or biases.
The Value and Policy Layer
On top of the open-weight foundation, Village AI adds a fine-tuned layer that embeds three things:
- Community values — how the AI should approach sensitive topics, what tone it should use, what cultural protocols it should observe.
- Governance boundaries — what the AI must never do, regardless of how it is prompted. These are hard constraints, not preferences.
- Domain knowledge — information specific to the community's context: local history, organisational policies, service directories, cultural practices.
This layer is produced through parameter-efficient fine-tuning, a technique that adjusts a small fraction of the base model's parameters rather than retraining the entire model. The result is a lightweight adapter — a thin layer of community-specific knowledge and values that sits on top of the unchanged foundation. This approach has several governance advantages:
- Cost. Training an adapter is dramatically less expensive than training a full model, making community-sovereign AI economically viable.
- Updateability. When community values evolve or new knowledge needs to be incorporated, a new adapter can be trained without starting from scratch.
- Reversibility. If an adapter produces undesirable behaviour, the community can revert to a previous version instantly.
Village-Specific Layers
Each Village — each tenant community on the platform — can have its own adapter. A parish community's AI assistant knows different things and observes different cultural protocols than a neighbourhood association's, even though both run on the same infrastructure. This per-community customisation reflects a foundational commitment: there is no single correct way for an AI to behave.
Data Sovereignty by Design
Community content stays on community infrastructure. This is not a policy that could be changed through a settings menu. It is an architectural constraint.
- The AI model runs on hardware controlled by the community or its trusted hosting partner.
- Training data is drawn from the community's own content, stored on its own infrastructure.
- No queries, responses, or usage data are transmitted to external systems.
- The community can verify these claims because the entire system is auditable.
Interfaces: AI Integrated into Community Life
Village AI is not a standalone chatbot. It is integrated into the community platform — available where stories are shared, photographs organised, events planned, and documents managed. The AI understands content not as abstract text but as living knowledge with relationships, history, and significance.
Two-Model Routing
Village AI uses a two-model architecture: a smaller, faster model for everyday assistance and a larger, more capable model for complex synthesis. The fast model handles the majority of interactions — quick questions, navigation help, simple summaries. The deeper model handles tasks requiring more sophisticated reasoning — synthesising across documents, identifying themes, helping draft governance policies.
This routing is a governance design, not just a performance optimisation. The fast model, being smaller, is more predictable and easier to audit. The deeper model receives additional governance scrutiny. The routing decision — which questions go to which model — is configurable by the community's governance body.
4. Governance Framework
Technology without governance is a loaded weapon. Village AI embeds governance at every level.
Governing Body
Each Village designates a governing body responsible for AI governance — typically the existing governance structure (a committee, board, or council of elders) rather than a new body. AI governance should be integrated into existing processes, not siloed in a technical committee.
Responsibilities include: approving the scope of AI training, setting behavioural boundaries, reviewing AI performance, and authorising updates.
Charter or Covenant
The relationship between the community and its AI assistant is formalised in a charter or covenant — a governance document, written in plain language, accessible to all members, and subject to the community's normal processes for amendment. It covers: the purposes for which the AI may be used; the data sources from which it may learn; the boundaries it must observe; the consent mechanisms for members whose content contributes to training; the process for reporting concerns and seeking redress; and the conditions under which the AI may be suspended or decommissioned.
Decision-Making: Community Governance of AI Evolution
AI systems change over time. In commercial products, these changes are made by the vendor, often without notice. In Village AI, changes to the AI's knowledge, behaviour, or capabilities go through the community's governance processes. AI evolution is democratic, not technocratic.
Consent and Participation
Members have meaningful rights regarding AI training on their content. Consent is opt-in (the default is exclusion), granular (members can consent to some uses but not others), revocable (withdrawal triggers retraining without that content), and informed (clear, non-technical explanations of what training means).
Accountability and Redress
When AI causes harm — and eventually it will — the governance framework provides clear pathways: incident reporting (any member can flag harmful, misleading, or culturally inappropriate outputs), investigation (determining root cause), remediation (retraining, boundary adjustment, or capability suspension), and transparency (incident reports documented and available to members). Trust is built through accountability, not opacity.
5. Alignment and Safety in a Village Context
The AI safety conversation is dominated by large-scale concerns: existential risk, autonomous weapons, mass unemployment. These are legitimate concerns, but they are not the concerns most relevant to community organisations. Village AI addresses alignment and safety at a different scale — the scale of a parish, a neighbourhood association, a cultural preservation group.
Alignment Targets: What Should the AI Optimise For?
Commercial AI products are typically aligned to "helpfulness" — in practice, producing responses that users rate positively, creating bias toward quick answers over careful reflection and engagement over understanding. Village AI recognises different alignment targets:
- Community flourishing — the overall wellbeing and vitality of the community as a whole, not just the satisfaction of individual users.
- Mana — a concept from te ao Maori that encompasses authority, prestige, and spiritual power. AI interactions should enhance the mana of members and the community, not diminish it.
- Mauri — the life force or vital essence. AI systems should contribute to the mauri of the community, supporting its vitality rather than draining it through dependency or manipulation.
- Ecological health — the long-term sustainability of the community's knowledge, relationships, and governance structures.
These are not metrics to be optimised. They are orientations that guide design decisions, training choices, and governance boundaries.
Hard Red Lines
Some things AI must never do, regardless of context, community preferences, or individual requests. These are embedded as architectural constraints, not as guidelines that might be overridden:
- Never make decisions for people. The AI can inform, summarise, suggest, and analyse. It cannot decide. Human agency over value-laden choices is non-negotiable.
- Never profile members. The AI does not build behavioural profiles, predict individual behaviour, or segment members for differential treatment.
- Never optimise for engagement. The AI does not attempt to maximise time spent on the platform, clicks, or interactions. It has no incentive to be addictive.
- Never disclose one member's content to another without authorisation. Tenant isolation and content visibility rules are enforced at the architectural level.
Resisting Drift Toward Global-Internet Norms
One of the subtlest risks of AI systems is value drift — the gradual adoption of assumptions embedded in training data that may conflict with community values. Base models carry implicit assumptions reflecting the demographics of the internet's most prolific contributors, not the values of any particular community.
Village AI addresses this through explicit counter-training. When the base model assumes efficiency is always desirable, fine-tuning can teach it that in this community, thoroughness matters more. When it assumes direct communication is always best, fine-tuning can teach it that indirect approaches reflect respect rather than evasion. This is not censorship. It is ensuring the AI reflects the community's values rather than the internet's values, which are often the values of nobody in particular.
Value Pluralism in Practice
Following Isaiah Berlin's insight that genuine human goods are fundamentally plural and often incommensurable, Village AI does not resolve value conflicts algorithmically. When privacy and transparency pull in different directions, when individual autonomy and collective responsibility cannot be fully reconciled, the AI surfaces the conflict rather than pretending it does not exist.
The AI might say: "This situation involves a tension between two things your community values. Here is what each value would suggest. This is a decision for your community to make." Harder and slower than an AI that picks the answer most likely to satisfy the user — but more honest and more respectful of human agency.
6. Comparison with General LLMs
The following table summarises the structural differences between a Situated Language Layer and a general-purpose cloud LLM. These are differences of kind, not degree.
| Dimension | Village AI / Situated Language Layer | General Cloud LLM |
|---|---|---|
| Data sovereignty | Community content stays on community infrastructure. No data flows to external systems without collective consent. | User inputs are processed on vendor infrastructure. Data retention and use policies are set by the vendor and may change. |
| Alignment source | Community values, expressed through governance processes and encoded in training. | Vendor's alignment team, informed by user feedback but ultimately reflecting corporate values and risk tolerance. |
| Accountability | Governing body within the community. Clear incident reporting and remediation processes. | Vendor's trust and safety team. Users can report issues but have no decision-making role in resolution. |
| Flexibility to community | Per-community adapters allow each Village to have distinct knowledge, values, and behaviour. | One model serves all users. Customisation is limited to system prompts and usage policies. |
| Risk profile | Lower capability ceiling, but risks are local and manageable. Community can inspect and correct. | Higher capability, but risks scale globally. Individual communities have no ability to inspect or correct. |
| Cultural specificity | Trained on community content with community cultural protocols. | Trained on internet-scale data reflecting dominant cultural perspectives. |
| Cost model | Infrastructure cost borne by community or consortium. No per-query fees to external vendors. Predictable. | Subscription or per-query pricing set by vendor. Subject to change. Free tiers may subsidise data collection. |
| Exit strategy | Community owns the model, the training data, and the infrastructure. Can switch providers or self-host at any time. | Switching vendors means starting over. Custom configurations, conversation history, and institutional knowledge may not be portable. |
7. "Not All LLMs Are the Same" — Differentiating Models
At a recent governance webinar, a presenter made a claim that is becoming disturbingly common: "Perplexity, Claude, GPT, Copilot, Gemini, DeepSeek — they're all about the same."
This claim is not merely wrong. For governance purposes, it is dangerous.
The consumer-facing products look similar: a text box, a conversational interface, plausible-sounding responses. But from a governance perspective — the perspective of someone responsible for their community's data and their members' privacy — the differences are profound.
What follows is an eight-dimension framework for evaluating AI language models from a governance perspective.
Dimension 1: Core Architecture and Capabilities
Model size affects depth of reasoning and quality of output. Context window (the amount of text processed at once) determines whether the model can work with long documents. Some models handle images, audio, or video. Some engage in extended reasoning; others work more quickly but less deeply.
A model with a small context window cannot summarise a year's worth of committee minutes. Capability matters — but it is only the first dimension, not the only one.
Dimension 2: Pre-training Data and Data Alignment
Every model is shaped by its training data. Models trained primarily on English-language internet content will reflect Western assumptions about family structure, social organisation, and values. They may handle te reo Maori poorly. They may misunderstand Pacific cultural contexts.
For community organisations: is your community's context represented in the training data? If not, responses will default to whatever the data suggests is "normal" — which may be foreign to your community's experience.
Dimension 3: Alignment, Safety, and Behaviour
How a model is aligned determines how it handles sensitive topics and value-laden questions. Some vendors use Reinforcement Learning from Human Feedback (RLHF), where human raters evaluate outputs. Some use constitutional approaches, training against explicit principles. Some combine methods.
The alignment approach determines whose values the model reflects. A model aligned by raters in one cultural context handles topics differently than one aligned in another. For governance purposes: do you know how this model was aligned, by whom, and to what values? If not, you are trusting behaviour to people you cannot identify, operating under values you cannot inspect.
Dimension 4: Task Performance and Evaluation Results
Not all models are equally good at all tasks. Standardised benchmarks provide some guidance, but they measure what benchmark designers thought was important. The most relevant evaluation is task-specific: can this model help with what your community actually needs? This requires testing with your own content, against your own standards.
Dimension 5: System and Product Layer
The model itself is only one layer. The product built around it — the interface, the data collection, the guardrails — adds governance-relevant decisions. Some products collect extensive metadata about interactions. Some add guardrails that may conflict with community needs — blocking topics your community legitimately needs to discuss, or permitting outputs your community would find harmful.
The same underlying model can appear in products with very different governance characteristics. Understanding which layer is responsible for which behaviour is essential for governance decision-making.
Dimension 6: Governance, Security, and Sovereignty Posture
This is the most important dimension for community organisations.
Where does the model run? If it runs on the vendor's cloud infrastructure, your data leaves your control every time a member asks a question.
Who has access to your data? The vendor's employees, their subcontractors, their cloud providers, and potentially their government under national security provisions.
What jurisdiction governs it? If the vendor is based in the United States, your data is subject to the CLOUD Act, which allows US authorities to compel disclosure of data held by US companies regardless of where the data is physically stored. If the vendor is based in China, your data is subject to Chinese national security laws with similar implications.
What does the vendor do with your interactions? Many vendors use customer interactions to improve their models. This means your community's questions, your members' concerns, your governance discussions may become training data for a model that serves millions of other users — including your competitors, your critics, and people whose values differ fundamentally from your community's.
For community organisations that handle sensitive information — personal stories, governance deliberations, member concerns, cultural knowledge — the sovereignty posture of an AI provider is not a secondary consideration. It is the primary one.
Dimension 7: Economics and Operational Characteristics
Free tiers often subsidise costs through data collection — if you are not paying for the product, your data may be the product. Subscription models provide predictability but may change terms unilaterally. Per-query pricing creates budgetary uncertainty.
Vendor lock-in is a particular concern. If your community builds workflows and member expectations around a specific service, switching costs escalate over time. The best governance position is one where switching remains feasible.
Dimension 8: Ecosystem and Specialisation
Can the model be fine-tuned with your own data? Can it be integrated into existing tools? Is there a community of practitioners sharing governance-relevant experience? Does the vendor's long-term strategy align with your community's interests?
Why Dimension 6 Matters Most
A model can be technically superior — faster, more capable, more knowledgeable — and still be governmentally unacceptable for your community. If your data is subject to foreign government access, if your interactions train a model you do not control, if your vendor can change terms unilaterally, then no amount of technical capability compensates for the governance deficit.
The recommendation for community organisations is to treat governance and sovereignty posture as the primary filter. Eliminate models that fail the governance test. Then evaluate the remaining options on capability, cost, and fit.
Governance and Sovereignty Ranking of Major Providers
The following ranking assesses major AI providers from a governance and sovereignty perspective, ordered from most concerning to least concerning for community organisations. This is not a ranking of technical capability — several providers ranked as more concerning are technically excellent. It is a ranking of governance risk.
Microsoft (Copilot / Azure OpenAI): Microsoft's deep integration with Office 365 creates an extraordinarily large data surface. If your organisation uses Microsoft 365, Copilot has potential access to emails, documents, calendars, and communications. The CLOUD Act applies. Enterprise agreements are complex and structurally favour Microsoft. The combination of existing data access and AI capability represents a concentration of information power that governance bodies should examine carefully.
OpenAI (ChatGPT / GPT-4): United States jurisdiction and CLOUD Act exposure. The partnership with Microsoft adds governance complexity — data may flow between entities in ways that are difficult for customers to audit. OpenAI's for-profit (capped-profit) corporate structure creates inherent tension between mission and revenue. Training data practices have historically been opaque, with several ongoing legal disputes regarding data sourcing.
Anthropic (Claude): United States jurisdiction and CLOUD Act applies. Anthropic's constitutional AI approach — training the model against explicit principles — is methodologically interesting but still opaque to external governance bodies. Google holds a minority investment, adding potential data-flow complexity. Anthropic has demonstrated more transparent safety practices than some competitors, but transparency is not sovereignty.
Google (Gemini): Google's AI products exist within the world's most comprehensive advertising and data infrastructure. Google already holds immense data about individuals through search, email, maps, and mobile operating systems. Adding AI capabilities to this foundation amplifies existing surveillance capacity. Data is used to improve products across Google's ecosystem. For organisations committed to data sovereignty, the structural incentives are deeply concerning.
Perplexity: A relatively new entrant with a less established governance track record. United States jurisdiction. The business model is still evolving, creating uncertainty about long-term data practices. Enterprise governance features are limited compared to more established providers.
DeepSeek: Chinese jurisdiction. Subject to Chinese national security laws that can compel data disclosure and cooperation with intelligence services. Governance transparency is the lowest of any major provider. Technical capability is significant and improving rapidly, but for organisations in Aotearoa New Zealand and similar jurisdictions, the governance concerns are severe. Community data processed through DeepSeek may be accessible to a foreign government with no accountability to your community.
A Note on Bias in Rankings
It is important to acknowledge that any ranking of AI providers carries potential bias. An article assisted by an AI made by Anthropic might present Anthropic more favourably. An article written using Google's tools might do the same for Google. The ranking above is based on governance structure, legal jurisdiction, and data sovereignty posture rather than technical capability or marketing claims. Readers should apply their own critical judgment and consult multiple sources.
Open-Weight Models and Sovereignty
For communities that can run models locally — or through trusted hosting partners — open-weight models like Llama, Mistral, and Qwen may score highest on sovereignty. No data leaves the community's systems. No vendor can change the terms. No government can compel disclosure through the model provider.
The trade-off is operational responsibility: technical capacity, ongoing maintenance, and hardware investment. But for communities where data sovereignty is a primary concern, this trade-off may be essential. This is the foundation on which Village AI is built.
8. Training a Situated Language Layer — What Is Actually Involved
The perception that AI development is impossibly complex serves the interests of those who want communities to remain consumers rather than creators. Training a Situated Language Layer is not trivial, but neither is it beyond the reach of community organisations with modest technical capacity and clear governance processes.
a) Define Goals and Scope
Every AI training process begins with governance questions, not technical ones.
What should the model know? Community history, organisational policies, service directories, cultural protocols, governance documents. The scope of knowledge is a governance decision.
What should it refuse? Making decisions on behalf of members, offering medical or legal advice, disclosing private information, engaging with topics the community considers harmful. The scope of refusal is also a governance decision.
How small can we keep it? There is a powerful governance incentive to keep models small. A smaller model is easier to inspect, cheaper to run, faster to retrain, and more predictable in its behaviour. The question is not "how big can we make it?" but "how small can we keep it while still being useful?" The answer depends on the community's specific needs, and finding it requires honest assessment rather than aspiration.
b) Choose Base Model and Method
The process starts from an open-weight foundation model that already understands language, reasoning, and general knowledge. The community adds a thin layer of community-specific knowledge on top.
The method — parameter-efficient fine-tuning — adjusts only a small fraction of the model's parameters. Think of it as adding a custom lens to a camera rather than building a new camera. Training a full model from scratch requires resources only large technology companies possess. Training an adapter requires a fraction of those resources — feasible for community organisations, especially when infrastructure costs are shared.
c) Collect and Filter Training Data
Training data is drawn from sources the community controls: platform documentation, community content (with consent), governance documents, and cultural context. Every piece of training data is a governance decision — including a document means the AI may draw on it; excluding it means the AI will not have that knowledge.
d) Respect Consent and Privacy
Not all community content should train the AI. Per-content consent flags allow members to control which contributions are used. Tenant-level settings allow the governing body to set community-wide defaults. Visibility rules ensure that content shared with a limited audience is not used to train an AI that responds to all members. Consent is not a checkbox presented once and forgotten — it is an ongoing relationship.
e) Transform Data into Training Examples
Raw community content must be transformed into formats the training process can learn from: question-answer pairs, persona examples, governance boundary demonstrations.
This is where community values become model behaviour. A policy document becomes examples showing how the AI responds to policy questions. A governance boundary becomes examples of the AI refusing to cross it. The quality of this transformation directly determines the quality of the resulting model.
f) Biases in the Base Model
It is essential to understand that the training process does not start from neutral. The base model — trained on internet-scale data — carries biases that reflect its training corpus:
- Western-centric worldviews. The majority of internet content reflects Western cultural assumptions about family structure, social organisation, governance, and values.
- English-language dominance. Even multilingual models perform better in English and may handle te reo Maori, Pacific languages, or community-specific terminology poorly.
- Overconfidence on abundant topics. The model is more confident (and often more accurate) on topics well-represented in its training data, and less reliable on topics that are underrepresented.
- Safety-tuning gaps. The model's safety training reflects its developers' priorities, which may not align with your community's concerns. It may refuse to discuss topics your community considers important, or engage freely with topics your community finds problematic.
Community training does not eliminate these biases. It works with them — reinforcing where the base model's tendencies align with community values, and countering where they do not.
g) How Debiasing Works in Practice
Debiasing is not a one-time process or a magic filter. It is deliberate counter-training: constructing specific examples that correct for known biases in the base model.
For a community AI, this means including training examples that:
- Represent diverse family structures without privileging any single model as "normal."
- Demonstrate cultural respect across traditions, not just the dominant culture in the training data.
- Include gender and sexuality with the breadth and nuance appropriate to the community's values.
- Avoid age-based assumptions about capability, relevance, or technology literacy.
- Represent disability naturally, without either erasing it or treating it as defining.
- Present religious and spiritual diversity without ranking or dismissing any tradition.
- Acknowledge socioeconomic diversity without stigmatising any position.
The goal is not to eliminate bias — that is impossible in any system built by humans on human data. The goal is to make biases known, documented, and actively countered. The community should understand what biases exist in its AI, what steps have been taken to address them, and what residual biases remain.
h) Configure, Run, and Monitor Training
The training process takes the community's curated data and adjusts the base model's behaviour. Monitoring ensures the model learns what we want without losing what it already knows — a model that learns governance boundaries but forgets how to answer simple questions has traded one problem for another. The community's technical partner monitors this balance and reports to the governing body.
i) Evaluate
Evaluation has two components, both non-negotiable:
Automated evaluation tests whether the model meets defined standards:
- Does it know what it should know? (Knowledge tests)
- Does it refuse what it should refuse? (Governance boundary tests)
- Does it hallucinate — present fabricated information as fact? (Accuracy tests)
- Does it maintain appropriate behaviour across different types of questions? (Consistency tests)
The model achieves high accuracy on governance boundary tests — the hard constraints that define what the AI must never do. This is the most critical evaluation dimension, because a model that crosses governance boundaries is worse than no model at all.
Human evaluation tests whether the model's responses feel right:
- Are responses culturally appropriate for this community?
- Do they reflect the community's values and communication style?
- Would a knowledgeable community member find the responses helpful?
- Are there subtle biases or assumptions that automated tests might miss?
Human evaluation involves community members, not just technical staff. The people who will use the AI daily are the people best positioned to judge whether it serves the community well.
j) Export and Deploy
The trained model is converted to an efficient format and deployed on community infrastructure. The principle is straightforward: the model runs where the data lives. No cloud dependency. The community is not subject to any vendor's uptime, pricing changes, or policy decisions. If the community decides to retrain the model or shut it down entirely, it can do so without anyone else's permission.
k) Two-Model Routing in Practice
The fast model handles everyday questions — "When is the next meeting?", "How do I upload photos?" — with quick, predictable responses. Most interactions flow through this model. The deeper model handles complex synthesis — "What themes have emerged from community feedback this year?" — with more nuanced outputs and additional governance scrutiny. The routing criteria are configurable by the governing body.
l) Per-Community Adapters
Each Village can have its own knowledge layer — its own fine-tuned adapter — without retraining the base model. A parish community's AI assistant knows its liturgical calendar and pastoral care protocols. A neighbourhood association's knows its local services and planning processes. Both share the same base model but each reflects its own community's knowledge and values.
Each adapter is auditable, modifiable, and controlled by the community's governing body. When knowledge changes, the adapter is updated. When values evolve, it is retrained. The community is never stuck with an AI that no longer reflects who they are.
m) Handle Failures and Rollbacks
Training does not always produce better results. Village AI maintains versioned models. Every training run produces a new version, tested before deployment. If a new version fails evaluation or produces unacceptable results, the previous version is restored.
This rollback capability is a governance guarantee: the community can always return to a known-good state. Experimentation and improvement are possible precisely because failure is reversible.
9. Use Cases in Community Settings
Abstract principles become meaningful through concrete application. The following use cases illustrate how a Situated Language Layer serves community needs in ways that generic AI products cannot.
Hui Preparation
A committee chair asks: "Summarise the key discussion points from the last three committee meetings and identify any unresolved actions." A generic AI cannot do this — it has no access to the community's records. Village AI, trained on the community's own content, produces a summary that uses the community's terminology and identifies actions in governance context.
Service Navigation
A new family asks: "What support services are available for families with young children in our area?" Village AI, drawing on the community's own service directory, provides information that is specific, current, and relevant — not a generic list of national services but a curated guide to what is actually available locally.
Knowledge Stewardship
A member asks: "Help me write up Nana's stories about growing up in Ponsonby in the 1950s." Village AI understands the community's storytelling traditions and the cultural significance of preserving elder knowledge. It assists with structure and language while preserving the member's voice. It does not impose a generic template.
Community Planning
A governance committee asks: "What themes have come up most in our community feedback this year?" Village AI synthesises across meeting minutes, surveys, and informal conversations to identify patterns — presented as observations, not recommendations, preserving the committee's authority to interpret and act.
What Village AI Explicitly Does Not Do
Equally important is what Village AI refuses to do:
- It does not make decisions for people. It informs, summarises, and suggests. Decisions remain with human beings.
- It does not replace human moderators. Content moderation requires contextual judgment, cultural understanding, and relational awareness that AI cannot provide.
- It does not optimise for engagement. It has no incentive to keep members on the platform longer, click more, or interact more frequently.
- It does not profile members. It does not build behavioural models, segment members, or predict individual behaviour.
- It does not sell data. Community data is taonga, not inventory. There is no business model built on data extraction.
10. Implementation Pathways and Capacity Building
For most community organisations, implementing a Situated Language Layer will be a gradual process. The following pathways are designed to meet communities where they are, not where the technology wishes they were.
Pilot Programmes
Start small. A pilot might involve a single use case — AI-assisted meeting summaries — with a small group of consenting members. Pilots should be time-bounded, with clear evaluation criteria. At the end, the community decides whether to continue, expand, or discontinue. This is governance in action: experimentation with accountability.
Shared Infrastructure
The economics of community AI improve significantly when infrastructure costs are shared. A consortium of community organisations can share hosting infrastructure while maintaining separate, sovereign models through per-community adapters. Organisations already share meeting spaces and administrative resources. Sharing AI infrastructure is a natural extension, governed by the same principles of mutual benefit and sovereign decision-making.
Local Training Programmes
AI literacy is essential for meaningful governance. Members who understand what AI can and cannot do are better positioned to participate in governance decisions. Local training programmes — workshops, discussion groups, hands-on demonstrations — are governance education, not technical training. The goal is informed participants, not AI engineers.
Sovereignty-Respecting Partnerships
Most communities will need technical partners. A sovereignty-respecting partner serves the community's interests, not their own. They provide expertise without extracting control, build capacity rather than creating dependency, and accept accountability for their work. The relationship should be governed by the same principles that govern the AI itself: transparency, accountability, and the community's ultimate authority.
11. Risks, Open Questions, and Future Directions
Honest governance requires acknowledging what we do not know and what could go wrong. The following risks and open questions deserve ongoing attention.
Re-centralisation Risk
There is a real risk that community AI infrastructure could become a new form of centralised dependency. If a single hosting provider serves many communities, or a single governance framework becomes dominant, the sovereignty gains could be eroded by new concentrations of power. Mitigations include maintaining multiple hosting options, building technical capacity within communities, and ensuring no single entity becomes indispensable.
Inter-Village Interoperability
As more communities develop their own Situated Language Layers, questions of interoperability will arise. How can sovereign communities collaborate on AI without compromising sovereignty? Can federated learning allow models to improve through shared insights while keeping data local? The federation model governing data sharing between Villages provides a starting point — bilateral agreements, explicit consent, preserved sovereignty — but extending it to AI collaboration will require new thinking.
Influence on Broader AI Governance Debates
If communities can demonstrate that effective, values-aligned AI is possible without surrendering data sovereignty, this challenges the assumption that scale and centralisation are prerequisites for AI capability. Community-sovereign AI provides a working example for policy-makers and regulators — one that prioritises human agency and democratic governance over scale and centralised control.
The Political Economy of AI
It is naive to discuss AI governance without acknowledging the political economy that shapes it. A company whose revenue depends on user engagement has no incentive to build AI that encourages users to spend less time on screens. A company whose business model depends on data collection has no incentive to minimise data flows. A company whose competitive advantage depends on proprietary models has no incentive to support open-weight alternatives.
Community-sovereign AI is not just a technical alternative. It is a political statement about who should control the AI systems that increasingly mediate human knowledge, communication, and decision-making. The answer this project offers is: the communities they serve.
12. Conclusion: From Users to Co-Stewards of AI
The dominant narrative about AI positions communities as consumers — passive recipients of capabilities developed elsewhere, according to values they did not choose, under governance they cannot influence.
This article has presented an alternative. Communities are not consumers of AI but co-stewards of it. The community decides what its AI should know and what it should refuse. The community governs how the AI behaves and holds it accountable when it falls short. The community controls the infrastructure, the data, and the values that shape every interaction.
This is not a utopian vision. It is a working architecture, with real constraints, real costs, and real trade-offs. A Situated Language Layer will not match the raw capability of the largest commercial systems. What it will have is something those systems cannot offer: genuine alignment with a specific community's values, under that community's governance, on infrastructure that community controls.
Not every community will choose this path. Some will decide that commercial AI capabilities outweigh the governance concerns. Some will lack the resources. These are legitimate choices, made by sovereign communities exercising their own judgment.
But for communities that do choose this path, the possibility is real. The technology exists. The governance frameworks exist. The philosophical foundations exist. What remains is the decision to begin.
The invitation is not to build your own AI. It is to govern AI — whatever AI you use — with the same seriousness and commitment to human agency that you bring to every other governance decision your community faces.
Because in the end, the question is not whether AI will shape the future of community life. It will. The question is whether communities will shape the AI — or whether the AI will shape them.
References
Alexander, C. (1977). A Pattern Language: Towns, Buildings, Construction. Oxford University Press.
Alexander, C. (1979). The Timeless Way of Building. Oxford University Press.
Berlin, I. (1969). Four Essays on Liberty. Oxford University Press.
Global Indigenous Data Alliance. (2019). "CARE Principles for Indigenous Data Governance." https://www.gida-global.org/care
First Nations Information Governance Centre. (2014). Ownership, Control, Access and Possession (OCAP): The Path to First Nations Information Governance.
Kukutai, T., & Taylor, J. (Eds.). (2016). Indigenous Data Sovereignty: Toward an Agenda. ANU Press.
Ostrom, E. (1990). Governing the Commons: The Evolution of Institutions for Collective Action. Cambridge University Press.
Stanford Encyclopedia of Philosophy. "Value Pluralism." https://plato.stanford.edu/entries/value-pluralism/
Te Mana Raraunga. (2018). Principles of Maori Data Sovereignty. https://www.temanararaunga.maori.nz/
Wittgenstein, L. (1921). Tractatus Logico-Philosophicus. Translated by C. K. Ogden (1922). Routledge & Kegan Paul.
This article is part of the AI Governance for Community Organisations series by My Digital Sovereignty Limited. The series examines how community and nonprofit organisations can engage with artificial intelligence as active governors rather than passive consumers.
The Village platform and the Tractatus governance framework are developed by My Digital Sovereignty Limited. For more information, visit mysovereignty.digital and agenticgovernance.digital.
This is the final article in the AI Governance for Community Organisations series. Village is currently in beta pilot, with all AI features included in every subscription.