Giesen Labs

The Open Source Question: Who AI Really Belongs To

Open source and edge-deployed AI have the potential to democratize access to one of the most powerful technologies ever created. But the question of who benefits — and who's at risk — demands a more honest conversation than the field is currently having.

There's a version of the future where the most powerful AI tools in the world sit behind the paywalls and API keys of a handful of large American technology companies. Where access to AI is a function of credit card limits and corporate subscriptions. Where the intelligence layer of the global economy is controlled by a small number of entities, and the rest of the world is simply a customer.

And there's another version — messier, more complicated, harder to govern — where AI models are openly available, deployable on affordable hardware, running offline in clinics without internet connections, in classrooms in communities that can't afford cloud subscriptions, on devices that don't report back to anyone.

The open source and edge AI movements represent a genuine fork in the road. Not a technical fork — a philosophical one. The question isn't whether open models are more capable than closed ones. The question is what kind of world we're building, and for whom.

The Promise Is Real

Let's start with what open source and edge AI actually offer, because the case is substantial. When models are openly available — weights, architecture, training methodology — they can be audited. Researchers can probe them for bias. Regulators can inspect them. Developers in contexts very different from Silicon Valley can adapt them. The transparency that open source demands is one of the most powerful tools we have for accountability in AI.

This matters enormously in a field where the most capable systems are currently black boxes. When a closed model makes a consequential decision — in a hiring pipeline, a medical context, a loan application — there is often no recourse, no audit trail, no independent verification of what the system was actually optimizing for. Open source doesn't solve this problem entirely, but it creates the conditions under which it can be solved.

Edge deployment adds another dimension. Processing AI locally — on device, without a network connection — means sensitive data stays where it belongs. It means a medical AI in a rural clinic isn't uploading patient information to a foreign cloud server. It means a family using an AI companion isn't feeding every conversation to a data center. It means the experience of using AI doesn't require infrastructure that billions of people don't have reliable access to.

The latency benefits are real too. Local inference is fast — often under 10 milliseconds — which unlocks applications that cloud round-trips simply can't support. But the latency argument, while technically compelling, is secondary. The primary case for edge AI is about sovereignty: data sovereignty, infrastructure sovereignty, and ultimately, technological sovereignty.

"AI must correct its structural biases and actively contribute to the expansion of rights." — UNESCO, 2025

Researchers at UNESCO have framed this clearly: equitable AI requires not just access to tools, but the ability to audit, adapt, and govern them. Open source is a necessary — though not sufficient — condition for that kind of participation. The alternative, as the same body notes, risks "perpetuating extractivist logics and reinforcing historical patterns of exclusion."

Access Is Not Empowerment

But here's where the conversation needs to get more honest. Democratizing access to AI is not the same thing as democratizing empowerment. Making something available is not the same as making it useful — or safe — for everyone.

The infrastructure paradox is real: the communities most likely to benefit from locally-deployed, offline-capable AI are often the same communities least equipped to deploy and maintain it. Hardware costs money. Technical expertise requires education. The gap between "this model is available for download" and "this model is serving the needs of an underserved community" is filled with resources that don't distribute evenly.

As researchers studying the democratization of generative AI have noted, access without literacy and infrastructure can actually widen existing socioeconomic gaps rather than close them. Technology that disproportionately benefits those already equipped to use it is not democratizing — it's just redistributing advantage within the already-advantaged.

Training data is another dimension. Open models trained primarily on English-language, western-origin data carry embedded biases that don't dissolve when the model is deployed elsewhere. A freely available model that systematically underperforms for speakers of minority languages or reflects the cultural assumptions of its training set is not a democratizing force — it's a global distribution of locally-calibrated values.

The Dual-Use Reality

The risks of open source AI are real, and it's important not to hand-wave them. The same properties that make open models valuable for accountability and adaptation — unrestricted access, modifiable weights, global availability — also make them available to actors with destructive intent.

The Global Center for the Responsibility to Protect has catalogued the threat landscape clearly: open models can be repurposed for deepfake generation, automated phishing, disinformation campaigns, and evasion of content detection systems. The FBI has documented cases of cybercriminals using open-source AI to develop malware. The regulatory challenge is compounded by the distributed, borderless nature of open-source communities — there is no single actor to hold accountable when a freely available model is weaponized.

The R Street Institute's 2025 mapping of the open-source AI debate describes the core tension well: the same transparency and accessibility that enable collaborative improvement also enable adversarial exploitation. Data poisoning, adversarial attacks, and model manipulation are all harder to govern in decentralized, open ecosystems.

None of this is an argument for closing off AI development. The historical record on closed, centralized control of powerful technologies is not encouraging. Concentration doesn't eliminate risk — it relocates it, and it adds the specific risk of capture: when a small number of actors control a transformative technology, the question of whose values it encodes becomes existentially important.

The choice isn't between open AI that's dangerous and closed AI that's safe. It's between different distributions of risk, different concentrations of power, and different visions of who the technology ultimately serves.

A More Honest Framework

What the field needs — and largely lacks — is a framework for thinking about open source AI that doesn't collapse into either techno-optimism or security panic. Both camps tend to argue from their preferred conclusion and work backward to the evidence.

A more useful framework would start with specificity. Not "is open source AI good or bad?" but: which models, deployed in which contexts, with what governance, serving whose needs? The risk profile of a small language model running offline on a community health device is categorically different from the risk profile of a frontier-scale generalist model with unrestricted weights available to anyone.

Hybrid approaches are emerging. Meta's Llama framework — requiring users to apply for access, with licenses that prohibit high-risk applications — attempts to thread the needle between openness and oversight. Tiered access models, where foundational capabilities are broadly available but advanced features require vetting, represent another direction. These are imperfect solutions, but they're evidence that the binary between "fully open" and "fully closed" is a false choice.

The edge AI trajectory is particularly promising here. Capable models that run on affordable hardware, without requiring cloud connectivity, offer a path to genuine democratization that doesn't depend on centralized gatekeepers. The challenge is ensuring that the models being deployed are aligned with the values of the communities using them — not just technically capable, but genuinely beneficial.

This is where the design question becomes paramount. Open source addresses who can access AI. Edge deployment addresses who can run it without infrastructure dependency. But neither addresses, on its own, the question of what the AI is designed to do — what it's optimizing for, whose interests it serves, what it considers a good outcome.

What Good Looks Like

The communities most in need of AI's capabilities — those without reliable internet, those in healthcare deserts, those speaking languages underrepresented in training data — deserve AI that was built with their context in mind. Not adapted as an afterthought. Not a global model with a localization layer dropped on top. Purpose-built.

That requires more than open weights. It requires open benchmarks that evaluate AI across dimensions of human benefit, not just capability. It requires evaluation frameworks that ask not just "does this model perform well?" but "for whom, and at whose expense?" It requires the AI industry to take seriously the distributional question — not just aggregate performance, but who the performance distribution actually covers.

Open source is a precondition for this kind of accountability. Edge deployment is a precondition for this kind of access. But they're conditions of possibility, not solutions. The work of actually building AI that serves long-term human benefit — across geographies, languages, income levels, and infrastructure realities — is a design challenge. It doesn't get solved by releasing model weights. It gets solved by a field that's asking better questions.

The question of who AI belongs to is, at its root, a question of who it was built for. Open source and edge AI give us the tools to expand that circle substantially. Whether we use those tools to actually do it is still an open question — in every sense of the word.

Sources