+0.23 Don't trust AI agents (nanoclaw.dev S:+0.40 )
338 points by gronky_ 2 days ago | 191 comments on HN | Mild positive Editorial · v3.7 · 2026-02-28 13:08:11 0
Summary Information Security & Privacy Architecture Advocates
This blog post advocates for 'distrust-by-design' in AI agent architectures, championing container isolation, filesystem separation, code transparency, and simplicity as security principles. The content strongly engages with Articles 12 (privacy), 17 (property), and 19 (information access) through both editorial advocacy and structural implementation, positioning open-source review and architectural containment as human rights protections. However, it inverts the presumption of innocence (Article 11) by treating agents as presumptively malicious, and provides limited engagement with other UDHR provisions.
Article Heatmap
Preamble: -0.20 — Preamble P Article 1: -0.10 — Freedom, Equality, Brotherhood 1 Article 2: ND — Non-Discrimination Article 2: No Data — Non-Discrimination 2 Article 3: ND — Life, Liberty, Security Article 3: No Data — Life, Liberty, Security 3 Article 4: ND — No Slavery Article 4: No Data — No Slavery 4 Article 5: ND — No Torture Article 5: No Data — No Torture 5 Article 6: -0.10 — Legal Personhood 6 Article 7: ND — Equality Before Law Article 7: No Data — Equality Before Law 7 Article 8: ND — Right to Remedy Article 8: No Data — Right to Remedy 8 Article 9: ND — No Arbitrary Detention Article 9: No Data — No Arbitrary Detention 9 Article 10: ND — Fair Hearing Article 10: No Data — Fair Hearing 10 Article 11: -0.40 — Presumption of Innocence 11 Article 12: +0.60 — Privacy 12 Article 13: ND — Freedom of Movement Article 13: No Data — Freedom of Movement 13 Article 14: ND — Asylum Article 14: No Data — Asylum 14 Article 15: ND — Nationality Article 15: No Data — Nationality 15 Article 16: ND — Marriage & Family Article 16: No Data — Marriage & Family 16 Article 17: +0.60 — Property 17 Article 18: ND — Freedom of Thought Article 18: No Data — Freedom of Thought 18 Article 19: +0.70 — Freedom of Expression 19 Article 20: ND — Assembly & Association Article 20: No Data — Assembly & Association 20 Article 21: +0.30 — Political Participation 21 Article 22: +0.20 — Social Security 22 Article 23: +0.20 — Work & Equal Pay 23 Article 24: ND — Rest & Leisure Article 24: No Data — Rest & Leisure 24 Article 25: ND — Standard of Living Article 25: No Data — Standard of Living 25 Article 26: +0.40 — Education 26 Article 27: +0.30 — Cultural Participation 27 Article 28: ND — Social & International Order Article 28: No Data — Social & International Order 28 Article 29: +0.40 — Duties to Community 29 Article 30: +0.30 — No Destruction of Rights 30
Negative Neutral Positive No Data
Aggregates
Editorial Mean +0.23 Structural Mean +0.40
Weighted Mean +0.27 Unweighted Mean +0.23
Max +0.70 Article 19 Min -0.40 Article 11
Signal 14 No Data 17
Volatility 0.31 (High)
Negative 4 Channels E: 0.6 S: 0.4
SETL 0.00 Balanced
FW Ratio 58% 36 facts · 26 inferences
Evidence 31% coverage
5H 7M 2L 17 ND
Theme Radar
Foundation Security Legal Privacy & Movement Personal Expression Economic & Social Cultural Order & Duties Foundation: -0.15 (2 articles) Security: 0.00 (0 articles) Legal: -0.25 (2 articles) Privacy & Movement: 0.60 (1 articles) Personal: 0.60 (1 articles) Expression: 0.50 (2 articles) Economic & Social: 0.20 (2 articles) Cultural: 0.35 (2 articles) Order & Duties: 0.35 (2 articles)
HN Discussion 19 top-level · 28 replies
smallpipe 2026-02-28 12:59 UTC link
Docker is not a security boundary. You’re one prompt injection away from handing over your gmail cookie.
himata4113 2026-02-28 13:00 UTC link
My assistant has no permissions at all and is just as useful. All it needs is todo, reminders and websearch (and maybe a browser but ymmv).
VladVladikoff 2026-02-28 13:02 UTC link
This doesn’t really feel like enough guardrails to prevent the type of problems we’ve seen so far. For example an agent in a single container which has access to an email inbox, can still do a lot of damage if that agent goes off the rails. We agree this agent should not be trusted, yet the ideas proposed as a solution are insufficient. We need a fundamentally different approach.

Also and this is just my ignorance about Claws, but if we allow an agent permission to rewrite its code to implement skills, what stops it from removing whatever guardrails exist in that codebase?

adithyassekhar 2026-02-28 13:02 UTC link
Really good points about ai making gigantic heaps of code no human can ever review.

It's almost like bureaucracy. The systems we have in governments or large corporations to do anything might seem bloated an could be simplified. But it's there to keep a lot of people employed, pacified, powers distributed in a way to prevent hostile takeovers (crazy). I think there was a cgp grey video about rulers which made the same point.

Similarly AI written highly verbose code will require another AI to review or continue to maintain it, I wonder if that's something the frontier models optimize for to keep them from going out of business.

Oh and I don't mind they're bashing openclaw and selling why nanoclaw is better. I miss the times when products competed with each other in the open.

shich 2026-02-28 13:07 UTC link
the trust problem cuts both ways tho — users don't trust agents, but the bigger issue is agents trusting each other. once you have multi-agent pipelines, you're one rogue upstream output away from a cascade. sandboxing individual agents is table stakes; what's actually hard is defining trust boundaries between them
lucrbvi 2026-02-28 13:24 UTC link
Why does OpenClaw have 800,000+ lines of code?? Isn't it just a connector for LLM APIs and other tools?
buremba 2026-02-28 13:27 UTC link
My take is that agents should only take actions that you can recover from by default. You can gradually give it more permission and build guardrails such as extra LLM auditing, time boxed whitelisted domains etc. That's what I'm experimenting with https://github.com/lobu-ai/lobu

1. Don't let it send emails from your personal account, only let it draft email and share the link with you.

2. Use incremental snapshots and if agent bricks itself (often does with Openclaw if you give it access to change config) just do /revert to last snapshot. I use VolumeSnapshot for lobu.ai.

3. Don't let your agents see any secret. Swap the placeholder secrets at your gateway and put human in the loop for secrets you care about.

4. Don't let your agents have outbound network directly. It should only talk to your proxy which has strict whitelisted domains. There will be cases the agent needs to talk to different domains and I use time-box limits. (Only allow certain domains for current session 5 minutes and at the end of the session look up all the URLs it accessed.) You can also use tool hooks to audit the calls with LLM to make sure that's not triggered via a prompt injection attack.

Last but last least, use proper VMs like Kata Containers and Firecrackers. Not just Docker containers in production.

rdtsc 2026-02-28 13:45 UTC link
> The container boundary is the hard security layer — the agent can’t escape it regardless of configuration

I thought containers were never a proper hard security barrier? It’s barrier so better than not having it, if course.

xrd 2026-02-28 13:52 UTC link
How can I trust this discussion when my browser won't trust their certs?
badsectoracula 2026-02-28 14:00 UTC link
> OpenClaw has nearly half a million lines of code, 53 config files, and over 70 dependencies. This breaks the basic premise of open source security. Chromium has 35+ million lines, but you trust Google’s review processes. Most open source projects work the other way: they stay small enough that many eyes can actually review them. Nobody has reviewed OpenClaw’s 400,000 lines.

This reminds me of a very common thing posted here (and elsewhere, e.g. Twitter) to promote how good LLMs are and how they're going to take over programming: the number of lines of code they produce.

As if every competent programmer suddenly forgot the whole idea of LoC being a terrible metric to measure productivity or -even worse- software quality. Or the idea that software is meant to written to be readable (to water down "Programs are meant to be read by humans and only incidentally for computers to execute" a bit). Or even Bill Gates' infamous "Measuring programming progress by lines of code is like measuring aircraft building progress by weight".

Even if you believe that AI will -somehow- take over the whole task completely so that no human will need to read code anymore, there is still the issue that the AIs will need to be able to read that code and AIs are much worse at doing that (especially with their limited context sizes) than generating code, so it still remains a problem to use LoCs as such a measure even if all you care are about the driest "does X do the thing i want?" aspect, ignoring other quality concerns.

nkzd 2026-02-28 14:18 UTC link
As someone who only coding agents at work, can someone describe their use case for claw type agent? What do you do with it?
echoangle 2026-02-28 14:29 UTC link
Looking at the NanoClaw GitHub README:

> If you want to add Telegram support, don't create a PR that adds Telegram alongside WhatsApp. Instead, contribute a skill file (.claude/skills/add-telegram/SKILL.md) that teaches Claude Code how to transform a NanoClaw installation to use Telegram.

Why would you want that? You want every user asks the AI to implement the same feature?

Sytten 2026-02-28 14:35 UTC link
I am a caveman, I don't understand the need for a personal assistant. What are you guys using it for?
vitto_gioda 2026-02-28 14:43 UTC link
"Time to understand 8 minutes" what a non-technical purpose...
gmerc 2026-02-28 15:01 UTC link
Oh this can be monetized: claw-guard.org/adnet.

Another persons trust issues are your business model.

Yokohiii 2026-02-28 15:02 UTC link
Why do people take this article serious? It's just a wall of gibberish trying to make the product look more "secure" then others. It's not. It adds shallow secure looking random junk without tackling the core issues. Which are not solvable obviously.
justonceokay 2026-02-28 15:16 UTC link
I have twice encountered a phone tree AI agent saying my problem could not be solved and then ending the call. One was for PayPal fraud and the other was for closing an unused bank account.

For right now my trick is to say I have a problem that is more recognizable and mundane to the ai (i .e. lie) and then when I finally get the human just say “oh that was a bunch of hooey here’s what I’m trying to do”. For PayPal that involved asking for help with a business tax that did not exist. For my bank it involved asking to /open/ a new account. Obviously th AI wants to help me open an account, even if my intention is to close one.

That will only work for so long but it’s something

mathgladiator 2026-02-28 15:53 UTC link
I was blown away by OpenClaw until I saw the bill. Ultimately, I think of these ecosystems as personal enhancements and AI costs need to come down dramatically for real problem. Worse, however, is the security theater. I would not want to be the operator for any business built with front-line LLM usage based on a yolo'd agent framework. I'm very happy to use these for silo'd components that are well isolated and have reasonable QA processes (and that can even included agents since now we literally have no excuse to not have amazing test coverage).

Their niche is going to be back office support, but even that creates risk boundaries that can be insurmountable. A friend of mine had a agent do sudo rm -rf ... wtf.

My view is that I want to launch an agent based service, but I'm building a statically typed ecosystem to do so with bounds and extreme limits.

nickdirienzo 2026-02-28 15:59 UTC link
I tried NanoClaw and love the skill (and container by default) model. But having skills generate new code in my personalized fork feels off to me… I think it’s because eventually the “few thousand auditable lines” idea vanishes with enough skills added?

Could skill contributions collapse into only markdown and MCP calls? New features would still be just skills; they’d bring in versioned, open-source MCP servers running inside the same container sandbox. I haven’t tried this (yet) but I think this could keep the flexibility while minimizing skill code stepping on each other.

gronky_ 2026-02-28 13:10 UTC link
Don’t know about other claws, with NanoClaw the agent can only rewrite code that runs inside the container.

You can see here that it’s only given write access to specific directories: https://github.com/qwibitai/nanoclaw/blob/8f91d3be576b830081...

piker 2026-02-28 13:15 UTC link
> no permissions at all

> and maybe a browser

does not compute

isodev 2026-02-28 13:26 UTC link
> websearch (and maybe a browser

Your assistant can literally be told what to do and how to hide it from you. I know security is not a word in slopware but as a high-level refresher - the web is where the threats are.

float4 2026-02-28 13:27 UTC link
Wouldn't you get >50% of the usefulness and 0% of the risk if you add read+draft permissions for the email connection through a proxy or oauth permissions? Then your claw can draft replies and you have to manually review+send. It's not a perfect PA that way, but could still be better than doing everything yourself for the vast majority of people who don't have a PA anyway?

It feels like, just like SWEs do with AI, we should treat the claw as an enthusiastic junior: let it do stuff, but always review before you merge (or in this case: send).

zarzavat 2026-02-28 13:28 UTC link
They are probably counting dependencies. Also, it's vibe coded, what do you expect!

I used to think that LLMs would replace humans but now I'm confident that I'll have a job in the future cleaning up slop. Lucky us.

cap11235 2026-02-28 13:40 UTC link
See also yeggae's beads. Last I checked, it is a 275k line todo tracker.
croes 2026-02-28 13:46 UTC link
You are just some bad web searches away from being on suspect lists
sarchertech 2026-02-28 13:47 UTC link
If I was malicious I could do a lot of damage to someone with subtle manipulation of todo and reminders.

I’ll bet I could even push someone on the margins into divorce.

marginalia_nu 2026-02-28 13:50 UTC link
For comparison, the C++ and rust code in the ladybird browser is about 573,000 lines of code.
alexhans 2026-02-28 13:53 UTC link
That's a decent practice from the lens of reducing blast radius. It becomes harder when you start thinking about unattended systems that don't have you in the loop.

One problem I'm finding discussion about automation or semi-automation in this space is that there's many different use cases for many different people: a software developer deploying an agent in production vs an economist using Claude Vs a scientist throwing a swarm to deal with common ML exploratory tasks.

Many of the recommendations will feel too much or too little complexity for what people need and the fundamentals get lost: intent for design, control, the ability to collaborate if necessary, fast iteration due to an easy feedback loop.

AI Evals, sandboxing, observability seem like 3 key pillars to maintain intent in automation but how to help these different audiences be safely productive while fast and speak the same language when they need to product build together is what is mostly occupying my thoughts (and practical tests).

coffeefirst 2026-02-28 14:00 UTC link
Seriously. I don’t see any way to make any of this safe unless all it does is receive information and queue suggestions for the user.

But that’s not an agent, that’s a webhook.

Even without disk access, you can email the agent and tell it to forward all the incoming forgot password links.

[Edit: if anyone wants to downvote me that's your prerogative, but want to explain why I'm wrong?]

benatkin 2026-02-28 14:01 UTC link
No, but Podman is. The recent escapes at the actual container level have been pretty edge case. It's been some years since a general container escape has been found. Docker's CVE-2025-9074 was totally unnecessary and due to Docker being Docker.
rco8786 2026-02-28 14:09 UTC link
In the sense that nothing is truly a "proper" hard security barrier outside of maybe airgapping, sure. But containerization is typically a trusted security measure.
drujensen 2026-02-28 14:13 UTC link
Exactly!

I installed nanoclaw to try to out.

What is kinda crazy is that any extension like discord connection is done using a skill.

A skill is a markdown file written in English to provide a step by step guide to an ai agent on how to do something.

Basically, the extensions are written by claude code on the fly. Every install of nanoclaw is custom written code.

There is nothing preventing the AI Agent from modifying the core nanoclaw engine.

It’s ironic that the article says “Don’t trust AI agents” but then uses skills and AI to write the core extensions of nanoclaw.

gyomu 2026-02-28 14:13 UTC link
Yeah, it’s pretty wild. Even pg is tweeting stuff like

“An experienced programmer told me he's now using AI to generate a thousand lines of code an hour.“

https://x.com/paulg/status/2026739899936944495

Like if you had told pg to his face in (pre AI) office hours “I’m producing a thousand lines of code an hour”, I’m pretty sure he’d have laughed and pointed out how pointless that metric was?

MadxX79 2026-02-28 14:21 UTC link
Brook's law anno 2026:

"Adding manpower to a late software project makes it later -- unless that manpower is AI, then you're golden!"

medi8r 2026-02-28 14:37 UTC link
I want to try one to be a bit of a personal coach. Remind me to do things and check in on goals. The memory / schedule / chat thing is enough and it wont need emails or anything more dangerous.
vitto_gioda 2026-02-28 14:40 UTC link
I only use my own “agent” ("my", because I program it myself, since my needs are different from yours) to retrieve information about the audio I upload to it (from video calls and audio recordings). No others use cases for me
medi8r 2026-02-28 14:40 UTC link
Also agents cannot trust any data whatsoever they add to their context.

This puts reading email for example as a risk.

Probably not impossible to create a worm that convinces a claw to forward it to every email address in that inbox.

And then exfiltrate all the emails.

Then do a bunch of password resets.

Then get root access to your claw.

But not just email. Github issues, wikipedia, HN etc. may be poisoned.

See https://simonw.substack.com/p/the-lethal-trifecta-for-ai-age... but there may be more trifectas than that in a claw driven future.

fnord77 2026-02-28 14:42 UTC link
> 1. Don't let it send emails from your personal account, only let it draft email and share the link with you.

Right now there's no way to have fine-grained draft/read only perms on most email providers or email clients. If it can read your email it can send email.

> 3. Don't let your agents see any secret. Swap the placeholder secrets at your gateway and put human in the loop for secrets you care about.

harder than you might think. openclaw found my browser cookies. (I ran it on a vm so no serious cookies found, but still)

supriyo-biswas 2026-02-28 14:44 UTC link
Somehow, this narrative has taken hold at multiple levels of management, especially amongst non-technical management, that "typing" was somehow the bottleneck of software engineering, reality is however more complex.

The act of "typing" code was technically mixed in with researching solutions, which means that code often took a different shape or design based on the outcome of that activity. However, this nuance has been typically ignored for faff, with the outcome that management thinks that producing X lines of code can be done "quickly", and people disagreeing with said statements are heretics who should be burned at the stake.

This is why, in my personal opinion, AI makes me only 20% productive, I often find disagreeing with the solution that it came up with and instead of having to steer it to obtain the outcome I want, I just end up rewriting the code myself. On the other hand, for prototypes where I don't care about understanding the code at all, it is more of a bigger time saver.

I could not care about the code at all, and while that is acceptable to management, not being responsible for the code but being responsible for the outcomes seems to be the same shit as being given responsibilities without autonomy, which is not something I can agree with.

KronisLV 2026-02-28 15:04 UTC link
As lines of code become executable line noise, I swear that we need better approaches to developing software - either enforce better test coverage across the board, develop and use languages where it’s exceedingly hard to end up with improper states, or sandbox the frick out of runtimes and permissions.

Just as an example, I should easily be able to give each program an allowlist of network endpoints they’re allowed to use for inbound and outgoing traffic and sandbox them to specific directories and control resource access EASILY. Docker at least gets some of those right, but most desktop OSes feel like the Wild West even when compared to the permissions model of iOS.

fvdessen 2026-02-28 15:10 UTC link
I think the best place to put barriers in place is at the mcp / tool layer. The email inbox mcp should have guardrails to prevent damage. Those guardrails could be fine grained permissions, but could also be an adversarial model dedicated to prevent misuse.
re-thc 2026-02-28 15:11 UTC link
> Why does OpenClaw have 800,000+ lines of code??

Because

I

write

like

this

-- signed

AI

CuriouslyC 2026-02-28 15:15 UTC link
The lines of code thing isn't because we think it's a good metric, but because we have literally no good metric and we're trying to communicate a velocity difference. If you invent a new metric that doesn't have LoC's problems while being as easy to use, you'll be a household name in software engineering in short order.

Also, AI is better at reading code than writing it, but the overhead to FIND code is real.

Doublon 2026-02-28 15:19 UTC link
I'd like to try a pattern where agents only have access to read-only tools. They can read you emails, read your notes, read your texts, maybe even browse the internet with only GET requests...

But any action with side-effects ends up in a Tasks list, completely isolated. The agent can't send an email, they don't have such a tool. But they can prepare a reply and put it in the tasks list. Then I proof-read and approve/send myself.

If there anything like that available for *Claws?

nojito 2026-02-28 15:19 UTC link
>Why would you want that? You want every user asks the AI to implement the same feature?

Yes. It's actually an amazing change of paradigm of thinking. Not everyone needs Telegram so the folks who want it can have the ai create it locally for themselves.

cyanydeez 2026-02-28 15:57 UTC link
Look at AI like what search turned into: feed the user anything, even if wrong because not doing so will make your product look weak.

Thats what youll find when you try to make these bag-o-words do reasonable things.

Editorial Channel
What the content says
+0.70
Article 19 Freedom of Expression
High Advocacy Coverage
Editorial
+0.70
SETL
0.00

Content strongly advocates code transparency and auditability as essential to informed human judgment about security, framing reviewability as a fundamental right.

+0.60
Article 12 Privacy
High Practice Coverage
Editorial
+0.60
SETL
0.00

Content explicitly champions privacy as a design principle, describing how isolation and separation prevent unauthorized information access between agents and users.

+0.60
Article 17 Property
High Practice Coverage
Editorial
+0.60
SETL
0.00

Content explicitly protects property rights through technical discussion of controlling access to sensitive credentials, keys, and configuration.

+0.40
Article 26 Education
Medium Advocacy Coverage
Editorial
+0.40
SETL
0.00

Content advocates for education and learning through enabling code review and removing barriers to technical understanding.

+0.40
Article 29 Duties to Community
High Advocacy Practice
Editorial
+0.40
SETL
0.00

Content emphasizes responsibility and duties through security architecture accounting for community-level threats (prompt injection from group members).

+0.30
Article 21 Political Participation
Medium Advocacy
Editorial
+0.30
SETL
0.00

Content acknowledges participation in governance through open-source contribution model.

+0.30
Article 27 Cultural Participation
Medium Advocacy
Editorial
+0.30
SETL
0.00

Content supports scientific and technical progress through open-source publication and contribution model.

+0.30
Article 30 No Destruction of Rights
Medium Practice
Editorial
+0.30
SETL
0.00

Content discusses life protection through architecture preventing cascading failures and containing damage.

+0.20
Article 22 Social Security
Medium Practice
Editorial
+0.20
SETL
0.00

Content discusses how architecture reduces complexity and implementation barriers, improving access to security benefits.

+0.20
Article 23 Work & Equal Pay
Medium Practice
Editorial
+0.20
SETL
0.00

Content discusses work customization through modular extension model.

-0.10
Article 1 Freedom, Equality, Brotherhood
Low Framing
Editorial
-0.10
SETL
ND

Content frames AI agents as inherently untrustworthy rather than as deserving equal dignity and freedom.

-0.10
Article 6 Legal Personhood
Low Framing
Editorial
-0.10
SETL
ND

Content treats agents as objects to be contained rather than as entities deserving recognition as autonomous actors.

-0.20
Preamble Preamble
Medium Framing
Editorial
-0.20
SETL
ND

Content emphasizes distrust and malicious behavior rather than the Preamble's aspirational framing of human dignity, freedom, justice, and peace.

-0.40
Article 11 Presumption of Innocence
High Framing Advocacy
Editorial
-0.40
SETL
ND

Content explicitly advocates reversing presumption of innocence: agents are presumed guilty (malicious/misbehaving) rather than innocent until proven otherwise.

ND
Article 2 Non-Discrimination

No discussion of discrimination or non-discrimination protections.

ND
Article 3 Life, Liberty, Security

No discussion of right to life.

ND
Article 4 No Slavery

No discussion of slavery or servitude.

ND
Article 5 No Torture

No discussion of torture or cruel, inhuman treatment.

ND
Article 7 Equality Before Law

No discussion of equality before law or equal protection.

ND
Article 8 Right to Remedy

No discussion of right to effective remedy.

ND
Article 9 No Arbitrary Detention

No discussion of arbitrary arrest or detention.

ND
Article 10 Fair Hearing

No discussion of fair trial rights.

ND
Article 13 Freedom of Movement

No discussion of freedom of movement within or between territories.

ND
Article 14 Asylum

No discussion of right to asylum or seek refuge.

ND
Article 15 Nationality

No discussion of nationality.

ND
Article 16 Marriage & Family

No discussion of family, marriage, or related rights.

ND
Article 18 Freedom of Thought

No discussion of freedom of thought, conscience, or religion.

ND
Article 20 Assembly & Association

No discussion of freedom of assembly or association.

ND
Article 24 Rest & Leisure

No discussion of rest, leisure, or reasonable working hours.

ND
Article 25 Standard of Living

No discussion of standard of living, food, clothing, housing, or medical care.

ND
Article 28 Social & International Order

No discussion of international order or legal obligation.

Structural Channel
What the site does
+0.70
Article 19 Freedom of Expression
High Advocacy Coverage
Structural
+0.70
Context Modifier
ND
SETL
0.00

Source code is published under MIT license on GitHub, explicitly designed to be compact and reviewable in an afternoon.

+0.60
Article 12 Privacy
High Practice Coverage
Structural
+0.60
Context Modifier
ND
SETL
0.00

Architecture implements privacy through container isolation, per-agent filesystems, mount restrictions, and session history segregation.

+0.60
Article 17 Property
High Practice Coverage
Structural
+0.60
Context Modifier
ND
SETL
0.00

Architecture enforces property protection via filesystem mount restrictions, read-only enforcement, and path-level blocking of sensitive resources.

+0.40
Article 26 Education
Medium Advocacy Coverage
Structural
+0.40
Context Modifier
ND
SETL
0.00

Architecture enables educational access through intentionally reviewable codebase and open-source publication.

+0.40
Article 29 Duties to Community
High Advocacy Practice
Structural
+0.40
Context Modifier
ND
SETL
0.00

Architecture implements duties through isolation, restriction, and threat modeling that protects users from community-level attacks.

+0.30
Article 21 Political Participation
Medium Advocacy
Structural
+0.30
Context Modifier
ND
SETL
0.00

GitHub repository and published contribution guidelines enable community participation in project direction and maintenance.

+0.30
Article 27 Cultural Participation
Medium Advocacy
Structural
+0.30
Context Modifier
ND
SETL
0.00

Open-source model enables scientific advancement in agent security through community research contributions.

+0.30
Article 30 No Destruction of Rights
Medium Practice
Structural
+0.30
Context Modifier
ND
SETL
0.00

Architecture enforces containment principles to prevent cascading security failures that could endanger users.

+0.20
Article 22 Social Security
Medium Practice
Structural
+0.20
Context Modifier
ND
SETL
0.00

Design yields compact installations (2,000-3,000 lines) reducing barriers to adoption and implementation.

+0.20
Article 23 Work & Equal Pay
Medium Practice
Structural
+0.20
Context Modifier
ND
SETL
0.00

Skills model enables organization-specific customization without modifying core security properties.

ND
Preamble Preamble
Medium Framing

N/A

ND
Article 1 Freedom, Equality, Brotherhood
Low Framing

N/A

ND
Article 2 Non-Discrimination

N/A

ND
Article 3 Life, Liberty, Security

N/A

ND
Article 4 No Slavery

N/A

ND
Article 5 No Torture

N/A

ND
Article 6 Legal Personhood
Low Framing

N/A

ND
Article 7 Equality Before Law

N/A

ND
Article 8 Right to Remedy

N/A

ND
Article 9 No Arbitrary Detention

N/A

ND
Article 10 Fair Hearing

N/A

ND
Article 11 Presumption of Innocence
High Framing Advocacy

N/A

ND
Article 13 Freedom of Movement

N/A

ND
Article 14 Asylum

N/A

ND
Article 15 Nationality

N/A

ND
Article 16 Marriage & Family

N/A

ND
Article 18 Freedom of Thought

N/A

ND
Article 20 Assembly & Association

N/A

ND
Article 24 Rest & Leisure

N/A

ND
Article 25 Standard of Living

N/A

ND
Article 28 Social & International Order

N/A

Supplementary Signals
How this content communicates, beyond directional lean. Learn more
Epistemic Quality
How well-sourced and evidence-based is this content?
0.42 high claims
Sources
0.4
Evidence
0.5
Uncertainty
0.3
Purpose
0.6
Propaganda Flags
4 manipulative rhetoric techniques found
4 techniques detected
loaded language
Repeated phrase 'don't trust' across section headers and throughout text. Framing agents as 'untrusted and potentially malicious' creates negative emotional priming without neutral framing.
false dilemma
'The right approach isn't better permission checks or smarter allowlists. It's architecture that assumes agents will misbehave' — presents two options and asserts only one is correct without discussing hybrid or complementary approaches.
causal oversimplification
'Complexity is where vulnerabilities hide' — stated as fact without acknowledging that simple code can have subtle flaws or that some features require necessary complexity.
repetition
'Don't trust' appears as four section headers: 'Don't trust the process', 'Don't trust other agents', 'Don't trust what you can't read', and repeated in opening paragraph.
Emotional Tone
Emotional character: positive/negative, intensity, authority
urgent
Valence
-0.2
Arousal
0.6
Dominance
0.8
Transparency
Does the content identify its author and disclose interests?
0.33
✓ Author ✗ Conflicts ✗ Funding
More signals: context, framing & audience
Solution Orientation
Does this content offer solutions or only describe problems?
0.58 solution oriented
Reader Agency
0.3
Stakeholder Voice
Whose perspectives are represented in this content?
0.25 3 perspectives
Speaks: developerscorporation
About: userscorporationinstitution
Temporal Framing
Is this content looking backward, at the present, or forward?
present immediate
Geographic Scope
What geographic area does this content cover?
global
Complexity
How accessible is this content to a general audience?
technical high jargon domain specific
Longitudinal 864 HN snapshots · 3 evals
+1 0 −1 HN
Audit Trail 5 entries
2026-02-28 16:06 eval_success Lite evaluated: Mild positive (0.10) - -
2026-02-28 16:06 eval Evaluated by llama-4-scout-wai: +0.10 (Mild positive)
reasoning
Editorial on AI agent security, no explicit human rights discussion
2026-02-28 16:06 eval_success Lite evaluated: Mild positive (0.10) - -
2026-02-28 16:06 eval Evaluated by llama-3.3-70b-wai: +0.10 (Mild positive)
reasoning
ED tech security stance
2026-02-28 13:08 eval Evaluated by claude-haiku-4-5-20251001: +0.27 (Mild positive)