feat: LinkedIn, Dev.to, onboarding emails, landing page

LinkedIn: 7 posts (founder story, psychology, market thesis, culture, contrarian, sovereignty, education)
Dev.to: 3 articles (Pi setup guide, architecture deep-dive, psychology cross-post)
Onboarding: 3-email post-deploy sequence (welcome → expand → identity)
Landing page: 'Own Your Stack' full page copy with A/B test plan

RoadChain-SHA2048: acf9a01550f3cb91
RoadChain-Identity: alexa@sovereign
RoadChain-Full: acf9a01550f3cb91373a2a43d02efc1a6ed557778584f96778e6db55f59d3f9be7579ac754e7851ae2384c4909ab2e3c354f15b72918a6c65b22a2003531290b41e80a375437714826a03df976f32e0b2e3f8124c0a3c1e9ae9ffeac6f9f1d4c21196c3e2afd67f01ec224425e75c23d1a89020123953925768d54384553962622d1527ca1a18501defbb22c1415e165a9d8201087ca9c99fb684648edb0c7ba709697602a9b6938325a01d19ec71178899568cbf5e01c14257be63aef06244157e9bb14320a2cc9cfbf8a610b66f262a08b3f234f0ba649cdc550ffab24088829a7a01519e546f467541f493cc87b6f7b90a352f69d9c8f88d228af2a486359
This commit is contained in:
2026-03-14 20:10:58 -05:00
parent 8ddac49849
commit 30967e4410
4 changed files with 912 additions and 0 deletions

View File

@@ -0,0 +1,276 @@
# BlackRoad LinkedIn Posts
**Principle:** Authority + Value-Expressive + Central Route (LinkedIn audience processes deeply)
**Tone:** Professional but not corporate. Founder voice. Data-driven.
---
## Post 1: The Founder Story
```
I left finance to build AI infrastructure on Raspberry Pis.
Not because I couldn't get a cloud budget.
Because I did the math.
One NVIDIA H100 on AWS: $3.90/hour.
Run it 24/7 for a year: $33,696.
For one GPU.
Two Hailo-8 accelerators: $198 total.
52 trillion operations per second.
Run them forever. No bill. No vendor.
I have a Series 7. A Series 24. A Series 65. A Series 66.
I passed the same licensing exams Wall Street requires — then used that financial rigor to ask a simple question: "Why are we renting compute we could own?"
The answer is that nobody told small teams they could own it. The cloud providers certainly didn't.
So I built BlackRoad OS: self-hosted AI that runs on $400 of Raspberry Pi hardware. Five nodes. Sixteen models. Thirty websites. A billing system. A git server. Fifty AI skills.
Monthly cloud bill: $0.
The self-hosted cloud market hit $18.48 billion in 2025 (Grand View Research). Edge AI is growing at 21.7% CAGR.
This isn't contrarian. This is where the math points.
#SelfHosted #EdgeAI #AI #Infrastructure #Founder
```
---
## Post 2: The Psychology Angle
```
I studied the Psychology of Advertising at the University of Minnesota (JOUR 4251, Dr. Claire Segijn).
One stat changed how I build marketing:
80% of advertisements are misunderstood by their audience.
Not rejected. Not ignored. Misunderstood.
The audience processes the ad and walks away believing something the advertiser never intended.
Four mechanisms cause this:
1. Omitted comparisons — "The best AI platform" (better than what?)
2. Pragmatic inference — "May be the best" (may also be the worst)
3. Juxtaposition — "Smart people choose X" (implies causation)
4. Affirmation of consequent — "If you want Y, you need X" (false logic)
So we made a rule at BlackRoad:
Every claim must be literally, specifically, and verifiably true.
"52 TOPS of neural inference" — true, measured.
"16 Ollama models on a Pi 5" — true, run `ollama list`.
"$0/month cloud bill" — true, there is no cloud.
We even killed our own "30K agents" marketing copy when we realized it was aspirational, not factual.
The truth effect (Fennis & Stroebe) says: the more people see a claim, the more true it seems. This works on lies AND truths.
So we repeat truths. Obsessively. With sources.
It's slower than hype marketing. It's also why our users stay.
#MarketingPsychology #Advertising #Transparency #AI
```
---
## Post 3: The Market Thesis
```
Three markets are converging on the same conclusion:
1. Self-hosted cloud: $18.48B in 2025, growing 11.9% CAGR → $49.67B by 2034
2. Edge AI: $24.91B in 2025, growing 21.7% CAGR → $118.69B by 2033
3. AI inference: $106.15B in 2025, growing 19.2% CAGR → $254.98B by 2030
The conclusion: inference belongs on the edge.
Training requires cloud-scale compute. Always will.
But inference — running models, answering queries, classifying data, powering agents — runs cheaper, faster, and more privately on hardware you own.
A Hailo-8 accelerator costs $99 and delivers 26 TOPS.
A cloud GPU costs $3.90/hour — you exceed the Hailo's price in 26 hours.
94% of IT leaders fear vendor lock-in (Parallels 2026). 42% are moving workloads back on-premises.
This isn't a prediction. This is happening right now, backed by $150B+ in market activity.
BlackRoad OS sits at the intersection: self-hosted AI inference on commodity edge hardware.
The shift doesn't require faith. It requires arithmetic.
Sources: Grand View Research, MarketsandMarkets, Parallels
#EdgeAI #SelfHosted #Infrastructure #AI #MarketAnalysis
```
---
## Post 4: The Hiring/Culture Post
```
Our entire AI infrastructure runs on 5 Raspberry Pis.
Here's what that says about how we build:
1. We solve problems with architecture, not budget.
A $55 computer is a datacenter node if you know Docker, WireGuard, and DNS.
2. We own everything.
Git server. Billing system. AI models. DNS filtering. Auth. Search.
Nothing is rented. Nothing can be revoked.
3. We measure in watts, not invoices.
46 watts total. That's less than a light bulb.
Not because we're cheap. Because efficiency is a design value.
4. We verify before we claim.
Every stat in our marketing is sourced. Every number is measured.
We killed our own copy when it wasn't accurate.
5. We document everything.
207 repos on Gitea. Every decision, every config, every architecture choice.
If I get hit by a bus, the system runs itself.
This is the culture that builds BlackRoad OS.
We're not hiring right now. But when we do, this is what we look for:
People who'd rather own $400 of hardware than rent $33,696 of someone else's.
#Culture #Startup #Engineering #AI #SelfHosted
```
---
## Post 5: The Contrarian Take
```
Unpopular opinion: most AI startups are paying 100x too much for inference.
Not training. Inference.
Training a foundation model requires H100 clusters, thousands of GPUs, millions of dollars. Fair.
But RUNNING a model? Answering a query? Classifying a document? Embedding text for search?
That runs on a $99 accelerator plugged into a $55 computer.
The AI inference market is $106 billion. Most of it is cloud inference — metered by the hour, billed by the token, scaled by the credit card.
But inference doesn't need the cloud. Inference needs:
- A model (free, open-source, download it)
- Hardware (Hailo-8: $99, Pi 5: $55)
- Software (Ollama: free, one command)
Total: $154. Runs forever.
The cloud GPU business model depends on you not knowing this.
AWS doesn't want you to run `ollama serve` on a Raspberry Pi. That's a $33,696/year customer they lose.
The edge AI market is growing at 21.7% because the secret is getting out.
BlackRoad OS is how we prove it works — 16 models, 50 skills, 30 websites, $0/month.
Not a pitch. A proof.
#AI #Inference #EdgeComputing #SelfHosted #Startups
```
---
## Post 6: The Data Sovereignty Angle
```
Your AI vendor's privacy policy is not a technical guarantee.
It's a promise. Promises change.
Technical guarantees look like this:
- Pi-hole DNS filtering blocks 120+ tracking domains at the network level
- WireGuard encrypts all inter-node traffic
- No port is open to the internet (Cloudflare Tunnels for ingress only)
- Models run on hardware in your physical possession
- Data never transits a third-party network
"Never leaves your network" is a physics statement, not a policy statement.
The difference matters when:
- A vendor changes their ToS (they will)
- A government requests data (they can)
- A breach exposes your prompts (it happens)
- An acquirer gets your usage data (it's an asset)
Self-hosted AI is not about distrust. It's about architecture.
The same way HTTPS doesn't mean "I don't trust the internet" — it means "I'm not relying on trust when I can use encryption."
Self-hosted means: I'm not relying on policy when I can use physics.
#DataSovereignty #Privacy #AI #Security #SelfHosted
```
---
## Post 7: The Education Flex
```
Most marketing teams optimize for clicks.
We optimize for accurate comprehension.
Sounds the same. It's not.
Optimizing for clicks means:
- Sensational headlines
- Vague claims that imply more than they state
- CTAs designed to create urgency
- Metrics: impressions, CTR, conversions
Optimizing for comprehension means:
- Headlines with verified stats and named sources
- Claims that are literally, specifically true
- CTAs that reduce uncertainty instead of creating urgency
- Metrics: reply rate, deploy rate, retention
Why? Because the psychology (ELM, Fennis & Stroebe) says:
Central-route persuasion — where the audience thinks carefully — produces attitude change that is DEEP and LASTING.
Peripheral-route persuasion — where the audience uses shortcuts — produces change that is REAL but TEMPORARY.
Clicks are peripheral. Deploys are central.
We'd rather have 100 users who understand what BlackRoad does than 10,000 who clicked a flashy ad and bounced.
That's not idealism. It's customer acquisition cost math.
#Marketing #Psychology #ContentStrategy #AI #Startup
```
---
## Posting Schedule
| Day | Post Type | Frequency |
|-----|-----------|-----------|
| Monday | Market data / thesis | Weekly |
| Wednesday | Technical / architecture | Weekly |
| Friday | Culture / founder story | Biweekly |
| Varies | Contrarian take | Monthly |
| Varies | Psychology / education | Monthly |
## Engagement Rules
- Reply to every comment within 4 hours
- Like every comment (signal that you're paying attention)
- Never argue — redirect with data
- If someone asks a question you can't answer honestly, say "I don't know" (builds more credibility than guessing)
- Cross-reference blog posts when relevant (drives owned traffic)