Files
alexa-amundson-resume/roles/10-edge-computing-engineer.md
Alexa Amundson ec7b1445b5 kpi: auto-update metrics 2026-03-13
RoadChain-SHA2048: c645c1292ab1555e
RoadChain-Identity: alexa@sovereign
RoadChain-Full: c645c1292ab1555ebe6982915536d1c94701ff6bb16c20ed6ef4144eb50c9f984b4bfe5b9902109e8defd958d6be43ced8ec11cf95d6241536cd4da0b75f8fb48cbeb1b9f450c8f665b73d39e837d23e73e2ba4201af4dc40c02a34283efb04b39c612083465536f194f16adfadb1b56f714a65b918f40750f54eebf7724236861de173ec31963ff3b1b988d712be7e5acc3fe391eb804d3fdcfb9ccf77afc732660d23fff801f894318327eabf775eb4f4e67f7f22d07f23b0e17f6594cfe95b83b275fb7baaa97115e86562604fc5b47cc8024574b61396924e0ee2b7e454b0a1480c3076c7ad72408ceb4a75360d2d49c7d805c37ac5315af00e4a8ca2262
2026-03-13 23:16:12 -05:00

2.2 KiB
Raw Blame History

Alexa Amundson

Edge Computing Engineer

amundsonalexa@gmail.com | github.com/blackboxprogramming


Summary

Cloud inference is someone else's computer running your data. Deployed 27 AI models on-device across 5 Pi nodes with 52 TOPS acceleration, built a WiFi mesh for local connectivity, and kept it all running with self-healing automation.


Experience

BlackRoad OS | Founder & Edge Computing Engineer | 2025Present

The Vision: AI at the Edge, Not in the Cloud

  • 27 Ollama models (48.1 GB) running on 3 Pi 5 nodes — inference happens on-premise, data never leaves the network
  • 2x Hailo-8 NPUs (52 TOPS total) for hardware-accelerated inference — PCIe integration, driver management, firmware updates
  • 4 custom fine-tuned CECE models — personality, voice, and domain expertise that can't be replicated with off-the-shelf models

The Network: Mesh Connectivity Without Internet

  • RoadNet WiFi mesh: 5 APs on channels 1/6/11, 5 subnets (10.10.x.0/24), NAT through wlan0 — devices connect to fleet directly
  • WireGuard mesh for encrypted node-to-node communication. Tailscale overlay (9 peers) for remote management from anywhere
  • Pi-hole DNS for local resolution + custom zones (.cece, .blackroad) — edge services discoverable by name, not IP

The Challenge: Keeping Edge Alive

  • Edge hardware fails differently than cloud — SD cards degrade, power supplies sag, thermal throttling kills inference mid-response
  • Self-healing autonomy on every node. Power monitoring every 5 minutes. Automatic service restarts. Temperature alerts before shutdown

Technical Skills

Raspberry Pi, Hailo-8, Ollama, WireGuard, WiFi mesh, Pi-hole, Docker, Linux


Metrics

Metric Value Source
Fleet Nodes live fleet.sh — SSH probe to all nodes
Nodes Online live fleet.sh — SSH probe to all nodes
AI Models live services.sh — ollama list via SSH
Avg Temp live fleet.sh — /sys/class/thermal via SSH
Tailscale Peers live services.sh — tailscale status via SSH
Fleet Storage (GB) live fleet.sh — df via SSH