Reference Document

Dark Factory Transition Plan: 90 Days to Profitability

This document is the operating plan to move Pinehaven Ventures toward a dark factory structure where an engineering background is leveraged to design and automate AI-driven workflows — not to vibe code, but to build repeatable systems that drive each venture toward profitability.

Last updated: March 3, 2026. Evidence items below are presented as reported public claims and should be periodically re-validated.

This is engineering, not vibe coding

The goal of this plan is not to prompt AI into existence and hope for the best. Vibe coding — generating code without understanding what it does — breaks down the moment complexity increases, edge cases appear, or systems need to integrate.

Pinehaven's approach is the opposite: use a background in engineering to design the automation itself. That means writing precise specifications, building structured pipelines, defining acceptance scenarios, and evaluating outcomes with the same rigor an engineer applies to any system they build.

Vibe coding

  • Prompt and pray
  • Accept whatever the model produces
  • No structured evaluation
  • Breaks at scale

Engineered automation

  • Precise specs with acceptance criteria
  • Structured agent pipelines
  • Scenario-based evaluation at every stage
  • Designed by an engineer to scale reliably

What dark factories are

In the StrongDM framing, a tiny team can run a software factory where no human writes code and no human reviews code line-by-line. Humans produce precise Markdown specs, agents implement and test against behavioral scenarios, and humans only approve what ships based on outcomes.

  • 1.Humans define the spec and success criteria.
  • 2.AI agents build, test, and package shippable artifacts.
  • 3.Humans evaluate behavior, business impact, and risk before release.

Evidence frontier teams are ahead

  • Claude Code has publicly claimed that around 90% of its own codebase was generated by Claude Code itself.
  • Internal reports have suggested multiple teams operating in the 70-90% AI-written range, with some workflows approaching nearly all generated implementation.
  • Public claim: project lead Boris Cherny reportedly has not personally written code in more than two months.

Evidence many developers may be slower

  • METR 2025 randomized controlled trial (experienced OSS developers, familiar codebases) found tasks completed 19% slower with AI tools.
  • Participants expected to be 24% faster before starting, then still believed they were 20% faster after completion.
  • Interpretation: adoption alone is insufficient. Workflow design and evaluation architecture determine whether AI creates speed or drag.

Core argument: workflow and org design are the bottlenecks

Frontier teams did not just add better tools. They changed the unit of management from code production to specification precision and outcome evaluation. Teams that keep traditional code-review-heavy structures often stay stuck at partial gains.

  • Spec-writing discipline replaces ad hoc ticket grooming.
  • Scenario-based evaluation replaces line-by-line diff review for most changes.
  • Engineering management shifts from implementation supervision to flow control.
  • Coordination becomes the dominant source of friction.

Portfolio this plan is built around

These are the four live ventures included in the 90-day model. The goal across all ventures is profitability — driven by automated workflows that reduce cost-to-ship and accelerate customer acquisition.

Power Digital Intelligence

Institutional data center market intelligence and queue analytics

Primary buyers: Data center developers, institutional investors, utilities

Growth motion: Account-based outbound + investor/developer demos

View product page

Power Queue Tracker

ERCOT queue monitoring, digests, alerts, and exports

Primary buyers: Developers, consultants, and energy-adjacent teams

Growth motion: Digest-led self-serve conversion + enterprise upsell

View product page

ReelPost.ai

AI video generation and multi-platform short-form video publishing

Primary buyers: Creators, small businesses, and agencies

Growth motion: Creator funnel + agency multi-account upgrades

View product page

Crypto Transaction Log

Cross-exchange crypto import, organization, and exports

Primary buyers: Crypto investors and active traders

Growth motion: Demo-led acquisition + premium upgrade automation

View product page

Five Levels of Vibe Coding (Dan Shapiro, early 2026)

This framework is useful context, but Pinehaven's operating model is not about climbing the vibe coding ladder. It is about applying engineering discipline to AI automation — the human designs the system, the agents execute within it.

Level 0

Spicy Autocomplete

AI suggests next lines, but the human still writes the software.

Level 1

Coding Intern

AI handles scoped tasks; human reviews everything and owns architecture.

Level 2

Junior Developer

AI can make multi-file changes; the human still reads every diff. Most AI-native developers appear to be here.

Level 3

Developer as Manager

AI implements and the human mostly directs work plus reviews diffs. Many teams plateau here.

Level 4

Developer as PM

The human writes a clear spec, leaves, then returns to evaluate outcomes. Diff-reading mostly disappears.

90-day transition phases

Days 1-30

Build the factory rails
  • -Standardize Markdown spec templates with acceptance scenarios and business metrics.
  • -Set up agent pipeline (spec intake -> implementation -> scenario run -> deployment candidate).
  • -Ship at least 10 non-critical features fully through the spec-to-scenario path.
  • -Instrument baseline metrics: cycle time, rework rate, scenario pass rate, and deployment frequency.

Days 31-60

Move core workflows to dark-factory execution
  • -Route at least 50% of product roadmap items through spec-first automation.
  • -Replace line-by-line code review with scenario result review for non-critical changes.
  • -Launch weekly decision review: approve/hold/reject based on behavioral outcomes only.
  • -Connect release telemetry to business signals (trial starts, conversion, expansion, churn).

Days 61-90

Scale output and lock in profitability momentum
  • -Run 70%+ of implementation volume via dark-factory path (exception: security/compliance hot spots).
  • -Achieve sub-48-hour median spec-to-production cycle time for repeatable feature classes.
  • -Operate a single scorecard that ties delivery throughput to customer acquisition and retention.
  • -Formalize operating design for post-90-day operation: staffing model, incentives, and cadence.

Dark-factory sprint backlog by venture (Days 1-90)

Each product gets a dedicated spec backlog so execution speed maps to measurable business outcomes instead of generic feature output.

VentureDays 1-30Days 31-60Days 61-90Core KPI
Power Digital IntelligenceSpec packs for watchlist onboarding, queue-delta alerts, and proposal generation from live market data.Automate investor-facing intelligence report generation with scenario checks for data accuracy.Ship enterprise-ready API onboarding and expansion playbooks to increase annual contract close velocity.Annual licenses closed and time-to-first-insight under 24 hours
Power Queue TrackerBuild digest personalization and one-session watchlist setup with behavior-based acceptance tests.Deploy alert-quality scoring and queue anomaly detection to raise trust and reduce false positives.Roll out team collaboration and annual prepay experiments optimized for Team and Enterprise plans.Net new subscribers, activation in first 24 hours, and logo churn
ReelPost.aiCreate onboarding flow that gets first generated video posted in under 10 minutes for new trials.Launch auto A/B caption-hashtag generation and schedule reliability scenarios for multi-account posting.Ship agency operations dashboard and usage-driven upgrade nudges for Pro and Agency growth.Trial-to-paid conversion and weekly retained posting accounts
Crypto Transaction LogImprove CSV/XLS import success and standardize demo-to-signup conversion scenarios on web/mobile.Release tax export wizard, transaction auto-tagging, and premium prompt timing experiments.Introduce retention loops (portfolio digests, reminders) plus premium annual prepay packaging tests.Free-to-premium conversion, D30 retention, and import completion rate

Operating scorecard

Update this table weekly. If any metric is marked Off Track for two consecutive weeks, trigger a corrective action sprint. The focus is on operational excellence — profitability follows from consistently shipping quality at speed.

MetricBaselineDay 30Day 60Day 90CurrentStatus
Roadmap volume shipped via dark-factory path0%30%50%70%+Update weeklyOn Track
Median spec-to-production cycle timeUnknown (measure in week 1)<120 hours<72 hours<48 hoursUpdate weeklyOn Track
Scenario pass rate (first full run)Unknown (measure in week 1)85%+92%+95%+Update weeklyOn Track
Post-release defects per releaseUnknown (measure in week 1)<3<2<1Update weeklyOn Track
Customer acquisition trend (all ventures)Current baselinePositive week-over-weekSustained growthPath to profitability confirmedUpdate weeklyAt Risk
Paid subscriber retentionCurrent baselineBaseline establishedImproving trendHealthy retention ratesUpdate weeklyOn Track

Legacy migration approach

  • Start with repetitive, low-regret workflows before critical billing paths.
  • Create scenario harnesses that test behavior, not implementation details.
  • Freeze unstable interfaces while agents converge on repeatable output quality.
  • Use parallel runs (legacy flow vs factory flow) before full cutover.

Organizational redesign requirements

  • Shift sprint planning to spec planning with explicit behavioral acceptance rules.
  • Reduce default diff review and increase scenario quality audits.
  • Raise the bar for engineering roles toward product judgment, systems thinking, and evaluation rigor.
  • Adopt tiny, high-output teams with direct ownership of business outcomes.

Weekly operating rhythm

  1. Monday: prioritize spec backlog by expected business impact.
  2. Daily: run automated spec-to-ship pipeline and review scenario outcomes.
  3. Wednesday: inspect at-risk metrics and launch corrective experiments.
  4. Friday: scorecard review, release decisions, and customer acquisition trends.
  5. End of week: publish a one-page update with wins, misses, and next actions.

Unresolved tension to manage

The frontier is accelerating while the middle of the market risks falling behind. Success over the next 90 days depends less on model access and more on whether the engineering foundation is strong enough to consistently produce clear specs, evaluate outcomes honestly, and redesign operating structure around coordination efficiency. The differentiator is not AI access — it is the engineering discipline applied to the automation workflow itself.

Next action: run week-1 baseline instrumentation

Confirm baseline cycle time, defect rate, scenario coverage, and customer funnel health before launching full dark-factory rollout.