A.R.C. app on wooden surface with kitchen interior

A.R.C. Archive. Ready. Cloud.

A computer vision system for residential asset documentation and insurance gap analysis. Designed, engineered, and shipped by a single builder.

Field Home Inventory Computer Vision Insurance Technology

Author Jeremy Prasatik Published: 2024 Status: V1 Live In market

Classification Product Design Brand Identity Full-Stack Engineering Go-to-Market

Abstract

Home inventory is a solved problem that nobody has solved well. The average American household contains approximately 300,000 items with a combined insurable value that most homeowners have never calculated. Existing documentation tools are spreadsheets with better packaging. Manual entry, manual categorization, manual everything. The math is predictable: 60% of homeowners are underinsured because they've never cataloged what they own.

A.R.C. applies computer vision to the problem. Point a camera at a room. The system identifies objects, estimates replacement value, assigns categories, and archives everything against a structured database. The financial layer calculates total documented assets against insurance policy limits and flags coverage gaps before a claim becomes necessary.

The entire product was designed, engineered, branded, and brought to market by one person working nights and weekends alongside a full-time creative director role. Python backend. Streamlit frontend. OpenAI Vision API for object recognition. Deployed on Vercel with Supabase handling data persistence. Concept to live product in weeks, not quarters.

SECTION 02: PROBLEM STATEMENT

The Documentation Barrier

60% of American homeowners are underinsured because they've never cataloged what they own. The tools that exist haven't changed the math. Manual entry remains the barrier.

Based on industry estimates of homeowner documentation rates and average coverage gaps.

A.R.C. app — smartphone mockup
A.R.C. app — room scanning interface
A.R.C. app — item detail view

The insurance industry operates on a fundamental asymmetry. Carriers know exactly what they'll pay on a policy. Homeowners rarely know what they'd need to claim. This gap widens with every purchase, every gift, every inherited piece that enters a home without documentation.

Standard homeowner's policies cover personal property at 50-70% of the dwelling coverage amount. A home insured at $400,000 carries roughly $200,000-$280,000 in personal property coverage. Whether that number is adequate depends entirely on whether the homeowner knows what they own and what it costs to replace. Most don't.

The documentation process is the barrier. Open a spreadsheet. Walk room to room. Describe each item. Research replacement values. Photograph everything. Attach receipts. The estimated time to properly inventory an average home: 40+ hours. The percentage of homeowners who complete this process: single digits.

The market has tried. Apps exist. They fall into two categories.

The first is the glorified spreadsheet. Manual entry fields. Manual photo attachment. Manual value assignment. The app adds a database and maybe a cloud sync, but the work is identical to the spreadsheet it replaced. The friction that prevents documentation remains fully intact.

The second is the insurance carrier tool. Built by or for specific insurers, locked to their ecosystem, designed primarily to streamline claims processing rather than empower the homeowner. The interface reflects the priority: functional, utilitarian, built for adjusters who already know what they're looking at.

Neither category addresses the core problem. Documentation is tedious because identification and valuation require human judgment on every single item.

Computer vision changes the input. Instead of describing what you own, you show it. The system observes, identifies, and classifies. The human role shifts from data entry to data review. Confirmation instead of creation.

This reframes the entire experience. The 40-hour inventory becomes a room-by-room scan measured in minutes. The barrier drops from prohibitive to trivial.

A.R.C. was built on this premise. Reduce the input friction to nearly zero. Let the technology handle observation. Let the human handle judgment.

A.R.C. mobile app screens overview

Every room tells a story. I built the system that remembers it.

A.R.C. app lifestyle scenes — mobile interfaces in context
SECTION 03: METHODOLOGY / HOW IT WORKS

System Architecture & Recognition Engine.

A single photograph triggers a six-stage recognition pipeline.

The image passes through vision processing, object identification, value estimation, and archival. Each stage feeds the next. Each decision point governed by confidence thresholds. Processing time measured under typical indoor lighting conditions.

Image Capture

Image Capture

User photographs a room or individual item using their device camera. No special hardware. No calibration. Standard smartphone optics.

Vision Processing

Vision Processing

OpenAI Vision API receives the image and returns structured analysis. Object identification, material detection, style classification, condition assessment, estimated era or manufacture period.

Value Estimation

Value Estimation

Identified objects are matched against market replacement data. The system estimates current replacement cost, not depreciated value or original purchase price. Replacement cost is the insurance-relevant metric.

Concept ValidationWk 1-2ArchitectureWk 3-4Interface Design + BuildWk 5-6Financial LayerWk 7-8Brand + Go-to-MarketWk 9-10Low Confidence → User ReviewMultiple Items → Split DetectionAmbiguous Category → Suggest Top 3Bad Lighting → Re-scan PromptPartial View → Merge RequestHigh Value → Documentation FlagDuplicate → Match & MergeCoverage Gap → Alert TriggerUserHomeRoomItemPhotoCategoryMaterialStyleEraConditionValueReceiptDocumentAppraisalCoveragePolicyGapReportPDF ExportConfidenceThresholdPythonStreamlitOpenAI Vision APISupabase PostgreSQLVercelClaude CodeSupabase AuthSupabase Storage01ImageCapture02VisionProcessing03ObjectIdentification04ValueEstimation05ArchiveEntry06FinancialAnalysisa.r.c.Ring IndexRecognition PipelineTechnology StackData ModelAI Decision PointsDevelopment TimelineData CategoriesUser / SpacesItems / AssetsAI AttributesFinancialInsurance
SECTION 03: METHODOLOGY / HOW IT WORKS

Classification System

Taxonomy designed for insurance relevance, not retail categorization. Each category maps to standard personal property claim classifications.

Sub-categories provide the granularity needed for accurate valuation without requiring specialized knowledge from the user. Aligned with classifications used by major U.S. carriers.

73ITEMS DOCUMENTEDAverage per home$49,630TOTAL VALUETracked assets$680AVG ITEM VALUEAcross categories13CATEGORIESClassification depthLivingBedroomBathKitchenHallwayLaundryGarageFurniture$680Electronics$425Artwork$580Appliances$890Fixtures$310Textiles$185Collectibles$695Vehicles$5,000Jewelry$1,200Living 90%Bedroom 80%Bedroom 90%Kitchen 80%Living 90%Hallway 70%Kitchen 90%Garage 80%Kitchen 90%Bath 80%Laundry 90%Living 80%Garage 90%Bedroom 80%Garage 90%Kitchen 10%Bedroom 80%Laundry 70%CaptureClassifyValidateValueArchiveMonitorInventory Analysis — Value × Room × Density

Scroll to explore →

A.R.C. classification system interface
SECTION 05: FINANCIAL INTELLIGENCE

Insurance Gap Analysis

Where the product stops being an inventory tool and becomes a risk management system.

Documented asset value compared against user-reported policy limits. Gap shown as a specific dollar amount before a disaster reveals it.

Every documented item contributes to a running total. That total is compared against the user's stated policy limit for personal property coverage. The math is simple. The insight is not.

Most homeowners set their personal property coverage when they purchase the policy and never revisit it. Meanwhile, the contents of their home change continuously. New furniture. Upgraded appliances. Gifts. Inherited pieces. A home that was adequately covered five years ago may be $50,000 underinsured today without the homeowner knowing.

A.R.C. makes the gap visible. Not as an abstract concept. As a specific dollar amount tied to specific documented items in specific rooms.

$49,630Documented valueTotal assets$38,000Policy limitCurrent coverage$11,630Coverage gapUnderinsured by23%Gap ratioOf total valueValuePolicy limit$0k$12k$25k$37k$50kAssetsPolicyFurnitureElectronicsArtworkAppliancesFixturesTextilesCollectiblesOther$12.6k$8.3k$7.5k$6.9k$5.0k$3.4k$3.3k$49,63046%47%95%71%92%87%86%$38,000By typeFurniture$12,580Electronics$8,320Artwork$7,540Appliances$6,890Fixtures$4,960Textiles$3,420Collectibles$3,280Other$2,640

Scroll to explore →

Insurance Coverage Analysis
SECTION 06: BRAND SYSTEM

Visual Identity & Design Language

The brand had to solve a tension. Home inventory sounds like a chore. Insurance analysis sounds like a meeting with your agent. Neither association invites engagement. The visual identity needed to make documentation feel like something worth doing, not something you should get around to eventually.

Editorial warmth applied to utility software. Magazine sensibility meets insurance rigor.

Chromatic brand circle

Primary

#B1BC94

RGB 177/188/148

Warm Register

#C4A265

Photography tones

Ground

#000000

Structure, text

Brand philosophy

The solution was editorial warmth applied to utility software. Magazine sensibility meets insurance rigor. The interface treats data as something worth designing, not just storing. Asset cards that feel like collection entries. Room views that read like curated galleries. Financial summaries that carry the weight of their content without the sterility of a spreadsheet.

The same philosophy extends to the documentation experience itself. Scanning a room should feel considered, not clinical. Reviewing your inventory should feel like browsing a personal archive, not auditing a warehouse. The brand language exists to make the practical feel purposeful.

Ogg Primary Typeface

Warm, editorial, slightly elevated. Carries the brand's emotional register. Headlines, feature names, moments of consideration.

Avenir Next Secondary Typeface, Medium

Clean, neutral, highly legible. Carries the product's utility layer. Data labels, navigation, body text, interface clarity.

Avenir Next Secondary Typeface, Demi Bold

Structural emphasis. Section labels, key data points, navigational hierarchy. Weight that signals importance without shouting.

Ogg Regular

Brand grey RGB: 177/188/148 #B1BC94 CMYK 34/16/50/0

Avenir Next Medium

Brand grey RGB: 177/188/148 #B1BC94 CMYK 34/16/50/0

Avenir Next Demi Bold

Brand grey RGB: 177/188/148 #B1BC94 CMYK 34/16/50/0

Avenir Next Regular

Brand grey RGB: 177/188/148 #B1BC94 CMYK 34/16/50/0

A.R.C. logo construction guidelines
A.R.C. app interface components
SECTION 07: BUILD METHODOLOGY

Solo Engineering Concept to Deployment

One person. Ten weeks. Concept to live product in the App Store.

AI-assisted development with Claude Code as the primary environment. Python backend, Streamlit frontend, deployed on Vercel. No engineering team, no QA department.

No engineering team. No product manager assigning tickets. No design review board. No QA department. One person identifying the problem, designing the solution, writing the code, testing the output, fixing what broke, and shipping the result.

This isn't a limitation narrative. It's a velocity argument. The feedback loop between identifying a problem and deploying a fix is measured in hours, not sprints. A UX friction point noticed during testing gets resolved in the same session. A feature idea that shows up during development gets prototyped immediately. The distance between intention and execution is as short as it can possibly be.

The tradeoff is real. Solo development means every decision is a prioritization decision. What ships now versus what ships later. What gets refined versus what gets functional. V1 is an honest assessment of those tradeoffs: comprehensive in scope, considered in design, pragmatic in implementation.

Claude Code served as the primary development environment. The workflow: describe the intended behavior in natural language. Review the generated code. Test the output. Refine through conversation. Ship.

This approach inverts the traditional bottleneck. The constraint is no longer syntax knowledge or framework expertise. It's clarity of intention. Knowing exactly what the product should do matters more than knowing exactly how to make it do it.

The same AI-assisted methodology that powers A.R.C.'s computer vision also powered its creation. A product built with AI, built to use AI, built by someone who understands both sides of that equation.

Week 1-2: Core concept validation. Can computer vision reliably identify household items from standard smartphone photographs? Testing across lighting conditions, angles, room types. The answer was yes, with caveats that informed the UX design.

Week 3-4: Product architecture. Database schema. User flow. Room and item data models. Authentication. Storage. The foundational decisions that everything else builds on.

Week 5-6: Interface design and implementation. Simultaneously designing and building. The luxury of solo development: no handoff gap between design intent and code reality.

Week 7-8: Financial layer. Insurance gap calculations. Policy limit comparisons. The feature that transforms a documentation tool into a risk management product.

Week 9-10: Brand identity. Visual system. Marketing site. Go-to-market preparation. Launch. Weeks, not months. Not quarters. Not fiscal years. Weeks.

Five A.R.C. application screens showing dashboard, item entry, room view, project selection, and document library
SECTION 08: PRODUCT INTERFACE

Application Views & Data Architecture

Every view designed to make the practical feel purposeful. Documentation as curated archive.

Dashboard, room, item detail, and report views. All screens reflect V1 production application with representative usage data.

Dashboard View

Dashboard View

The home screen presents aggregate intelligence. Total items documented. Total estimated value. Category breakdown. Coverage status. Recent activity. The information hierarchy prioritizes financial awareness: what you own, what it's worth, whether you're covered.

Room View

Room View

Each documented room functions as a contained archive. Items displayed as cards with thumbnail, name, category, and value. Sortable by value, category, or date added. The room becomes a gallery of your own possessions, organized for both browsing and analysis.

Document AI

Document AI

Upload a receipt, appraisal, warranty, or insurance document. AI extracts relevant details: purchase date, amount, vendor, coverage terms. The extracted data associates with the corresponding item automatically when possible, or prompts the user for assignment.

Reports

Reports

Generate PDF summaries for insurance review, estate planning, or personal reference. Configurable by room, category, or full home. Includes item photographs, descriptions, values, and aggregate statistics. Formatted for professional presentation to agents or advisors.

8-12HOURS MANUALTraditional inventory~30MINUTES A.R.C.Same 73 items16-24×FASTERDocumentation speed4VIEW TYPESDashboard to reportsPhasesCapture180mOrganize120mValue90mDocument60mReport30mReview40m8-12 hrs~30 minViewsDashboardOverviewValue summaryCategory scanCoverage checkRoom ViewNavigate roomsItem cardsSort/filterPhoto browseDocument AIReceipt scanData extractItem matchAuto-fileReportsPDF generateRoom filterPhoto compileFormatSpeedManual time distribution (minutes per phase)A.R.C.16-24×faster

Scroll to explore →

SECTION 09: USAGE DATA

Field Observations

Traditional manual inventory of a 73-item home: estimated 8-12 hours. A.R.C. documentation of the same scope: under 30 minutes. Reduction factor: 16-24x.

Metrics reflect V1 usage since launch. Early-stage numbers. Presented without inflation.

DEVELOPMENT TIMELINE

10 weeks, concept to launch

12345678910WeekPhaseConcept ValidationArchitectureInterface Design + BuildFinancial LayerBrand + Go-to-MarketCommitsStackSwift / SwiftUIPython / MLCore MLFirebaseVision APIStripeCloudKitFeaturesMarket researchDB schemaML pipelineCamera flowValuation engineGap analysisApp Store prepEffortLinesSwiftPythonConfig

Scroll to explore →

A.R.C. app lifestyle — smartphone and tablet
SECTION 10: CLOSING

Currently in Market V1 Live

Built because the person who made it needed it. A renovated house, years of collected objects, and nothing documented.

V1 live and in market. V2 roadmap includes native architecture migration, enhanced scanning precision, and deeper financial analysis.

Services

Product Design

Brand Identity

Full-Stack Engineering

Go-to-Market Strategy

Stack

Python

Streamlit

OpenAI Vision API

Supabase

Vercel

Claude Code

Links

arcready.app

heythere@arcready.app

A single person with design experience, AI tools, and a real problem to solve shipped a complete product. Not a prototype. Not a demo. A live application in the App Store with paying users and a roadmap.

V2 migrates the full stack to a native architecture. More precise scanning. Faster processing. Deeper financial analysis. The foundation built in V1 supports everything planned without a rebuild.

The larger point extends beyond this specific product. The tools exist now for designers who think in systems to build the systems they think about. The gap between vision and execution isn't technical anymore. It's about willingness to ship. A.R.C. shipped.