Layout analysis, motion detection, and quality evaluation — indexed as searchable patterns across 26 MCP tools with sub-millisecond search.
N 35.6762
E 139.6503
100m
10.66ms
search latency
768d
vector dimensions
27
database models
16
search indexes
Scroll to explore
N 35.6762
Chapter I
What we analyze
7 analysis domains
analyze any site with MCP tools
structured data + vector embeddings
01domain
Structure
section classification
Layout Analysis
Every webpage has structure. Hero sections, feature grids, CTAs, pricing tables, footers -- Reftrix classifies each by type and encodes them as searchable vector embeddings. Query your own analyzed sites with sub-millisecond semantic retrieval.
hero / feature / cta / testimonial / pricing / footer / navigation / gallery -- classified via multilingual ML model, similarity search at < 11ms
Stored as 768D vector embeddings in pgvector
02domain
Motion
animation capture
Motion Detection
CSS transitions, JS animation libraries, and WebGL shaders -- three detection channels capture every animation on a page. Duration, easing curve, trigger condition, and library origin are all recorded.
Gradients, textures, and color fields extracted from every layer. Seven distinct background types cataloged from inline styles, stylesheet sources, and external CSS -- each encoded for pattern matching.
Every design conveys intent. Professional gravitas, playful energy, minimal clarity -- mood, tone, and emotional register are classified into sixteen distinct profiles via AI vision analysis and CSS heuristics.
Screenshots become searchable vectors. DINOv2 ViT-B/14 encodes each section and part as a 768-dimensional visual embedding -- enabling 'find by appearance' queries that text search cannot match.
DINOv2 ViT-B/14 visual embedding (768D L2-normalized) -- section + part dual embeddings -- cosine similarity with HNSW indexing
Visual embeddings alongside text embeddings
06domain
Parts
specimen collection
Part Decomposition
Buttons, headings, images, navigation bars, forms -- each UI element is extracted, classified into 16 part types, and dual-embedded with both text and visual vectors for precision component search.
16 part types (button / heading / image / navigation / card / form / icon / link / footer / ...) -- text + visual dual embedding -- boundingBox + computedStyles
12,784 parts extracted across analyzed sites
07domain
Responsive
cross-section survey
Responsive Mapping
Every site renders differently at 375px, 768px, and 1440px. Multi-viewport screenshots are captured, Pixelmatch computes visual diffs, and layout shifts are quantified across breakpoints.
Run page.analyze on any URL. Reftrix extracts layout structure, motion patterns, background designs, and design narratives -- encoding each as a 768-dimensional vector embedding for semantic similarity search. All data persisted to PostgreSQL + pgvector with HNSW indexing.
768
dimensions
<11ms
P95 search
27
DB models
26
MCP tools
DIM 768
Chapter II
Visual intelligence
find by appearance
DINOv2 + text dual embedding
Every section and UI part is captured, cropped, and processed through DINOv2 ViT-B/14 to produce a 768-dimensional visual embedding -- enabling similarity search by appearance, not just text.
visual
search mode
Find by appearance
Upload or describe a visual pattern. Reftrix searches across all analyzed sections using cosine similarity on DINOv2 embeddings, returning visually similar designs regardless of their text content.
60/40
vector / full-text
Hybrid RRF search
Combine visual and text signals with Reciprocal Rank Fusion. 60% vector weight and 40% full-text weight produce results that match both appearance and semantic meaning.
0.8
confidence threshold
Preference profiling
As you provide feedback on search results, Reftrix learns your design preferences. A 2-factor confidence model reranks future results based on mood category coverage and interaction history.
Part-level precision
Every UI element is extracted and classified into 16 part types. Each part carries dual embeddings -- text and visual -- for precision search. Compare styles, layout, interaction, and accessibility across parts.
16
part types
12,784
parts extracted
dual
embeddings
button
heading
image
navigation
card
form
icon
link
input
hero-image
logo
badge
avatar
video
divider
footer
E 139.6503
Chapter III
The instruments
Model Context Protocol 14 tools / 3 categories / Claude integration input validation / stdio transport
Unified analysis in a single pass. Extracts layout structure, motion patterns, background designs, and quality scores simultaneously. AI vision analysis for scroll-based design detection is enabled by default. Supports async mode for heavy sites and auto_timeout adjusts to page complexity.
Five phases transform unstructured web pages into a searchable vector space -- capture, map, observe, interpret, and index across 27 database models, 16 search indexes, and 768-dimensional dual embeddings.
0
Capture
Playwright + lazy scroll
27
DB models
Playwright captures full-page HTML and screenshots. Lazy scroll triggers IntersectionObserver-based loading before capture. HTML sanitized via DOMPurify. SSRF protection blocks private IPs. External CSS resolved and inlined.
PlaywrightDOMPurify 3.3SSRF ProtectionLazy Scroll
1
Map
Section + part classification
19
section types
Sections classified into 19 types via Vision LLM. Post-processor rules: split oversized sections (>10,000px), merge 3+ consecutive same-type, absorb empty unknowns. Parts extracted into 16 UI component types with bounding boxes.
Vision LLMRule 1-4 PostProcessor16 Part TypesJSDOM
2.5
Observe
Motion + quality + responsive
15px
scroll resolution
Scroll Vision analysis detects CSS/JS/WebGL animations at 15px frame resolution. Quality evaluation scores originality, craftsmanship, and contextuality. Multi-viewport responsive analysis captures layout shifts across breakpoints.
Scroll VisionQuality ScorerPixelmatch3 Viewports
4
Interpret
Narrative + mood profiling
16
mood categories
Vision LLM classifies design narrative into 16 mood categories. DB analysis: professional (141 sites), elegant (119), premium (103). Confidence-scored mood profiles with tone, color, and typography signals.