Skip to content

Vision Extraction Reference

Reference documentation for the Vision Extraction system, including all status indicators, metrics, configuration variables, and output formats.


The Vision Extraction page displays the current configuration:

Status

Enabled or Disabled

Shows whether vision extraction is currently active

Vision Model

Model name (e.g., qwen2-vl)

The Ollama model being used for image analysis

Ollama Server

Server URL

The Ollama instance processing images


MetricDescription
PropertiesTotal number of properties in the database
With ImagesProperties that have at least one photo
Total ImagesTotal number of property photos across all listings
DescriptionsNumber of images with AI-generated descriptions
Visual EmbeddingsNumber of images with visual similarity embeddings
CoveragePercentage of images that have been indexed (Descriptions / Total Images × 100)
Coverage % = (Images with Descriptions / Total Images) × 100

Target: 100% coverage for complete searchability


Jobs progress through these states:

StatusBadge ColorDescription
pendingGrayJob queued but not started
runningBlueCurrently processing images
completedGreenSuccessfully finished
failedRedEncountered critical error
cancelledYellowStopped by user action

OptionTypeDefaultDescription
LimitIntegernull (all)Maximum number of properties to process
Skip visual embeddingsBooleanfalseGenerate descriptions only, skip image similarity embeddings

Ranked list of extracted feature tags with counts:

FieldTypeDescription
FeatureStringTag name (e.g., “hardwood floors”)
TypeStringCategory (material, view, style, appliance)
CountIntegerNumber of images with this feature

Classification breakdown showing photo distribution:

  • Kitchen
  • Living Room
  • Bedroom
  • Bathroom
  • Exterior
  • Dining Room
  • Office
  • Garage
  • Other

ColumnData TypeDescription
StatusStringJob completion state (completed, failed, cancelled)
StartedDateTimeTimestamp when job began (ISO 8601)
DurationTimeTotal processing time (HH:MM:SS)
PropertiesIntegerNumber of properties processed
ImagesStringFormat: “successful / failed” (e.g., “1,234 / 5”)
TagsIntegerTotal feature tags created across all images
Triggered ByStringUser or system identifier that started the job

Environment variables controlling vision extraction:

VariableTypeDefaultDescription
OLLAMA_BASE_URLURL-URL of your Ollama server (required)
VISION_MODELStringqwen2-vlModel name to use for image analysis
VISION_CONCURRENCYInteger4Number of parallel image processing tasks
Terminal window
OLLAMA_BASE_URL=https://ollama.supported.systems
VISION_MODEL=qwen2-vl
VISION_CONCURRENCY=8

{
"image_id": "property-123-photo-1",
"description": "Modern kitchen featuring granite countertops, stainless steel appliances, and pendant lighting over a large island",
"room_type": "kitchen",
"features": [
"granite countertops",
"stainless steel appliances",
"pendant lighting",
"kitchen island"
],
"generated_at": "2024-12-28T10:30:00Z"
}
{
"image_id": "property-123-photo-1",
"embedding": [0.023, -0.451, 0.782, ...], // 768-dimensional vector
"embedding_model": "qwen2-vl",
"generated_at": "2024-12-28T10:30:00Z"
}

Default processing constraints:

Limit TypeDefault ValueAdjustable Via
Concurrent Requests4VISION_CONCURRENCY
Max Batch SizeUnlimitedJob “Limit” option
Timeout per Image30 secondsServer configuration