CLIP
ClipEmbed
Computes a CLIP embedding for each image in the dataset using the openai/clip-vit-large-patch14
model.
CLIP embedding is quite fast and well worth it, as it unlocks a lot of functionality.
This node computes CLIP embeddings efficiently in batch, and supports multi-GPU operation. Several other
nodes depend on CLIP embeddings, for example FuzzyDedupe
and SmartHumanFilter
.
You may also use beprepared to compute CLIP embeddings for your own purposes, which can be consumed out
of the image.json
sidecars in the output directory.
Parameters
batch_size
(default: 128): The number of images to process in parallel.target_property
(default: "clip"): The property to store the CLIP embedding in.
Output properties
image.{target_property}
: The CLIP embedding for the image.
Example
dataset >> ClipEmbed