Our video annotation solution combines machine learning and human-generated training data labels to track objects moving through space and time up to 100 times faster than human-only solutions.
On the first frame of a video, a human labeler annotates the objects in question. Functionally, this step is like a typical image annotation workflow. What makes this solution truly powerful is what comes next:
Using a deep learning ensemble model, our solution predicts where all annotated objects move in subsequent frames. Each individual label persists, even if there are dozens of instances of the same class. Instead of relabeling the entire image from scratch, a human labeler simply corrects the annotation if necessary, dragging or resizing the persisted label to squarely fit around the annotated object.
Our solution allows you to create an ontology of up to 255 classes, specific to your use case. We also support multiple instances in each class so you can label everything you need.
Once an annotator has labeled each object, an ensemble deep learning model persists those labels onto subsequent frames. This frees up annotators to make small corrections instead of labeling every object again, making our solution up to 100 times faster than human-only approaches.
We automatically parse your video into frames and reassemble it when the annotation is complete. You provide the video URL and customize your ontology, and our platform takes care of the rest.
Interested in trying out our Machine Learning Assisted Video Object Tracking solution?
Talk with one of our experts and we’ll help you get set up.