Hey there,
I’m working on a project to seamlessly integrate Blender’s motion tracking with ML video annotation.
Overview:
Harnessing tracking markers as primary annotation points, we aim for dynamic annotations that adjust with tracked objects, optimizing the video annotation process for ML.
Technical Approach:
- Tracking Markers: Direct utilization of
MovieClipTrackingTrack
andMovieClipTrackingMarker
. - Custom Properties: Embedding annotation metadata directly on markers.
- Visualization: Real-time annotation visualization tools in the Movie Clip Editor.
- Data Management: Systems for I/O of tracking and annotation data, targeting Pascal VOC format compatibility.
Aspirations:
- ML Model Comparison: Load and juxtapose overlays from different ML training sessions.
- ML Output Refinement: In-Blender tools for editing and refining ML model outputs.
- Interactive Annotation: Exploit motion tracking for on-the-fly annotation adjustments.
- Collaborative Annotation: Features to bolster team-based annotation workflows.
- Object & Pose Integration: Seamlessly import and refine object detection and pose detection results.
- Segment Anything: integrating automatic object segmentation and pytorch with Blender.
- Pie Menus: Intuitive pie menus for rapid label management, editing, and selection.
- Streamlined UI: A user interface tailored for efficiency, minimizing clicks and maximizing productivity.
- Extensibility: A modular design to allow future enhancements and integration with other ML tools.
Seeking Feedback On:
- Challenges: Potential pitfalls or challenges with this approach?
- Overlap: Know of any Blender projects or tools that might intersect with this?
- Technical Recommendations: Pointers on APIs, data structures, or best practices, especially around motion tracking?
Your insights would be invaluable. Thanks for your time – Blend on!
-Don