face_gesture_detector 0.1.0-dev
face_gesture_detector: ^0.1.0-dev copied to clipboard
A Flutter widget that translates camera frames and facial landmarks into high-level semantic callbacks, analogous to how GestureDetector translates pointer events into touch gestures.
Changelog #
All notable changes to this project will be documented in this file.
The format loosely follows Keep a Changelog and the project adheres to Semantic Versioning.
0.1.0-dev #
Added #
- Data Model:
FaceFrame,FaceBlendshape(52 values),PoseAngles,FaceLandmark,ImageQualityMetrics - Details objects:
FaceDetectedDetails,QualityDetails,DistanceDetails,PoseDetails,BlinkDetails,SmileDetails,MouthDetails,BrowDetails,HeadTurnDetails,HeadNodDetails - Configuration:
FaceGestureConfigurationwith calibrated defaults - Platform interface:
FaceGestureDetectorPlatform+MethodChannelFaceGestureDetector+processFrame() - Recognizers:
FacePresenceRecognizer,QualityGateRecognizer,PoseRecognizer,BlinkRecognizer,SmileRecognizer,MouthRecognizer,BrowRecognizer,HeadTurnRecognizer,HeadNodRecognizer,RawFrameRecognizer - Widgets:
RawFaceGestureDetector(Layer 2),FaceGestureDetector(Layer 1 facade) - Controller:
FaceGestureDetectorControllerwith pause/resume/reset - Native Android:
FaceLandmarkerEngine(MediaPipe LIVE_STREAM + GPU delegate),EulerAngleCalculator(pitch/yaw/roll from 4×4 matrix),FrameDecoder(NV21 → MPImage),SingleSlotFrameBuffer(lock-free backpressure),FaceFrameStreamHandler(EventChannel main-thread emission),face_landmarker.taskmodel bundled - Example app: MVP with all callback families and event log
- Tests: 117 Dart unit tests, 14 Android JVM unit tests, 3 integration tests on device