Today, I kicked off a brand-new computer vision project:
Real-Time Daily Activity Recognition based on webcam input and CNNs.
It’s my first time working on a project that’s meant to react live to human actions — like brushing teeth, eating, or typing — and I’m already hooked on the challenge.
brushing_teeth, eating, drinking, typing, reading, sleeping, walking, washing_face"person reading book" instead of "reading" to increase qualityPIL.Image.verifyimages/ directory (to later split into train/val)"person brushing teeth" yield way better results than vague ones like "brushing"ImageDataGeneratortrain/val with proper ratio (80/20)MobileNetV2 as a baseIt’s exciting to move beyond static classification and into something reactive.
This project feels a bit more personal — it’s recognizing the kind of daily behaviors I actually do.
Tomorrow I’ll dive into model training.
For now, the dataset is ready — and so am I.
“You can’t train a model on data you haven’t gathered.
Today was about gathering.”