Today, I kicked off a brand-new computer vision project:
Real-Time Daily Activity Recognition based on webcam input and CNNs.
It’s my first time working on a project that’s meant to react live to human actions — like brushing teeth, eating, or typing — and I’m already hooked on the challenge.
brushing_teeth
, eating
, drinking
, typing
, reading
, sleeping
, walking
, washing_face
"person reading book"
instead of "reading"
to increase qualityPIL.Image.verify
images/
directory (to later split into train/val
)"person brushing teeth"
yield way better results than vague ones like "brushing"
ImageDataGenerator
train/val
with proper ratio (80/20)MobileNetV2
as a baseIt’s exciting to move beyond static classification and into something reactive.
This project feels a bit more personal — it’s recognizing the kind of daily behaviors I actually do.
Tomorrow I’ll dive into model training.
For now, the dataset is ready — and so am I.
“You can’t train a model on data you haven’t gathered.
Today was about gathering.”