Skip to content
JP.

> initializing portfolio...

> loading projects [6/6]

> calibrating formation control...

> ready.

All Projects
COMPUTER VISIONHack-A-Bot 2025 — 3rd Place

AI Camera System

PythonComputer VisionAI Prototyping
AI Camera System

Overview

Conceptualised and built the system architecture for an AI-powered classroom assistant at Hack-A-Bot 2025, achieving 3rd place. I was the originator of the core concept: using pose estimation to simultaneously measure attendance and hand-raise behaviour — merging two previously separate manual tasks into one AI pipeline. I wrote the key conceptual skeleton of the code, handled final debugging, and resolved a critical compilation error with the Sony IMX500 AI Camera's pose-capture library interface. One teammate built the JSON/Flask web dashboard; another designed the CAD casing for the Pi and AI Camera.

Key Highlights

3rd place at University of Manchester Hackathon 2025
Led system architecture design and detection pipeline
Real-time student attendance and attentiveness tracking
Built end-to-end in under 24 hours

Gallery

The team at UoM Hackathon 2025 — built the AI classroom assistant in under 24 hours, achieved 3rd place
1 / 1

The team at UoM Hackathon 2025 — built the AI classroom assistant in under 24 hours, achieved 3rd place

Deep Dive

The Concept — My Idea

During brainstorming, I proposed using pose estimation to simultaneously measure attendance and hand-raise behaviour — merging two previously separate tasks into one AI pipeline. Traditional classrooms handle attendance via roll calls and engagement monitoring subjectively. Our system automates both in real-time through a single camera feed.

The core innovation: replacing two separate manual processes (roll call + subjective observation) with a single AI camera pipeline that does both in real-time.
The core innovation: replacing two separate manual processes (roll call + subjective observation) with a single AI camera pipeline that does both in real-time.

System Pipeline

The system runs in six stages: the Sony IMX500 AI Camera captures live video with its on-board neural processor; PoseNet runs on-device inference to extract 17 body keypoints per person; a confidence filter (threshold > 0.3) removes partial detections; the hand-raise detector checks if wrist keypoints are above shoulder keypoints; metrics compute attendance percentage, question count, and understanding count; finally, a Flask JSON endpoint at port 5050 serves live data while the annotated frame displays skeleton overlays with green circles above students with raised hands.

Complete 6-stage pipeline — from IMX500 capture through PoseNet inference, confidence filtering, hand-raise detection, metrics computation, to Flask JSON output.
Complete 6-stage pipeline — from IMX500 capture through PoseNet inference, confidence filtering, hand-raise detection, metrics computation, to Flask JSON output.

Hand-Raise Detection Logic

The detection is elegantly simple: PoseNet provides 17 keypoints per person, including wrist and shoulder positions. If either wrist keypoint is above its corresponding shoulder keypoint (lower Y coordinate in image space), the student is flagged as having their hand raised. A green circle is drawn above the head of detected students. This binary check runs at frame rate with negligible computational cost since the heavy lifting is done by PoseNet's on-device neural processor.

PoseNet 17-keypoint skeleton — hands down (left) vs hand raised (right). Detection checks if wrist Y < shoulder Y in image coordinates.
PoseNet 17-keypoint skeleton — hands down (left) vs hand raised (right). Detection checks if wrist Y < shoulder Y in image coordinates.

The Critical Bug Fix

There was a critical error preventing compilation — an issue with the Sony IMX500 AI Camera's pose-capture library interface. I found and resolved it, getting the final code running in time for the demo. This was the make-or-break moment of the hackathon — without the fix, we would have had no working demo.

Interactive Lab

Hand Detection — Classroom AI!

In 12 seconds, click all students with raised hands. After scanning, see your Precision (no false detections) and Recall (found all hands) — the same metrics the hackathon AI camera used.