# Copyright 2025-2026 Yakhyokhuja Valikhujaev # Author: Yakhyokhuja Valikhujaev # GitHub: https://github.com/yakhyo """Age and gender prediction on detected faces. Usage: python tools/attribute.py --source path/to/image.jpg python tools/attribute.py --source path/to/video.mp4 python tools/attribute.py --source 0 # webcam """ from __future__ import annotations import argparse import os from pathlib import Path from _common import get_source_type import cv2 from uniface.attribute import AgeGender from uniface.detection import SCRFD, RetinaFace from uniface.draw import draw_detections def draw_age_gender_label(image, bbox, sex: str, age: int): """Draw age/gender label above the bounding box.""" x1, y1 = int(bbox[0]), int(bbox[1]) text = f'{sex}, {age}y' (tw, th), _ = cv2.getTextSize(text, cv2.FONT_HERSHEY_SIMPLEX, 0.6, 2) cv2.rectangle(image, (x1, y1 - th - 10), (x1 + tw + 10, y1), (0, 255, 0), -1) cv2.putText(image, text, (x1 + 5, y1 - 5), cv2.FONT_HERSHEY_SIMPLEX, 0.6, (0, 0, 0), 2) def process_image( detector, age_gender, image_path: str, save_dir: str = 'outputs', threshold: float = 0.6, ): """Process a single image.""" image = cv2.imread(image_path) if image is None: print(f"Error: Failed to load image from '{image_path}'") return faces = detector.detect(image) print(f'Detected {len(faces)} face(s)') if not faces: return draw_detections(image=image, faces=faces, vis_threshold=threshold, corner_bbox=True) for i, face in enumerate(faces): result = age_gender.predict(image, face) print(f' Face {i + 1}: {result.sex}, {result.age} years old') draw_age_gender_label(image, face.bbox, result.sex, result.age) os.makedirs(save_dir, exist_ok=True) output_path = os.path.join(save_dir, f'{Path(image_path).stem}_age_gender.jpg') cv2.imwrite(output_path, image) print(f'Output saved: {output_path}') def process_video( detector, age_gender, video_path: str, save_dir: str = 'outputs', threshold: float = 0.6, ): """Process a video file.""" cap = cv2.VideoCapture(video_path) if not cap.isOpened(): print(f"Error: Cannot open video file '{video_path}'") return fps = cap.get(cv2.CAP_PROP_FPS) width = int(cap.get(cv2.CAP_PROP_FRAME_WIDTH)) height = int(cap.get(cv2.CAP_PROP_FRAME_HEIGHT)) total_frames = int(cap.get(cv2.CAP_PROP_FRAME_COUNT)) os.makedirs(save_dir, exist_ok=True) output_path = os.path.join(save_dir, f'{Path(video_path).stem}_age_gender.mp4') fourcc = cv2.VideoWriter_fourcc(*'mp4v') out = cv2.VideoWriter(output_path, fourcc, fps, (width, height)) print(f'Processing video: {video_path} ({total_frames} frames)') frame_count = 0 while True: ret, frame = cap.read() if not ret: break frame_count += 1 faces = detector.detect(frame) draw_detections(image=frame, faces=faces, vis_threshold=threshold, corner_bbox=True) for face in faces: result = age_gender.predict(frame, face) draw_age_gender_label(frame, face.bbox, result.sex, result.age) cv2.putText(frame, f'Faces: {len(faces)}', (10, 30), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 255, 0), 2) out.write(frame) if frame_count % 100 == 0: print(f' Processed {frame_count}/{total_frames} frames...') cap.release() out.release() print(f'Done! Output saved: {output_path}') def run_camera(detector, age_gender, camera_id: int = 0, threshold: float = 0.6): """Run real-time detection on webcam.""" cap = cv2.VideoCapture(camera_id) if not cap.isOpened(): print(f'Cannot open camera {camera_id}') return print("Press 'q' to quit") while True: ret, frame = cap.read() if not ret: break frame = cv2.flip(frame, 1) faces = detector.detect(frame) draw_detections(image=frame, faces=faces, vis_threshold=threshold, corner_bbox=True) for face in faces: result = age_gender.predict(frame, face) draw_age_gender_label(frame, face.bbox, result.sex, result.age) cv2.putText(frame, f'Faces: {len(faces)}', (10, 30), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 255, 0), 2) cv2.imshow('Age & Gender Detection', frame) if cv2.waitKey(1) & 0xFF == ord('q'): break cap.release() cv2.destroyAllWindows() def main(): parser = argparse.ArgumentParser(description='Run age and gender detection') parser.add_argument('--source', type=str, required=True, help='Image/video path or camera ID (0, 1, ...)') parser.add_argument('--detector', type=str, default='retinaface', choices=['retinaface', 'scrfd']) parser.add_argument('--threshold', type=float, default=0.6, help='Visualization threshold') parser.add_argument('--save-dir', type=str, default='outputs', help='Output directory') args = parser.parse_args() detector = RetinaFace() if args.detector == 'retinaface' else SCRFD() age_gender = AgeGender() source_type = get_source_type(args.source) if source_type == 'camera': run_camera(detector, age_gender, int(args.source), args.threshold) elif source_type == 'image': if not os.path.exists(args.source): print(f'Error: Image not found: {args.source}') return process_image(detector, age_gender, args.source, args.save_dir, args.threshold) elif source_type == 'video': if not os.path.exists(args.source): print(f'Error: Video not found: {args.source}') return process_video(detector, age_gender, args.source, args.save_dir, args.threshold) else: print(f"Error: Unknown source type for '{args.source}'") print('Supported formats: images (.jpg, .png, ...), videos (.mp4, .avi, ...), or camera ID (0, 1, ...)') if __name__ == '__main__': main()