Skip to content

Scripts for using Eulerian Video Magnification and neural networks as a baby breathing monitor

Notifications You must be signed in to change notification settings

miguelfreitas/eulerian-baby-monitor

Repository files navigation

eulerian-baby-monitor

Scripts for using Eulerian Video Magnification and neural networks as a baby breathing monitor.

Disclaimer: this is a functional set of scripts to implement a baby breathing monitor using IP camera and Linux computer, however it is not intended for end-users. Setting up the whole thing requires some programming skills and possibly retraining the neural network model with your data. I'm just documenting what I've hacked together to use with my newborn baby - I'm not providing support.

How it works

  1. Video from IP camera is streamed using RTSP protocol
  2. ffmpeg converts H264 RTSP streamed video to a v4l2 loopback device
  3. The forked branch of Realtime-Video-Magnification project performs realtime motion video magnification and outputs a "motion scalar" to stdout
  4. Motion data is piped into pipeInMagPredictAlarm.py script which runs a previously trained Keras model to predict either 1.0 (baby is breathing) or 0.0 (no baby / no breathing)
  5. In case of a long sequence of "no breathing" of ~30 seconds the alarm sounds (currently mplayer playing a sound, may be changed in alarm-start.sh)
  6. Alarm may be turned on/off from smartphone by acessing a simple web page (quite useful if computer is far away from the sleeping room)

Random notes

Video from IP camera

  • Set the IP camera for maximum bitrate/quality as possible since compression artifacts may cause false movement detection. I also use GOP size (intra-frame period) to 1. Alternatively one may try using MJPEG instead of MPEG4/H264 etc (I couldn't figure out how to get MJPEG from my camera).
  • Load kernel module: sudo modprobe v4l2loopback
  • ffmpeg command: ffmpeg -i rtsp://user:[email protected]:8554/profile0/media.smp -f v4l2 /dev/video0 (The actual URL is camera specific)

Training

The model is trained using treinaRede.py script. This script should be run with a "dataset params" file, which first line defines the global parameters:

inputLen (in frames),batchSize,modelNum

And further text lines will add previously acquired data for training of each tag:

tag(0 or 1),numRandomSlicesForTraining,dataFile

  • Current model uses inputLen=125 (5 seconds of 25 fps) and modelNum=2 (3 levels of "relu" with 0.2 dropout factor and sigmoid activation).
  • Model was trained using a single data value per frame (the "motion" scalar which is the sum of motion of all pixels). Since lighting conditions may affect the general noise level of the image (compression artifacts and stuff) the Realtime-Video-Magnification branch now outputs 2 values per line: overall motion and the standard deviation of pixels' motion. Scripts have not yet been updated to make use of this new data.
  • If camera or lightning conditions change the model may need to be retrained to fix false positive/negatives. I have found useful to save data of every night and eventually retrain the model to fix the misdetections.

Motion detection and alarm

Every night I use the following command:

./Realtime-Video-Magnification/src/rvm | python3 pipeInMagPredictAlarm.py --outputDate --alarm

This will execute RVM program (I still need to go through GUI menus to open camera and select Haar Wavelet magnification) and pipe motion data into script. The script will execute the previously trained model, creating a backup of the motion data name as current date and time and call the external alarm if required.

usage: pipeInMagPredictAlarm.py [-h] [--input INPUT] [--output OUTPUT]
                                [--outputDate] [--plot] [--alarm]
                                [--clipPlotVals]

Sample baby's sleeping motion data

Whole night baby's movement and breathing

Whole night baby's movement and breathing data. The blue line (vals) is the acquired data which is actually fed into the neural network model for prediction. The green line (pred) is the output of the model ranging from 1.0 (breathing or moving) to 0.0 (no movement / no baby). avgPred and medPred are smoothed versions of predicted pred data. Red dots (alarm) show whenever the alarms starts (1) or stops (0).

Detail of irregular (but normal) newborn's breathing pattern

Detail of a newborn's irregular breathing pattern. In this case the baby stopped breathing for about 5 seconds. This is perfectly normal (literature says that up to 10 seconds of no breathing is commonly seen in newborns).

About

Scripts for using Eulerian Video Magnification and neural networks as a baby breathing monitor

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published