danger ball - rutgers ece · fig 4: video streams captured by the raspberry pi using the zoneminder...

1
Danger Ball Shaun Kotikalapudi, Eddie Lee, Ben Montone, Neeraj Venkatesan Advisor: Professor Christopher Rose Motivation Hardware Components Implementation Future Work Acknowledgements Inertial Measurement Units(IMU) Ultrasonic Sensors Cameras(Webcams) Arduino Uno(Microcontroller) Raspberry Pi(Mini-Computer) Wireless Router Fig5: ArduIMU V3 IMU Fig 6: MaXbotix Ultrasonic Distance Sensor Fig 7: Logitech C210 Webcam Fig 8: Arduino Uno R3 microcontroller Fig 9: Raspberry Pi Mini-Computer Fig 10: D-Link Wireless N150 Router Christopher Rose John Scafidi Steve Orbine John McCarthy WINLAB Develop a portable sensor package that helps layout the interiors of hazardous buildings. System Architecture Sensor package comprises of two distance sensors, two Inertial measurement unit(IMU), two video sensors(webcams), a microcontroller (Arduino) and a minicomputer The data collected by the sensors through the microcontroller and minicomputer are processed on the minicomputer and are transmitted wirelessly to an external processing station The processing station would fuse the data coming out of the sensor package and attempt to provide a coherent representation of the interior of the room. Ultrasonic Sensors The two ultrasonic sensors stream continuous distance measurements through the Arduino. Distance data is packaged into a .dat files comprising of two columns of data, one for each sensor The data coming out of the ultrasonic sensors is then plotted in real-time simultaneously for both sensors as shown in Fig. 7. Video Sensors Implemented simultaneous video stream capture through ZoneMinder application on Raspberry Pi. Set up a dedicated network server on the raspberry pi through a wireless router. The video streams captured by the Raspberry can be accessed wirelessly through an external computer as shown in Fig. 8. IMUs The two IMUs stream continuous accelerometer and gyroscope data through FTDI cable. Accelerometer and gyroscope data is packaged into .dat files comprising of 6 columns of data. The data coming of the IMUs is then plotted in real time simultaneously for both the IMUs as shown in Fig. 7. Fuse accelerometer and gyroscope data to gain a better understanding of sensor orientation using either the Kalman filter or the Complementary filter. Using the two video sensors to implement a panoramic image stitching algorithm and create a depth map with the distance sensors Researching portability (encasing and power source) Fig 3: Real-time plots of data coming out of the IMUs and ultrasonic sensors. The graphs in the first row are the x, y, z accelerometer data for the two IMUs. The graphs in the second row are the x, y, z gyroscope data for the two IMUs. The graph in the last row is the distance data coming out of the two ultrasonic sensors Fig 4: Video streams captured by the Raspberry pi using the ZoneMinder surveillance application. The video streams are being viewed wirelessly on an external computer through a custom network server set up using the wireless router Fig. 1: Proposed System Architecture Fig. 2: Demonstration System Architecture

Upload: others

Post on 22-May-2020

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Danger Ball - Rutgers ECE · Fig 4: Video streams captured by the Raspberry pi using the ZoneMinder surveillance application. The video streams are being viewed wirelessly on an external

Danger Ball Shaun Kotikalapudi, Eddie Lee, Ben Montone, Neeraj Venkatesan

Advisor: Professor Christopher Rose

Motivation   Hardware  Components  Implementation  

Future  Work  

Acknowledgements  

Ø Inertial Measurement Units(IMU) Ø Ultrasonic Sensors Ø Cameras(Webcams) Ø Arduino Uno(Microcontroller) Ø Raspberry Pi(Mini-Computer) Ø Wireless Router

Fig5: ArduIMU V3 IMU

Fig 6: MaXbotix Ultrasonic Distance

Sensor

Fig 7: Logitech C210 Webcam

Fig 8: Arduino Uno R3

microcontroller

Fig 9: Raspberry Pi

Mini-Computer

Fig 10: D-Link Wireless N150

Router

Ø Christopher Rose Ø John Scafidi Ø Steve Orbine Ø John McCarthy Ø WINLAB

Ø Develop a portable sensor package that helps layout the interiors of hazardous buildings.

System  Architecture  Ø Sensor package comprises of two

distance sensors, two Inertial measurement unit(IMU), two video sensors(webcams), a microcontroller (Arduino) and a minicomputer

Ø The data collected by the sensors through the microcontroller and minicomputer are processed on the minicomputer and are transmitted wirelessly to an external processing station

Ø The processing station would fuse the data coming out of the sensor package and attempt to provide a coherent representation of the interior of the room.

Ø Ultrasonic Sensors Ø The two ultrasonic sensors stream continuous distance measurements through the

Arduino. Ø Distance data is packaged into a .dat files comprising of two columns of data, one

for each sensor Ø The data coming out of the ultrasonic sensors is then plotted in real-time

simultaneously for both sensors as shown in Fig. 7. Ø Video Sensors

Ø Implemented simultaneous video stream capture through ZoneMinder application on Raspberry Pi.

Ø Set up a dedicated network server on the raspberry pi through a wireless router. Ø The video streams captured by the Raspberry can be accessed wirelessly through

an external computer as shown in Fig. 8. Ø IMU’s

Ø The two IMU’s stream continuous accelerometer and gyroscope data through FTDI cable.

Ø Accelerometer and gyroscope data is packaged into .dat files comprising of 6 columns of data.

Ø The data coming of the IMU’s is then plotted in real time simultaneously for both the IMU’s as shown in Fig. 7.

Ø Fuse accelerometer and gyroscope data to gain a better understanding of sensor orientation using either the Kalman filter or the Complementary filter.

Ø Using the two video sensors to implement a panoramic image stitching algorithm and create a depth map with the distance sensors

Ø Researching portability (encasing and power source)

Fig 3: Real-time plots of data coming out of the IMU’s and ultrasonic sensors. The graphs in the first row are the x, y, z accelerometer data for the two IMU’s. The graphs in the second row are the x, y, z gyroscope data for the two IMU’s. The graph in the last row is the distance data coming out of the two ultrasonic sensors

Fig 4: Video streams captured by the Raspberry pi using the ZoneMinder surveillance application. The video streams are being viewed wirelessly on an external computer through a custom network server set up using the wireless router

Fig. 1: Proposed System Architecture

Fig. 2: Demonstration System Architecture