Professional Documents
Culture Documents
5, OCTOBER 2016
767
I. I NTRODUCTION
2327-4662 2015 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
768
769
Fig. 2. EZ-VSN software architecture. On the left system components communicating with messages: the NPS, the NNS, the IP subsystem and the AM
subsystem. On the right the system pipeline in the ATC case.
Fig. 3. Message header format (in bytes). SN: sequential number, unique for
each message; NumPkts: number of packets a message is made of; PktId: packet
id, between 0 and NumPkts-1; MT: message type; SId: source node unique ID;
DId: destination node unique ID; LT: link type identifying the transmission
technology used; PayloadSize: size (in bytes) of the message payload.
770
Available: http://opencv.org/
Available: http://libjpeg-turbo.virtualgl.org/
4 [Online]. Available: www.boost.org
3 [Online].
771
Fig. 4. EZ-VSN GUI screenshot in a two camera setup. (a) Camera view in CTA mode: the compressed image is displayed. (b) Camera view in ATC mode:
keypoints are displayed. (c) Performance monitor: in the upper part the energy consumption, in the lower part the frame rate. (d) Camera settings.
Available: http://www.qt.io/
772
Available: http://elinux.org/Beagleboard:BeagleBone\_Capes
Available: http://www.radiumboards.com/HD\_Camera\_Cape\_
for\_BeagleBone\_Black.php
7 [Online].
TABLE I
C OMPARISON OF P OWER C ONSUMPTION IN D IFFERENT P OWER M ODES
773
TABLE II
EZ-VSN N ODE P OWER C ONSUMPTION IN D IFFERENT
H ARDWARE C ONFIGURATION (W)
B. Energy Profiling
To assess the energy performance of an EZ-VSN node, we
performed several experimental measurements coupling the
BeagleBone Black with different image acquisition and wireless communication devices and analyzing the different power
modes available on the ARM Cortex-A8 microprocessor. All
the power consumption tests were performed using an Adafruit
INA219 dc current sensor connected over I2C to an Arduino
Mega 2560. The power to the device under test is provided by
a third party 5 V 3 A external supply. The 0.1 measurement
resistor of the INA219 is placed in series between the external
power supply and the +5-V pin of the BeagleBone. The INA219
onboard ADC measures the dropout voltage across the resistor
and the voltage supplied to the device under test, thus giving
enough information to compute the current drawn by the device
and its power consumption.
As a first test, we compared the energy performance of the
BeagleBone Black with the one of the Raspberry PI 1 Model
B, a $20 USD, credit-card-sized platform which features a 700MHz ARM microprocessor and may thus constitute a viable
alternative for being used as a basis for the EZ-VSN node.
Table I reports the average power consumption for the two
tested architectures, in different activity modes when no external devices are attached. As one can see, the BeagleBone Black
is a clear winner for a number of reasons: first, it exhibits lower
power consumption in all activity modes; second, differently
from the Raspberry PI, it supports a sleep mode where the
power consumption is as low as 260 mW; finally, it features
a more powerful processor with tuneable base-clock frequency.
Having said that, we provide in Table II a detailed analysis of
the BeagleBone Black power consumption when coupled with
different devices, and in different configurations of microprocessor speed. The operating system running on the BeagleBone
Available: http://rcn-ee.net/deb/wheezy-armhf/v3.8.13-bone72/
774
V. A PPLICATION S CENARIOS
The flexibility of the proposed framework can be leveraged
to implement several different VSN applications. In particular, we focus on visual analysis applications, where the content
acquired by camera nodes is processed and analyzed to extract
high-level semantic concepts. In this section, we provide two
representative examples that have been implemented using the
EZ-VSN framework: object recognition and parking lot monitoring. In addition to describing how each application scenario
was implemented, we also exploit the framework to assess and
compare the performance of the novel ATC paradigm and the
traditional CTA approach. Such comparison demonstrates the
flexibility of the proposed system in implementing and evaluating practical solutions for VSNs. The following scenarios
and the corresponding performance evaluation have been carried out with an EZ-VSN camera node in the configuration that
requires the lowest amount of energy, that is coupled with the
RadiumBoard HD camera cape and the IEEE 802.15.4 compliant TelosB dongle. On the latter, we relied on the CSMA/CA
protocol with data acknowledgment provided by TinyOS at the
MAC layer and a routing protocol with ad hoc static routes
hard-coded on each node of the VSN.
A. Object Recognition
Many monitoring applications for VSNs require to recognize a particular object. Upon recognition other actions may be
performed, such as tracking the particular object or launching
an alarm. Object recognition is typically performed following
these steps: first, local visual features are extracted from the
image under processing. Such features are then matched against
a database of labeled features extracted from known objects to
find the most similar one. The particular type of features to be
used and how matching is performed can vary from case to case,
but the general process remains the same. It is clear that, in the
VSNs scenario, deciding where to perform feature extraction
plays a crucial role. In traditional systems, the CTA paradigm
is followed, i.e., the acquired image is first compressed relying
on a proper algorithm (e.g., JPEG), and then transmitted wirelessly to a central controller which performs feature extraction
and matching. Conversely, in the ATC paradigm, the camera
node may directly perform features extraction on the acquired
image, and then transmit a compressed version of such features
to the central controller for further matching. We implemented
both the CTA and ATC paradigms for object recognition using
EZ-VSN. The following paragraphs give details on the role of
each node in the VSN, depending on the chosen paradigm.
1) Compress-Then-Analyze: Upon receiving a STARTCTA
command from the central controller, a camera node acquires a
new image and compresses it with JPEG. The bitstream generated by the JPEG encoder is then encapsulated in either 802.11
or 802.15.4 frames (depending on the communication technology chosen from the GUI) and transmitted to the sink node in a
DATACTA message. Note that several application parameters,
such as the image resolution, the JPEG compression quality factor and the wireless technology to adopt for image transmission
are all contained in the STARTCTA message and can, therefore, be changed in real time from the GUI. Upon receiving the
775
Fig. 5. (a) Power consumption of the proposed VSN for object recognition in CTA mode and (b) ATC mode at a target application MAP of 0.58. For CTA, this
corresponds to a JPEG quality factor equal to 1. The same accuracy is reached in ATC with the transmission of 100 BRISK descriptors, reducing the transmission
time and the corresponding overall energy consumption.
Fig. 6. (a) Camera view representing the same sample object in CTA mode. (b) ATC mode. (c) Image reconstructed from local features.
776
Fig. 8. Parking lot monitoring: a feature vector is computed by means of a histogram of pixel hue values. A linear SVM is employed to classify each parking
lot as either vacant or occupied.
Fig. 9. Camera view in parking lot monitoring mode for (a) CTA paradigm and
(b) ATC paradigm.
Fig. 10. Parking lot monitoring application. Accuracy versus energy in ATC
and CTA mode. The energy is measured integrating power over time during
acquisition, extraction (ATC only), coding, and transmission phases. For the
ATC case 6,12,23,45,90,180, bins are used. For the CTA case, JPEG quality
factor is set to 7,10,20,30,40.
VI. C ONCLUSION
Reliable architectures for testing VSN solutions are key to
the success and the dissemination of such technology. In this
paper, we have presented EZ-VSN, a complete and flexible
framework to allow quick experimentation in the field of VSNs.
The proposed system is composed of: 1) a low-cost, low-power
hardware platform supporting both IEEE 802.11 and IEEE
802.15.4 wireless technology; 2) the software needed to operate the hardware platform, including a set of image processing
and networking primitives based on widely accepted and opensource C++ standards; and 3) a cross-platform GUI to control in
real time the performance of the VSN. We have demonstrated
the flexibility of EZ-VSN implementing two different application scenarios and evaluating the benefits of the novel ATC
paradigm compared to the traditional CTA one. The complete
framework is made publicly available for research purposed
at www.greeneyesproject.eu. As future research directions, we
plan to use the framework to test in a real-life scenario several
recently proposed solutions in the field of VSNs, including distributed extraction of visual features from overlapping fields of
view, and joint optimized multiview encoding and routing of
local features.
R EFERENCES
[1] E. Bentley, L. Kondi, J. Matyjas, M. Medley, and B. Suter, Spread spectrum visual sensor network resource management using an end-to-end
cross-layer design, IEEE Trans. Multimedia, vol. 13, no. 1, pp. 125131,
Feb. 2011.
[2] C. Li, J. Zou, H. Xiong, and C. W. Chen, Joint coding/routing optimization for distributed video sources in wireless visual sensor networks,
IEEE Trans. Circuits Syst. Video Technol., vol. 21, no. 2, pp. 141155,
Feb. 2011.
[3] N. Kouadria, N. Doghmane, D. Messadeg, and S. Harize, Low complexity DCT for image compression in wireless visual sensor networks,
Electron. Lett., vol. 49, no. 24, pp. 15311532, Nov. 2013.
[4] P. Wang, R. Dai, and I. Akyildiz, A spatial correlation-based image
compression framework for wireless multimedia sensor networks, IEEE
Trans. Multimedia, vol. 13, no. 2, pp. 388401, Apr. 2011.
[5] S. Paniga, L. Borsani, A. Redondi, M. Tagliasacchi, and M. Cesana,
Experimental evaluation of a video streaming system for wireless multimedia sensor networks, in Proc. 10th IFIP Annu. Mediterr. Ad Hoc Netw.
Workshop (Med-Hoc-Net), Jun. 2011, pp. 165170.
777
[6] A. Redondi, L. Baroffio, M. Cesana, and M. Tagliasacchi, Compressthen-analyze vs. analyze-then-compress: Two paradigms for image analysis in visual sensor networks, in Proc. IEEE 15th Int. Workshop
Multimedia Signal Process. (MMSP), Sep. 2013, pp. 278282.
[7] B. Girod et al., Mobile visual search, IEEE Signal Process. Mag.,
vol. 28, no. 4, pp. 6176, Jul. 2011.
[8] L. Baroffio, M. Cesana, A. Redondi, and M. Tagliasacchi, Performance
evaluation of object recognition tasks in visual sensor networks, in Proc.
26th Int. Teletraffic Congr. (ITC), Sep. 2014, pp. 19.
[9] L. Baroffio, M. Cesana, A. Redondi, M. Tagliasacchi, and S. Tubaro,
Coding visual features extracted from video sequences, IEEE Trans.
Image Process., vol. 23, no. 5, pp. 22622276, May 2014.
[10] I. Akyildiz, T. Melodia, and K. Chowdhury, Wireless multimedia sensor networks: Applications and testbeds, IEEE Proc., vol. 96, no. 10,
pp. 15881605, Oct. 2008.
[11] K. Abas, C. Porto, and K. Obraczka, Wireless smart camera networks for
the surveillance of public spaces, Computer, vol. 47, no. 5, pp. 3744,
May 2014.
[12] A. Seema and M. Reisslein, Towards efficient wireless video sensor
networks: A survey of existing node architectures and proposal for a flexiWVSNP design, IEEE Commun. Surv. Tuts., vol. 13, no. 3, pp. 462486,
Third Quart. 2011.
[13] S. Hengstler, D. Prashanth, S. Fong, and H. Aghajan, Mesheye: A
hybrid-resolution smart camera mote for applications in distributed intelligent surveillance, in Proc. 6th Int. Symp. Inf. Process. Sensor Netw.
(IPSN07), Apr. 2007, pp. 360369.
[14] W.-C. Feng, E. Kaiser, W. C. Feng, and M. L. Baillif, Panoptes:
Scalable low-power video sensor networking technologies, ACM Trans.
Multimedia Comput. Commun. Appl., vol. 1, no. 2, pp. 151167,
May 2005. [Online]. Available: http://doi.acm.org/10.1145/1062253.
1062256
[15] D. Lymberopoulos and A. Savvides, XYZ: A motion-enabled, power
aware sensor node platform for distributed sensor network applications,
in Proc. 4th Int. Symp. Inf. Process. Sensor Netw. (IPSN05), Los Angeles,
CA, USA, 2005, pp. 449454.
[16] M. Rahimi et al., Cyclops: In situ image sensing and interpretation in wireless sensor networks, in Proc. 3rd Int. Conf. Embedded
Netw. Sensor Syst. (SenSys05), 2005, pp. 192204 [Online]. Available:
http://doi.acm.org/10.1145/1098918.1098939
[17] R. Kleihorst, A. Abbo, B. Schueler, and A. Danilin, Camera mote
with a high-performance parallel processor for real-time frame-based
video processing, in Proc. IEEE Conf. Adv. Video Signal Based Surveill.
(AVSS07), Sep. 2007, pp. 6974.
[18] A. Kandhalu, A. Rowe, and R. Rajkumar, DSPCAM: A camera
sensor system for surveillance networks, in Proc. 3rd ACM/IEEE
Int. Conf. Distrib. Smart Cameras (ICDSC09), Aug 2009,
pp. 17.
[19] L. Nachman, J. Huang, J. Shahabdeen, R. Adler, and R. Kling, Imote2:
Serious computation at the edge, in Proc. Int. Wireless Commun. Mobile
Comput. Conf. (IWCMC08), Aug 2008, pp. 11181123.
[20] J. Schiller, A. Liers, and H. Ritter, Scatterweb: A wireless sensornet platform for research and teaching, Comput. Commun.,
vol. 28, no. 13, pp. 15451551, Aug. 2005 [Online]. Available:
http://dx.doi.org/10.1016/j.comcom.2004.12.044
[21] A. Redondi, M. Cesana, M. Tagliasacchi, I. Filippini, G. Dan, and
V. Fodor, Cooperative image analysis in visual sensor networks, Ad Hoc
Netw., vol. 28, pp. 3851, 2015.
[22] E. Eriksson, G. Dan, and V. Fodor, Real-time distributed visual feature extraction from video in sensor networks, in Proc. IEEE Int. Conf.
Distrib. Comput. Sensor Syst. (DCOSS), May 2014, pp. 152161.
[23] D. G. Lowe, Distinctive image features from scale-invariant keypoints,
Int. J. Comput. Vis., vol. 60, no. 2, pp. 91110, 2004 [Online]. Available:
http://dx.doi.org/10.1023/B:VISI.0000029664.99615.94
[24] H. Bay, A. Ess, T. Tuytelaars, and L. Van Gool, Speeded-up robust features (surf), Comput. Vis. Image Understand., vol. 110, no. 3, pp. 346
359, Jun. 2008 [Online]. Available: http://dx.doi.org/10.1016/j.cviu.2007.
09.014
[25] E. Rublee, V. Rabaud, K. Konolige, and G. Bradski, ORB: An efficient alternative to sift or surf, in Proc. Int. Conf. Comput. Vis.
(ICCV11), 2011, pp. 25642571 [Online]. Available: http://dx.doi.org/
10.1109/ICCV.2011.6126544
[26] L. Baroffio, A. Canclini, M. Cesana, A. Redondi, and M. Tagliasacchi,
Briskola: Brisk optimized for low-power arm architectures, in Proc.
IEEE Int. Conf. Image Process. (ICIP), Oct. 2014, pp. 56915695.
[27] A. Redondi, L. Baroffio, J. Ascenso, M. Cesana, and M. Tagliasacchi,
Rate-accuracy optimization of binary descriptors, in Proc. 20th IEEE
Int. Conf. Image Process. (ICIP), Sep. 2013, pp. 29102914.
778
Matteo Cesana received the M.S. degree in telecommunications engineering and Ph.D. degree in information engineering from the Politecnico di Milano,
Milan, Italy, in 2000 and 2004, respectively.
He is currently an Assistant Professor with
the Dipartimento di Elettronica, Informazione
e Bioingegneria, Politecnico di Milano. From
September 2002 to March 2003, he was a Visiting
Researcher with the Department of Computer
Science, University of California at Los Angeles
(UCLA), Los Angeles, CA, USA. His research
interests include design, optimization, and performance evaluation of wireless
networks with a specific focus on wireless sensor networks and cognitive radio
networks.
Dr. Cesana is an Associate Editor for the Ad Hoc Networks Journal
(Elsevier).