stfnets: learning sensing signals from the time-frequency ...lusu/papers/the experimental results...

11
STFNets: Learning Sensing Signals from the Time-Frequency Perspective with Short-Time Fourier Neural Networks Shuochao Yao 1 , Ailing Piao 2 , Wenjun Jiang 3 , Yiran Zhao 1 , Huajie Shao 1 , Shengzhong Liu 1 , Dongxin Liu 1 , Jinyang Li 1 , Tianshi Wang 1 , Shaohan Hu 4 , Lu Su 3 , Jiawei Han 1 , Tarek Abdelzaher 1 1 University of Illinois at Urbana-Champaign 2 University of Washington 3 State University of New York at Buffalo 4 IBM 1 {syao9, zhao97, hshao5, sl29, dongxin3, jinyang7, tianshi3, hanj, zaher}@illinois.edu 2 [email protected] 3 {wenjunji, lusu}@buffalo.edu 4 [email protected] ABSTRACT Recent advances in deep learning motivate the use of deep neural networks in Internet-of-Things (IoT) applications. These networks are modelled after signal processing in the human brain, thereby leading to significant advantages at perceptual tasks such as vision and speech recognition. IoT applications, however, often measure physical phenomena, where the underlying physics (such as inertia, wireless signal propagation, or the natural frequency of oscillation) are fundamentally a function of signal frequencies, offering better features in the frequency domain. This observation leads to a funda- mental question: For IoT applications, can one develop a new brand of neural network structures that synthesize features inspired not only by the biology of human perception but also by the fundamental nature of physics? Hence, in this paper, instead of using conven- tional building blocks (e.g., convolutional and recurrent layers), we propose a new foundational neural network building block, the Short- Time Fourier Neural Network (STFNet). It integrates a widely-used time-frequency analysis method, the Short-Time Fourier Transform, into data processing to learn features directly in the frequency do- main, where the physics of underlying phenomena leave better foot- prints. STFNets bring additional flexibility to time-frequency analysis by offering novel nonlinear learnable operations that are spectral- compatible. Moreover, STFNets show that transforming signals to a domain that is more connected to the underlying physics greatly simplifies the learning process. We demonstrate the effectiveness of STFNets with extensive experiments on a wide range of sensing inputs, including motion sensors, WiFi, ultrasound, and visible light. STFNets significantly outperform the state-of-the-art deep learning models in all experiments. A STFNet, therefore, demonstrates su- perior capability as the fundamental building block of deep neural networks for IoT applications for various sensor inputs 1 . CCS CONCEPTS Computing methodologies Machine learning; Com- puter systems organization Embedded and cyber-physical systems; Human-centered computing Ubiquitous and mobile computing. 1 Code is available on https://github.com/yscacaca/STFNets This paper is published under the Creative Commons Attribution 4.0 International (CC- BY 4.0) license. Authors reserve their rights to disseminate the work on their personal and corporate Web sites with the appropriate attribution. WWW ’19, May 13–17, 2019, San Francisco, CA, USA © 2019 IW3C2 (International World Wide Web Conference Committee), published under Creative Commons CC-BY 4.0 License. ACM ISBN 978-1-4503-6674-8/19/05. https://doi.org/10.1145/3308558.3313426 KEYWORDS Deep learning; time frequency analysis; Internet of Things; IoT ACM Reference Format: Shuochao Yao, Ailing Piao, Wenjun Jiang, Yiran Zhao, Huajie Shao, Shengzhong Liu, Dongxin Liu, Jinyang Li, Tianshi Wang, Shaohan Hu, Lu Su, Jiawei Han, Tarek Abdelzaher. 2019. STFNets: Learning Sensing Signals from the Time-Frequency Perspective with Short-Time Fourier Neural Net- works. In Proceedings of the 2019 World Wide Web Conference (WWW ’19), May 13–17, 2019, San Francisco, CA, USA. ACM, New York, NY, USA, 11 pages. https://doi.org/10.1145/3308558.3313426 1 INTRODUCTION Motivated by the needs of IoT applications, this paper presents a principled way of designing deep neural networks that learn (from IoT sensing signals) features inspired by the fundamental properties of the underlying domain of measurements; namely, properties of physical signals. We refer by IoT applications to those where sensors measure some physical quantities, generating (possibly complex and multi-dimentional) time-series data, typically reflecting some under- lying physical process. The human brain (whose wiring inspires the structure of conventional neural networks) extracts features well- suited for external perceptual tasks, which explains the great success of such networks at those tasks. In contrast, the internal physical processes underlying sensor measurements in IoT systems have prop- erties (such as physical intertia, characteristics of wireless signal propagation, and signal resonance) that depend more on signal fre- quency, motivating feature extraction in the frequency domain. It is no coincidence that much of classical signal processing literature works by transforming time-series data to the frequency domain first. To help capture signatures of internal physical processes the way a brain captures their externally perceived properties, this pa- per develops a new neural network block designed specifically for learning in the frequency domain. The design of neural network structures greatly influences effi- ciency of signal modelling and ease of extraction of hidden patterns. Convolutional neural networks (CNNs) for image recognition, for ex- ample, align perfectly with biological studies of the visual cortex [13] and with domain knowledge in digital image processing [7]. We thus ask a fundamental question: what structures are well-suited for the domain of physical sensor measurements, which we henceforth call the domain of IoT? Previous research on customizing deep learning models to the needs of IoT applications [16, 26, 31] mainly focused on designing neural network structures that integrate conventional deep learning components, such as convolutional and recurrent layers, to extract spatial and temporal properties of inputs. On the other hand, since the 2192

Upload: others

Post on 07-Sep-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: STFNets: Learning Sensing Signals from the Time-Frequency ...lusu/papers/The experimental results validate the design settings of STFNets and illustrate their superior accuracy compared

STFNets: Learning Sensing Signals from the Time-FrequencyPerspective with Short-Time Fourier Neural Networks

Shuochao Yao1, Ailing Piao2, Wenjun Jiang3, Yiran Zhao1, Huajie Shao1, Shengzhong Liu1,Dongxin Liu1, Jinyang Li1, Tianshi Wang1, Shaohan Hu4, Lu Su3, Jiawei Han1, Tarek Abdelzaher11University of Illinois at Urbana-Champaign 2University of Washington 3State University of New York at Buffalo 4IBM

1{syao9, zhao97, hshao5, sl29, dongxin3, jinyang7, tianshi3, hanj, zaher}@[email protected] 3{wenjunji, lusu}@buffalo.edu [email protected]

ABSTRACTRecent advances in deep learning motivate the use of deep neuralnetworks in Internet-of-Things (IoT) applications. These networksare modelled after signal processing in the human brain, therebyleading to significant advantages at perceptual tasks such as visionand speech recognition. IoT applications, however, often measurephysical phenomena, where the underlying physics (such as inertia,wireless signal propagation, or the natural frequency of oscillation)are fundamentally a function of signal frequencies, offering betterfeatures in the frequency domain. This observation leads to a funda-mental question: For IoT applications, can one develop a new brand ofneural network structures that synthesize features inspired not onlyby the biology of human perception but also by the fundamentalnature of physics? Hence, in this paper, instead of using conven-tional building blocks (e.g., convolutional and recurrent layers), wepropose a new foundational neural network building block, the Short-Time Fourier Neural Network (STFNet). It integrates a widely-usedtime-frequency analysis method, the Short-Time Fourier Transform,into data processing to learn features directly in the frequency do-main, where the physics of underlying phenomena leave better foot-prints. STFNets bring additional flexibility to time-frequency analysisby offering novel nonlinear learnable operations that are spectral-compatible. Moreover, STFNets show that transforming signals toa domain that is more connected to the underlying physics greatlysimplifies the learning process. We demonstrate the effectivenessof STFNets with extensive experiments on a wide range of sensinginputs, including motion sensors, WiFi, ultrasound, and visible light.STFNets significantly outperform the state-of-the-art deep learningmodels in all experiments. A STFNet, therefore, demonstrates su-perior capability as the fundamental building block of deep neuralnetworks for IoT applications for various sensor inputs 1.

CCS CONCEPTS• Computing methodologies → Machine learning; • Com-puter systems organization→ Embedded and cyber-physicalsystems; • Human-centered computing → Ubiquitous andmobile computing.

1Code is available on https://github.com/yscacaca/STFNets

This paper is published under the Creative Commons Attribution 4.0 International (CC-BY 4.0) license. Authors reserve their rights to disseminate the work on their personaland corporate Web sites with the appropriate attribution.WWW ’19, May 13–17, 2019, San Francisco, CA, USA© 2019 IW3C2 (International World Wide Web Conference Committee), published underCreative Commons CC-BY 4.0 License.ACM ISBN 978-1-4503-6674-8/19/05.https://doi.org/10.1145/3308558.3313426

KEYWORDSDeep learning; time frequency analysis; Internet of Things; IoTACM Reference Format:Shuochao Yao, Ailing Piao, Wenjun Jiang, Yiran Zhao, Huajie Shao,Shengzhong Liu, Dongxin Liu, Jinyang Li, Tianshi Wang, Shaohan Hu, LuSu, Jiawei Han, Tarek Abdelzaher. 2019. STFNets: Learning Sensing Signalsfrom the Time-Frequency Perspective with Short-Time Fourier Neural Net-works. In Proceedings of the 2019 World Wide Web Conference (WWW ’19),May 13–17, 2019, San Francisco, CA, USA. ACM, New York, NY, USA, 11 pages.https://doi.org/10.1145/3308558.3313426

1 INTRODUCTIONMotivated by the needs of IoT applications, this paper presents aprincipled way of designing deep neural networks that learn (fromIoT sensing signals) features inspired by the fundamental propertiesof the underlying domain of measurements; namely, properties ofphysical signals. We refer by IoT applications to those where sensorsmeasure some physical quantities, generating (possibly complex andmulti-dimentional) time-series data, typically reflecting some under-lying physical process. The human brain (whose wiring inspires thestructure of conventional neural networks) extracts features well-suited for external perceptual tasks, which explains the great successof such networks at those tasks. In contrast, the internal physicalprocesses underlying sensor measurements in IoT systems have prop-erties (such as physical intertia, characteristics of wireless signalpropagation, and signal resonance) that depend more on signal fre-quency, motivating feature extraction in the frequency domain. It isno coincidence that much of classical signal processing literatureworks by transforming time-series data to the frequency domainfirst. To help capture signatures of internal physical processes theway a brain captures their externally perceived properties, this pa-per develops a new neural network block designed specifically forlearning in the frequency domain.

The design of neural network structures greatly influences effi-ciency of signal modelling and ease of extraction of hidden patterns.Convolutional neural networks (CNNs) for image recognition, for ex-ample, align perfectly with biological studies of the visual cortex [13]and with domain knowledge in digital image processing [7]. We thusask a fundamental question: what structures are well-suited for thedomain of physical sensor measurements, which we henceforth callthe domain of IoT?

Previous research on customizing deep learning models to theneeds of IoT applications [16, 26, 31] mainly focused on designingneural network structures that integrate conventional deep learningcomponents, such as convolutional and recurrent layers, to extractspatial and temporal properties of inputs. On the other hand, since the

2192

Page 2: STFNets: Learning Sensing Signals from the Time-Frequency ...lusu/papers/The experimental results validate the design settings of STFNets and illustrate their superior accuracy compared

physics of measured phenomena are best expressed in the frequencydomain, decades of research on signal processing developed powerfultechniques for time-frequency analysis of signals, including motionsensor signals [12, 23], radio frequency signals [19, 25], acousticsignals [3, 8], and visible light signals [17]. A popular transformthat maps time-series measurements to the frequency domain is theShort-Time Fourier Transform (STFT). We, therefore, propose a newneural network model, namely, Short-Time Fourier Neural Networks(STFNets) that operate directly in the frequency domain.

One potential approach for learning in the frequency domainmight simply be to convert sensing signals into the frequency do-main first, and then apply conventional neural network components,possibly extending them to support operations on complex-numbersso they can represent frequency-domain quantities [24]. These ap-proaches miss two key opportunities for improvement, describedbelow, that we take advantage of in this work. As a result, our workleads to more accurate results, as shown in the evaluation section.The two reasons that account for our improvements are as follows.

First, different from traditional neural networks, where the inter-nal representations constitute features with no physical meaning,the internal representations in STFNet leverage frequency domain se-mantics that encode time and frequency information. All operationsand learnable parameters we propose are explicitly made compatiblewith the basic properties of spectral data, and align correspond-ing frequency and time components. In our design, we categorizespectral manipulations into three main types: filtering, convolution,and pooling. Filtering refers to the general spectral filtering andglobal template matching operation; convolution refers to the localmotif detection including shift detection and local template detec-tion; and pooling refers to dimension reduction over the frequencydomain. We then design the spectral-compatible parameters andoperating rules for these three manipulation categories respectively,which have shown superior performance in evaluations comparedto the application of conventional neural networks in the domain ofcomplex-numbers.

Second, transforming signals to the frequency domain is governedby the uncertainty principle [22]. The transformed representationcannot achieve both a high frequency resolution and a high timeresolution at the same time. In STFT, the time-frequency resolutionis controlled by the length of the sliding window (the length of thepart of the time-series being converted at a time). With a longerwindow, we can obtain a finer-grained frequency representation.However, we then cannot achieve a time resolution smaller than thewindow size. The uncertainty principle causes a dilemma in tradi-tional time-frequency analysis. One often needs to guess the besttime-frequency resolution using trial and error. In STFNet, we cir-cumvent this dilemma by simultaneously computing multiple STFTswith different time-frequency resolutions. The representations withdifferent time-frequency resolutions are then mutually enhanced in adata-driven manner. The network then automatically learns the bestresolution or resolutions, where the most useful features are present.STFNet defines a formal way to extract features from multiple time-frequency transformations with the same set of spectral-compatibleoperations and parameters, which greatly reduces model complexitywhile improving accuracy.

We demonstrate the effectiveness of STFNet through extensiveexperiments with various sensing modalities. During the evaluation,

we focus on device-based and device-free human activity recognitionwith a broad range of sensing modalities, including motion sensors(accelerometer and gyroscopes), WiFi, ultrasound, and visible light.The experimental results validate the design settings of STFNets andillustrate their superior accuracy compared to the state-of-the-artdeep learning frameworks for IoT applications.

Broadly speaking, the main contributions of this paper to thegeneral research landscape of deep learning and IoT are twofold:

(1) STFNet presents a principled way of designing neural net-works that reveal the key properties of physical processesunderlying the sensing signals from the time-frequency per-spective.

(2) STFNet unveils the benefit of incorporating domain-specificanalytic modelling and transformation techniques into theneural network design.

The rest of paper is organized as follows. Section 2 introducesrelated work on deep learning in the context of mobile sensing aswell as deep learning for spectral representations. We introduce thedetailed technical design of STFNet in Section 3. The evaluation ispresented in Section 4. Finally, we discuss the results in Section 5and conclude in Section 6.2 RELATEDWORKThe impressive achievements in image classification using deepneural networks at the turn of the decade [15] precipitated a re-emergence of interest in deep learning. Deep neural networks haveachieved significant accuracy improvements in a broad spectrum ofareas, including computer vision [11, 21], natural language process-ing [1, 4], and network analysis [14, 18].

Recent efforts applied deep learning in the context of IoT. In orderto improve the predictive accuracy of IoT applications, researchersemployed deep learning to model complicated sensing tasks [16, 26].In order to improve system efficiency at executing neural networkson low-end IoT devices, efforts have been made to compress modelparameters and/or structures in a manner that does not entail (almostany) accuracy loss [2, 10, 27, 32]. Recent work in the context of IoTalso addressed mathematical foundations for quantifying confidencein deep learning predictions to support mission-critical applications.The work produced deep neural networks that offer well-calibrateduncertainty estimates in results [5, 6, 28, 30]. Finally, the challenge ofinsufficient labeling of IoT data was addressed by introducing semi-supervised approaches for deep learning that allow neural networktraining using mostly unlabeled data [29]. However, none of theaforementioned IoT-inspired efforts addressed the customization oflearning machinery to a different signal space inspired by the physicsof measured processes; namely, the frequency domain.

To fill the above gap, recent work in machine learning focused onextending deep neural networks to complex numbers and spectralrepresentations. Trabelsi et al. propose deep complex networks, inves-tigating the complex-value neural network structure [24]. However,they mainly concentrate on the problems of initialization, normaliza-tion, and activation functions when extending real-valued operationsdirectly into the complex-value domain. Their designs focus moreon complex-value representations than spectral representations, anddo not take the properties of spectral data into consideration. Rippelet al. study spectral representations for convolutional neural net-works [20]. However, their study focuses on spectral parametrizing

2193

Page 3: STFNets: Learning Sensing Signals from the Time-Frequency ...lusu/papers/The experimental results validate the design settings of STFNets and illustrate their superior accuracy compared

Multi-resoultuion short-time Fourier Transform

X

Frequency components

Tim

e c

hu

nks

STFNet-Filtering STFNet-Convolution STFNet-Pooling

Hologram interleaving

Inverse short-time Fourier Transform

STFNet Block

Figure 1: The overview design of STFNet block.of standard CNNs, instead of designing operations customized forspectral data. In addition, their work treats input data fully from thefrequency perspective instead of the time-frequency perspective. Yaoet al. propose a neural network that takes short-time Fourier trans-formed data as inputs [26]. Yet their design uses traditional CNNsand RNNs, combining the real and imagery parts of complex-valueinputs as additional features.

To the best of our knowledge, STFNet is the first work that in-tegrates neural networks with traditional time-frequency analysis,and designs fundamental spectral-compatible operations for Fourier-transformed representations. Our study shows that the approachleads to improved accuracy compared to the state of the art. It impliesthat integrating neural networks with domain-inspired transforma-tion techniques (in our case, the Fourier Transform of physical time-series signals) projects input signals into a space that significantlyfacilitates the learning process.

3 SHORT-TIME FOURIER NEURALNETWORKS

We introduce the technical details of STFNets in this section. Weseparate the technical descriptions into six parts. In the first twosubsections, we provide some background followed by a high-leveloverview of STFNet components, including (i) hologram interleav-ing, (ii) STFNet-filtering, (iii) STFNet-convolution, and (iv) STFNet-pooling. In the remaining four subsections, we describe the technicaldetails of each of these components, respectively.

3.1 Background and STFNet OverviewIoT devices sample the physical environment generating time-seriesdata. Discrete Fourier Transform (DFT) is a mathematical tool thatconverts n samples over time (with a sampling rate of fs ) into a ncomponents in frequency (with a frequency step of fs/n). The moresamples are selected, the finer the component resolution is in fre-quency. We can always transform the whole sequence of data withDFT, achieving a high frequency resolution. However, we then loseinformation on signal evolution over time, or the time resolution. Inorder to solve this problem, Short-Time Fourier Transform (STFT)divides a longer time signal into shorter segments of equal lengthand computes DTF separately on each shorter segment. By losing acertain degree of frequency resolution, STFT helps us regain the timeresolution to some extent. In choosing n, there arises a fundamen-tal trade-off between the attainable time and frequency resolution,which is called the uncertainty principle [22]. For the purposes of

Filt

er

2

Mu

tua

l Sp

ect

ral

R

efi

ne

me

nt

Filt

er

1

Filt

er

4

Mu

tua

l Sp

ect

ral

R

efi

ne

me

nt

Filt

er

3

Inverse STFT

Dat

a T

Ime

Se

rie

s

Dim

en

sio

n #1

Dat

a T

Ime

Se

rie

s

Dim

en

sio

n #2

Dimension #1 data flow

Dimension #2 data flow

Inverse STFT

Inverse STFT

Inverse STFT

Inverse STFT

Inverse STFT

Inverse STFT

Inverse STFT

STFT HologramEnchanced

Hologram

Nonlinear

Activation Function

Figure 2: Data Flow within a block of STFNet.learning to pedict a given output, the optimal trade-off point dependson the time and frequency granularity of the features that best deter-mine the outputs we want to reproduce. The goal of STFNets is thusto learn frequency domain features that predict the output, while atthe same time learn the best resolution trade-off point in which therelevant features exist.

The building component of an STFNet is an STFNet block, shownin Figure 1. An STFNet block is the layer-equivalent in our neu-ral network. The larger network would normally be composed bystacking such layers. Within each block, STFNet circumvents theuncertainty principle by computing multiple STFT representationswith different time-frequency resolutions. Collectively, these rep-resentations constitute what we call the time-frequency hologram.And we call an individual time-frequency signal representation, ahologram representation. They are then used to mutually enhanceeach other by filling-in missing frequency components in each.

Candidate frequency-domain features are then extracted fromthese enhanced representations via general spectral manipulationsthat come in two flavors; filtering and convolution. They representglobal and local feature extraction operations, respectively. The fil-tering and convolution kernels are learnable, making each STFNetlayer a building block for spectral manipulation and learnable fre-quency domain feature extraction. In addition, we also design a newmechanism, called pooling, for frequency domain dimensionalityreduction in STFNets. Combinations of features extracted using theabove manipulations then pass through activation functions andan inverse STFT transform to produce (filtered) outputs in the timedomain. Stacking STFNet blocks has the effect of producing pro-gressively sharper (i.e., higher order) filters to shape the frequencydomain signal representation into more relevant and more fine-tunedfeatures.

Figure 2 gives an example of an SFTNet block that accepts asinput a two-dimensional time-series signal (e.g., 2D accelerometersdata). Each dimension is then transformed to the frequency domainat four different resolutions using STFT, generating four differentinternal nodes, each representing the signal in the frequency do-main at a different time-frequency resolution. Collectively, the fourrepresentations constitute the hologram. In the next step, mutualenhancements are done improving all representations. Each repre-sentation then undergoes a variety of alternative spectral manipu-lations (called “filters" in the figure). Two filters are shown in thefigure for each dimension. The parameters of these filters are theweights multiplied by the frequency components of the filter input; adifferent weight per component. These parameters are what the net-work learns. Note that, a filter does not change the time-frequency

2194

Page 4: STFNets: Learning Sensing Signals from the Time-Frequency ...lusu/papers/The experimental results validate the design settings of STFNets and illustrate their superior accuracy compared

resolution of the corresponding input. Filter outputs of the sametime-frequency resolution are then combined additively across alldimensions and passed through a non-linear activation function (asin a conventional convolutional neural network). An inverse STFTbrings each such combined output back to the time domain, whereit becomes an input to the next STFNet block. (Alternatively, theinverse STFT can be applied after dimension combination and beforethe activation function.) Hence, each output time-series is producedby applying spectral manipulation and fusion to one particular time-frequency resolution of all input time-series. Once converted to thetime domain, however, the output time-series can be resampled in thenext block at different time-frequency resolutions again. The goal ofSTFNet is to learn the weighting of different frequency componentswithin each filter in each block such that features are produced thatbest predict final network outputs.

3.2 STFNet Block FundamentalsIn this subsection, we introduce the formulation of our design ele-ments wihin each STFNet block. In the rest of this paper, all vectorsare denoted by bold lower-case letters (e.g., x and y), while matricesand tensors are represented by bold upper-case letters (e.g., X andY). For a vector x, the jth element is denoted by x[j]. For a tensor X,the t th matrix along the third axis is denoted by X[·, ·,t ], and otherslicing denotations are defined similarly. We use calligraphic lettersto denote sets (e.g., X and Y). For set X, |X| denotes the cardinality.

We denote the input to the STFNet block as X ∈ RT×D , wherewe divide the input D-dimension time-series into windows of sizeT samples. We call T the signal length and D the signal dimension.Since we concentrate on sensing signals, we assume that all the rawand internal-manipulated sensing signals are real-valued in timedomain.

As shown in Figure 1, the input signal X first goes through amulti-resolution short-time Fourier transform (Multi_STFT), whichis a compound traditional short-time Fourier transform (STFT), toprovide a time-frequency hologram of the signal. STFT breaks theoriginal signal up into chunks with a sliding window, where slidingwindowW(t ) with width τ only has non-zero values for 1 ≤ t ≤ τ .Then each chunk is Discrete-Fourier transformed,

STFT(τ ,s ) (X)[m,k,d ] =T∑t=1

X[t,d ] ·W(t − s ·m) · e−j2πkτ (t−s ·m ), (1)

where STFT(τ ,s ) (X) ∈ CM×K×D denotes the short-time Fouriertransform with width τ and sliding step s . M denotes the numberof time chunks. K denotes the number of frequency components.Since input signal X is real-valued, its discrete Fourier transformis conjugate symmetric. Therefore, we only need the ⌊τ/2⌋ + 1 fre-quency components to represent the signal, i.e., K = ⌊τ/2⌋ + 1. Inthis paper, we focus on sliding chunks with rectangular window andno overlaps to simplify the formulation, i.e., s = τ andM = T /τ . Wetherefore denote of short-time Fourier transform as STFT(τ ) (X).

The Multi_STFT operation is composed of multiple short-timeFourier transform with different window widths T = {τi }. The win-dow width, τi , determines the time-frequency resolution of STFT.Larger τi provides better frequency resolution, while smaller τi pro-vides better time resolution. In this paper, we set the window widthsto be powers of 2, i.e., τi = 2pi ∀pi ∈ Z+0 , to simplify the design later.

Hologram interleaving

fs/2fs/40 fs/20 fs/20fs/4 fs/4fs/8 3fs/8 fs/8 3fs/8

A en!on Merge for four elementsA en!on Merge for two elements

Figure 3: The design of hologram interleaving.We can thus formulate Multi_STFT as:

Multi_STFT(T ) {X} ={STFT(τi ) (X)

}for 2pi ∈ T . (2)

Next, according to Figure 1, the multi-resolution representationsgo into the hologram interleaving component, which enables therepresentations to compensate and balance their time-frequencyresolutions with each other. The technical details of the holograminterleaving component are introduced in Section 3.3.

The STFNet block thenmanipulates multiple hologram representa-tions with the same set of spectral-compatible operation(s), includingSTFNet-filtering, STFNet-convolution, and STFNet-pooling. We willformulate these operations in Section 3.4, 3.5, and 3.6, respectively.

Finally, the STFNet block converts the manipulated frequencyrepresentations back into the time domain with the inverse short-time Fourier transform. The resulting representations from differentviews of the hologram are weighted and merged as the input “sig-nal" for the next block. Since we merge the output representationsfrom different views of the hologram, we reduce the output featuredimension of STFNet-filtering and convolution operations by thefactor of 1/|T | to prevent the dimension explosion.

3.3 STFNet Hologram InterleavingIn this subsection, we introduce the formulation of hologram inter-leaving. Due to the Fourier uncertainty principle, the representationsin time-frequency hologram either have high time resolution or highfrequency resolution. The hologram interleaving tries to use repre-sentations with high time resolution to instruct the representationswith low time resolution to highlight the important components overtime. This is done by two steps:

(1) Revealing the mathematical relationship of aligned time-frequency components among different representations inthe time-frequency hologram.

(2) Updating the original relationship in a data-driven mannerthrough neural-network attention components.

We start from the definition of time-frequency hologram, gen-erated by Multi_STFT defined in (2). Note that, the window widthset T is defined as {2pi }, ∀pi ∈ Z+0 . Without loss of generality, anillustration of multi-resolution short-time Fourier transformed repre-sentations with input signal having length 16 and signal dimension3 as well as T = {4, 8, 16} are illustrated in Figure 3.

In order to find out the relationship of aligned time-frequencycomponents, we start with the frequency-component dimension.Since different representations only change the window width τiof STFT but not the sampling frequency fs of input signal, thesefrequency components represent frequencies from 0 to fs/2 (Nyquistfrequency) with step fs/τi . Then we can first obtain the relationshipof frequency ranging steps among different representations,

2195

Page 5: STFNets: Learning Sensing Signals from the Time-Frequency ...lusu/papers/The experimental results validate the design settings of STFNets and illustrate their superior accuracy compared

∀pi > pj ,fs /τjfs /τi

= 2pi−pj ∈ Z+0 . (3)

Therefore, a low frequency-resolution representation (with win-dow width 2pj ) can find their frequency-equivalent counterparts forevery 2pi−pj frequency components in a high frequency-resolutionrepresentation (with window width 2pi ). The upper part of Figure 3provides a simple illustration of such relationship. In the followinganalysis, we will use the original index k and corresponding fre-quency k · fs/τi interchangeably to recall the frequency componentfrom the time-frequency hologram STFT(τ ) (X)[m,k,d].

Next, we analyze the relationship over the time-chunk dimension,when two representations have frequency-equivalent components.Note that time chunks in STFT(τ ) (X) are generated by sliding rect-angular window without overlap. Based on (1), for representationshaving window widths τi = 2pi and τj = 2pj (pi > pj ),

STFT(τi ) (X)[m,2pi −pj k,d ] =2pi (m+1)∑t=2pim+1

X[t,d ] · e−j 2π 2pi −pj k

2pi (t−m·2pi ),

=

2pi −pj (m+1)−1∑mj=2

pi −pjm

2pj (mj+1)∑t=mj+1

X[t,d ] · e−j 2πk

2pj(t−m·2pj )

,

=

2pi −pj (m+1)−1∑mj=2

pi −pjm

STFT(τj ) (X)[mj ,k,d ] .

(4)

Therefore, given the equivalent frequency component, a time com-ponent in low time-resolution representation (with window width2pi ) is the sum of 2pi−pj aligned time components of the high time-resolution representation (with window width 2pj ). As a toy examplein Figure 3, the first row of the middle tensor is equal to the sum offirst two rows of the left tensor for frequencies 0, fs/4, and fs/2. Therow of the right tensor is equal to the sum of four rows of the lefttensor for frequencies 0, fs/4, and fs/2. The row of the right tensoris equal to the sum of two rows of the middle tensor for frequenciesfs/8 and 3fs/8, etc.

According to the analysis above, the high frequency-resolutionrepresentations lose their fine-grained time resolutions at certainfrequencies by summing the corresponding frequency componentsup over a range of time. However, the high time-resolution represen-tations preserve these information.

The idea of hologram interleaving is to replace the sum operationin high frequency-resolution representation with a weighted mergeoperation to highlight the important information over time. Fora certain frequency component, the weight of merging is learntthrough the most fine-grained information preserved in the time-frequency hologram. In this paper, we implement theweightedmergeoperation as a simple attentionmodule. For amerging input z ∈ CS×1,where S is the number of elements to be merged, the merge operationis formulated as:

a = softmax( |Wmz |),

y = S × a⊺z,(5)

where | · | is the piece-wise magnitude operation for complex-numbervector; andWm ∈ C

S×S is the learnable weight matrix. Notice thatthe final merged result is rescaled by the factor S to imitate the “sum"property of Fourier transform.

STFNet-filtering

fs/2fs/40 fs/20 fs/20fs/4 fs/4fs/8 3fs/8 fs/8 3fs/8

Linear/Spectral interpolation

WfFilter Kernel

Figure 4: The STFNet-filtering operation.3.4 STFNet-Filtering OperationStarting from this subsection, we will introduce our three spectral-compatible operations in STFNet. In each subsection, the introduc-tion includes two main parts: 1) the basic formulation of proposedspectral-compatible operation, and 2) extending a single operationto multi-resolution data.

Spectral filtering is a widely-used operation in time-frequencyanalysis. The STFNet-filtering operation replaces the traditional man-ually designed spectral filter with a learnable weight that can updateduring the training process. Although the spectral filtering is equiv-alent to the time-domain convolution according to convolution the-orem 2, the filtering operation helps to handle the multi-resolutiontime-frequency analysis, and facilitates the parameterization andmodelling. We denote the input tensor as X ∈ CM×K×D , whereM is the number of time chunk, K frequency component number,and D input feature dimension. The STFNet-filtering operation isformulated as:

Y[m,k, ·] = X[m,k, ·]Wf [k, ·, ·], (6)

where Wf ∈ CK×D×O is the learnable weight matrix, O the output

feature dimension, and Y ∈ CM×K×O the output representation.The function of STFNet-filtering operation is providing a set of

learnable global frequency template matchings over the time. How-ever, it is not straightforward to extend the matching operation to therepresentations with different time-frequency resolutions. Althoughwe can create multipleWf with different frequency resolutions K ,it can introduce unnecessary complexity and redundancy.

STFNet-filtering solves this problem by interpolating the fre-quency components in weight matrix. As we mentioned in Sec-tion 3.3, data in hologram with different frequency resolutions havethe same frequency range (from 0 to fs/2) but different frequencysteps (fs/τ ). Therefore, STFNet-filtering operation only has oneweight matrixWf with K = ⌊τ/2⌋ +1 frequency components. Whenthe operation input has K ′ = ⌊τ ′/2⌋ + 1 frequency components withK ′ < K , we can subsample the frequency components inWf . WhenK ′ > K , we interpolate the frequency components of Wf . STFNetprovides two kind of interpolation methods: 1) linear interpolationand 2) spectral interpolation.

The linear interpolation generates the missing frequency com-ponents in extended weight matrix W′f ∈ C

K ′×D×O from the twoneighbouring frequency components inWf :

kl =⌊k ′

ττ ′

⌋kr = kl + 1,

W′f [k′, ·, ·] =Wf [kl , ·, ·]

(kr − k ′

ττ ′

)+Wf [kr , ·, ·]

(k ′

ττ ′− kl).

(7)

The spectral interpolation utilizes the relationship betweendiscrete-time Fourier transform (DTFT) and discrete Fourier trans-form (DFT). For a time-limited signal (with length τ ), DTFT regards2https://en.wikipedia.org/wiki/Convolution_theorem

2196

Page 6: STFNets: Learning Sensing Signals from the Time-Frequency ...lusu/papers/The experimental results validate the design settings of STFNets and illustrate their superior accuracy compared

Spectral PaddingSpectral PaddingSpectral Padding

fs/2fs/40 fs/20 fs/20fs/4 fs/4fs/8 3fs/8 fs/8 3fs/8

Dilated Convolution

Wc

STFNet-Convolution

Convolution Kernel

Figure 5: The STFNet-convolution operationwith dilated con-figuration.it as a infinite-length data with zeros outside the time-limited range,while DFT regards it as a τ -periodic data. As a result, DTFT gener-ates a continuous function over the frequency domain, while DFTgenerates a discrete function. Therefore, DFT can be regarded as asampling of DTFT with step fs/τ . In order to increase the frequencyresolution of Wf , we can increase the sampling step from fs/τ tofs/τ

′, which is called spectral interpolation. Spectral interpolationcan be done through zero padding in the time domain [22],

W′f [·,d,o] = DFT(ZeroPadτ ′−τ IDFT

(Wf [·,d,o]

)), (8)

where ZeroPadt denotes padding t zeros at the end of sequence, andIDFT(·) denotes the inverse discrete Fourier transform. Please notethat, if we pad infinite zeros to the IDFT result, then DFT turns intoDTFT. An simple illustration of STFNet-filtering operation is shownin Figure 4.

3.5 STFNet-Convolution OperationIn this subsection, we introduce our design of STFNet-convolutionoperation. Other than filtering operation that handles global patternmatching, we still need the convolution operation to deal with lo-cal motifs in the frequency domain. We denote the input tensor asX ∈ CM×K×D , whereM is the number of time chunk, K number offrequency component, and D input feature dimension. The convo-lution operation involves two steps: 1) padding the input data, and2) convolving with kernel weight matrix Wc ∈ C

1×S×D×O , where Sis the kernel size along the frequency axis and O is still the outputfeature dimension.

Without the padding step, the output of convolution operationwill shrink the number of frequency components, which may breakthe underlying structure and information in the frequency domain.Therefore, we need to pad extra “frequency component" to keep theshape of output tensor unchanged compared to that of the inputdata. In the deep learning research, padding zeros is a commonpractice. Zero padding is reasonable for inputs such as images andsignal in the time domain, meaning no additional information in thepadding range. However, padding zero-valued frequency componentintroduces additional information in the frequency domain.

Therefore, STFNet-convolution operation proposes the spectralpadding for time-frequency analysis. According to the definition ofDFT, transformed data is periodic within the frequency domain. Inaddition, if the original signal is real-valued, then the transformeddata is conjugate symmetric within each period. Previously, we cutthe number of frequency components of a τ -length signal to K =⌊τ/2⌋ + 1 for reducing the redundancy. In the spectral padding, weadd these frequency components back according to the rule

X[·,τ−k, ·] = X[·,−k, ·] = X∗[·,k, ·], (9)

STFNet-Pooling

fs/2fs/40 fs/20 fs/20fs/4 fs/4fs/8 3fs/8 fs/8 3fs/8

Figure 6: The low-pass STFNet-pooling operation.where X∗ denotes complex conjugation. In addition, the number ofpadding before and after the input tensor is same as the previouspadding techniques.

Then we can define the basic convolution operation in STFNetY = SpectralPad(X) ⊛Wc , (10)

where SpectralPad(·) denotes our spectral padding operation, and⊛ denotes the convolution operation.

Next, we discuss the way to share the kernel weight matrixWc with multi-resolution data. Other than interpolating the kernelweight matrix as shown in (7) and (8), we propose another solutionfor the STFNet-convolution operation. The convolution operationconcerns more about the pattern of relative positions on the fre-quency domain. Therefore, instead of providing additional kerneldetails on fine-grained frequency resolution, we can just ensure thatthe convolution kernel is applied with the same frequency spac-ing on representations with different frequency resolutions. Suchidea can be implemented with the dilated convolution [33]. If Wcis applied to a input tensor with K = ⌊τ/2⌋ + 1 frequency compo-nents, for a input tensor with K ′ = ⌊τ ′/2⌋+1 frequency components(τ ′ > τ ), the dilated rate r is set to τ ′/τ − 1. A simple illustration ofSTFNet-convolution with dilated configuration is shown in Figure 5.

3.6 STFNet-Pooling OperationIn order to provide a dimension reduction method for sensing serieswithin STNet, we introduce the STFNet-pooling operation. STFNet-pooling truncates the spectral information over time with a pre-defined frequency pattern. As a widely-used processing technique,filtering zeroes unwanted frequency components in the signal. Var-ious filtering techniques have been designed, including low-passfiltering, high-pass filtering, and band-pass filtering, which serveas templates for our STFNet-pooling. Instead of zeroing unwantedfrequency components, STFNet-pooling removes unwanted compo-nents and then concatenates the left pieces. For applications withdomain knowledge about signal-to-noise ratio over the frequencydomain, specific pooling strategy can be designed. In this paper, wefocus on low-pass STFNet-pooling as an illustrative example.

To extend the STFNet-pooling operation to multiple resolutionsand preserving spectral information, we make sure that all repre-sentations have the same cut-off frequency according to their ownfrequency resolutions. A simple example of low-pass STFNet-poolingoperation is shown in Figure 6. We can see that our three tensors aretruncated according to the same cut-off frequency, fs/4.

4 EVALUATIONIn this section, we evaluate the STFNet with diverse sensing modu-larities. We focus on the device-based and device-free human activityrecognitions with motion sensors (accelerometer and gyroscope),WiFi, ultrasound, and visible light. We first introduce the experimen-tal setting, including data collection and baseline algorithms. Next,we show the performance metrics of leave-one-user-out evaluationof human activity recognition with different modularities. Finally, weanalyze the effectiveness of STFNet through several ablation studies.

2197

Page 7: STFNets: Learning Sensing Signals from the Time-Frequency ...lusu/papers/The experimental results validate the design settings of STFNets and illustrate their superior accuracy compared

4.1 Experimental SettingsIn this subsection, we first introduce detailed information of thedataset we used or collected for each evaluation task. Thenwe specifythe way to test the performance of evaluation task.

Motion Sensor: In this experiment, we recognize human activitywith motion sensors on smart devices. We use the dataset collectedby Allan et al. [23]. This dataset contains readings from two mo-tion sensors (accelerometer and gyroscope). Readings were recordedwhen users executed activities scripted in no specific order, whilecarrying smartwatches and smartphones. The dataset contains 9 vol-unteers, 6 activities (biking, sitting, standing, walking, climbStair-up,and climbStair-down). We align two sensor readings, linear interpo-late two readings by 100Hz, and segment them into non-overlappingdata samples with time interval 5.12s. Therefore, each data sampleis a 512 × 6 matrix, where both accelerometer and gyroscope havereadings on x , y, and z axis.

WiFi: In this experiment, we make use of Channel State Infor-mation (CSI) to analyze human activities. CSI refers to the knownchannel properties of a communication link, which can be affected bythe presence of humans and their activities. We employ 11 volunteers(including both men and women) as the subjects and collect CSI datafrom 6 different rooms in two different buildings. In particular, webuild a WiFi infrastructure, which includes a transmitter (a wirelessrouter) and two receivers. We use the tool to report CSI values of30 OFDM subcarriers [9]. The experiment contains 6 activities (wip-ing the whiteboard, walking, moving a suitcase, rotating the chair.sitting, as well as standing up and sitting down). We linearly interpo-late the CSI data with a uniform sampling period, and down-samplethe measurements into 100Hz. Then we segment the down-sampledCSI data into non-overlapping data samples with time interval 5.12s.Therefore, each data sample is a 512 × 30 matrix, where each CSImeasurement has readings from 30 subcarriers.

Ultrasound: In this experiment, we conduct human activityrecognition based on ultrasound. We employ 12 volunteers as thesubjects to conduct the 6 different activities. The activity data arecollected from 6 different rooms in two different buildings. The trans-mitter is an iPad on which an ultrasound generator app is installed,and it can emit an ultrasound signal of approximately 19 KHz. Thereceiver is a smartphone and we use the installed recorder app tocollect the sound waves. We demodulate the received signal withcarrier frequency 19KHz, and down-sample the measurement into100Hz. Then we segment the down-sampled ultrasound data intonon-overlapping data samples with time interval 5.12s. Therefore,each sample is a 512 × 1 matrix.

Visible light: In this experiment, we capture the human activityin the visible light system. We build an optical system using pho-toresistors to capture the in-air body gesture, which can detect theilluminance change (lux) caused by the body interaction. In the ex-periment, there are three light conditions (natural mode, warmmode,and cool mode) and 4 hand gestures (drawing an anticlockwise circle,drawing a clockwise circle, drawing a cross, and shaking hand sideto side). We employ 6 volunteers as the subjects and each of themperforms 20 trials of every gesture under a given lighting condition.We linearly interpolate and down-sample the measurements into25Hz. Then we segment the data into non-overlapping data sampleswith time interval 5.12s. Therefore, each sample is a 128 × 6 matrix,where each measurement contains readings from 6 CdS cells.

Table 1: Illustration of models with two sensor inputs.STFNet-Filter/Conv DeepSense/ComplexNet

Sensor Data 1 Sensor Data 2 Chunked ChunkedSensor Data 1 Sensor Data 2

STFNet1-1 STFNet1-2 Conv Layer1-1 Conv Layer1-2STFNet2-1 STFNet2-2 Conv Layer2-1 Conv Layer2-2STFNet3-1 STFNet3-2 Conv Layer3-1 Conv Layer3-2

STFNet-pooling Max poolingSTFNet4 Conv Layer4STFNet5 Conv Layer5STFNet6 Conv Layer6Averaging GRUSoftmax Softmax

Testing: In the whole evaluation, to illustrate the generalizationability of STFNet and other baseline models, we perform leave-one-user-out cross validation for every task. For each time, we choose thedata from one user as testing data with the left as training data. Wethen compare the performance of models according to their accuracyand F1 score with 95% confidence interval.

4.2 Models in ComparisonIn order to evaluate, when compared to conventional deep learningcomponents (i.e., convolutional and recurrent layers), whether ourproposed STFNet component is better at decoding information andextracting features from sensing inputs, we substitute componentsin the state-of-the-art neural network structure for IoT applicationswith STFNet. In the whole evaluation, we choose DeepSense as thestate-of-the-art structure, which has shown signifiant improvementson various sensing tasks [26]. The illustration of structures of fivecomparing models with two sensor inputs are shown in Table 1.Detailed information of our comparing models are listed as follows,

(1) STFNet-Filter: This model integrates the proposed STFNetcomponent and the DeepSense structure. Within the STFNetcomponent, we use the STFNet-filtering operation designedin Section 3.4. The intuition of DeepSense structure is to firstperform local processing within each sensor and then per-form global sensor fusion over multiple sensors. In this model,we replace all convolutional layers used in local/global sen-sor data processing with our time-frequency analyzing com-ponent, STFNet. Since our model has already incorporatedtime-domain analysis within the STFNet component throughmulti-resolution processing, we replace the Gated RecurrentUnits (GRU) with simple feature averaging time at last.

(2) STFNet-Conv: This model is almost the same as the STFNet-Filter, except that we use the STFNet-convolution operationdesigned in Section 3.5.

(3) DeepSense-Freq: This model is the original DeepSense [26].It divides the input sensing data into chunks, and processeseach chunk with DFT. It treats the real and imagery parts ofdiscrete Fourier transformed time chunks as the additionalfeature dimensions. This is the state-of-the-art deep learningmodel for sensing data modelling and IoT applications.

(4) DeepSense-Time: This model is almost the same as theDeepSense-Freq, except that it directly takes the chunkedraw sensing data without DFT as input.

(5) ComplexNet: This model is a complex-value neural net-work [24] that can operate on complex-value inputs. Insteadof using simple CNN and RNN structure as originally pro-posed [24], we cheat in their favor by using the DeepSense

2198

Page 8: STFNets: Learning Sensing Signals from the Time-Frequency ...lusu/papers/The experimental results validate the design settings of STFNets and illustrate their superior accuracy compared

Accuracy F1 Score0.75

0.8

0.85

0.9

0.95 STFNet-FilterSTFNet-ConvDeepSense-FreqDeepSense-TimeComplexNet

Figure 7: The accuracy and F1 score with 95% confidence in-terval for motion sensors.

Accuracy F1 Score0.65

0.7

0.75

0.8

0.85

0.9STFNet-FilterSTFNet-ConvDeepSense-FreqDeepSense-TimeComplexNet

Figure 8: The accuracy and F1 score with 95% confidence in-terval for WiFi.

structure, which improves the performance in all tasks. Thenetwork inputs are chunked sensing data with DFT.

4.3 EffectivenessIn this section, we discuss about the effectiveness of our proposedSTFNet based on extensive experiments and diverse sensing modali-ties, compared with other state-of-the-art deep learning models.

As we mentioned in Section 4.1, all models are evaluated throughleave-one-user-out cross validation with accuracy and F1 score ac-companied by the 95% confidence interval. STFNet-based models(STFNet-Filter and STFNet-Conv) take a sliding window set for multi-resolution short-time Fourier transform. We choose the set to be{16, 32, 64, 128} for activity recognition with motion sensors, WiFi,and ultrasound; and choose set to be {8, 16, 32, 64} for activity recog-nition with visible light. DeepSense-based models (DeepSense-Freqand DeepSense-Time) need a sliding window for chunking inputsignals. In the evaluation, we cheat in their favor by choosing thebest-performing window size from {8, 16, 32, 64, 128} according tothe accuracy metric. In addition, we consistently configure STFNet-filtering operation with linear interpolation, and STFNet-convolutionoperation with spectral padding. We will show further evaluationson multi-resolution operations and the effects of diverse operationsettings in Section 4.4.

4.3.1 Motion Sensors. For device-based activity recognition withmotion sensors, there are 9 users. The accuracy and F1 score withthe 95% confidence interval for leave-one-user-out cross validationare illustrated in Figure 7. STFNet based models, i.e., STFNet-Filterand STFNet-Conv, outperform all other baseline models with a largemargin. The confidence interval lower bound of STNet-Filter andSTFNet-Conv is even better than the confidence interval upper boundof DeepSense-Freq and DeepSense-Time. STFNet-Filter performsbetter than STFNet-Conv in this experiment, indicating that differentactivities have distinct global profiling patterns with motion sen-sor readings in the frequency domain, even among different users.

Accuracy F1 Score0.65

0.7

0.75

0.8

0.85

0.9STFNet-FilterSTFNet-ConvDeepSense-FreqDeepSense-TimeComplexNet

Figure 9: The accuracy and F1 score with 95% confidence in-terval for Ultrasound.

Accuracy F1 Score0.65

0.7

0.75

0.8

0.85

0.9

0.95

1

STFNet-FilterSTFNet-ConvDeepSense-FreqDeepSense-TimeComplexNet

Figure 10: The accuracy and F1 score with 95% confidence in-terval for Visible light.STFNet-Filter is able to learn the accurate global frequency profiling,which makes it the top-performance model in this task. In addition,compared to ComplexNet, STFNet based models show clear improve-ments. Therefore, using just complex-value neural network for sens-ing signal is far from enough. The multi-resolution processing andoperations that are spectral-compatible are all crucial designs.

4.3.2 WiFi. For device-free activity recognition with WiFi signal,there are 11 users. The accuracy and F1 score with the 95% confi-dence interval for leave-one-user-out cross validation are illustratedin Figure 8. STFNet based models still outperform all others with aclear margin, illustrating the effectiveness of principled design ofSTFNet from time-frequency perspective. DeepSense-Freq outper-forms DeepSense-Time in this experiment, which means that evenhaving time-frequency transformation as pre-processing can help.The complex-value network, ComplexNet, performs worse than itsreal-value counterpart, DeepSense-Freq. This indicates that blindlyprocessing time-frequency representations without preserving theirphysical meanings can even hurt the final performance. STNet-Convperforms better than STNet-Filter in the WiFi experiment, indicatingthat local shiftings in the frequency domain are more representativefor diverse activities profiled with WiFi CSI.

4.3.3 Ultrasound. There are 12 users in device-free activity recogni-tion with ultrasound experiment. The accuracy and F1 score withthe 95% confidence interval for leave-one-user-out cross validationare illustrated in Figure 9. STFNet based models still significantlyoutperforms all other baselines. An interesting observation is thatComplexNet performs even worse than both DeepSense-Freq andDeepSene-Time, which again validates the importance of designingneural networks for sensing signal with multi-resolution processingas well as preserving the time and frequency information.

4.3.4 Visible Light. There are 6 users in the experiment of device-free activity recognition with visible light. The accuracy and F1 scorewith the 95% confidence interval are illustrated in Figure 10. Except

2199

Page 9: STFNets: Learning Sensing Signals from the Time-Frequency ...lusu/papers/The experimental results validate the design settings of STFNets and illustrate their superior accuracy compared

Motion WiFi Ultrasound Light0.65

0.7

0.75

0.8

0.85

0.9

0.95

1

Acc

urac

y

STFNe−FilterSTFNet−Single−FilterSTFNet−ConvSTFNet−Single−ConvDeepSense−Freq

(a) Accuracy with 95% confidence interval.

Motion WiFi Ultrasound Light0.65

0.7

0.75

0.8

0.85

0.9

0.95

1

F1 S

core

STFNe−FilterSTFNet−Single−FilterSTFNet−ConvSTFNet−Single−ConvDeepSense−Freq

(b) F1 score with 95% confidence interval.

Figure 11: Multi-Resolution v.s. Single-Resolutionfor the DeepSense-Time, all othermodels can can achieve an accuracyof approximately 90% or higher. STFNet based models still do thebest. There is no significant difference between STFNet-Filter andSTFNet-Conv, which indicates that measured visible light readingshave quite clean representations in the frequency domain.

4.4 Ablation StudiesIn the previous section, we illustrate the performance of STFNetcompared to other state-of-the-art baselines. In this section, we focusmainly on the STFNet design. We conduct several ablation studiesby deleting one designing feature from STFNet at a time.

4.4.1 Multi-Resolution v.s. Single-Resolution. First, we validate theeffectiveness of our design of multi-resolution processing in STFNetblock. As shown in Figure 1, this includes multi-resolution STFT,hologram interleaving, and weights sharing techniques in STFNet-Filtering and STFNet-Convolution operations. In this experiment,we add two more baseline models, STFNet-Single-Filter and STFNet-Single-Conv, generated by deleting the multi-resolution processingin STFNet-Filter and STFNet-Conv respectively. These two modelspick the best-performing window size from {8, 16, 32, 64, 128} ac-cording to the accuracy metric. The results for all four tasks areillustrated in Figure 11, where DeepSense-Freq severs as a decentperformance low-bound. The design of multi-resolution processingsignificantly impacts the performance of STFNet. STFNet-Single-Filter and STFNet-Single-Conv show clear performance degradationcompared to their multi-resolution counterparts. In addition, STFNet-Single-Filter and STFNet-Single-Conv still consistently outperformDeepSense-Freq with a clear margin. This is because our other de-signed operations, including STFNet-Filtering, STFNet-Convolution,STFNet-Pooling still facilitate the learning in time-frequency domain.

4.4.2 Spectral Padding v.s. Zero Padding. Next, we validate our de-sign of spectral padding in the STFNet-Convolution operation asshown in Figure 5. In this experiment, we add a new baseline al-gorithm, STFNet-Conv-zPad, by replacing spectral padding withtraditional zero padding in the STFNet-Conv. The accuracy and F1score of all four tasks are shown in Figure 12. Here, DeepSense-Freq

Motion WiFi Ultrasound Light0.65

0.7

0.75

0.8

0.85

0.9

0.95

1

Acc

urac

y

STFNet−ConvSTFNet−Conv−zPadDeepSense−Freq

(a) Accuracy with 95% confidence interval.

Motion WiFi Ultrasound Light0.65

0.7

0.75

0.8

0.85

0.9

0.95

1

F1 S

core

STFNet−ConvSTFNet−Conv−zPadDeepSense−Freq

(b) F1 score with 95% confidence interval.Figure 12: Spectral Padding v.s. Zero Padding

is still treated as a performance low-bound. By comparing STFNet-Conv-zPad and STFNet-Conv, we can see that spectral padding con-sistently helps improving the model performance. In most cases, theimprovement is limited. However, in the case of visible light, spectralpadding significantly improves both accuracy and F1 score. There-fore, designing neural network by preserving the time-frequencysemantics of sensing signal is an important rule to follow.

4.4.3 Linear Interpolation v.s. Spectral Interpolation. Then, we com-pare our two designs of weight interpolation method in the STFNet-Filtering operation, linear interpolation and spectral interpolation,as shown in Figure 4. The STFNet-Filter defined in Section 4.2 useslinear interpolation, so we rename it as STFNet-Filter-LinearInptin this experiment. We add a new baseline model called STFNet-Filter-SpectralInpt by using spectral interpolation instead of linearinterpolation in STFNet-Filter. The results of all four tasks are illus-trated in Figure 13. In general, the performance of two design choicesare almost the same. At most of time, linear interpolation performsslightly better. In addition, we recommend using linear interpolation,since its implementation is easier,

4.4.4 STFNet Pooling v.s. Mean/Max Pooling. Finally, we validateour design of STFNet-Pooling (low-pass deisgn) as shown in Figure 6.In this experiment, we add two new baseline algorithms, STFNet-Filter-mPad and STFNet-Conv-mPad, by replacing STFNet-Pooling inSTFNet-Filter and STFNet-Conv with traditional max/mean poolingin the time domain (through choosing the one has better accuracy).The results are illustrated in Figure 14. In all settings, STFNet-Poolingshows better performance. In some cases, the improvement are sig-nificant. We believe that STFNet-Pooling can achieve even betterperformance if given the detailed signal-to-noise ratio over the fre-quency domain for each specific sensor. Then we can employ otherpooling strategies instead of the low-pass design.

5 DISCUSSIONThis paper provides a principled way of designing neural networksfor sensing signals inspired by the fundamental nature of the under-lying physical processes. STFNet, operates directly in the frequency

2200

Page 10: STFNets: Learning Sensing Signals from the Time-Frequency ...lusu/papers/The experimental results validate the design settings of STFNets and illustrate their superior accuracy compared

Motion WiFi Ultrasound Light0.65

0.7

0.75

0.8

0.85

0.9

0.95

1

Acc

urac

y

STFNe−Filter−LinearInptSTFNet−Filter−SpectralInptDeepSense−Freq

(a) Accuracy with 95% confidence interval.

Motion WiFi Ultrasound Light0.65

0.7

0.75

0.8

0.85

0.9

0.95

1

F1 S

core

STFNe−Filter−LinearInptSTFNet−Filter−SpectralInptDeepSense−Freq

(b) F1 score with 95% confidence interval.Figure 13: Linear Interpolation v.s. Spectral Interpolation

domain, in which the measured physical phenomena are best ex-posed. We propose three types of learnable frequency manipulationsthat are able to operate on multi-resolution representations, whilepreserving the underlying time-frequency information. Althoughextensive experiments have illustrated the superior performanceof STFNet, further research is needed to better understand designchoices for neural networks from the time-frequency perspective.

One challenge is to explore the possibility of integrating neuralnetworks with other time-frequency transformations. In this pa-per, STFNet focuses on the short-time Fourier Transform. However,STFT is the most basic one. There are plenty of other transforma-tion basis functions in traditional time-frequency analysis. How tonaturally integrate them with neural network while keeping the un-derlying physical meaning within transformed representations? Howto choose or design the most suitable transformation basis functionsthat meet the corresponding mathematical requirements? Answersto these questions can greatly impact the way researchers designneural networks for sensing signal processing. Another challengeis to empower the frequency manipulations to have heterogeneousbehaviours over the time. In STFNet, all designed operations arelearnable frequency manipulations, which perform identically overtime. In order to fully exploit the potential of time-frequency analysis,further research is needed on designing time-varying time-frequencymanipulations, that adapt to current temporal patterns.

Furthermore, a better experimental and theoretical understand-ing is needed of the basic settings of neural networks to supportcomputation in time-frequency domain. For traditional real-valueneural networks, researchers have good intuitions about the basicconfigurations of initialization, activation functions, dropout andnormalization techniques, and optimization methods. However, forneural network in the time-frequency domain, our understanding islimited. Although the reseach community started to study the basicsettings of neural networks with complex values [24], the current un-derstanding remains preliminary. Time-frequency analysis can haveoperations in both the real and complex domains. At the same time,the underlying time-frequency information within the internal repre-sentations can make the related studies even more complicated. We

Motion WiFi Ultrasound Light0.65

0.7

0.75

0.8

0.85

0.9

0.95

1

Acc

urac

y

STFNe-FilterSTFNet-Filter-mPoolSTFNet-ConvSTFNet-Conv-mPoolDeepSense-Freq

(a) Accuracy with 95% confidence interval.

Motion WiFi Ultrasound Light0.65

0.7

0.75

0.8

0.85

0.9

0.95

1

F1 S

core

STFNe-FilterSTFNet-Filter-mPoolSTFNet-ConvSTFNet-Conv-mPoolDeepSense-Freq

(b) F1 score with 95% confidence interval.Figure 14: STFNet Pooling v.s. Mean/Max Pooling

believe that this understanding will greatly facilitate future designof deep learning systems for IoT.

In addition, outside the IoT context, there exists a large numberof transformations and dimension reduction techniques, such asSVD and PCA, that have made great impact in revealing usefulfeatures of complex phenomena. Our study of deep learning withSTFT suggests that integrating deep neural networks with othercommon transformations may facilitate learning in domains wheresuch transformations reveal essential features of the input signaldomain. Future work is needed to explore this conjecture.6 CONCLUSIONIn this paper, we introduced STFNet, a principled way of design-ing neural networks from the time-frequency perspective. STFNetendows time-frequency analysis with additional flexibility and capa-bility. In addition to just parameterizing the frequency manipulationswith deep neural networks, we bring two key insights into the de-sign of STFNet. On one hand, STFNet leverages and preserves thefrequency domain semantics that encode time and frequency in-formation. On the other hand, STFNet circumvents the uncertaintyprinciple through multi-resolution transform and processing. Evalua-tions show that STFNet consistently outperforms the state-of-the-artdeep learning models with a clear margin under diverse sensingmodalities, and our two designing insights significantly contributeto the improvement. The designs and evaluations of STFNet unveilthe benefits of incorporating domain-specific modeling and transfor-mation techniques into neural network design.ACKNOWLEDGMENTSResearch reported in this paper was sponsored in part by NSF undergrants CNS 16-18627 and CNS 13-20209 and in part by the ArmyResearch Laboratory under Coop- erative Agreements W911NF-09-2-0053 andW911NF-17-2-0196. The views and conclusions contained inthis document are those of the authors and should not be interpretedas representing the official policies, either expressed or implied, ofthe Army Research Laboratory, NSF, or the U.S. Govern- ment. TheU.S. Government is authorized to reproduce and distribute reprintsfor Government purposes notwithstanding any copyright notationhere on.

2201

Page 11: STFNets: Learning Sensing Signals from the Time-Frequency ...lusu/papers/The experimental results validate the design settings of STFNets and illustrate their superior accuracy compared

REFERENCES[1] Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2014. Neural machine

translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473(2014).

[2] Sourav Bhattacharya and Nicholas D Lane. 2016. Sparsification and separation ofdeep learning layers for constrained resource inference onwearables. In Proceedingsof the 14th ACM Conference on Embedded Network Sensor Systems CD-ROM. ACM,176–189.

[3] Ke-Yu Chen, Daniel Ashbrook, Mayank Goel, Sung-Hyuck Lee, and Shwetak Patel.2014. AirLink: sharing files between multiple devices using in-air gestures. In Pro-ceedings of the 2014 ACM International Joint Conference on Pervasive and UbiquitousComputing. ACM, 565–569.

[4] Ronan Collobert, Jason Weston, Léon Bottou, Michael Karlen, Koray Kavukcuoglu,and Pavel Kuksa. 2011. Natural language processing (almost) from scratch. Journalof Machine Learning Research 12, Aug (2011), 2493–2537.

[5] Yarin Gal and Zoubin Ghahramani. 2015. Dropout as a Bayesian approximation.arXiv preprint arXiv:1506.02157 (2015).

[6] Yarin Gal and Zoubin Ghahramani. 2016. A theoretically grounded application ofdropout in recurrent neural networks. In Advances in neural information processingsystems. 1019–1027.

[7] Rafael C Gonzalez, Richard E Woods, et al. 2002. Digital image processing.[8] Sidhant Gupta, Daniel Morris, Shwetak Patel, and Desney Tan. 2012. Soundwave:

using the doppler effect to sense gestures. In Proceedings of the SIGCHI Conferenceon Human Factors in Computing Systems. ACM, 1911–1914.

[9] Daniel Halperin, Wenjun Hu, Anmol Sheth, and David Wetherall. 2011. Toolrelease: Gathering 802.11 n traces with channel state information. ACM SIGCOMMComputer Communication Review 41, 1 (2011), 53–53.

[10] Song Han, Huizi Mao, and William J Dally. 2015. Deep compression: Compressingdeep neural networks with pruning, trained quantization and huffman coding.arXiv preprint arXiv:1510.00149 (2015).

[11] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residuallearning for image recognition. In Proceedings of the IEEE conference on computervision and pattern recognition. 770–778.

[12] Samuli Hemminki, Petteri Nurmi, and Sasu Tarkoma. 2013. Accelerometer-basedtransportation mode detection on smartphones. In Proceedings of the 11th ACMConference on Embedded Networked Sensor Systems. ACM, 13.

[13] David H Hubel and Torsten N Wiesel. 1968. Receptive fields and functionalarchitecture of monkey striate cortex. The Journal of physiology 195, 1 (1968),215–243.

[14] Thomas N Kipf and Max Welling. 2016. Semi-supervised classification with graphconvolutional networks. arXiv preprint arXiv:1609.02907 (2016).

[15] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. 2012. Imagenet classifica-tion with deep convolutional neural networks. In Advances in neural informationprocessing systems. 1097–1105.

[16] Nicholas D Lane, Petko Georgiev, and Lorena Qendro. 2015. DeepEar: robustsmartphone audio sensing in unconstrained acoustic environments using deeplearning. In Proceedings of the 2015 ACM International Joint Conference on Pervasiveand Ubiquitous Computing. ACM, 283–294.

[17] Tianxing Li, Qiang Liu, and Xia Zhou. 2016. Practical human sensing in thelight. In Proceedings of the 14th Annual International Conference on Mobile Systems,Applications, and Services. ACM, 71–84.

[18] Bryan Perozzi, Rami Al-Rfou, and Steven Skiena. 2014. Deepwalk: Online learningof social representations. In Proceedings of the 20th ACM SIGKDD internationalconference on Knowledge discovery and data mining. ACM, 701–710.

[19] Qifan Pu, Sidhant Gupta, Shyamnath Gollakota, and Shwetak Patel. 2013. Whole-home gesture recognition using wireless signals. In Proceedings of the 19th annualinternational conference on Mobile computing & networking. ACM, 27–38.

[20] Oren Rippel, Jasper Snoek, and Ryan P Adams. 2015. Spectral representationsfor convolutional neural networks. In Advances in neural information processingsystems. 2449–2457.

[21] Karen Simonyan and Andrew Zisserman. 2014. Very deep convolutional networksfor large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014).

[22] Julius Orion Smith. 2007. Mathematics of the discrete Fourier transform (DFT): withaudio applications. Julius Smith.

[23] Allan Stisen, Henrik Blunck, Sourav Bhattacharya, Thor Siiger Prentow,Mikkel Baun Kjærgaard, Anind Dey, Tobias Sonne, and Mads Møller Jensen. 2015.Smart devices are different: Assessing andmitigatingmobile sensing heterogeneitiesfor activity recognition. In Proceedings of the 13th ACM Conference on EmbeddedNetworked Sensor Systems. ACM, 127–140.

[24] Chiheb Trabelsi, Olexa Bilaniuk, Ying Zhang, Dmitriy Serdyuk, Sandeep Subrama-nian, João Felipe Santos, Soroush Mehri, Negar Rostamzadeh, Yoshua Bengio, andChristopher J Pal. 2017. Deep complex networks. arXiv preprint arXiv:1705.09792(2017).

[25] Wei Wang, Alex X Liu, Muhammad Shahzad, Kang Ling, and Sanglu Lu. 2015.Understanding and modeling of wifi signal based human activity recognition. InProceedings of the 21st annual international conference on mobile computing andnetworking. ACM, 65–76.

[26] Shuochao Yao, Shaohan Hu, Yiran Zhao, Aston Zhang, and Tarek Abdelzaher. 2017.Deepsense: A unified deep learning framework for time-series mobile sensing dataprocessing. In Proceedings of the 26th International Conference on World Wide Web.International World Wide Web Conferences Steering Committee, 351–360.

[27] Shuochao Yao, Yiran Zhao, Huajie Shao, ShengZhong Liu, Dongxin Liu, Lu Su,and Tarek Abdelzaher. 2018. Fastdeepiot: Towards understanding and optimizingneural network execution time on mobile and embedded devices. In Proceedings ofthe 16th ACM Conference on Embedded Networked Sensor Systems. ACM, 278–291.

[28] Shuochao Yao, Yiran Zhao, Huajie Shao, Aston Zhang, Chao Zhang, Shen Li, andTarek Abdelzaher. 2018. Rdeepsense: Reliable deep mobile computing models withuncertainty estimations. Proceedings of the ACM on Interactive, Mobile, Wearableand Ubiquitous Technologies 1, 4 (2018), 173.

[29] Shuochao Yao, Yiran Zhao, Huajie Shao, Chao Zhang, Aston Zhang, Shaohan Hu,Dongxin Liu, Shengzhong Liu, Lu Su, and Tarek Abdelzaher. 2018. Sensegan:Enabling deep learning for internet of things with a semi-supervised framework.Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies2, 3 (2018), 144.

[30] Shuochao Yao, Yiran Zhao, Huajie Shao, Chao Zhang, Aston Zhang, Dongxin Liu,Shengzhong Liu, Lu Su, and Tarek Abdelzaher. 2018. Apdeepsense: Deep learninguncertainty estimation without the pain for iot applications. In 2018 IEEE 38thInternational Conference on Distributed Computing Systems (ICDCS). IEEE, 334–343.

[31] Shuochao Yao, Yiran Zhao, Aston Zhang, Shaohan Hu, Huajie Shao, Chao Zhang,Lu Su, and Tarek Abdelzaher. 2018. Deep Learning for the Internet of Things.Computer 51, 5 (2018), 32–41.

[32] Shuochao Yao, Yiran Zhao, Aston Zhang, Lu Su, and Tarek Abdelzaher. 2017.Deepiot: Compressing deep neural network structures for sensing systems witha compressor-critic framework. In Proceedings of the 15th ACM Conference onEmbedded Network Sensor Systems. ACM, 4.

[33] Fisher Yu and Vladlen Koltun. 2015. Multi-scale context aggregation by dilatedconvolutions. arXiv preprint arXiv:1511.07122 (2015).

2202