wireless user positioning via synthetic data augmentation

1
Dataset Cyclical Learning Rate (CLR) Ensembling Visit signet.dei.unipd.it and lttm.dei.unipd.it to know more about our research groups Results Wireless User Positioning via Synthetic Data Augmentation and Smart Ensembling Umberto Michieli, Paolo Testolina, Mattia Lecci, Michele Zorzi University of Padova – Padova, Italy – {umberto.michieli, paolo.testolina, mattia.lecci, zorzi}@dei.unipd.it Mattia Lecci’s activity was financially supported by Fondazione Cassa di Risparmio di Padova e Rovigo under the grant “Dottorati di Ricerca 2018” Autoencoder Pre-training Models Several models from the literature were tested: U-Net [1]; VGG [2]; ResNet50 [3]; DenseNet [custom]; Others. The architectures were adjusted to the different nature of our problem, enhancing their performance. Cyclical Learning Rate can train deep networks faster and better; CLR works especially well with Convolutional Layers; Regularization is necessary to avoid strong overfitting; CLR allows to find better minima; One-cycle policy performed best for us. To exploit the correlation between the received signal power and the RX-TX distance, the following architecture is proposed: A deep network, chosen among the best-performing ones, extracts the information from the raw channel matrix; A shallower network conveys the power information to the output layers, giving it a higher relevance on determining the position. The respective outputs are therefore concatenated and fed to a last, dense output layer. In the novel framework we proposed, first an autoencoder learns a compressed representation of the channel matrix using the synthetic data, in a completely unsupervised manner. Thus, positions (labels) are not needed during the first phase of the training. After this phase, only the encoder part of the network is reused and the acquired knowledge is transferred to the real data domain. We used an implementation of the 3GPP TR 38.901 indoor channel model to obtain a large number of synthetic, unlabeled channel matrices, making it possible for the autoencoder to understand the underlying correlations within a channel matrix. The given data are not normalized Pathloss yields important distance information; Frequency and spatial correlations have to be taken into account; Models from the literature tend to have lots of parameters; The low number of given samples might be problematic for deep-learning approaches; For this reason we keep a large test set. TOTAL Train Validation Test 17486 11366 2623 3497 (a) One-cycle policy (b) Triangular policy (c) Vanishing Triangular policy Autoencoders: although the framework seems to be valid, further work is needed to achieve higher precision. VGG: the most promising base architecture, future work should investigate the reasons why this specific convolutional model outperforms all the others. Ensemble: the side information on the SNR can enrich the pretrained base network boosting the performance. CLR: beside speeding up the training, CLR can help reaching better minima. To completely exploit the information contained in the received power and deliver it to the output layer, a simple multilayer perceptron was used. [1] Ronneberger O. et al., "U-Net: Convolutional Networks for Biomedical Image Segmentation", MICCAI 2015. [2] Simonyan K., Zisserman A., "Very Deep Convolutional Networks for Large-Scale Image Recognition", ICLR 2015. [3] He et al., "Deep Residual Learning for Image Recognition", CVPR 2016. Fig.: Smoothed training curves for VGG-based networks Fig.: Performance of some of the tested networks Fig.: Visualization of our best model’s errors Quivers are resized for aesthetic reasons.

Upload: others

Post on 08-Apr-2022

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Wireless User Positioning via Synthetic Data Augmentation

Dataset

CyclicalLearningRate(CLR)

Ensembling

Visitsignet.dei.unipd.it andlttm.dei.unipd.ittoknowmoreaboutourresearchgroups

Results

WirelessUserPositioningviaSyntheticDataAugmentationandSmartEnsembling

UmbertoMichieli,PaoloTestolina,MattiaLecci,MicheleZorziUniversityofPadova– Padova,Italy– {umberto.michieli, paolo.testolina, mattia.lecci, zorzi}@dei.unipd.it

MattiaLecci’s activitywasfinanciallysupportedbyFondazioneCassa diRisparmio diPadovaeRovigo underthegrant“Dottorati diRicerca 2018”

AutoencoderPre-training

Models

Several models from the literature were tested:• U-Net [1];• VGG [2];• ResNet50 [3];• DenseNet [custom];• Others.The architectures were adjusted to the differentnature of our problem, enhancing theirperformance.

• Cyclical Learning Rate can train deepnetworks faster and better;

• CLR works especially well withConvolutional Layers;

• Regularization is necessary to avoid strongoverfitting;

• CLR allows to find better minima;

• One-cycle policy performed best for us.

To exploit the correlation between thereceived signal power and the RX-TXdistance, the following architecture isproposed:

• A deep network, chosen amongthe best-performing ones, extractsthe information from the rawchannel matrix;

• A shallower network conveys thepower information to the outputlayers, giving it a higher relevanceon determining the position.

The respective outputs are thereforeconcatenated and fed to a last, denseoutput layer.

In the novel framework we proposed, first an autoencoder learns a compressedrepresentation of the channel matrix using the synthetic data, in a completelyunsupervised manner. Thus, positions (labels) are not needed during the first phase ofthe training.After this phase, only the encoder part of the network is reused and the acquiredknowledge is transferred to the real data domain.

We used an implementation of the 3GPP TR 38.901 indoor channel model to obtain alarge number of synthetic, unlabeled channel matrices, making it possible for theautoencoder to understand the underlying correlations within a channel matrix.

• Thegivendataarenotnormalized• Pathlossyieldsimportantdistanceinformation;

• Frequencyandspatialcorrelationshavetobetakenintoaccount;

• Modelsfromtheliteraturetendtohavelotsofparameters;• Thelownumberofgivensamplesmightbeproblematicfordeep-learningapproaches;

• Forthisreasonwekeepalargetestset.

TOTAL Train Validation Test

17486 11366 2623 3497

(a)One-cyclepolicy

(b)Triangularpolicy (c)VanishingTriangularpolicy

Autoencoders: although the frameworkseems to be valid, further work is neededto achieve higher precision.

VGG: the most promising basearchitecture, future work shouldinvestigate the reasons why this specificconvolutional model outperforms all theothers.

Ensemble: the side information on the SNRcan enrich the pretrained base networkboosting the performance.

CLR: beside speeding up the training, CLRcan help reaching better minima.

To completely exploit the informationcontained in the received power and deliverit to the output layer, a simple multilayerperceptron was used.

[1] Ronneberger O. et al., "U-Net: ConvolutionalNetworks for Biomedical Image Segmentation", MICCAI2015.[2] Simonyan K., Zisserman A., "Very DeepConvolutional Networks for Large-Scale ImageRecognition", ICLR 2015.[3] He et al., "Deep Residual Learning for ImageRecognition", CVPR 2016.

Fig.:SmoothedtrainingcurvesforVGG-basednetworks

Fig.:Performanceofsomeofthetestednetworks

Fig.:Visualizationofourbestmodel’serrorsQuiversareresizedforaestheticreasons.