Sie sind auf Seite 1von 536

User Manual

Version 6.20

2012 by Anite Finland Ltd. All rights reserved. This manual, as well as the software described in it, is furnished under license and may only be used or copied in accordance with the terms of such license. The information in this manual is intended for informational use only and is subject to change without notice. Anite Finland Ltd assumes no responsibility or liability for any errors or inaccuracies that may appear in this user manual. Except as permitted by such license, no part of this publication may be reproduced or transmitted in any form or by any means, electronic, mechanical, recording, or otherwise, without the prior written permission of Anite Finland Ltd. Windows XP Pro, Windows Vista, Windows 7, and Windows Network Monitor are registered trademarks of the Microsoft Corporation. MapInfo is a registered trademark of the MapInfo Corporation. Any other trademarks or service marks contained herein are the property of their respective owners. Revision 6.20.01 Last Edited April 2012

CON T ENTS

CONTENTS QUICK GUIDE


SETTING UP THE SYSTEM

9
9

BEFORE YOU BEGIN


END-USER LICENSE AGREEMENT NOTES

10
10 10

INSTALLING NEMO ANALYZE


HARDWARE AND SOFTWARE REQUIREMENTS NEMO ANALYZE INSTALLATION COPY PROTECTION License Server (Floating License) Copy Protection Modules Technical Support Expiration

11
11 11 16 16 21 24

STARTING NEMO ANALYZE USING NEMO ANALYZE WORKSPACE


EXPORTING/IMPORTING FAVORITE PARAMETERS WORKSPACE FILTERS

25 26 27
29 31

CREATING THE DATABASE


ADDING FILES TO THE DATABASE Migrating Nemo Analyze database to a new laptop Importing Image Files as Maps ADDING CSV FILES TO THE DATABASE Opening CSV Files from the Database Opening CSV Files without Database Import Creating Custom Queries for CSV Files ORGANIZE MEASUREMENTS Drag & Drop Folders Search Folders Hiding Folder Content Query Folders JOINED MEASUREMENTS

33
33 34 35 36 41 44 46 48 51 52 58 59 61 64

ADDING DESCRIPTIONS TO MEASUREMENT FILES AND FOLDERS DELETE FOLDER CONTENTS RETRIEVING ORIGINAL FILES FROM THE DATABASE DATABASE CLEANUP 65 66 67

VIEWING MEASUREMENT DATA


PARAMETER TREE Statistics/Statistics with filters over Parameter

68
68 71

Statistics by/Statistics with Filters by: No Grouping Parameter Launchpad Change Defaults VIEWING MEASUREMENT DATA IN GRAPHS

71 77 79 81

VIEWING MEASUREMENT DATA ON MAP 83 Changing default map 83 Opening Measurement Files on Map 84 Opening BTS Files on Map 88 Viewing Notification Icons on Map 91 Use Case 1.: Viewing Cell Footprints and RSCP Footprints93 Use Case 2.: Viewing Uplink Voice Quality Server Data 95 Use Case 3.: Viewing IP/UDP Packet Trace Data 99 PARAMETER FILTERING 107 Filtering Based on Polygon Area 110 Use Case 4.: Global Parameter Filtering Based on a Secondary Parameter FILTERING DATA AND CREATING CUSTOM KPIS 120

113

DRILL-DOWN 125 Drilling Down from Map Data View 128 Drilling Down from Troubleshooting Parameters in Map Data View

130

DATA VIEWS

133

GRAPHS 133 Graph Tools Menu 133 Graph Popup Menu 134 Change Graph Type 134 Reset 134 Query 135 Pick Parameter 135 Add Function 136 Add Reference Line 136 Tool 137 Mode 137 Side Panel 139 Line Graph Filled 142 3D Mode 142 Group Values 143 Graph Properties 144 Layer Properties 145 Use Case 5.: Multiple Graph Layers 148 Use Case 6.: Notification Icons in Graphs 151 Use Case 7.: Correlating Parameters Using Color Grids and Surface Graphs GRIDS Side Panel Row Details Export Data to Grid Properties Use Case 8.: Color Sets in Grids Use Case 9.: Play Audio Sample 155 156 158 158 159 161 164

153

CON T ENTS

Use Case 10.: Using L3 and RRC Message Search Parameters 166 MAPS 169 Viewing on Live Map 169 Google Street View 172 Map Tools Menu 173 Map Popup Menu 175 Add 175 Find 177 Generate Color Set 178 Export to KML File 178 Side Panel 180 Tool 183 MapX 183 Map Properties 185 Route Properties 186 Draw Route as Symbols 188 BTS Files 190 Use Case 11.: Adding Map Layers and Saving Layer Combinations as Geosets 198 Use Case 12.: Coloring Routes Based on BTS Coverage201 Use Case 13.: Performing Area Binning 202 Use Case 14.: Comparing Two Groups of Measurements from the Same Route on Map Use Case 15.: Displaying Base Station Cell Beam Range on Map217 Use Case 16.: Synchronizing Base Station Map Overlay with Grid Rows 223 Use Case 17.: Using BTS Reference Parameters 226 Use Case 18.: Displaying Base Station Connections on Map Based on Any Parameter SPREADSHEETS 234 Editing Cell Format 238 Filtering Data 242 Creating Formulas 243 Adding Functions 244 Use Case 19..: Retrieving Data from Minimized Data Sets246 Use Case 20.: Creating Reports and Report Templates Using Spreadsheets PROCESSING UPLINK VOICE QUALITY DATA INDOOR MEASUREMENTS NUMERICAL DATA VIEWS INFO VIEWS TIMELINE VIEW Highlight Parameter in Timeline View Notifications in Timeline View Range Selection in Timeline View OTHER VIEWS Network Parameters Measurement Settings Properties Query Clipboard Database Loader 254 255 258 261 262 263 265 266 267 267 268 269 270 271

212

229

247

Window Browser Messenger Output MANUAL LAYOUT EDITOR Creating a Layout Adding Content to Manual Layout Editor Data Views

272 273 274 275 275 277

WORKBOOKS
ADDING DATA VIEWS ADDING PAGES SAVING A WORKBOOK EXPORTING WORKBOOKS Exporting Workbooks as PDF Files Exporting Workbooks as Image Files COPYING A WORKBOOK WORKBOOK PROPERTIES PAGE PROPERTIES EXAMPLE WORKBOOK

279
280 284 285 287 287 289 290 291 291 292

REPORTS
WORKBOOKS SPREADSHEETS REPORTING WITH MICROSOFT EXCEL REPORTING WITH CRYSTAL REPORTS Generating Reports Generating Reports from Multiple Files Exporting Reports Configuring Reports

295
295 295 296 306 306 309 310 312

CREATING CUSTOM QUERIES


SQL QUERIES Getting started Nemo Analyze Database Schema Views Automatically joined tables Database timestamps Filtering log files Value enumeration Connections Correlating tables based on time Guidelines for creating SQL queries SQL VS. KPI Workbench QUERY MANAGER Pick Measurement Parameter Generic Query Wizard Manual Query Correlate Parameters DATABASE BROWSER

314
315 315 315 319 322 323 324 325 327 328 330 332 333 337 342 344 344

CON T ENTS

CUSTOM KPI WORKBENCH 345 Designing Custom KPIs and Adding Input Data Sets 346 Combining Input Data Sets 349 Adding Operations 366 Running and Testing KPIs 397 Saving KPIs as Components 398 Saving Custom KPIs 399 Reopening Custom KPIs for Editing 400 Defining KPI Execution Method and Value Constants 402 Use Case 21.: Creating Complex Filters Using Multiple Conditions 404 Use Case22.: Creating a KPI for Dropped Calls Resulting from a Missing Handover ANALYTICS (S)FTP LOG FILE AUTOLOAD 438 440

412

OTHER TASKS

442

EDITING COLOR SETS 442 Importing Color Sets 443 Automatic Generation of Color Sets 444 Use Case 23.: Automatic Generation of Color Set for a Value Range Use Case 24.: Creating a Color Set 452 Use Case 25.: Creating and Applying a Color Set on Map454 Use Case 26.: Creating and Applying a Color Set in Grid458 SCHEDULING EVENTS Report Events Workbook Events Load Folder Events KPI Threshold Alarm Events Defining Recurring Events TASK MANAGER AND PROBLEM REPOSITORY Configuring and Editing a Task MANAGING USERS AND USER GROUPS User Groups VIEWING SERVER STATISTICS Managing Tasks EDITING OPERATOR SETTINGS EDITING CDMA OPERATOR SETTINGS CONFIGURING NOTIFICATION ICONS Configuring Notifications Using the Parameter Tree Notification Configuration EDITING MENUS EDITING QUERY AND WORKBOOK GROUPS CONFIGURING SYSTEM SETTINGS Options Environment Options Database Options Report Options Graph 462 464 468 469 470 473 474 475 479 481 483 485 487 489 491 491 493 494 500 501 501 507 509 510

448

Options BTS IMPORTING CUSTOM SETTINGS EXPORTING CUSTOM SETTINGS IMPORTING MAPINFO POLYGONS

514 516 517 519 521

VIEWING LOG ON SYSTEM PERFORMANCE AND SQL FUNCTIONS

USER INTERFACE
MENUS Analyze Menu File Menu View Menu Tools Menu Fleet Menu Layout Menu Help Menu

523
523 523 524 524 525 526 526 526

SHORTCUT KEYS FAQ NEMO ANALYZE SUPPORT


INTERNET SUPPORT AND NEMO USER CLUB PHONE, EMAIL, AND FAX SUPPORT

527 528 529


529 529

APPENDIX 1
MAKING MAPINFO MAPS Registering a Raster Map

531
531 531

APPENDIX 2
END-USER LICENSE AGREEMENT

533
533

QU IC K GU ID E

QUICK GUIDE
The instructions given in this quick guide have been designed to get you started in a few easy steps. Please, note that these instructions do not give a full picture of the software. Consequently, every step has a cross-reference to the place in the manual where the topic in question will be explained in more detail.

SETTING UP THE SYSTEM


From Nemo Analyze 6.0 onwards the software is protected with a copy protection module and it is not possible to install Nemo Analyze without it. Software-based copy protection licenses are no longer valid. 1. Installing the software. Check the manual for hardware requirements if necessary (p.11). Run the Nemo Analyze installation program Nemo Analyze.x.xx.xx.exe from the Nemo Analyze CDROM. During the installation, Nemo Analyze will ask you to insert the copy protection module on the computer on which you are running Nemo Analyze. Please insert the copy protection module and wait until Windows has completed installing the device drivers. Press OK to continue the setup. When the installation is finished, restart the computer. 2. Loading files to database. Start by loading your measurement and BTS files in the Nemo Analyze database (p. 33). This may take a few minutes, but once files have been added, they will remain in the database until you remove them. 3. You are now ready to start analyzing. You can, for example, generate a report from the data (p.295) and view measurements in graphs and maps (p.68).

BEFORE YOU BEGIN


Nemo Analyze is a powerful post-processing and reporting tool for planning, optimizing, and maintaining cellular networks. Nemo Analyze serves many purposes from network overview to problem solving and report generation. Some of its features include: Post-processing of Nemo Outdoor, Nemo Autonomous, Nemo Compact-i, and Nemo Handy measurement files Full support for the latest Nemo measurement file format in real time SQL interface to the database engine User-defined KPIs with custom queries Open ODBC interface for third-party software Fully customizable user interface Statistical analysis and reporting Predefined, customizable report templates

END-USER LICENSE AGREEMENT


The software described in this book is furnished under a license agreement and may be used only in accordance with the terms of that agreement. For full details of this agreement, refer to page 533.

NOTES
Nemo Analyze users must be appropriately trained and should be familiar with the signaling behind wireless technologies depending on their usage needs. The user is expected to have basic knowledge on mobile networks and their terminology. A working knowledge of computers and Microsoft Windows is required for using the Nemo Analyze software. The user should know how to use the mouse as well as standard Windows menus and commands. To review these techniques, see your Microsoft Windows documentation.

IN ST AL L ING N EMO AN AL YZ E

11

INSTALLING NEMO ANALYZE


HARDWARE AND SOFTWARE REQUIREMENTS

PC with Windows XP Professional/Windows Vista/ Windows 7 Pentium III processor or better, minimum 1 GHz, preferably 2 GHz 1 GB RAM minimum 1 GB of free hard disk space for installation and use; 20 GB recommended Nemo Analyze copy protection module One USB port for copy protection module Display resolution 1400 x 900 recommended Internet Explorer 8 or newer with Windows XP Professional Internet Explorer 9 with Windows Vista/Windows 7 Crystal Reports Professional edition for creating custom-made report templates (optional)

NEMO ANALYZE INSTALLATION


Installing Nemo Analyze
1. Microsoft Network Monitor 3.3 must be installed BEFORE Nemo Analyze to enable complete IP/UDP trace details to be displayed in Nemo Analyze. For installing the Microsoft Network Monitor 3.3, see Installing Microsoft Network Monitor 3.3 on page 99. If Microsoft Network Monitor 3.3 is nevertheless installed after Nemo Analyze, Nemo Analyzes Microsoft Network Monitor 3.3 functionality can be activated by copying the file NMAPI.dll from the folder C:\Program Files\Microsoft Network Monitor 3\ to the folder C:\Program Files\Anite\Nemo Analyze. If Microsoft Network Monitor 3.3 is not found on the PC or Nemo Analyze has been installed before Microsoft Network Monitor 3.3, Nemo Analyze will display only a limited set of IP/UDP trace details. 2. To begin Nemo Analyze installation, insert the setup CD-ROM into the CD-ROM drive (e.g., drive D). Note: The user performing the installation must have administration rights. Installation without administration rights could result in a database error. If using Windows 7, start installation by right-clicking the setup executable and select "Run as administrator" from the menu. Note: Before installation, uninstall the previous version of Nemo Analyze using Add/Remove Programs from the Windows Control Panel. Note: Before installation, if using Windows Vista, it is recommended that User Account Control be disabled. Disable User Account Control by browsing to Control Panel | User Accounts and setting User Account Control (UAC) to OFF. It is also recommended that all Windows Firewall profiles (i.e. Domain Profile, Private Profile, Public Profile) in Windows Firewall with Advanced Security | Properties are set to OFF.

3. Select Run from the Windows Start menu. Type D:\Nemo Analyze\ x.xx.xx.exe and press the RETURN key. This command starts the Nemo Analyze installation software. Follow the instructions given by the installation program.

4. The Analyze Setup Prerequisites dialog opens. Click Install.

Note: If you have not uninstalled the previous version of Nemo Analyze, the installation procedure will be interrupted. Uninstall the previous version of Nemo Analyze using Add/Remove Programs from the Windows Control Panel. Once the previous version has been uninstalled, restart the installation procedure.

5. Once the installation of Nemo Analyze Setup Prerequisites has been completed, the Welcome to the InstallShield Wizard for Nemo Analyze dialog opens. Click Next.

IN ST AL L ING N EMO AN AL YZ E

13

6. The License Agreement dialog opens. Select I accept the terms of the license agreement, and click Next.

7. The Customer Information dialog opens. Enter User Name and Company Name, and click Next.

8. The Choose Destination Location dialog opens. The default directory for installation is C:\Program Files\Anite\Nemo Analyze. Accept this by clicking Next or click the Change button to select another location.

9. The Choose Working Directory dialog opens. The default directory for Nemo measurement data is C:\Nemo Tools. Accept this by clicking Next or click the Change button to select another location.

IN ST AL L ING N EMO AN AL YZ E

15

10.The Ready to Install the Program dialog opens. Select Install to begin installation.

During the installation, Nemo Analyze will ask you to insert the copy protection module on the computer on which you are running Nemo Analyze. Please insert the copy protection module and wait until Windows has completed installing the device drivers. Press OK to continue the setup. When the installation is finished, restart the computer.

11.After the installation, restart the computer.

When starting, Nemo Analyze checks if the database is correct, i.e. has the same version as Analyze. If the database does not match, a warning is displayed. The warning does not prevent you from using Analyze but may cause it not to function properly.

COPY PROTECTION
For Nemo Analyze you can select between two copy protection options: a license server or a copy protection module.

License Server (Floating License)


Licensing is separate from the product itself, therefore four options exist: Analyze Standalone with copy protection module Analyze Standalone with floating license copy protection Client/server (Enterprise) Analyze with copy protection module Client/server (Enterprise) Analyze with floating license copy protection

The license server option enables the use of floating license, allowing more efficient usage of purchased licenses. When using the floating license, the users do not need to have a physical copy protection module attached into their PCs. Instead, there is only one physical copy protection module in the floating license server containing multiple licenses. When Nemo Analyze is started, it reserves a license over the network from the floating license server. Floating License can be used both with Standalone and Client/Server:

IN ST AL L ING N EMO AN AL YZ E

17

Unlimited amount of users in different physical locations can install Nemo Analyze, floating license only limits the amount of concurrent users. If the usage pattern of Nemo Analyze is bursty among the users, certain amount of licenses can easily serve 2-4 fold amount of actual users, since all users are typically never using the SW at the same time. Continuous network connection is needed when floating license is used. In floating license, there is one physical copy protection module in the floating license server containing multiple licenses. Therefore, Nemo Analyze users do not have physical copy protection module attached into their PCs. Note that the use of network license with the Nemo Analyze Enterprise Edition requires your system to have separate hardware for the Nemo Analyze database server, the license server, and the Nemo Analyze Clients. For database server and license server system requirements, see Nemo Analyze Server Administration Guide. The license server, however, does not have to be dedicated for the license server, i.e. the server can run other services as well.

Installing the floating license server


Nemo Analyze requires copy protection key during the installation process. In the case of floating license server, The HASP control center run-time SW must be installed to the floating license server first. The server setup can be found from the Nemo Analyze installation CD or downloaded from the Nemo User Clubs Nemo Analyze section. 1. Plug in the floating license dongle to the license server computer. You can verify that the installation is OK by going to address http://localhost:1947/ in the floating license server computer with a web browser.

2. You are now able install Nemo Analyze. It should find the floating license dongle automatically from the network. When installing Analyze you can go to address http://localhost:1947/ with a web browser on that same computer you are installing Analyze. You should see the floating license dongle available over the network in the "Hasp Keys" and "Products" sections:

If you don't see the dongle in HASP control center, try going to the Configuration section and Access to Remote License Managers page. In the Search Parameters box type the IP address of the license server. Click Submit and wait for a while - the dongle should appear in HASP Keys section. Also check firewall settings of the license server.

IN ST AL L ING N EMO AN AL YZ E

19

3. Start Analyze - it will complain about key not being found. Just type the IP address of the license server into Analyze. Now Analyze starts with the floating license. The license server uses a TCP/UDP port to communicate with the computers running Nemo Analyze. Please make sure that communication is allowed in the following port: TCP/UDP 1947.

Configuring the floating license server.


The Floating license server can be configured via web browser. The address of the control center is http://localhost:1947/ when accessed from the floating license computer itself. The server can also be accessed over the network: Error! Hyperlink reference not valid.. Currently used licenses can be monitored and disconnected under Sessions. One user utilizes multiple sessions. To force license release of a user, all sessions of the machine must be disconnected. Note that sessions do not map one to one to licenses. One user has multiple sessions, but only one license is consumed per one machine.

Under Configuration, access restrictions for users, logging settings, and remote access setting to the Control center can be configures. Detailed description of the license management interface can be found under Help in the side panel of the Admin Control Center.

IN ST AL L ING N EMO AN AL YZ E

21

No Valid License Detected


If an error message Error: No valid license detected. appears when the application is started, either the dongle key is not plugged in or the dongle key is an unsupported model. Please follow the instructions for how to read the license information from the existing dongle key. To be able to use the Nemo Analyze application supported by the old dongle key, it is recommended that you uninstall Nemo Analyze version 5.80. After un-installation reboot the computer and re-install the previous version. The example above is seen, e.g. when the SPS 00868 dongle key is in use.

Copy Protection Modules


From Nemo Analyze 6.0 onward, only hardware-based copy protection is used. A USB copy protection dongle will be delivered in the Nemo Analyze package. Plug in the copy protection dongle in a USB port on your PC. Note that the new dongle introduced in Nemo Analyze 6.00 will not be backwards compatible with the older versions of Nemo Analyze. A new dongle with a valid technical support and maintenance agreement option will be in use with Nemo Analyze 6.00 and later versions. If you have problems starting Nemo Analyze, and if the following error messages appear, please contact Nemo Technical Support at Nemo.Helpdesk@anite.com. Note: Copy protection modules are not needed when using floating license.

Connecting the copy protection module in the USB port:


Plug the copy protection module into your computer's USB port. Do not plug the USB Port copy protection module in any other port. It may damage your computer.

No Valid Technical Support Agreement


If an error message saying the version of Nemo Analyze is not supported by the technical support appears, it means that the dongle key type is correct, but that the dongle key is missing valid technical support & maintenance agreement information. Please follow the instructions for how to read the license information from the existing dongle key with the Remote Update Utility. To be able to use the Nemo Analyze application supported by the dongle key without information about the technical support agreement, it is recommended that you uninstall Nemo Analyze version 5.80. After un-installation reboot the computer and re-install the previous version.

The old dongle key

The new dongle key

The dongles are updated using the Remote Update Utility software. Before updating Nemo Outdoor application please make sure that new dongle key is available. Note that you must use the latest version of the software which comes with Nemo Analyze version 6.00, or it can alternatively be downloaded from Nemo User Club. See the instructions below for updating your Nemo Outdoor dongle.

Updating the dongle (Nemo Analyze version 6.00 and later)


1. Start the SecureUpdate by selecting Start | All Programs | Nemo Tools | Remote Update Utility. 2. The Collect Key Status Information tab opens.

IN ST AL L ING N EMO AN AL YZ E

23

3. Connect your copy protection module 4. Click Collect Information and select a location where the C2V file is saved. 5. Send the C2V file to Nemo Technical Support at nemo.helpdesk@anite.com. Please add the following information to the e-mail using the following subject Nemo Analyze dongle update: Company name First name Surname Street address Zip/ Postal Code Country Telephone number Mobile number

6. After you have received the V2C file containing the update, select the Update File in the Apply License Update tab and click Apply Update.

7. The dongle is now updated. Nemo Outdoor can be installed and is ready for use.

Technical Support Expiration


The new dongle keys used with Nemo Analyze 6.00 and later versions include a compulsory technical support and maintenance agreement option. The technical support expiration date defines the date the agreement ends. After this date it is not possible to start the new version of the Nemo Analyze application. Example: A customer has purchased Nemo Analyze version 6.00 with a 1-year technical support maintenance agreement in October 2011. The customer is entitled for free updates until October 2012. Nemo Outdoor versions released after October 2012 cannot be used with the dongle key without renewing the technical support & maintenance agreement. After the technical support & maintenance agreement is renewed, the dongle key can be updated remotely and the customer can update the application to the latest version.

ST AR T ING N EM O ANAL YZ E

25

STARTING NEMO ANALYZE


After running Setup, you will see a new Nemo Tools program group under the Programs selection in the Windows Start menu.

To start Nemo Analyze:


1. Double-click the Nemo Analyze.exe icon in My Computer or select it from the Start menu (StartProgramsNemo ToolsNemo Analyze). 2. When Nemo Analyze is starting, hold down the Shift key to prevent Nemo Analyze from loading a previously used workspace. 3. When the program starts, the window below will open. In the Welcome to Nemo Analyze workbook you will find useful information about Nemo Analyze, links to related documents, and shortcuts to loading measurement and BTS files.

Ribbonbar

Workspace

Workbook area Parameter tree

USING NEMO ANALYZE


In Nemo Analyze you can view measurement data in multiple ways. The most convenient way, however, is to focus on the Workspace user interface since it is quick and easy. For example, in the workspace you can double-click a parameter, and Nemo Analyze opens the parameter automatically in the default data view. Furthermore, you can go to View | Workbook | Add Data View, choose a particular data view, select a measurement, and drag a parameter from the parameter tree in the data view. You can also view several parameters in the same data view by dragging them in it. If you drag a parameter in an empty workbook, a dialog box appears and asks you to select a data view type. If you drag a measurement file to a workbook with a data view, a dialog box emerges asking you to select the parameter you wish to view. The filter field above the parameter tree is useful when looking for a specific parameter in a long list. All data views are synchronized with each other based on time. Consequently, if you open two or more data views on the same measurement file but on different parameters simultaneously, and switch from one point of time in one data view to another point in time, the data in the other data views change accordingly.

CR EAT ING TH E D ATABAS E

27

WORKSPACE
The Workspace forms the basis of the user interface in Nemo Analyze. All measurements and parameters can be found in the workspace, and from there it is convenient to drag them in various data views. See also Using Nemo Analyze (p.26). The Workspace view is by default attached to the left-hand side of the Analyze main window. You can also drag it elsewhere in the window, and if you want to dock it back to the side of the main window, double-click on the Workspace. The Workspace view consists of five pages: Measurements, Base Stations, Maps, Polygons, Data Source Files, and Reports pages, respectively. You can switch between these pages via the icon tabs at the bottom of the Workspace view. The Measurements page in the Workspace is divided into three sections: Folders, Folder Contents, and the Parameter tree. The folders include the All Measurements folder, which shows all the files that are currently in the database, and other user-defined folders. The current database connection is also identified next to the All Measurements folder, e.g. Analyze Local Database. The Folder Contents displays a list of all measurements in the folder selected in the Folders section. The name of the measurement device is also displayed when the mouse is hovered over a measurement file. The Parameter Tree shows all available parameters and KPIs for the selected file, measurement, or folder, depending on the selections in the other two sections. The KPIs can be displayed in graphs, maps, and other data views, and statistics can also be calculated from the parameter tree.

You can add parameters to Favorites by right-clicking on a parameter and selecting Add to Favorites from the popup menu. To toggle between the Favorites view and the Parameter Tree view, click the Favorites button in the upper right corner of the Parameter Tree.

CR EAT ING TH E D ATABAS E

29

EXPORTING/IMPORTING FAVORITE PARAMETERS


It is possible to export and import favorite (starred) parameters by exporting/importing identifiers of all favorite parameters. To Export favorite parameters, select File | Settings.

The Save As dialog opens. Name the file and click Save. The Export Settings dialog opens.

Select the parameter favorites you want to export and click OK. To import parameter favorites, select File | Settings.

The Open dialog opens. Open the export file (*aex) where you have saved the exported favorite parameters by selecting Open.

Select the parameter favorites you want to import and click OK. A dialog box appears asking if you want to replace existing favorites.

If user selects Yes, all favorites are cleared and new ones are imported. If user selects No, existing favorites are remembered and imported ones are added.

CR EAT ING TH E D ATABAS E

31

WORKSPACE FILTERS
Workspace filters enable the user to locate measurement files, parameters, maps, base stations, etc. more quickly. Below you can find some usage examples for the filters. First, above the measurement files (see the red box) you can find the filter for measurement files. Type the name of the measurement file in the filter field to find the desired measurement file.

Second, above the parameter tree you can see the filter field for parameters. Type in the name of the parameter, and only parameters matching the name will be viewed in the parameter tree.

Third, in the Base Stations view of the Workspace, the filter can find base stations based on any content in the BTS file, allowing base stations to be filtered based on e.g. channel number.

CR EAT ING TH E D ATABAS E

33

CREATING THE DATABASE


When you start Nemo Analyze for the first time, the database is empty. Adding files - measurements, BTS files, and maps - to the database may take a while, but once files have been added, they will remain in the database until you remove them.

ADDING FILES TO THE DATABASE


Select File | Measurement | Open [Measurement /Map/BTS/Report/Data Source File].

The Open dialog opens.

Select the file(s) you want to add to the database and click Open. You can select several files by holding down the Shift key.

From the Files of type drop down menu you can define the type of file that you want to open. Supported network measurement tools of Nemo Analyze include Nemo Outdoor, Nemo Autonomous, Nemo Compact-i, Nemo Handy, TEMS Investigation, and TEMS Pocket, EADS REMS TETRAPOL and R&S ROMES. Measurement data in CSV (Character-Separated Value) ASCII format is also supported. Marker file and map are loaded automatically from Handy embedded indoor file when downloading the file to Analyze. Note that adding files to the database may take several minutes depending on the number and the size of the files. Note that all the files need to have unique file names. Note that the file names of Nemo measurement files must follow the format filename.devicenumber.nmf (for example, t5gsm.1.nmf). Device number signifies the number of the device in case of simultaneous measurements with multiple devices. Device number must always be separated with an extra dot in the file name. With one device, the value is 1. Note that the marker file and map are loaded automatically from Handy embedded indoor file when downloading the file to Analyze. The Database Loader (see page 271) displays the progress of file upload, file conversion, and file upload queue. The added measurement files will appear in the All Measurements folder in the Workspace | Folder Contents view in the left-hand side of the Nemo Analyze main window.

Migrating Nemo Analyze database to a new laptop


When changing your laptop, it is possible to migrate existing Nemo Analyze database to a new laptop once you have Analyze already installed on it. To achieve this, copy the following two subfolders from Nemo Tools folder on the C: the Nemo Analyze, with the exception of the Logs folder under Nemo Analyze (DO NOT copy the Logs folder or the database might get corrupted), and the Nemo Analyze Datastore

CR EAT ING TH E D ATABAS E

35

Importing Image Files as Maps


In addition to adding map files in the database through File | Measurement | Open Map, you can import image files to be used as maps. Select File | | Import | Image as map. In the Open dialog, select the image that you want to use as a map. It can be, for example, a floorplan from a test site. After selecting the image file you need to define either the width and length of the map area in meters or the GPS coordinates for two map corners. After you click OK in the Map Import Parameters dialog, the map is added in the database and it will appear in the Maps folder in the workspace.

You can also create black and white copies of raster maps by right-clicking on a map in the workspace and selecting Create Black & White Copy. The map is saved with the same name with BW added at the end of the file name.

ADDING CSV FILES TO THE DATABASE


In addition to files produced with Nemo tools, you can import any character-separated value (CSV) ASCII data into the Nemo Analyze database. The data can be post-processed and visualized using the data views available in Nemo Analyze. Select View | Tools | Database manager.

The Database Manager dialog opens.

CR EAT ING TH E D ATABAS E

37

To create a new CSV import template, select the folder Database Management | CSV Templates from the directory tree, right-click on the Template field right of the directory tree, and select Add Template from the popup menu. The CSV Import Wizard CSV Template File dialog opens.

Template file enables you to select a CSV file to serve as a basis for the import template. Click the browse button to browse for CSV files.

Extension enables you to define the file type the import template will apply to. Note: It is recommended that the extensions of CSV files with different content structures are renamed to provide each type with a unique extension. For instance, four different types of CSV files, all originally with the extension .csv, could be renamed as .csv1, .csv2, .csv3, and .csv4.

Once Template file and Extension have been defined, click Next. The CSV Import Wizard CSV Import dialog opens.

Ignore defines the parts of the imported data set that are to be ignored when uploading the data, i.e. the parts that do not contain relevant data. Column row defines the number of top rows to be ignored. Rows allows you to define specific rows to be ignored. Rows starting with allows you to define rows starting with a specific string to be ignored. Columns allows you to divide a CSV data set into columns. Fixed lines defines the content of each column based on the number of characters in a CSV string. For example, the setting 12, 22, 32 defines that the first twelve characters (including spaces) belong to column 1, the next 22 characters to column 2, and the next 32 characters to column 3. Fixed lines should only be used if there are no delimiters in the data set. Delimiter defines the character that separates columns in the CSV data set. You can use one of the default delimiters, i.e. comma, dot, semicolon, <TAB>, or <SPACE>, or you can define your own delimiter character by entering a character to the Delimiter field. Strip leading and trailing quotation marks removes leading and trailing quotation marks from each column of the imported data set. File preview displays a preview of the CSV file contents. Once the Ignore and Columns settings have been defined, click Next. The CSV Import Wizard Database Schema Definition dialog opens.

CR EAT ING TH E D ATABAS E

39

Table name defines the name of the template. Columns enables you to select which columns are included in the imported data set and to define the properties of each column using Column, Column type and Format. Decimal separator defines the decimal separator as comma or dot. Column defines the name of the column selected in the Columns control. Column type defines the data type of the column selected in the Columns control. Keyword allows you to define the content type of columns containing some general data types, such as time and positioning data. Based on these defined content types, the data can be correlated with measurement files of other formats. Format defines the required syntax for the column selected in the Columns control. To define a custom time column syntax element-by-element, click the button. The Time Format dialog opens. To add syntax elements (e.g. YYYY) and delimiters (e.g. ;), doubleclick the relevant element. Custom syntax is displayed in the Format field of the Time Format dialog.

Data preview displays a preview of the data set structure. Once the properties of each column have been defined, click Finish.

The new template is displayed in the Template field of the User Manager window.

To open a CSV file using an import template, select Database Management | CSV Templates from the directory tree, right-click on the Template field right of the directory tree, and select Import Template from the popup menu. You can also open a CSV files using an import template by selecting File | Measurement | Open Measurement.

CR EAT ING TH E D ATABAS E

41

The Open dialog opens.

Select a CSV import template (e.g. Template1) by using the Files of type drop-down menu, browse to a folder containing measurement files with the extension defined in the template, select a file, and click Open. For Opening CSV Files from the Database, see below. To export a CSV template, right-click an existing template and select Export Template.

Choose under which name the exported template will be saved as from the file dialog.

Opening CSV Files from the Database


All CSV files within the database are displayed on the Measurements page of the Workspace. Files displays all the CSV files stored in the database.

Parameters displays all the parameters that are available for the file. The User folder displays userdefined custom parameters. For creating custom queries for CSV files, click here. To open a parameter in its default view, select a CSV file in the Files view and double-click a parameter (e.g. Jerkiness [%]) in the Parameters view. If you want to open the parameter in some other type of view, right-click on a parameter (e.g. Jerkiness [%]) in the Parameters view and select the view type from the popup menu. The Choose Graph Type dialog opens.

Select a graph type. The Select Columns dialog opens.

Define the x and y axes and click OK. The data view opens displaying the selected parameter data on the selected CSV file.

CR EAT ING TH E D ATABAS E

43

Opening CSV Files without Database Import


Optionally, character-separated value (CSV) data can also be opened and viewed in various data views without importing the data into the database. Using this method, the CSV data is loaded only in the workspace. Select File | Open | Measurement | Open Data Source File, and select the file in the Open dialog.

In the Analyze Wizard, you need to define file import parameters. After you have made the settings, click Next and finally Finish.
With Skip first you can define how many lines are skipped at the beginning of the file. Delimiter defines how columns in the source file are separated. Columns displays a list of columns detected in the source file. Column type defines the data type of each column. Format defines the data format of each column.

CR EAT ING TH E D ATABAS E

45

Column aliases define how data is displayed in the data views. You can define a parameter for the axes of each view type.

The imported file is added to the Data Source Files page in the Workspace. You can view the data by right-clicking on the file and selecting Open in Data View.

Creating Custom Queries for CSV Files


Correlating and creating custom queries for CSV files is possible using SQL. The Database browser can be used in examining the table structure of the database. To open the Database browser, select View | Database browser in the Ribbonbar.

The Database browser view opens. CSV import templates (e.g. Template1) can be found in the folder Views | User.

CR EAT ING TH E D ATABAS E

47

The CSV import template folder displays both the columns that were imported and some extra columns containing database structure data, such as the_file_title and the_file_extension.

If multiple CSV files have been imported to the database using a particular template, each column in the directory structure will display all rows from the corresponding columns of all the files that match the template. In order to make it possible to identify the file to which each row belongs, the column the_file_title displays the name of the file from which the row was retrieved.

ORGANIZE MEASUREMENTS
By default, all added measurement files go into the All Measurements folder.

CR EAT ING TH E D ATABAS E

49

If you have a lot of files, it may be convenient to categorize measurement files into custom-made subfolders. In the Organize Measurements dialog you can conveniently handle and organize large sets of measurement files. Go to View in the Ribbonbar and select Organize Measurements, or rightclick on the All Measurements folder in the Workspace and select Organize.

In the Organize Measurements dialog you can create new folders and subfolders, make joined measurements, and remove files from drag & drop folders. Right-click on a folder in the left-hand panel to create new folders. See the following chapters for information on the different folder types.

Select measurements, right-click on them, and select Join to create joined measurements. See page 61 for more information on joined measurements.

You can also organize measurements in the Workspace. Right-click on the All Measurements item and select Add Folder.

There are three types of folders you can add: Drag & Drop, Search, and Query.

CR EAT ING TH E D ATABAS E

51

Drag & Drop Folders


Drag & Drop folders are folders where you can drag files from the All Measurements folder. You can organize files, for example, by technology, time or place. When you drag and drop files from the All Measurements folder to the custom-made sub-folders, the files are only copied, i.e., the files will appear in two folders. The drag & drop folders are identified with a symbol.

The example shows a Workspace with two custom-made drag & drop folders named October 2006 and CDMA2000. The CDMA2000 folder has two files, and the October 2006 folder has four files. The same files are also in the All Measurements folder.

Search Folders
Search folders are folders where files are copied based on user-defined search criteria. The search folders are identified with a symbol. To create a new search folder, right-click on the All Measurements folder, and select Add Folder | Search from the popup menu.

If you have an existing search folder and you have loaded new files to the database, you must refresh the folder to display also the newly added files. To refresh a folder, right-click on the folder, and select Refresh from the popup menu.

CR EAT ING TH E D ATABAS E

53

The Search Folder Properties dialog opens, allowing you to define search criteria for the folder using the Measurement, Notifications, Network and System tabs. You can combine multiple search criteria and create search folders for, e.g., WiMAX scanners from specific dates and from specific polygon-defined geographical area. Name defines the search folder name.

Title, extension and description options allow you to search for specific measurement files based on the filename, file extension, and file description. Start date and End date allow you to define a time range from which all measurement files will be searched. Search last allows you to search a defined number of the most recent measurement files based on the following parameters: Days, Weeks, and Months finds measurement files from an N number of previous days, weeks, and months. Added finds an N number of measurement files added to the database the most recently Measurements finds measurement files from an N number of the most recent measurement sessions. Duration allows you to search for specific measurement files based on their duration. Hour from N to N allows you to filter files based on the hour of day when they were recorded Device type enables you to search for mobile or scanner measurements. Area enables you to search for measurements made on a defined area. To define the area, click the Define Area button. This opens the Select Polygon Area dialog below.

Clicking the Define Area button on the Properties dialog Measurement tab (see above) opens the Select Polygon Area dialog.

With Map | Filename you can select or browse for the map file you want to use. With Area | Name you can select a polygon area you have previously saved. With the Save, Save As and Delete buttons you can save an edited polygon area, save and name a new polygon area and delete a previously saved polygon area. With Change Map you select another map. Reset Area removes the polygon selection.

CR EAT ING TH E D ATABAS E

55

When you right-click on the map, a popup menu opens. With Area selected, you can select a polygon area.

In the Events page you can select the events you want to include in the search.

With the Mobile country code option you can search for measurements made in a certain country. With the Mobile network code option you can search for measurements made in a certain network.

CR EAT ING TH E D ATABAS E

57

With the System option you can search for measurements made in a certain system. With the Band option you can search for measurements made in a certain band.

Finally click OK and the measurements that match the search criteria are copied to the search folder.

Hiding Folder Content


When there are 20000+ log files in the database, it is recommended to hide folder content to improve the responsiveness of the Nemo Analyze user interface. To hide folder content, right-click on the All Measurements folder and select Hide Folder Contents from the popup menu.

CR EAT ING TH E D ATABAS E

59

Query Folders
Query folders are folders where files are added based on user-defined criteria, that is, queries. Once the query is defined, Nemo Analyze will go through the measurement files in the database and automatically add all relevant files to the query folder. To create a new query folder, right-click on the All Measurements folder and select Add Folder | Query from the popup menu.

If you have an existing query folder and you have loaded new files to the database, you must refresh the folder to display also the newly added files. To refresh a folder, right-click on the folder, and select Refresh from the popup menu.

The Properties dialog opens. In the Query field, select the criteria according to which the files will be copied to the folder, e.g., Last Year. The folder will be named automatically, but if you wish to name the folder manually, clear the option and type a name. If you want the folder to be updated automatically when files are added or removed from the database, select the Auto-update option. Click OK.

CR EAT ING TH E D ATABAS E

61

Nemo Analyze creates the new folder, and the files that match the query criteria are copied to the folder. The query folders are identified with a symbol.

The example shows a Workspace with one custommade query folder named Last Year.

JOINED MEASUREMENTS
With Nemo Analyze it is possible to join separate measurement files into a single joined measurement. Joined measurements will be handled as one measurement file. Note that you cannot join overlapping measurement files. Files must have been recorded at different times.

Right-click on All Measurements and select Add Joined Measurement.

In the Add Joined Measurement dialog, type a name for the joined measurement, e.g., 18 Jun 2006.

CR EAT ING TH E D ATABAS E

63

The joined measurement appears in the Workspace | Folder Contents. You can now drag & drop measurement files into the folder in the Folder Contents section. Joined measurements are identified with a symbol.

It is also possible to remove all measurements from the joined measurement file by right-clicking on the joined measurement file and selecting Remove All Joined Measurements.

ADDING DESCRIPTIONS TO MEASUREMENT FILES AND FOLDERS


Custom descriptions can be added to measurement files and folders. These descriptions can be used, for instance, to further individualize particular files or folders for the purposes of later searches. To add a description, right-click on the file or folder, and select Set Description from the popup menu.

The Measurement Description dialog opens.

Enter the description to the Measurement file description field and click OK. The description is displayed as a tooltip when hovering over the measurement file or folder. Measurement files and folders can also be searched based on the description using Search Folders (see page 52).

CR EAT ING TH E D ATABAS E

65

To add a description to several measurement files or folders, right-click on the measurement files or folders while holding down the shift key, and select Set Description from the popup menu. To edit an existing description, right-click on the measurement file or folder and select Set Description from the popup menu.

DELETE FOLDER CONTENTS


To permanently delete the contents of a particular folder from the database, right-click on the folder and select Delete Folder Contents from the popup menu.

The following prompt appears.

Select Yes to permanently delete the measurement files from the database.

RETRIEVING ORIGINAL FILES FROM THE DATABASE


Note: This feature is supported by Nemo Analyze Enterprise Edition only. To retrieve original measurement files from the database, right-click on a measurement folder or file in the Workspace, and select Retrieve Original Files from the popup menu.

The Retrieve Files dialog opens.

Retrieve to folder defines the folder to which the retrieved files will be saved. Retrieving file displays the file that is currently being processed. File N/NNN displays the progress of the retrieval process. Retrieve begins the retrieval process. Define a target location for the retrieved files with the Retrieve to folder control and click the Retrieve button.

CR EAT ING TH E D ATABAS E

67

DATABASE CLEANUP
To clean up old files from the database, right-click on All Measurements in the Workspace and select Clean Up from the popup menu.

The Clean Up dialog opens.

Define the time range from which the measurements are to be deleted and click Delete.

VIEWING MEASUREMENT DATA


Measurement data can be viewed in different kinds of data views, such as graphs, grids, numerical views, and maps. A workbook is a collection of data views and Nemo Analyze offers some readymade workbooks for analyzing measurement data. You can also easily make your own workbooks and save them for later use.

PARAMETER TREE
The fastest way to view measurement data is through the Parameter Tree. To open a parameter in its default view, select the relevant measurement files in the Workspace and then double-click a parameter in the parameter tree.

M AN AG ING W ORKBO OKS

69

The selected parameter is opened in the default view. If you want to open the parameter in some other type of view, right-click on a parameter in the parameter tree and select the view type from the popup menu.

You can also drag and drop parameters in data views from the Parameter Tree. Nemo Analyze will display a green icon if the data view is suitable for the parameter, and a red icon if the data view is unsuitable. In the Parameter Tree you can search and filter parameters by typing the parameter name in the field at the top of the parameter tree view. The parameter tree will display only the defined parameters.

Some parameters require further definition before they can be viewed in any given data view. When opening these parameters in a data view, a dialog box appears asking for specifications applicable to the selected parameter. For example, if you want to view RX level full selected BSIC in a graph, in the Fill Parameters dialog you can select a BSIC value from the drop-down menu in the Value field. Click Finish, select the graph type, and the parameter is displayed in a graph according to the BSIC specifications.

The RX level full parameter values are shown according to BSIC value 18.

M AN AG ING W ORKBO OKS

71

Statistics/Statistics with filters over Parameter


In addition to running reports on measurement files, it is also possible to run statistics on a single parameter straight from the Parameter Tree. The statistics and statistics with filters can be run over a single measurement file, or over a folder containing several measurement files (e.g., the All Measurements folder). You can run parameter statistics over a single measurement file by accessing the Parameter Launchpad. To do this, select a measurement file in the Workspace, select a parameter in the Parameter Tree, and click on the button displayed at the bottom left corner of the Parameter Tree. For more information on the Parameter Launchpad, see page 77. Alternatively, you can run statistics over a single measurement file by selecting a measurement file in the Workspace, then right-clicking on a parameter in the Parameter Tree, and selecting Statistics By or Statistics With Filters By.

Statistics by/Statistics with Filters by: No Grouping


To run statistics over a single measurement file without grouping the data in any way, select a measurement file in the Workspace, right-click on a parameter in the Parameter Tree, and select Statistics by | No Grouping. To run statistics with filters over a single measurement file without grouping the data in any way, select a measurement file in the Workspace, right-click on a parameter in the Parameter Tree, and select Statistics With Filters By | No Grouping.

M AN AG ING W ORKBO OKS

73

Below is a workbook with statistics on RSCP best active set when run on a single file, based on No Grouping.

Benchmarking can be done by dragging parameter statistics on different files or folders in the same graph.

Statistics by: Fixed Geographical Bin Area


Statistics by Fixed Geographical Bin Area and Bounding Geographical Bin Area make it possible to view statistics on map in area bins. With Fixed Geographical Bin Area, the user can select an area (e.g. a portion of the measurement route) and perform area binning on it. With Bounding Geographical Bin Area, area binning will be automatically performed on the entire measurement route. To produce statistics by fixed geographical area, select a measurement, right-click on a parameter, and select Statistics by | Fixed Geographical Bin Area. The Select Rectangular Area dialog box appears.

Select a map in the Map | Filename field, and define the area on the map which you want to include in the query with your mouse. The area binning statistics are calculated using linear math. Because of this, it is possible that with low zoom levels and non-linear map projections the results do not plot correctly on the map. This is most apparent when using a map of the whole world or continent. With city level maps the results are generally very accurate. Click OK. The bins show the average value according to the color set in each bin.

M AN AG ING W ORKBO OKS

75

It is also possible to view statistics based on Minimum, Maximum, Sample count, Std. deviation and Variance. The statistics are by default calculated based on Average. Go to the side panel, right-click on the parameter in the Layers view, and select Properties from the popup menu.

The Properties dialog opens.

Select a statistics type from the Statistics drop-down menu and click OK. The area binning layer is drawn based on the selected statistics type.

Statistics by: Bounding Geographical Bin Area


Statistics by Fixed Geographical Bin Area and Bounding Geographical Bin Area make it possible to view statistics on map in area bins. With Fixed Geographical Bin Area (see page 74), the user can select an area (e.g. a portion of the measurement route) and perform area binning on it. With Bounding Geographical Bin Area, area binning will be automatically performed on the entire measurement route. To produce statistics by bounding geographical bin area, select a measurement, right-click on a parameter, and select Statistics by | Bounding Geographical Bin Area. It is also possible to view statistics based on Minimum, Maximum, Sample count, Std. deviation and Variance. The statistics are by default calculated based on Average. Go to the side panel, right-click on the parameter in the Layers view, and select Properties from the popup menu.

The Properties dialog opens.

Select a statistics type from the Statistics drop-down menu and click OK. The area binning layer is drawn based on the selected statistics type.

M AN AG ING W ORKBO OKS

77

Parameter Launchpad
To produce parameter statistics, such as Cumulation & density, Count, Average, Minimum, Maximum, Standard deviation, Variance, Mode, Median, Midrange, and Histogram, based on measurement data, select a parameter and click on the Tree. button displayed at the bottom left corner of the Parameter

The Parameter Launchpad view opens. Open in data view opens the parameter with the settings defined in the sections Presentation, Statistics and Details. Apply filters, if selected, opens a dialog for defining filters for the parameter after you have defined all the parameter settings and selected Open in data view. Open in defines the data view type used in displaying the parameter data. Statistics (see below) Details displays the parameter description if it exists. Change defaults enables you to change parameter default settings using the Change Defaults dialog (see p. 71).

From the Statistics submenu, you can open various parameter statistics in spreadsheet format. To include more than one parameter statistics type on the same spreadsheet, drag and drop each statistics type on the spreadsheet. Available statistics types include Cumulation & density, Count, Average, Minimum, Maximum, Standard deviation, Variance, Mode, Median, Midrange, and Histogram. Apply filters, if selected, opens a dialog for defining parameter filters once the parameter is dragged and dropped onto a data view. Weight by defines whether the statistics are to be weighted by distance, by time, or by sample (no weight). Group by defines how the statistics are to be grouped.

VIEW ING M EASUR EM EN T D ATA

79

Change Defaults
Through the Parameter Tree view you can change the default color set and default graph scales for each parameter. Right-click on a parameter and select Change Defaults. Alternatively, you can also change defaults using the Parameter Launchpad (see Parameter Launchpad on page 77).

The General tab of the Change Defaults dialog enables you to select a default color set for the parameter, select the image, and set the graph scale top and bottom values.

The View tab of the Change Defaults dialog enables you to set the default data view for the parameter.

The Drill down tab of the Change Defaults dialog enables you to select the drill down workbook that is being used by default when double-clicking on a parameter in the parameter tree.

VIEW ING M EASUR EM EN T D ATA

81

VIEWING MEASUREMENT DATA IN GRAPHS


This use case describes the steps for viewing measurement data in various graphs. It is assumed that there are measurement files in the database. Select a measurement in the workspace. You can locate a measurement file quickly by typing the name of the measurement file in the Filter field, which is marked by the red frame below. This filtering functionality for measurement files in the Workspace is especially useful with a large number of measurement files in the database, or in a specific folder. Select a measurement in the Workspace, right-click on a parameter and select Open | In, e.g. Graph.

In the Choose Graph Type dialog, select the graph type in which you want to show the parameter, in this case a line graph.

The BLER parameter is opened in a line graph.

VIEW ING M EASUR EM EN T D ATA

83

VIEWING MEASUREMENT DATA ON MAP


This use case describes the steps for viewing measurement data on map. For the purposes of this use case, it is assumed that there are measurement files with coordinates, BTS files, and corresponding maps in the database. For loading files to the database, see Adding Files to the Database on page 33.

Changing default map


To change default map (displayed in bold), right-click on the map file name on Workspace view and select Default Map from the menu.

Opening Measurement Files on Map


In the Maps Folder of the Workspace window, double-click on a map (.tab) or Geoset (.gst) file. The map file is loaded to the Analyze database. Now double-click on the map file in the Loaded view.

VIEW ING M EASUR EM EN T D ATA

85

The map view opens.

For adding more map layers to map view, see Use Case 11.: Adding Map Layers and Saving Layer Combinations as Geosets on page 198. Go to the Measurement file page in the Workspace view, and select a measurement file. Then select a parameter from the Parameter Tree, and drag it on the map.

VIEW ING M EASUR EM EN T D ATA

87

The measurement is shown on the map.

The route is opened on the map and colored using the default color set for the selected parameter. You can zoom in and out on the map with the mouse wheel. The side panel displays the selected color set.

Opening BTS Files on Map


Click the base station icon the Base Stations view. at the bottom of Workspace to switch from the Measurements view to

Drag a BTS file from the Workspace to the map.

Click OK when Nemo Analyze asks if the added BTS should be associated with the route. The base station icons are drawn.

VIEW ING M EASUR EM EN T D ATA

89

To change the parameter color set used for coloring the BTS icons, right-click on the BTS layer in the side panel under Layers, and select Properties.

The Properties dialog opens. Select the Color tab, change the setting of both Parameter and Color set drop-down menus, and click OK.

The BTS icons are colored based on the selected color set.

VIEW ING M EASUR EM EN T D ATA

91

Viewing Notification Icons on Map


Notification icons enable specific measurement events to be viewed on map. Right-click on the route layer in the side panel (or directly on the measurement route on map) and select Properties from the popup menu.

The Properties dialog opens. Go to the Notifications tab. Select all notifications that you are interested in. For example, we might want to view Cell reselections.

Select the notifications you want to be displayed and click OK.

If the selected notifications occur in the measurement file, they are drawn on the map as small icons.

You can now zoom in on the map and move the current location marker on the notifications. If all the different views (line graphs, grids, maps, etc.) are synchronized, all data views will jump to the same point in the measurement, allowing you to analyze all relevant parameter data before, during and after the notification was logged.

VIEW ING M EASUR EM EN T D ATA

93

Use Case 1.: Viewing Cell Footprints and RSCP Footprints


Nemo Analyze can automatically create a cell footprint or an RSCP footprint map plot for every cell measured. Cell/RSCP footprint is displayed for every cell whose signal has been among the three strongest at some point during the measurement session. The footprint of each cell is displayed on map on a separate page, allowing you to browse from footprint to another and immediately see both the footprint and the cell. Cell footprint/RSCP analysis can be performed on both UMTS scanner and mobile data. Analysis will not work properly if there will be hundreds of pages in the results. Therefore, it is advisable to use filters to limit the amount of results. To view cell/RSCP footprints, right-click a measurement file in Workspace | Measurements | Folder Contents, or a measurement folder in Workspace | Measurements | Folders, and select Analyses | Ec/NO Cell Footprints (mobile) / Ec/NO Footprints (scanner) / RSCP Cell Footprints (mobile) or RSCP Cell Footprints (scanner) from the popup menu.

A dialog box allowing you to choose between Scrambling code filter and Channel number filter appears. You are also able to select if the color legend and the entire route is shown.

Select the filter of your choice and click OK. The footprints view opens, displaying the footprint of each cell as a separate page.

You can browse from cell to cell by selecting pages from the tabs below the map view.

VIEW ING M EASUR EM EN T D ATA

95

Use Case 2.: Viewing Uplink Voice Quality Server Data


This use case describes the steps for viewing uplink voice quality server data together with mobile terminal (downlink) data from the same measurement session.

Step 1: Loading Voice Quality Measurement Data into the Database


To load an uplink voice quality measurement session into the database, select File | Measurement | Open Measurement from the Ribbonbar. The Open dialog opens.

Select both the server measurement file and the mobile terminal measurement file. You can do this (by holding down the Shift key while selecting the files. Once both files have been selected, click Open. Note: In order for the uplink server data to be displayed correctly, the database must contain both the server (UL) measurement file and the mobile terminal (DL) measurement file. Nemo Analyze will correlate these files automatically. The files are loaded into the database. You can monitor loading progress by using the Database Loader (see page 271).

Step 2: Running Queries on Uplink Voice Quality Server Data


Once both files have been loaded into the database, select the downlink mobile terminal file (i.e. do not select the uplink server file) in the Workspace, right-click on a relevant audio quality parameter, and right-click on a relevant audio quality parameter and select Open In | [Data view].

VIEW ING M EASUR EM EN T D ATA

97

A workbook is opened based on the measurement file.

VIEW ING M EASUR EM EN T D ATA

99

Use Case 3.: Viewing IP/UDP Packet Trace Data


Note: Microsoft Network Monitor 3.3 must be installed BEFORE Nemo Analyze to enable complete IP/UDP trace details to be displayed in Nemo Analyze. For installing the Microsoft Network Monitor 3.3, see Installing Microsoft Network Monitor 3.3 below. Note: If Microsoft Network Monitor 3.3 is nevertheless installed after Nemo Analyze, Nemo Analyzes Microsoft Network Monitor 3.3 functionality can be activated by copying the file NMAPI.dll from the folder C:\Program Files\Microsoft Network Monitor 3\ to the folder C:\Program Files\Anite\Nemo Analyze. Note: If Microsoft Network Monitor 3.3 is not found on the PC or Nemo Analyze has been installed before Microsoft Network Monitor 3.3, Nemo Analyze will display only a limited set of IP/UDP trace details. With IP packet capturing, network packets sent between IP addresses are stored in log files and can be post-processed with a third party application such as Ethereal. The IP packet data is stored in a separate file (.pcap) for each measurement terminal for which IP capturing is enabled. This use case describes the steps for viewing IP trace data based on a measurement file and the corresponding IP trace (.pcap) file. Both files must be from the same measurement session with identical time stamps. First, a relevant workbook (e.g. HSDPA full details) is opened on the measurement file. Next, a separate workbook on the IP trace file corresponding with the measurement file. The two workbooks are automatically synchronized, enabling the correlation of events on the parameter level with events on the IP trace level. Note: In order for the synchronization of the two files to work, the time stamps of the measurement file and the IP trace (.pcap) file must be identical, i.e. both files have to be from the same exact measurement session.

Step 1: Installing Microsoft Network Monitor 3.3


Go to Microsoft website at <http://go.microsoft.com/fwlink/?LinkID=103158&clcid=0x409> to download the free Network Monitor installation package. Select the NM33_x86.exe installation file and click the Download button. Once the installation file has been successfully downloaded, double-click on the file.

The Open File Security Warning dialog opens.

Select Run. The Microsoft Network Monitor 3.3 dialog box opens.

Click Yes to continue installation. The Microsoft Network Monitor 3.3 Setup Welcome dialog opens.

Click Next. The Microsoft Network Monitor 3.3 Setup End-User License Agreement dialog opens.

VIEW ING M EASUR EM EN T D ATA

1 01

Select I accept the terms in the License Agreement option and click Next.

The Microsoft Network Monitor 3.3 Setup Choose Setup Type dialog opens.

Select Typical.

The Microsoft Network Monitor 3.3 Setup Ready to Install dialog opens.

Select Install. Once the installation procedure has finished, the Microsoft Network Monitor 3.3 Setup Completing the Setup Wizard dialog opens.

Select Finish.

VIEW ING M EASUR EM EN T D ATA

1 03

Step 2: Open a Workbook Based on a Measurement File


Open a workbook based on a measurement file by right-clicking on the measurement file in the Device Menu and selecting a relevant workbook (e.g. HSDPA full details) from the popup menu.

A workbook is opened based on the measurement file.

Step 3: Load IP Trace Files (.pcap) into the Database


Next, select the IP Traces icon from the panel below the Parameter Tree. The IP Traces page of the Workspace opens. If there are no IP trace files on the IP Traces page, right-click on the page background and select Open from the popup menu.

The Open dialog opens.

Browse for .pcap files, select the files you want to load to the database, and click Open.

VIEW ING M EASUR EM EN T D ATA

1 05

Step 4: Open an IP Trace Information Workbook Based on the Corresponding IP Trace File (.pcap)
The IP Traces page displays all .pcap files stored in the database. Double-click on the IP trace file that corresponds with the measurement you opened the measurement file workbook based on in Step 1.

The IP trace file is opened in a separate IP trace information workbook that is automatically synchronized with the measurement file workbook.

Step 5: View Synchronized Measurement and IP Trace Data


Go to the measurement file workbook and select an event that is of interest to you (e.g. a point in time where Throughput suddenly drops).

Switch back to the IP trace workbook and observe the IP trace messages surrounding the selected event (i.e. the possible problem causes on the IP level).

VIEW ING M EASUR EM EN T D ATA

1 07

PARAMETER FILTERING
It is also possible to filter all parameters based on, for example, time and region. Select a measurement file, right-click a parameter in the parameter tree, and select Open With Filters in | [dataview]. Here we will use L3 signaling as an example.

An empty workbook, and the Fill Parameters dialog box to define the filtering criteria appear.

In the Fill Parameters dialog you can further limit the query results by adding different filters. Filters can be applied to parameters that are part of the query.

VIEW ING M EASUR EM EN T D ATA

1 09

In this case, Data transfer direction with the value 1 is added.

Click Next. A grid with the parameter according to the specifications appears.

Filtering Based on Polygon Area


With parameter filtering it is also possible to limit a query based on a user-defined area, i.e. by Polygon area selection. As a consequence, the results of a query will be limited to a user-defined part of a measurement route in the chosen data view. This is useful, for example, when you want the values of a parameter to be viewed with a value-based color set only on a part of the route on a map. In this case, the rest of the measurement route will be colored with a default color. In this example we will select Ec/N0 best active set as the parameter. Select a measurement file in the Workspace and type Ec/N0 best active set in the Parameter Tree filter field. Right-click on the parameter and select Open With Filters In | Map.

VIEW ING M EASUR EM EN T D ATA

1 11

An empty workbook and the Analyze Wizard - Fill Parameters dialog box appear. In the dialog box select Area under Name, and press the tab.

The Select Area dialog box appears. Select a map in the Filename field, or load one through the browse button. Next, define an area on the map using the left mouse button. The Save As button enables you to save the map with the selected area for later use. Click OK.

The Analyze Wizard Fill Parameters dialog box appears. Click Finish.

A new workbook opens. The measurement route on the map is colored with the Ec/N0 color set over the section of the route specified with the polygon area. The rest of the route is colored with the default color, here blue.

VIEW ING M EASUR EM EN T D ATA

1 13

Use Case 4.: Global Parameter Filtering Based on a Secondary Parameter


Service providers often define the extent of coverage area using a specific parameter as a threshold value. In Nemo Analyze, it is possible to use this threshold value for filtering out the irrelevant data coming from outside coverage area. By setting a threshold condition for a parameter value, you can define a global filter that will be applied to all subsequent Nemo Analyze operations (apart from Crystal Reports report templates). For the purposes of this use case, all data with Received Signal Code Power (RSCP) of -100 or higher will be considered measurement data from coverage area. The global filter created based on this condition will exclude all data with RSCP values lower than -100 from all subsequent Nemo Analyze operations. In other words, all Nemo Analyze operations will be filtered based on a secondary parameter, RSCP best active set. Filtering by secondary parameter enables the filtering of a primary parameter data set based on a selected secondary one. The resulting data set will contain only those values from the primary data set that coincide with the values in the secondary one. All parts of the primary data set that do not contain RSCP value of -100 or higher will not be considered measurement data from coverage area and thus will be excluded from the processed data set. To define a global filter, select Tools | Global Filters from the Ribbonbar.

The Global Filters dialog opens.

To add a new filter, click Edit. The Analyze Wizard Filters dialog opens.

To add a filter, click Add. An empty filter line is added.

Select <Secondary parameter> from the Name drop-down menu. Click the button in the Value column to browse for a secondary parameter.

VIEW ING M EASUR EM EN T D ATA

1 15

The Analyze Wizard Measurement Parameters dialog opens. Select the parameter RSCP best active set and click Next.

Another Analyze Wizard Filters dialog opens.

To add a filter, click Add. An empty filter line is added. Select the parameter Received signal code power from the Name drop-down menu. Next, select the operator >= (i.e. greater or equal) from the operator drop-down menu. Finally, enter the value -100 in the Value field.

Click Finish. This returns you to the first Analyze Wizard Filters dialog.

Click Finish.

VIEW ING M EASUR EM EN T D ATA

1 17

The Active global filters field of the Global Filters dialog now displays the defined filter.

All subsequent operations performed with Nemo Analyze during this session will be filtered based on the set threshold condition. To save the filter for use also during subsequent Nemo Analyze sessions, click Save. The Save Filter dialog opens.

Click OK. The Saved global filter sets field now displays the saved filter.

To activate a saved filter set, select a saved filter set in the Saved global filter sets and click the Set Active button. To remove a global filter, select the filter in the Saved global filter sets field and click Delete.

It is also possible to set a particular cell as filter by right-clicking on map on top of base station sector and selection Create Global Filter From Cell ID... From then on, all workbooks and queries will be

VIEW ING M EASUR EM EN T D ATA

1 19

limited in the way that results are returned only from those points where the cell in question has been serving.

FILTERING DATA AND CREATING CUSTOM KPIS


Custom KPIs can be created either by correlating parameters using the Query Manager or, to create more complex KPIs, by using the KPI Workbench (see p. 345). With the correlate parameters functionality it is possible to create events based on complex search criteria, create events showing problems and reasons for problems, and create a CSV table from the measurement data for export purposes. You can create new Correlate parameters KPIs in the Query manager. In the Ribbonbar, select View | Query manager | Add | Correlate parameters. The created KPI will be added to the User-branch in the Parameter tree in the Workspace. The correlate parameters functionality can also be used for creating a custom KPI by right-clicking on a device in the Workspace, and selecting Correlate Parameters, and Next. In this case the new KPI will not be saved. Right-click on a device in the Workspace, and select Correlate Parameters.

D AT A VI EW S

1 21

In the Correlate Parameters dialog, select the parameters that you want to correlate. You can add two or more parameters in the Selected view. There are three different Modes available: Show values when first parameter changes (left outer join), Show values when any of the parameters changes (outer join), and Show values when all parameters are valid (inner join). The Show values when first parameter changes (left outer join) mode creates an output row for every value of the first selected parameter. For other selected parameters, it gives the current, previous, or next value from the time stamp of the first parameter. This mode can used to find all events of a certain type in the measurement data, and the values of other parameters at the time the event occurred. For example, it can be used to find all dropped calls (the first parameter), the last TX power, Best active Ec/N0, and Best active RSCP values prior to each drop.

After selecting the parameters to be correlated, click Next. In this example you can find six different parameters in the Selected view, Call dropped, Percentage of DL power up commands, Percentage of UL power up commands, Ec/N0 best active set, and RSCP best active set.

Select the data view in which the data will be displayed, for example, a grid, and click Finish. The correlation data is opened in the selected data view.

The Show values when any of the parameters changes (outer join) mode creates an output row when any of the selected parameters changes. This mode can also be used when exporting data to Excel. In the example below, there are three parameters in the Selected view, RX level full, Ec/N0 best active set, and BLER. Click Next. The correlation data is opened in the selected data view.

D AT A VI EW S

1 23

As you can see, there are no values for RX level full and RAC in rows 4600-4616, and no values for Ec/N0 and BLER in rows 4614-4620.

The Show values when all parameters are valid (inner join) mode creates an output row when any of the selected parameters changes, and all the parameters are valid. This mode can be used when generating input for scatter graph, or when creating events based on custom search criteria. In the example below you can see a custom KPI: all rows where Ec/N0 < -7, BLER, and RSCP < -79 are shown.

D AT A VI EW S

1 25

DRILL-DOWN
The Troubleshooting toolkit with drill-down is an optional component of Nemo Analyze. When a query has been performed with one of the Troubleshooting parameters from the Parameter Tree, it is possible to drill down into further event detail from the data view created by the query. With pie charts, you can do this by double-clicking a sector of the pie chart or a sector color in the pie chart legend. As each sector represents a problem cause, double-clicking a sector will open a grid with a list of all problem events caused by that particular problem cause. Each drill-down from the same chart will open a new tab in the same window. These tabs are displayed on the left side of the window with the colors of the corresponding sectors.

Open tabs representing results of drill-downs from different sectors of this pie chart. The colors correspond with the pie chart sector colors.

For example, green color represents RACH Failure, unknown reason on the pie chart above. To drill down to the details of events where a RACH failure has occurred because of an unknown reason, double-click either the green sector on the chart or the text Failure, unknown reason on the chart legend.

This opens a grid where all failure events in the measurement data resulting from an unknown reason are gathered (see below). This grid will be represented by a green tab on the left of the window in the pie chart view.

Click blue arrow to return to the pie chart.

Click green tab to return to this grid from the pie chart.

Before drilling down into a single row of a grid, you can set a time range for log entries before and after the selected event that are to be included in the drill-down. To set a time range, select Tools | Drill-down from the Ribbonbar and enter the time range in seconds in the Before and After fields.

D AT A VI EW S

1 27

To drill down into a single row of a grid, either double-click or right-click on the selected row. Doubleclick will perform the drill-down if there are default drill-down parameters that are compatible with the measurement data on that particular row. If however there are not, you can either right-click on the selected row and select Drill-Down | Pick Parameter to drill down into the data on that row with other parameters, or open the row in a separate workbook by right-clicking on the row and selecting DrillDown | [workbook] (e.g. UMTS | UMTS Troubleshooting.)

Each drill-down opens the selected time range of measurement data with the selected new parameters in new tabs on the second tab row.

Drill-down tab The selected time range

Drilling Down from Map Data View


The Troubleshooting toolkit with drill-down is an optional component of Nemo Analyze. When a parameter is on opened on map, it is possible to drill down into further event detail by opening a workbook on a selected portion of the measurement route. To drill-down from map data view, select the route portion you want to see details on by left-clicking the position icon signifying the location of the measurement device in relation to the route.

To further define the extent or the time range of the target portion of the route (i.e. a time range surrounding the time stamp of the current location of the position icon), select Tools | Drill-down from the Ribbonbar, and enter the time range in seconds in the Before and After fields.

Right-click on the position icon, and select Drill-Down | [select system] | [Select workbook] from the popup menu.

D AT A VI EW S

1 29

The drill-down workbook is added as a tab to the original map data view.

To return to the map view, click the

tab in the upper left corner of the workbook.

Drilling Down from Troubleshooting Parameters in Map Data View


The Troubleshooting toolkit with drill-down is an optional component of Nemo Analyze. When a troubleshooting parameter is on opened on map, it is possible to drill down into further event detail on a failure event displayed on map. To open a troubleshooting query on map, select a measurement file in the Folder Contents section and a relevant failure-related troubleshooting parameter from under the Troubleshooting item in the Parameter Tree. Note that the Parameter Tree contains also versions of the failure parameters that do not belong to the Troubleshooting package and thus are not compatible with drill-down To search for failure-related troubleshooting parameters from the Parameter Tree, type the word drop or failure in the filter field above the Parameter Tree.

D AT A VI EW S

1 31

Right-click on the relevant failure parameter and select Open In | Map from the popup menu. The results of the troubleshooting query are opened on map.

To drill down to a failure event, double click on the event symbol. The drill-down workbook added as a tab to the original map data view.

To return to the map view, click the

tab in the upper left corner of the workbook.

D AT A VI EW S

1 33

DATA VIEWS
In this chapter are explained the various types of data views available in Nemo Analyze. In Nemo Analyze synchronization is used to keep the different data views in sync. When you click any view containing data, all the other views jump to the same point of time. Nemo Analyze also offers the convenience of the so-called drag & drop function. You can easily drag measurements or measurement parameters from the Workspace in the various data views. Thus, one data view can contain many different parameters from one or more measurement files. When you right-click on any of the data views, a popup menu will open up. The menu offers access to all functions and tools related to that data view. The contents of the menu vary depending on the data view type (graph, grid, map, etc.) but some of the items are common to all data views.

GRAPHS
Graph Tools Menu
Note: The Graph menu is displayed as part of the Ribbonbar only when a graph is open and in current use in Nemo Analyze. When you select another workbook or window, the Graph menu is closed automatically.

enables you to change the mouse cursor to an arrow and scroll the view. The same function Scroll is available in the graph side panel. Zoom enables you to zoom the view. Hold down the left mouse button and drag to the right to zoom in. Hold down the left mouse button and drag to the left to zoom out. The same function is available in the graph side panel. Zoom In and Zoom Out enable you to zoom in and out one step at a time. The same functions are available in the graph side panel. You can also zoom in and out on the view with the mouse wheel. Toggle 3D mode enables you to switch the 3D mode on/off. The same function is available in the graph toolbar. Note that this function is available for bar graphs only. Toggle side panel enables you to hide and display the side panel on the right-hand side of the grid. For more information, see Side Panel on page 139. Bin data enables you to define data bins for scatter graphs and color grids. In the Bin Data dialog, select the Enabled option to activate data binning. When data binning is activated, the scatter graph or color grid is divided into four data bins (defined by bin limits). For each data bin, the percentage of the samples out of all samples, or the number of samples in that bin is displayed.

Reference Lines | Horizontal and Reference Lines | Vertical enable you to add a vertical or horizontal reference line in the graph. You can move the line by dragging with your mouse. The same functions are available in the graph side panel.

Graph Popup Menu


From the graph popup menu you will find all graph-related tools and functions. Right-click on a graph to access the menu.

Change Graph Type


With the Change Graph Type function you can change the type of a graph. In the Choose Graph Type dialog, select the graph type, line, bar, scatter, pie, color grid, or surface graph, from the selection and the graph is opened

Reset
With the Reset function you can empty the data view. All the data and layout settings are removed.

D AT A VI EW S

1 35

Query
With the Query tool you can change the data displayed in the data view. With the Cut, Copy, and Paste functions you can copy or cut an existing query from one data view and paste it to another. This is especially useful if you want to open another view with the same data but with a different graph type or copy data from a graph to a grid. The copied queries are available in the Query Clipboard.

Pick Parameter
With the Pick Parameter tool you can add more data in data views. In graphs a new layer is added, in grids the new parameter is added in the same grid, and in maps a new route is added. Right-click on the graph and select Pick Parameter. In the Pick Parameter dialog, select the measurement file that you are currently viewing and then select the new parameter that will be displayed in the data view. Click OK to add the new parameter.

In the Filter field you can type the name of a specific parameter so that only that parameter will be shown. For example, in the picture below, you can see all the parameters starting with Missing Neighbor. This is useful for locating a parameter when the list is long.

Add Function
With the Add Function feature you can add some averaging functions in graphs.

Color defines the color of the function line. Source defines the parameter for which the averaging function is performed. Function defines what kind of averaging function is added. Average calculates the average value for the selected parameter. Exponential moving average calculates the average value using the following formula: 2 / Period value + 1. Period refers to the number of samples included in the calculation. High calculates the highest value for the selected parameter. Low calculates the lowest value for the selected parameter. Median calculates the median value for the selected parameter. Mode calculates the mode value, i.e., the most common value for the selected parameter. Moving average calculates the moving average value for the defined period (number of samples). Trend calculates the trend line for the selected parameter.

Period refers to the number of samples included in the calculation.

Add Reference Line


With the Add Reference Line tools you can add a vertical or horizontal reference line in the graph. You can move the line by dragging with your mouse. The same functions are available in the graph side panel.

D AT A VI EW S

1 37

Tool
Under the Tool item you will find two graph viewing tools: scroll and zoom. With the Scroll tool you can change the mouse cursor to an arrow and scroll the view. The same function is available in the graph side panel . With the Zoom tool you can zoom the view. Hold down the left mouse button and drag to the right to zoom in. Hold down the left mouse button and drag to the left to zoom out. The same function is available in the graph side panel . With the Zoom In and Zoom Out toolbar buttons you can zoom in and out one step at a time. The same functions are available in the graph side panel. You can also zoom in and out on the view with the mouse wheel.

Mode
When you have several layers open in a graph, you can change the layout of the page. Mode defines how the different layers in the graph are displayed. In Single mode, the layers are displayed overlapping each other as in the example below. The scale is displayed only for the active layer (highlighted with light-blue color in the Layers panel on the right).

In Stacked mode, the layers are displayed one layer after another and the scale is displayed for each parameter. You can change the order of layers in graph by right-clicking a layer in the Layers panel on the right and selecting Move Up or Move Down.

In Automatic mode, all layers that have the same Y-axis are automatically stacked into one graph (similar to Single mode). Layers that have different scale are shown in Stacked mode.

D AT A VI EW S

1 39

Side Panel
With the Side Panel option you can hide and display the side panel on the right-hand side of the graph.

The Tools panel contains tools for viewing graphs. See page 137 for more information on the zooming tools and page 136 on the reference line tools.

With the Bin data tool you can define data bins for scatter graphs and color grids. In the Bin Data dialog, select the Enabled option to activate data binning. When data binning is activated, the scatter graph or color grid is divided into four data bins (defined by bin limits). For each data bin, the percentage of the samples out of all samples, or the number of samples in that bin is displayed.

Bin limits defines the horizontal and vertical limits for the data bins. Output format defines if the data binning values are shown in percentages or the number of samples. In the example below, the horizontal bin limit is 50, vertical limit -12, and the output format is percentage. The percentage values are displayed in the corners.

D AT A VI EW S

1 41

The Layers panel displays the layers that are currently open in the view. When you add more parameters in the view, a new layer is created for each parameter. You can remove layers from the view by clicking the Close button (x) and hide layers by clearing the respective option. The Numerical Data panel displays selected information from the measurement.

Line Graph Filled


In line graphs, select the Filled option in the graph popup menu to show the graph as "filled".

3D Mode
With the 3D Mode function you can switch the 3D mode on/off. The same function is available in the graph toolbar . Note that this function is available for bar graphs only.

D AT A VI EW S

1 43

Group Values
Group Values enables you to group bars in bar graphs by parameter rather than by x axis values. To enable/disable Group Values, right-click on the bar graph and select Group Values from the popup menu. If Group Values is disabled, bars are sorted based on their x axis values. In the screenshot below, the bars are sorted based on their x axis values, i.e. Scrambling Code/Channel Number.

If Group Values is enabled, bars are sorted by parameter. In the screenshot below, the bars are sorted based on parameter, i.e. Ec/N0 Active Set and Ec/N0 Monitored Set).

Graph Properties
In the Graph Properties dialog you can edit the graph layout. These settings are the same for all graph types. Mode defines if the layers are stacked, in single, or in automatic mode. Background color defines the color for the graph background. Show side panel hides/displays the data panel on the right-hand side of the data view. Axes defines the scales for the left and right x axis.

The Layers tab displays a list of all layers open in the graph. You can add and remove layers. Double-click a layer name or select the layer from the list and click Modify to access the Layer Properties dialog.

D AT A VI EW S

1 45

Layer Properties
Right-click on a graph and select Properties. In the Graph Properties dialog, go to the Layers tab and double-click the layer name. The Layer tab in the Layer Properties dialog is common for all graph types but some of the options might be missing for some graph types. If there are graph type specific settings, they are found in the second tab. Note that all graph types do not have any special settings so there might be only one tab in the Layer Properties dialog. The graph type specific settings are explained in the following chapters.

Name defines the layer name. The name is displayed at the bottom of the data view. Color defines the color for the line or bar. Show marks displays the parameter values in small labels on the graph. Title defines a name for the axis. When the Scale automatically to values option is selected, the software checks what values exist for the parameter in the file and scales the axes accordingly. Alternatively you can define the Top and Bottom scales manually.

Parameter tab allows you to add filters, and modify existing filters and user-definable attributes. For instance, with N parameters (e.g. Ec/N0 Nth best), an N-based filter such as N = 1 is automatically set. To display the 2nd best cell instead of the default 1st best, change the filter value from N = 1 to N = 2.

Layer Properties for Line Graphs

Line width defines the line thickness in pixels. When the Hold value constant until next option is selected, the line is drawn at the same level until the value changes. See examples below.

In the example below, the Hold value constant until next option is not selected.

In the example below, the Hold value constant until next option is selected.

D AT A VI EW S

1 47

Layer Properties for Bar Graphs


Sorting defines the order in which the bars are displayed in the graph. The bars can be sorted in ascending or descending order based on the values on the X or Y axis.

Layer Properties for Scatter Graphs

Style defines the scatter shape.

Use Case 5.: Multiple Graph Layers


You can have as many layers as you like open in a graph. The example graph below has three layers open in stacked mode.

If you have many layers open, it may come in handy to resize the page to fit even more layers in the graph. Right-click on the graph and select Page | Properties.

D AT A VI EW S

1 49

In the Page Properties dialog, go to the Page tab. With the Fixed Size setting you can define the size of the page.

If the Fit to window option is selected, all the layers are displayed in the visible area of the page. This means that when a new layer is added, the space for each layer gets smaller. With the Fixed size option you can define an exact size in pixels for the page. This will give more space for each layer. With the scroll bar you can scroll up and down the page to view all layers.

When you have two layers in single mode you can select to view the scale for both of them. Right-click on the graph and select Properties.

In the Graph Properties dialog, go to the Graph tab. In the left and right Axes setting select the two layers.

In this graph the left axis displays scales for Ec/No best active set, and the right axis for RX level full.

D AT A VI EW S

1 51

Use Case 6.: Notification Icons in Graphs


You can view notification icons in graphs to notify you of special measurement events. Right-click on a graph and select Pick Parameter. From the Pick Parameter dialog, select Notifications and markers and click OK.

From the Notifications Properties dialog, select the notifications that you want to view in the graph and click OK.

The notification icons are displayed in the graph.

D AT A VI EW S

1 53

Use Case 7.: Correlating Parameters Using Color Grids and Surface Graphs
Color grid and surface graphs are graph types designed for correlating two parameters. For the purposes of this use case, we will describe correlating parameters in a color grid. However, all of this information is also applicable with surface graphs. In addition to the default parameter pairs (such as Rx quality sub vs. Rx level sub) that are suggested when you right-click on an empty color grid and select Pick parameter, you can also define your own parameter pairs. To do this, create a new color grid by selecting View | Add Workbook | Graph from the Ribbonbar and once the empty data view opens, right-click on it, select Change Graph Type from the popup menu and select Color grid from the available graphs. Alternatively, if you already have a graph open, you can right-click on it, select Change Graph Type from the popup menu and select Color grid.

Next, right-click on the empty color grid, select Correlate Parameters from the popup menu and a Correlate Parameters dialog opens.

Scope defines the measurement data included in the color grid. Color set defines the color set for the color grid.. X Parameter page defines the properties of X parameter. Parameter defines the X parameter. With Filters, you can define additional filters for the parameter. Minimum defines the minimum value for the parameter (see using filters). Maximum defines the maximum value for the parameter. Steps defines the number of steps on the color grid scales. Y Parameter page (see X parameter) After defining all the relevant values in the Correlate Parameters dialog, click OK and a color grid is created.

D AT A VI EW S

1 55

GRIDS
Grids can be used to display all sorts of data. The data is displayed in numerical format in userdefined columns. You can export grid data to MS Excel or to a text file. You can also use color sets to highlight certain data in the grid. The first three items in the grid popup menu, Reset, Query, and Pick Parameter, are the same as in the graph popup menu. See page 134 for more information on them. Play Audio Sample is displayed only with grids containing an audio quality sample file column.

Side Panel
With the Side Panel option you can hide and display the side panel on the right-hand side of the grid.

With the Search function, you can perform searches in grid views for text and values. Boolean operators are supported. You can highlight, filter in, or filter out matches.

Type in the search string in the Find field, select the columns that will be included in the search, and define if the search results should have matching case. Select the Highlight matches option if you want the matching cells highlighted. When the Filter in option is selected, only the search matches will be visible. When the Filter out option is selected, the search matches will be hidden.

D AT A VI EW S

1 57

With the Search decoded messages option, you can search for grid entries based on a value in the decoded entry data (displayed in the Information panel). In the example below, both Search decoded messages and Filter in options are selected. Thus, only the two grid rows containing the searched decoded information are displayed.

The Layers panel displays the layers that are currently open in the view. When you add more parameters in the view, a new layer is created for each parameter. You can remove layers from the view by clicking the Close button (x).

The Information panel displays the selected message in decoded format. If you want to view the decoded data in a separate view, double-click the message or right-click on the message and select Row Details.

Row Details
With the Row Details function you can decode grid data. The decoded data is opened in an Info View. You can also use double-click to decode grid data.

Export Data to
With the Export Data to function you can export grid data to MS Excel, text file, MapInfo Tab file, and Google KML-File. When you export data to a MapInfo Tab or Google KML-file, you can then open the file as a route on a map. Select the appropriate option, and the exported data is saved on a file. Below is an example of exported data in an Excel sheet.

If you want to export multiple parameters so that each parameter is in its own column in an Excel sheet, first correlate the parameters and view the results in a grid. Then export the data. Note that only visible data is exported. In the Grid Properties dialog, you can select the visible parameters, e.g., latitude and longitude. Additionally, you can copy and paste any query from any control into a grid control where it can be exported. Right-click on a control and select Query | Copy. Then go to the grid, right-click on the grid and select Query | Paste.

D AT A VI EW S

1 59

Grid Properties
In the Grid Properties dialog you can make various settings to the grid layout and select a color set.

From the Visible columns list, you can select the columns to be displayed. To change the order of columns, select the column you want to move and use the arrow buttons on the right to move it.

Add adds a parameter to the grid. Delete removes a parameter from the grid.

To define a color set for a grid column, select the column name, click the button that appears, and select a color set for the grid column from the drop-down menu. For more information, see Show heading hides/displays the column titles at the top of the grid view. Vertical text in heading option turns the column headings vertically. Show row numbers hides/displays the row numbers at the beginning of each row. Show time intervals hides/displays also the time interval in the Time column on each row. Font defines the font, font style, and font size used in the grid. Use Case 8.: Color Sets in Grids

Show heading hides/displays the column titles at the top of the grid view. Vertical text in heading option turns the column headings vertically. Show row numbers hides/displays the row numbers at the beginning of each row. Show time intervals hides/displays also the time interval in the Time column on each row. Font defines the font, font style, and font size used in the grid.

D AT A VI EW S

1 61

Use Case 8.: Color Sets in Grids


In addition to maps, you can use different color sets for different columns in grids to make the data presentation more visual. In grids the color sets work as bar graphs. This use case describes how to use a color set in a grid. Right-click on the grid and select Properties.

In the Properties dialog, select the Color Sets tab. Next, select the column you want to apply a color set to and click the button.

Select the appropriate color sets for the columns from the drop-down menu and click OK.

In the example below, Scr. Code and Ec/N0 columns are colored with the appropriate color sets.

The selected columns display the selected color sets. The size and color of the bar in each cell correlates with the parameter value.

D AT A VI EW S

1 63

Alternatively you can select the Color whole cell option in the Color Sets tab of the Grid Properties dialog. Using this setting, the whole cell will be colored and only the color of the cell will indicate the value.

Use Case 9.: Play Audio Sample


Nemo audio quality measurement files include a sample of received audio quality. This makes it possible to compare the audio quality of the received sample with that of the original sent sample. To play the audio sample, select an audio quality measurement from the measurements and the parameter Audio Sample File Name UL from the parameter tree. Right-click on the parameter and select Open In | Grid from the popup menu.

The measurement opens on a grid. Right-click on the row with the audio sample you want to hear and select Play Audio Sample from the popup menu.

D AT A VI EW S

1 65

Use Case 10.: Using L3 and RRC Message Search Parameters


To search values of particular type from decoded L3 and RRC signaling messages, first select a measurement file in Workspace | Measurements | Folder Contents. Then, select a search parameter depending on your need: Text search parameters (e.g. L3 signaling text search) are used for searching text strings value search parameters (e.g. L3 signaling value search) are used for searching numeric values

Double-click on the selected search parameter.

D AT A VI EW S

1 67

The Analyze Wizard Filters dialog opens.

Search string defines the string after which the searched values can be found in the searched data set. For instance, as LAC values are always in the format LAC: <numeric value>, the value search will employ LAC as the search string. Note that colon (:) character is not required in the search string. The search results will thus contain the numeric values that immediately follow the string LAC. If there are multiple matches for the search string inside a message, the first match is returned. Parameter name allows the user to define a name for the column that displays the searched values in the result data set (see below).

D AT A VI EW S

1 69

MAPS
Nemo Analyze incorporates two mapping methods; MapInfo-format (.tab) maps, based on the MapX engine, and KML format-based live maps. Maps can be used to display the measurement route, base station icons, and notification icons. The route can be colored based on parameter values.

Viewing on Live Map


Nemo Analyze allows you to view measurements on KML format-based live maps, Google Maps, and OpenStreetMaps. An internet connection is required for viewing measurements on live maps. The map tiles are retrieved as needed when map is opened, zoomed, or panned. To ensure the live map functionality to work appropriately, it is recommended to use Internet Explorer 8 or newer with Windows XP Professional and Internet Explorer 9 with Windows Vista/Windows 7 is used. Live maps can be used the same way as regular MapInfo maps. Data and base stations can be plotted on the map and data can be dragged to the map as in MapInfo maps. When playing back measurement data, the cursor is seen on the map and lines are drawn to the connected base stations. Live maps can be found from the Maps tab of the Workspace.

Maps can be used to display the measurement route, base station icons, and notification icons. The route can be colored based on parameter values. OpenStreetMap is included in the Nemo Analyze Professional feature package. OpenStreetMap is a free map resource available globally. The maps are updated with the same open source principle as, for example, Wikipedia. At the moment, the level of details in OpenStreetMap is similar to Google Maps with Street View for the most part. The following figure illustrates OpenStreetmap data overlay:

Google Maps is an optional feature. It incorporates a street view, a satellite view, a hybrid map of satellite and streets, and a physical map with terrain elevation illustrated.

D AT A VI EW S

1 71

Google Hybrid map with route overlay:

Google Street View


If a problem has occurred during data gathering, it is possible to check measurement route spots with the help of Google Street View. Google Street View enables exploring the route spots through 360degree street-level imagery. Select Maps page from Workspace and double-click Google Streets to open the Google Street View. After this, drag and drop a measurement file from Workspace | Measurements onto the map. Right-click on the route and select Show Street View On Route Point from the popup menu.

D AT A VI EW S

1 73

The view in split into two parts as illustrated below:

Hide Google Street View by right-clicking on the route and selecting Hide Street View from the pop up menu.

Map Tools Menu


Note: The Map menu is displayed as part of the Ribbonbar only when a map is open and in current use in Nemo Analyze. When you select another workbook or window, the Map menu is closed automatically.

The drop-down menu in Map | Tools | View enables you to select Arrow, Pan, Zoom In/Out, and Center tools. Arrow Pan enables you to select items on the map view. enables you to scroll the map. enable you to zoom the view. You can also zoom in and out on the

Zoom in and Zoom out views with the mouse wheel. Center

enables you to center the map based on a selected point on the map.

Save geoset enables the saving of customized settings as a MapInfo Geoset file (.gst).

Organize layers enables you to change the order of layers on the map. MapX properties enables you to customize the map view. For more information on MapX tools, see MapX on page 183. Toggle side panel grid. Measure distance enables you to hide and display the side panel on the right-hand side of the

enables you to measure the distance between two points selected on map.

Offset route enables you to move the measurement route to another location. Select the route you want to move and then click on the new location. Highlight value enables you to highlight selected values on map route. First open the Color tab of the Route Properties dialog and set Default Color as Mode and select a parameter from the Parameter drop-down menu. Then click on the route with the Highlight value tool and the Highlight Value dialog will open. The dialog displays all the values of the selected event type. From the list, pick the value that you would like to be highlighted and select a color for that value. Click OK and the value is colored in the selected value. View entire map enables you to zoom out quickly.

Auto centering enables you to force the map to be always centered on the current location. Area binning enables you to define an area on a map which you can run statistics on in the form of bins. The results are displayed on the map as a new semi-transparent layer. Polygon area selection enables you to specify an area of any shape, and run statistics over that area. The results are displayed in the statistics data view. Delta plotting enables you to compare the parameter values of two measurement groups from a same route by defining an area on a map (e.g. a portion of the measurement route). For more information on Delta Plotting, see the topic Use Case 14.: Comparing Two Groups of Measurements from the Same Route on Map.

D AT A VI EW S

1 75

Map Popup Menu


The first items in the map popup menu, Reset, Query, and Pick Parameter are the same as in the graph popup menu.

Add
Under the Add item you will find options you can add on a map view. With the Add Map Layer function you can add a map layer (.tab or .gst) in the same map view. With the Add Measurement function you can add another measurement file on the same map. Select the file from the Add Measurement dialog and click OK.

The new route will be displayed in addition to the existing one(s).

With the Add Frame and Add Comment functions you can add frames and comments on the map view. A frame is a resizable window that can contain any of the available data view types. For example, you can open the color legend in a frame.

A comment is a fixed-size window that is attached to certain coordinates on a map. If you scroll the map, the comment window will move as well. You can remove frames and comments by right-clicking on them and selecting Frame | Remove or Comment | Remove.

D AT A VI EW S

1 77

Find
If you require a more detailed map from the same location, right-click on the route and select Find | Map at Route or right-click anywhere on the map and select Find Map at Position. Note that the Find feature only works with raster maps. The Map Search dialog will be opened with the matching maps. You can choose whether the new map is displayed in a new workbook (Open) or as a layer (Add Layer) on the same map. Nemo Analyze will search for the maps from the location you specified for maps in the Options dialog box (View | Tools | Options | Environment | Default Paths).

In the example below, the new map is pasted on top of the existing map as a new layer.

Generate Color Set


See Automatic Generation of Color Sets

Export to KML File


To export a measurement route from Nemo Analyze map to Google Earth in KML format, in map data view, right-click on the map and select Export to KML file from the popup menu.

Alternatively, you can right-click on the route and select Export Data To | Google KML File from the popup menu.

The Save As dialog opens.

D AT A VI EW S

1 79

To save the route as a Google KML file, type in a file name in the File name field and click Save. To open a KML file with Google Earth, go to Google Earth and select File | Open from the Google Earth main menu. In the Open dialog, select a KML file and click Open. Google Earth displays the route complete with color sets, notification icons, etc.

In some cases, Google Earth may not be able to display the exported data in OpenGL mode. To switch to Direct X mode, select Tools | Options from the Google Earth main menu bar. The Google Earth Options dialog opens. Select Direct X in Graphics Mode section and click OK.

To apply the changes, restart Google Earth.

Side Panel
With the Side Panel option you can hide and display the side panel on the right-hand side of the map.

The Tools panel provides you with controls for manipulating the map view. The function of each tool is described below.

With the Arrow With the Pan

tool you can select items on the map view. tool you can scroll the map. tools you can zoom the view. You can also zoom in and out

With the Zoom In and Zoom Out on the views with the mouse wheel.

D AT A VI EW S

1 81

With the Center

tool you can center the map on the point where you click.

With the Area Binning tool you can define an area on a map which you can run statistics on in the form of bins. The results are displayed on the map as a new semi-transparent layer. With the Polygon Area Selection tool you can specify an area of any shape, and run statistics over that area. The results are displayed in the statistics data view. With the Delta Plotting tool you can and compare the parameter values of two measurement groups from a same route by defining an area on a map (e.g. a portion of the measurement route). For more information on Delta Plotting, see the topic Use Case 14.: Comparing Two Groups of Measurements from the Same Route on Map on p. 212. With the Distance tool you can measure the distance between two points.

With the Offset tool you can move the measurement route to another location. Click on the route you want to move, and then on the new location. With the Highlight tool you can highlight selected values on the map route. First select a scheme in the Route Properties, Color dialog). Then click on the route and the Highlight Value dialog will be opened. The dialog displays all the values of the selected event type. From the list, pick the value that you would like to be highlighted and select a color for that value. Click OK and the value is colored in the selected value. With the View Entire Map With the Organize Layers more information on the tool. tool you can zoom out quickly. tool you can change the order of layers on the map. See page 183 for

The BTS Filter drop-down menu allows you to select which BTS cells are displayed on the map.

The Layers panel displays the layers that are currently open in the view. When you add more routes on the map, a new layer is created. You can remove layers from the view by clicking the Close button (x) and hide layers by clearing the respective option. Right-click on the layer titles in the layer browser to access a set of map-related tools.

With the Zoom To tool you can zoom to the selected route. This is useful if you have several measurements open on the same map. With the Move To Top tool you can move a selected layer to top. With the Find tool you can search another map that matches the measurement route. Note that the Find feature only works with raster maps. See page 177 for more information on the Find tool. With the View In Live Map tool you can view measurements in live maps. With the Export Data To tool enables you to export data to MapInfo Tab-file/Google KML-File. With the Show BTS Connections tool you can enable base station connections to be displayed on the map. With the Add Task tool you can add a new task. With the Remove tool you can remove the selected layer. With the Properties option you can open the Route Properties dialog (see page 185 for more information). The Active Layer panel displays the name of the active layer and information on BTS sites to which the test device is currently connected. The Color Legends panel displays the color sets in use. Click the Add color set new color set in the Select Color Set dialog and click OK. button, select a

Click the Remove all color sets

button to remove all color sets.

D AT A VI EW S

1 83

Tool
Under the Tool item you will find a selection of tools for viewing the map.

These same functions are available in the map side panel.

MapX
Under the MapX item you will find tools related to MapX maps.

With the View Entire Map tool you can zoom out quickly. You can customize the map view using the MapX properties. With the Save Geoset function the customized settings can be saved in a MapInfo Geoset file (.gst) and loaded later on. With the Organize Layers tool you change the order of map layers. In the Layer Control dialog, change the order with the Up and Down buttons. Clear the Visible option if you want to hide a layer from the map.

Properties will open the MapX Properties dialog.

In the Layers tab the table a list of the current layers is shown. The order on the list defines the order in which the layers are displayed on a map; for example, in this case the Route layer is on the top, BTS sites under that, and World map at the bottom. This utility is useful if you want to have a bigger map on the bottom (e.g., a map of Europe) and a more detailed map of a smaller area (e.g., of Helsinki) on top of that. To add layers on the map click Add. An Open dialog box is opened where you can choose the map you want to add as a layer. The map is added on the list and with the Up and Down you can change the maps position on the list. By selecting and clearing the Visible option you can decide whether or not to display the layer on the map. To remove layers from the map, select the layer from the list and click the Remove button.

D AT A VI EW S

1 85

Map Properties
The Properties item will open the Map Properties dialog. In the Map Properties dialog you can define how the BTS icons are displayed. You can also select and clear the Auto centering tool.

Auto centering option keeps the map focus always on the measurement vehicle. Show route names option displays route names on the map. Show current position option displays the current position and direction on the map. Highlight active route option displays the currently selected route highlighted so it is easy to see on the map. Show scale bar option displays a distance scale bar on the map. Select the Draw line to active base station option to draw line from the serving BTS to the test vehicle. Display BTS overlay on top of other layers option keeps BTS layers always on top of the route layers.

Route Properties
In the Route Properties dialog you can modify the appearance and settings of the measurement route on the map. Right-click on the route and select Properties. On the last four pages you have the option to apply the new settings to all routes visible on the map. Name defines the title of the route layer. Draw Mode defines how the route is drawn (line or symbol). See page 188 for more information. Hide long distance lines option hides unusually long lines which probably mean that there is a problem with the GPS connection. Pixel Offset values enable you to move the route on the map horizontally and vertically. Thickness defines the route thickness in pixels. You can define, for example, a thicker line to be drawn when the test device is in high band.

Select measurement events that will be displayed on the map as icons. Select the Print description option to view event descriptions (e.g., marker numbers) on the map. Select the Hide overlapping descriptions option to hide the description text if the texts overlap each other. Pixel offset values enable you to change the position of notification icons on map horizontally and vertically.

D AT A VI EW S

1 87

Mode defines if the route is colored using the default color or using valuebased coloring. If you select color rotation, the color of the route changes whenever the value of the selected parameter changes. Parameter defines the parameter used in the coloring. Filters displays the parameters used in the query. Color set defines the colors for parameter values.

Parameter defines the parameter which will be displayed as numerical values along the measurement route. Font and Color define the text settings. Select the Hide overlapping values option to hide the values if the texts overlap each other.

In the BTS page you can select which lines are drawn to form a measurement route.

Draw Route as Symbols


You can draw the route on a map as symbols. In the Route Properties dialog, select Symbol in the Draw Mode field and click the button. In the Symbol Style dialog, define the appearance of the symbol and click OK.

Font defines the symbol font. Symbol defines the symbol used to draw the route. Color defines the symbol color. Rotation angle defines in which angle the symbol is drawn on the map. Background defines what kind of background is added for the symbol. Effects defines what kind of effects are applied to the symbol.

D AT A VI EW S

1 89

Below is an example of a map route drawn with symbols.

BTS Files
In the workspace you will find a page for Base Stations by clicking on the workspace panel. icon on the bottom of the

Files displays all BTS files stored in the database. File contents displays the contents (i.e. all base stations and the antennas of each station) of the BTS file selected in Files. Site information displays details of the base station/antenna selected in File contents. If there are multiple versions of the same BTS file, i.e. from different dates, the correct file is automatically selected based on the time of the measurement session if the file has been named according the naming convention filename_YYYY-MM-DD.nbf where filename must be identical with all versions of the same BTS file. The date suffix changes with the date (YYYY stands for year, MM for month, and DD for day) of the BTS file. Note: Change the filename format to filename_YYYY-MM-DD.nbf before you load the file to the database.

D AT A VI EW S

1 91

A BTS file needs to be activated in order for it to be used in saved workbooks and BTS ref queries. When opening a BTS file, a Set Active BTS File dialog opens. The user can activate a file by ticking a box next to the file.

Note: You can add custom site information columns to the BTS file. These columns will be displayed as part of site information in the Nemo Analyze UI. If there are multiple versions and the file has not been named according to the naming convention above, the relevant BTS file version must be activated. If none of the files are activated (i.e. designated as relevant files), reliable results cannot be guaranteed. To modify active BTS settings of a file that has already been added to the database, open the BTS page of the Workspace by clicking the Active from the popup menu. button, right-click on the relevant BTS file, and select

To display base stations as BTS icons on map, drag and drop a BTS file from Workspace | Base Stations | Files to a map. The green sectors represent the antenna directions.

D AT A VI EW S

1 93

Double-click on a BTS icon to view more information on that base station.

To connect the measurement route to the BTS, right-click on the route and select Properties. Go to the BTS tab, select the BTS file, and click OK. If the route is not associated with a BTS file, the line to the active base station is not drawn.

If you have several BTS files open on the map, you can define which base stations are displayed. Right-click on the map and select Properties. Select the Draw line to active system base stations to draw only the BTS icons from the system, for example, GSM that the mobile is currently using.

To connect the measurement route to the BTS, right-click on the route and select Properties. Go to the BTS tab, click Modify and select the BTS file to connect to, and click OK. If the route is not associated with a BTS file, the line to the active base station is not drawn.

D AT A VI EW S

1 95

BTS Properties
Right-click on the BTS file layer in the side panel and select Properties.

In the BTS Properties dialog you can edit the appearance of the BTS icons.

Draw line to active base station, option enables the drawing of a line from the serving BTS to the test vehicle.

Icon size defines the size of the BTS icon in pixels. Fixed defines the fixed size of BTS icons in pixels. Dynamic, max size defines a maximum size for dynamic BTS icons (i.e. icons which change size based on zoom level). Show site names option hides and displays the site names. Draw frame hides and displays the outlines of BTS icons. Site transparency defines the transparency of the BTS icons.

Show cell information option hides and displays the selected cell information. Hide cell texts on low zoom levels (performance optimization) option hides cell text when the map is zoomed out so that the map can be drawn faster and with less visual distractions. Use cell beam range from BTS file allows you to enable/disable the display of cell beam range based on cell beam range data in the BTS file. Beam transparency defines the transparency of the beam range layer.

D AT A VI EW S

1 97

Default color defines the color of the BTS icon. Mode (parameter/custom) enables the user to choose whether to use BTS Parameter or custom query option in the base station coloring. Technology allows the user to define the parameter coloring depending on the technology that has been used. BTS parameter defines the parameter used in the coloring. Mode (parameter/custom) enables the user to choose whether to use BTS Parameter or custom query option in the base station coloring. Custom query enables the user to select a custom query which defines how the coloring is done. Color set defines the colors for parameter values. Beam color defines the color that is used to draw a beam of a base station.

The Sites page displays a list of all sites and their cells in the BTS file.

Use Case 11.: Adding Map Layers and Saving Layer Combinations as Geosets
To add more map layers to the map view, drag and drop a .TAB file from the Workspace | Maps | Loaded view to an open map view.

For organizing map layers, see MapX on page 183. Once you have added the layers you need to the map, you can save the entire layer combination as a Geoset to enable direct access to the map in the future. To save the combination as a single Geoset, right-click on map view and select MapX | Save Geoset from the popup menu.

D AT A VI EW S

1 99

The Save As dialog opens.

Enter a name for the Geoset file and click Save. The Geoset Name dialog opens.

Enter a name for the Geoset and click OK. The entire combination of map layers can now be loaded into the database and opened as a single map.

D AT A VI EW S

2 01

Use Case 12.: Coloring Routes Based on BTS Coverage


This use case describes a situation where the user needs to view the coverage of a single base station on a map. It is assumed that there are measurement files with coordinates, BTS files, and corresponding maps in the database.

Step 1: Open a Measurement and a BTS file on a Map


Open a map file and drag and drop measurements on it. Do not open any parameters. Next, drag and drop BTS files on the map and associate them with the route(s). .

Step 2: Color Routes Based on BTS Coverage


To color route(s) based on base station coverage, right-click a base station and select Color Layers Based On Scrambling Code [number] from the popup menu.

The route will be colored based on the Ec/N0 or RSCP value of the selected base station.

To perform area binning on a selected part of the route, proceed to Use Case 13.: Performing Area Binning on page 202.

Use Case 13.: Performing Area Binning


Step 1: Select Bin Area
Click Area Binning icon on the Tools panel and select an area from the map. In the screenshot below, you can see the area selection surrounded by a thin dotted line.

Step 2: Select Measurement File(s)


Selecting an area opens the Analyze Wizard Select Measurement dialog below. In this dialog, you can add measurements you want to perform the area binning on. The measurement route on the map is included in the selected measurements by default. Folder defines the measurement folder whose contents are displayed in the field below. Using the Filter field, you can search for measurements. The Selected measurements field displays all the measurements that have been selected for area binning. With the button, you can select measurements With the button, you can clear measurements

After you have selected the measurements you want to perform the area binning on, click Next. This opens the Analyze Wizard Measurement parameters dialog. If you are performing an area binning for a single base station (see

D AT A VI EW S

2 03

Use Case 11.: Adding Map Layers and Saving Layer Combinations as Geosets To add more map layers to the map view, drag and drop a .TAB file from the Workspace | Maps | Loaded view to an open map view.

For organizing map layers, see MapX on page 2. Once you have added the layers you need to the map, you can save the entire layer combination as a Geoset to enable direct access to the map in the future. To save the combination as a single Geoset, right-click on map view and select MapX | Save Geoset from the popup menu.

The Save As dialog opens.

Enter a name for the Geoset file and click Save. The Geoset Name dialog opens.

Enter a name for the Geoset and click OK. The entire combination of map layers can now be loaded into the database and opened as a single map. Use Case 12.: Coloring Routes Based on BTS Coverage on page 198), type selected in the filter field (as only selected scrambling code and selected channel parameters are relevant for this type of area binning) and select only selected scrambling code or selected channel parameters. Otherwise, select the relevant parameter and click Next.

D AT A VI EW S

2 05

This opens the Analyze Wizard Filters dialog. If you are performing an area binning for a single base station (see Use Case 11.: Adding Map Layers and Saving Layer Combinations as Geosets To add more map layers to the map view, drag and drop a .TAB file from the Workspace | Maps | Loaded view to an open map view.

For organizing map layers, see MapX on page 2. Once you have added the layers you need to the map, you can save the entire layer combination as a Geoset to enable direct access to the map in the future. To save the combination as a single Geoset, right-click on map view and select MapX | Save Geoset from the popup menu.

The Save As dialog opens.

Enter a name for the Geoset file and click Save. The Geoset Name dialog opens.

Enter a name for the Geoset and click OK. The entire combination of map layers can now be loaded into the database and opened as a single map.

D AT A VI EW S

2 07

Use Case 12.: Coloring Routes Based on BTS Coverage on page 198), select a scrambling code or channel number value from the Value drop-down menu and click Finish. Otherwise, select the relevant filters and click Finish.

If you are performing an area binning for a single base station (see Use Case 11.: Adding Map Layers and Saving Layer Combinations as Geosets To add more map layers to the map view, drag and drop a .TAB file from the Workspace | Maps | Loaded view to an open map view.

For organizing map layers, see MapX on page 2. Once you have added the layers you need to the map, you can save the entire layer combination as a Geoset to enable direct access to the map in the future. To save the combination as a single Geoset, right-click on map view and select MapX | Save Geoset from the popup menu.

The Save As dialog opens.

Enter a name for the Geoset file and click Save. The Geoset Name dialog opens.

Enter a name for the Geoset and click OK. The entire combination of map layers can now be loaded into the database and opened as a single map.

D AT A VI EW S

2 09

Use Case 12.: Coloring Routes Based on BTS Coverage on page 198), clicking Finish opens another Analyze Wizard Filters dialog. Again, select the scrambling code or channel number value from the Value drop-down menu and click Finish.

The area binning statistics are by default weighted by time. Although area binning as an operation is location-based, each sample has its unique time and distance weights that can be used in calculating time and distance averages per bin. If you want to weight the results by distance, you must have Calculate statistics based on set to <Ask every time> in Options | Environment | Statistics (see page 501). In this case, the Choose Statistics Type dialog will open after clicking Finish in the Analyze Wizard - Filters dialog above, allowing you to select Distance as statistics type. After selecting the statistics type, click OK.

The area binning results are displayed on the map as a new layer.

Step 3: Adjusting Bin Size


To adjust bin size, right-click the area binning layer on the Layers panel and select Properties from the popup menu.

D AT A VI EW S

2 11

In the Properties dialog, adjust bin size by changing the X steps and Y steps values. The higher the value is, the smaller the bin size. When the bin size has been adjusted, click OK, and the area binning layer is displayed with the new bin size.

Name refers to the name of the base parameter. X steps and Y steps define the bin size. Scrambling code refers to the scrambling code of the base station. Draw method defines whether the bins are displayed in Fixed size irrespective of the zoom level, or in Actual size. With the Statistic drop-down menu, you can select the aggregate displayed on the map: Minimum, Maximum, Sample count, Std. deviation and Variance. By default, average per bin is displayed. Color set defines the color set that is used. Transparency defines the transparency of the area binning layer.

Step 4: Exporting Area Binning Data


To export area binning data to Excel, or to .txt or MapInfo .tab format, right-click the area binning layer on the Layers panel and select Export Data To | Excel or Export Data To | File from the popup menu. For more information on Export Data To function, see p. 158. For exporting area binning data to Google Earth in KML format, see Export to KML File.

Use Case 14.: Comparing Two Groups of Measurements from the Same Route on Map
This use case describes a situation where the user needs to compare the parameter values of two groups of measurements from the same route. It is assumed that there are measurement files with coordinates and corresponding maps in the database.

Step 1: Open a Measurement File on a Map


Open a map file and drag and drop measurements on it.

Step 2: Performing Delta Plotting on Map


Delta plotting makes it possible to display the difference between two groups of measurements from the same route on map.

Click the Delta Plotting icon

on the Tools panel and select an area from the map.

In the screenshot below, you can see the area selection surrounded by a thin dotted line.

Selecting an area opens the Delta Plotting dialog.

D AT A VI EW S

2 13

To select measurements for Measurement Group 1, click the Configure button and the Analyze Wizard Select Measurement dialog opens. In this dialog, you can select the measurements you want to perform the delta plotting on. You can have one to many measurements in each measurement group. A measurement group average is calculated from all measurements within a Measurement Group. The difference value plotted on map represents the difference between the two measurement group averages.

Folder defines the measurement folder whose contents are displayed in the field below. Using the Filter field, you can search for measurements.

The Selected measurements field displays all the measurements that have been selected for the measurement group With the With the button, you can select measurements button, you can clear measurements

Once you have selected the measurements you want to include in the Measurement Group 1, click Next. This opens the Analyze Wizard Measurement parameters dialog. Select the parameter based on which you want to compare the two measurement groups and click Next. Note that you have to select the same parameter for both measurement groups.

The Analyze Wizard Filters dialog opens.

D AT A VI EW S

2 15

Filter options are optional. Define filter options if necessary and click Finish.

Repeat the steps above to define measurements for Measurement Group 2.

Once you have selected measurements for both measurement groups, click OK in the Delta Plotting dialog.

From the Choose Statistics Type dialog you can choose whether statistics are weighted by time (Time), or distance (Distance), or whether they are not weighted at all (Sample). Although delta plotting as an operation is by default distance-based, each sample has its unique time and distance weights that can be used in calculating time and distance averages per bin. In delta plotting this average per bin is calculated for bins of both routes separately. To find out the difference between the bin values of one route and the bin values of the other, a subtraction is performed between the bin values of the routes. Depending on the operation you are performing, select either Time, Distance, or Sample and click OK. The delta plotting results are displayed on the map as a new layer.

D AT A VI EW S

2 17

Use Case 15.: Displaying Base Station Cell Beam Range on Map
To display the cell beam range of an individual cell on map, select the relevant cell with a left-click.

The beam range of the selected cell is displayed on the map as a sector. If the BTS file you are using does not contain cell beam range data, you can set the range in BTS options. Select View | Options to define base station display settings.

The Options dialog opens. Select BTS | General.

BTS texts options allows you to define the size and style of BTS text. Display BTS overlay on top of other layers sets BTS layers to be displayed over all other layers. Default BTS filter allows you to set default BTS filter that will automatically turn the chosen filter on if there are such base stations in the BTS file. In the Map window there is a BTS filter field that enables the user to choose which base stations are shown:

For UMTS and LTE options you must set a Carrier number which defines which part of the BTS filter is chosen.

D AT A VI EW S

2 19

Default enables you to define the default color for base stations when displayed in the map view. Default beam color enables you to define the default beam color. Default settings for BTS parameter, when selected, enables you to define default color sets for BTS parameters. BTS defines the parameter used in the coloring. Color set defines the colors for parameter values.

BTS Technology Based Settings enable you to choose the color and the size of the base station depending on which networks base station it is, or based on the carrier number. When the Use technology based settings (overrides icon size settings) option is selected, you are able to open the Set Technology Setting view by clicking Add. By clicking the Icon color option, the size and the color of the base stations changes on the map. If this option is not selected, only the size of the base station changes.

To edit cell-specific properties, right-clicking on a base station icon cell, and select Properties from the popup menu.

The Current tab of the Properties dialog opens.

Beam range defines the cell beam range in meters.

D AT A VI EW S

2 21

Draw line to active base station option enables the drawing of a line from the serving BTS to the test vehicle.

Icon size defines the size of the BTS icon in pixels. Fixed defines the fixed size of BTS icons in pixels. Dynamic, max size defines a maximum size for dynamic BTS icons (i.e. icons which change size based on zoom level). Show site names option hides and displays the site names. Draw frame hides and displays the outlines of BTS icons.

Site transparency defines the transparency of the BTS icons.

Show cell information option hides and displays the selected cell information. Hide cell texts on low zoom levels (performance optimization) option hides partly or completely the cell texts when the map is zoomed out so that the map can be drawn faster and with less visual distractions. The level is adjustable with a slider. Zoom levels are specific to used maps. Use cell beam range from BTS file allows you to enable/disable the display of cell beam range based on cell beam range data in the BTS file. Beam transparency defines the transparency of the beam range layer. Cell beam range can also be displayed by selecting a cell in Workspace | Base Stations | File Contents with a left-click.

D AT A VI EW S

2 23

Use Case 16.: Synchronizing Base Station Map Overlay with Grid Rows
Nemo Analyze allows the synchronization of grid data with base stations on map. To synchronize BTS map overlay with grid data, first open both a measurement file (p. 83) and a BTS file (p. 88) on map.

Next, split the data view vertically by right-clicking on the map and selecting Data View | Split | Vertically from the popup menu.

The data view is split into two sections.

To add a grid with BTS data to the empty section, drag and drop the BTS layer from the Layers panel on the right to the empty section.

D AT A VI EW S

2 25

The Select Data View dialog opens.

Select Grid and click OK. A grid with BTS data is now displayed.

The map now zooms automatically to the cell selected in the grid.

Use Case 17.: Using BTS Reference Parameters


Parameter tree contains a set of parameters with the parameter name prefix BTS reference. These parameters display information on the current serving and neighboring cell(s) based on a BTS reference file (i.e. BTS file). BTS reference parameters can be used, for instance, to display the name of the serving cell in a workbook together with recorded metrics, or as an input data set in KPI workbench to enable the grouping of aggregates based on serving cell, etc. For the BTS reference parameters to work correctly, the relevant BTS reference file must be present in the database. A BTS file needs to be activated in order for it to be used in saved workbooks and BTS ref queries. When opening a BTS file, a Set Active BTS File dialog opens. The user can activate a file by ticking a box next to the file.

D AT A VI EW S

2 27

If there are multiple versions of the same BTS file, e.g. from different dates, the relevant BTS file version must be activated. If none of the files are activated (i.e. designated as relevant files), reliable results cannot be guaranteed. To modify active BTS settings of a file that has already been added to the database, open the BTS page of the Workspace by clicking the button, right-click on the relevant BTS file, and select Activate from the popup menu.

Once the relevant BTS file has been activated, open the Measurements page of the Workspace by clicking the button.

Select a measurement or a measurement folder, and type the word reference to the filter above the Parameter Tree to list BTS reference parameters.

To open a parameter in its default view, select a measurement in the Workspace and then doubleclick a parameter in the parameter tree. If you want to open the parameter in some other type of view, right-click on a parameter in the parameter tree and select the view type from the popup menu. To run statistics based on the parameter, select the parameter in the Parameter Tree and click the click on the button displayed at the bottom left corner of the Parameter Tree. This opens the Parameter Launchpad. For more information on Parameter Launchpad, see page 77.

D AT A VI EW S

2 29

Use Case 18.: Displaying Base Station Connections on Map Based on Any Parameter
Nemo Analyze allows base station connections to be displayed on map based on any parameter containing BTS connection data. This can be useful, for instance, in terms of finding and visualizing cells causing pilot pollution. To display base station connections from terminal to individual cells on map based on any parameter, select a measurement file from Folder Contents, right-click on a parameter (e.g. Pilot pollution) in the Parameter Tree, and select Open In | Map from the popup menu.

The measurement route is displayed on map.

Click the Base Stations button in Workspace.

The Base Stations page opens.

D AT A VI EW S

2 31

Drag and drop a base station file from the Files field to the map. A dialog is displayed asking whether you would like to associate the open route with the BTS file.

Select Yes.

BTS overlay is displayed on map.

Right-click on the route layer displayed on the side panel and select Show BTS connections from the popup menu. The Route tab of the Route Properties dialog opens.

Select the BTS tab, clear the options Active/Serving Cells and Monitored/Neighbor Cells, select the [Parameter] BTS (e.g. Pilot Pollution BTS) option, and click OK.

D AT A VI EW S

2 33

Base station connections are displayed on map based on the selected parameter (e.g. Pilot pollution).

SPREADSHEETS
Nemo Analyze spreadsheets enable the processing of measurement data in spreadsheet format. As the basic functionality of Nemo Analyze spreadsheets is in line with the de facto standard followed by most spreadsheet applications on the market, this section will focus only on those features and usages that are specific to Nemo Analyze and to measurement data post-processing. Select View | Workbook | Spreadsheets in the Ribbonbar to open an empty spreadsheet.

D AT A VI EW S

2 35

You can drag and drop parameters on the spreadsheet from the Parameter Tree. Nemo Analyze will display a green icon if spreadsheet is a suitable data view for the parameter, and a red icon if the data view is unsuitable.

The query is opened on the spreadsheet.

Alternatively, you can open measurements with parameters on a spreadsheet straight from the Parameter Tree. Select a measurement folder or file in the Workspace and right-click a parameter in the Parameter Tree.

From the popup menu, select Spreadsheet.

D AT A VI EW S

2 37

The query is opened on a spreadsheet.

Editing Cell Format


To format cells containing parameter data, right-click on the cell in the top left corner of the parameter data set and select Format Cells from the popup menu.

The Query tab of the Format Cells dialog opens. Please note that the Query tab is not displayed if Format Cells is not accessed via a right-click on the cell in the top left corner of a parameter data set.

D AT A VI EW S

2 39

Show results, if selected, displays the parameter data set as a whole. If cleared, Show results minimizes the parameter data set into a single cell. Although minimized and not visible, functions can still be run on the entire parameter data set as each cell can still be referred to based on the original parameter data set structure. For more information on making references to a minimized parameter data set in a formula, see Use Case 19..: Retrieving Data from Minimized Data Sets on page 246. Show columns, if selected, displays data set column names. Short column names, if selected, displays short versions of data set column names. The Number tab of the Format Cells dialog allows you to define the format in which numbers are displayed within the selected cell(s).

General defines that the default settings for the content type are used. Number defines the cell content as numbers and allows you to set the number of decimals displayed. Percentage defines the cell content as a percentage and allows you to set the number of displayed decimals. Time defines the cell content as time and displays the value in the cell in hours, minutes, seconds, and milliseconds.

The Alignment tab of the Format Cells dialog allows you to define the alignment of text within cells.

Text alignment Horizontal defines the horizontal alignment of text within the selected cells as Default, Left, Center, or Right. Vertical defines the vertical alignment of text within cells as Top, Center, or Bottom. Text control Wrap text, if selected, enables word wrap within the selected cells. Merge cells enables you to merge selected cells into a single cell. Select the cells you want to merge, right-click on the selection, select Format Cells from the popup menu, and select Merge cells from the Alignment tab of the Format Cells dialog. The Font tab of the Format Cells dialog allows you to define the font and color of the text within cells.

D AT A VI EW S

2 41

Font defines the font of the text within cells. Color defines the font of the text within cells. The Border tab of the Format Cells dialog allows you to define borders for cells.

Border width defines the border thickness for cells as Thin (default), Medium, or Thick. Presets allows you to select which border groups are to be displayed. With None, no borders are displayed. With Outline, only the outlines of a cluster of cells are displayed as a border. With Inside, the borders of all cells within a cluster of cells are set as borders. Border buttons (i.e. Top, Middle, Bottom, Left, Middle, and Right) allows you to select manually which borders are to be displayed. The Patterns tab of the Format Cells dialog allows you to define cell background color.

Cell color defines the cell background color.

Filtering Data
To filter out unwanted data based on values in a particular column, select the contents of the columns you want to filter, right-click on the selection, and select AutoFilter from the popup menu.

An AutoFilter drop-down menu is created based on values in the selected column(s).

To filter out all rows that do not contain a particular column value, select a column value from the dropdown menu.

D AT A VI EW S

2 43

All rows that do not contain the selected value in the filter column will be filtered out. Selecting (All) from the drop-down menu displays all rows again. To remove an AutoFilter from a column, right-click the AutoFilter element at the column heading, and clear AutoFilter from the popup menu.

Creating Formulas
You can enter a formula in any cell on the spreadsheet. The creation of a formula always begins with the equal sign (=). The next element identifies the mathematical function, e.g. AVG stands for the function Average. For more functions, see Adding Functions on page 244. =AVG The rest of the formula defines the source data for the function and is placed within brackets. =AVG() If the cells containing the source data for the function are located on another sheet, it is first necessary to identify the location of the data by entering the sheet name followed by the exclamation mark (!). =AVG(Sheet 2!) If the cells containing the source data for the function are located within a separate data set on the same spreadsheet, it is first necessary to identify the location of the data by entering the data set name (e.g. the name of a query opened on the spreadsheet) preceded by the asterisk (*) and followed by the exclamation mark (!). =AVG(*BLER DL!)

Next, a reference to the actual cells is added. A cell is referred to by a combination of the column letter and the row number that intersect at the cells location, e.g. A1. To refer to a range of cells, first identify the start point of the range (i.e. the first cell in the range), and separate it from the end point (i.e. last cell in the range) with a colon (:). If the range covers multiple columns, the start and end point references should be to the cells in the upper left and lower right corners of the range. =AVG(*BLER DL!A1:D555) If the range covers all cells in a column, the range can be referred to by using just the column letter as both the start and end point, instead of referring to individual cells with row numbers. =AVG(*BLER DL!A:A) To refer to more than one cell range, separate each range with a semicolon (;). =AVG(*BLER DL!A1:B12;C10:D16) For information on the general format of spreadsheet formulas, see one of the numerous helps and guides for spreadsheet applications available on the internet.

Adding Functions
To perform a mathematical operation based on a group of cells, select the cells by holding the left mouse button down while moving the cursor over the selected cells. Once all intended cells have been selected, release the left mouse button and right-click over the selection. A popup menu opens. From the popup menu, select Insert Function | [function].

D AT A VI EW S

2 45

The function is added below the selected cells. The cell with the function displays the result of the function. The formula of the function is displayed in the formula field on the topmost bar of the spreadsheet view.

If you wish, you can relocate the function by cutting and pasting. If you do not find an appropriate function amongst those displayed in the popup menu, select Insert Function | More Functions instead.

The Insert Function dialog opens.

When you select a function from the list, the info field below displays a description of the function. Select a function from the list and click OK. The function is added below the selected cells, as discussed above.

Use Case 19..: Retrieving Data from Minimized Data Sets


It is possible to minimize parameter data sets on a spreadsheet to take up the space of only one cell. For minimizing data sets, see Editing Cell Format on page 238. To retrieve data from a minimized table, you first have to know which column in the maximized data set contains the data you want to retrieve. The first column in a separate data set is always considered to be Column A, the second Column B, etc., no matter how the data set is situated on the spreadsheet. Thus, if you are retrieving data from the first column of a minimized table, you are interested in the contents of Column A. The name of the query (i.e. the parameter data set that has been opened on the spreadsheet) has to be known as well in order for you to be able to refer to the right data set. The creation of a formula for data retrieval begins with the equal sign (=) to identify the string as a formula. Next, add the operator of the mathematical function you want to use. For instance, with the function Average, the operator is AVG. For more functions, see Adding Functions on page 244. Next, within brackets, define the data set on which the mathematical function is to be run. The data set definition string is generally of the following format: open bracket, asterisk, query name, exclamation mark, reference to the first cell in the range, colon, reference to the last cell in the range, and closed bracket. Thus, a formula for calculating an average based on values on rows 1-15 in Column A of the BLER DL data set would be of the following format: =AVG(*BLER DL!A1:A15). For more information on formulas, see Creating Formulas on page 243.

D AT A VI EW S

2 47

Use Case 20.: Creating Reports and Report Templates Using Spreadsheets
It is possible to create reports on spreadsheets and save them as report templates. These report templates can then be run on all measurement files that contain the events the report has been defined to look for. Select View | Workbook | Spreadsheets in the Ribbonbar to open an empty spreadsheet.

On the empty spreadsheet, create a layout for the report.

D AT A VI EW S

2 49

In Workspace, select a measurement and the parameter you want to add on the report. Click on the button to open the Parameter Launchpad.

In Parameter Launchpad, click on Statistics for a list of aggregates, select an aggregate (e.g. Count) from the list and drag and drop it to the cell you have designated for it in the spreadsheet layout.

Repeat the procedure with other relevant parameters until they are all on the spreadsheet.

If the report you are creating requires mathematical operations, select the cell you have designated for an operation and enter the appropriate formula into it. In the present example, the count of connected calls has to be divided by the count of call attempts to produce a ratio of successful calls. In order to present the ratio as a call success rate percentage, the result of the division has to be multiplied by hundred. At this point, the formula is of the following format: =((C3/C2)*100). For information on the general format of spreadsheet formulas, see one of the numerous helps and guides for spreadsheet applications available on the internet. Because division as an arithmetic operation always requires that the divisor (i.e. the cell C2) is not zero and because in a report of this type it is possible that the result of the count operation in C2 will be the value zero (i.e. if there are no call attempts), an additional condition following the format if C2 is zero, the result of the division is zero has to be created to prevent the error that would otherwise occur. In spreadsheet formulas, this will be expressed in the following format: =IF(C2;((C3/C2)*100);0).

D AT A VI EW S

2 51

If the result of the calculation is displayed in a format unsuitable for the value type (e.g. a percentage is displayed with multiple decimals), right-click on the cell containing the value, and select Format Cells from the popup menu. The Number tab of the Format Cells dialog opens.

Select a number format suitable for the value type, define the number of decimals to be displayed, and click OK. The report is now complete.

You can save the completed report as a template that can be run on any measurement file that contains the events required by the report. To save the report as a template, right-click on the spreadsheet and select Workbook | Save from the popup menu.

The Save Workbook dialog opens.

D AT A VI EW S

2 53

Filename defines the name of the report template. Popup menu defines the menus (e.g. Workspace Device Menu) in which the workbook will be saved. For more information on menus, see p.494. Groups defines the group (e.g. User) in which the workbook will be saved to. For more information on groups, see p. 500. To run the report on another measurement file, select a measurement from the Workspace section you defined as the location for the template (e.g. Workspace Device Menu) and right-click a measurement file.

From the popup menu, select the submenu you defined as the location for the template (e.g. User) and select the template. The report is run on the selected measurement file.

PROCESSING UPLINK VOICE QUALITY DATA


1. When measuring with Nemo Outdoor, make sure its clock is synchronized with the voice quality server clock. In Nemo Outdoor, select User Interface Measurements | Use time from GPS. In Nemo Server, select NTP (Network time protocol) from the NSM menu. Internet connection is needed in order for the server to use NTP time. If there is no internet connection available or NTP time cannot be set through, i.e. LAN, Nemo Outdoor clock and server clock must be synchronized manually. 2. Before measurement, the number of each servers phone line must be configured manually to the voice quality server. By doing this, the numbers will also be set to the server ends log file headers. 3. Load both Nemo Outdoor and server log files into Nemo Analyze. After this, you are able to use Nemo Outdoor files normally. If you want to see UL MOS for a specific Nemo Outdoor log file, select the file in question and UL MOS from the parameter tree. Nemo Analyze connects UL MOS values from the correct server file to the Nemo Outdoor file in question by comparing time and phone numbers.

D AT A VI EW S

2 55

INDOOR MEASUREMENTS
With Nemo Analyze you can also view indoor measurements on floor plans. Right-click on the measurement in the Workspace and select Adjust Coordinates.

In the Adjust Coordinates dialog, select the floor plan in the Map file field or click the Import Image as Map button to convert an image file into a map (see page 35). Click OK.

The selected floor plan is opened. First you need to place the coordinate markers in the right places. With the Arrow tool, click on the map and the active marker (highlighted with pink in the table on the right) will be placed there. The Longitude and Latitude values are updated as you place the markers on the map.

After you have placed all the markers, click the Update Coordinates button to save the marker coordinates.

Nemo Analyze will ask if you would like to open the route on a map. Click Yes, and the route is opened on a map.

D AT A VI EW S

2 57

When you open the indoor measurement on the floor plan the next time, the route is automatically drawn on the map. You can color the route and use all the features available for maps.

NUMERICAL DATA VIEWS


Numerical data views are useful, for example, for viewing benchmarking data from multi measurements: you can open several measurement files side by side and compare the values. You can open a numerical view, for example, by opening an empty workbook (View | Add data view| Workbook) and then adding a numerical view data view in the workbook: right-click on workbook, select Page | Add Data View | Numerical Data.

Drag a measurement or measurements in the view, and Nemo Analyze will display some network parameters for the file. If no data are shown on the table, the data are retrieved from the very beginning of the measurement. Double-click on the measurement file in the Workspace to view the Timeline view, and move the red time line forward to view captured data.

D AT A VI EW S

2 59

You can edit the contents of the numerical data view.

Right-click on the view and select Properties.

The Visible parameters list offers a list of parameters that can be displayed in the numerical data view. With the Move up and Move Down you can change the order in which the different parameters appear in the numerical data view.

Title defines the column title. Alignment defines how the text is aligned.

The Color Sets list enables you to select the appropriate color sets for the parameters from the dropdown menu.

D AT A VI EW S

2 61

INFO VIEWS
Info views are used to display more detailed information, for example, about signaling, BTS sites and layer messages.

TIMELINE VIEW
The Timeline view enables you to view and analyze a selected time range of a measurement file. For example, you can select to view data before and after certain events, e.g., call failures, to troubleshoot what may have caused the particular event. Or you can select a time range and generate a report for that range. Right-click on a measurement in the workspace and select Timeline or double-click a measurement to open the Timeline view. You can also open the Timeline view for multi measurements.

D AT A VI EW S

2 63

The Timeline view displays the start and finish time of the file. The red synchronization line marks the current point in the measurement file. Current time is displayed below the synchronization line.

Highlight Parameter in Timeline View


With Highlight Parameter you can visualize problem areas in the measurement. For example you can view certain events or parameters when their values are under/over certain user specified threshold. Right-click on the Timeline view and select Highlight Parameter.

In the Pick Parameter dialog, select the parameter you want to highlight and click OK. You can also find a given parameter from the list by using the filter field within the red box.

Double-click the Value field to define the event for the selected parameter and click Finish.

The parameter events will be highlighted.

D AT A VI EW S

2 65

Notifications in Timeline View


You can also view notifications in the Timeline view. Right-click on the view and select Properties. In the Notifications tab, select the notifications you want to view in the Timeline view, for example, attach failures, and BLER, and click OK.

The small red vertical bars represent the selected notifications. In the example below there are several attach attempt failures.

Range Selection in Timeline View


Now you can take a closer look at the problem spot by selecting a sample, a range, of the file around the attach failures. Select the starting point of the range by double-clicking with the left mouse button and the ending point with a left click.

To generate a report based on the selected time range, right-click on the range and select Range | Report | Open from the popup menu. In the Open dialog, select a report template (*.rpt) and click Open. To generate a workbook based on the selected time range, right-click on the range and select Range | Workbook | [workbook folder] | [workbook].

Depending on your selection, a report or a workbook is generated from the data in the selected range.

D AT A VI EW S

2 67

OTHER VIEWS
Network Parameters
The Parameters view displays some key network parameters. Right-click on a file in the workspace and select Network Parameters.

If the Parameters view is empty, right-click on the measurement in the Workspace to view the Timeline view. Move the red time line forward to a point in time where data have been gathered during the measurement.

Right-click on the view and select Pick Parameters to add more parameters. To remove parameters from the view, right-click on the view and select Properties. In the parameters page, clear the parameters you want to remove and click OK.

Measurement Settings
The Measurement Settings view displays some information about the settings that were used when the file was recorded. Right-click on a file in the workspace and select Measurement Settings.

D AT A VI EW S

2 69

Properties
The Properties view displays information about items that are selected in the main window. The view is by default docked to the right side of the main window. Move your mouse over the Properties tab for the window to appear. If the Properties tab is not visible, open the Properties view by selecting View | Properties. You can also drag the view elsewhere on the Analyze main window. If you want to dock it back to the side of the main window, double-click on the Properties view.

Query Clipboard
The Query Clipboard displays a list of queries that have been copied, and which can further be pasted into views. The view is by default docked to the right side of the main window. Move your mouse over the Query Clipboard tab for the hidden window to appear. If the Query Clipboard tab is not visible, enable the Query Clipboard by selecting View | Query clipboard. You can empty the clipboard by clicking the Clear button. You can also drag the view elsewhere on the Analyze main window. If you want to dock it back to the side of the main window, double-click on the Query Clipboard.

To paste a query into a view, right-click on a view and select Query | Paste. The Paste Query dialog displays the contents of the Query Clipboard. Select the query to be pasted and click OK.

D AT A VI EW S

2 71

Database Loader
The Database Loader displays the progress of file upload, file conversion, and file upload queue.

With Cancel and Cancel All buttons, you can abort the upload of either the currently uploaded measurement file or the upload of all measurement files in the upload queue.

Window Browser
The Window Browser displays a list of controls (workbooks, pages, data views) open in Nemo Analyze at a given moment. Right-click on the items to access the Properties dialogs for each control.

D AT A VI EW S

2 73

Messenger
The Messenger enables you to publicly chat with other users of the database. If you are logged to a database server with multiple users, all other current users are displayed as Current users.

To send a public message seen by all other users, type a message to the Message field and press Enter. The message is displayed in the Received messages field.

Output
The Output panel displays some program and error messages and the SQL statements that are being processed. The view is by default docked to the bottom of the main window. Move your mouse over the Output tab for the hidden window to appear. If the Output tab is not visible, open the Output view by selecting View | Output. Switch between the Log and the SQL pages by clicking on the tabs at the bottom of the view. You can also drag the Output view from the bottom of the Analyze main window elsewhere on the screen. If you want to dock it back to the bottom, double-click on the view.

You can clear the Output view by right-clicking on it, and selecting Clear. You can also save the contents of the Output view in a log file by selecting Write to File and type a name for the file. The log is saved in a text file (.txt). Select Copy to Clipboard to copy the contents on a clipboard to be pasted in another application.

D AT A VI EW S

2 75

MANUAL LAYOUT EDITOR


The Manual Layout Editor makes it possible to create custom layouts for workbook pages. Each Manual Layout Editor page can contain any number of user-defined data views laid out in a manner defined by the user.

Creating a Layout
To open Manual Layout Editor, select View | Workbook | Manual Layout Editor from the Ribbonbar. To add a layout editor data view to an open workbook, select Layout | Manual Layout Editor from the Ribbonbar.

An empty workbook opens. To add data views to the workbook, right-click on the workbook and select Insert Data View | [data view type] from the popup menu. Note that in order to add or edit data views in a Manual Layout Editor workbook, Edit Mode must be selected in the popup menu.

Once an empty data view has been created, the user can place it anywhere on the workbook window. The user can also change the size of the data view by simply holding down the left mouse button on the red border of the data view and dragging the selected side of the data view to the desired direction. The user can also create other empty data views, defining the layout for the workbook page before beginning to add content to the data views. However, data views can be added, edited and removed later on as well by right-clicking on the Manual Layout Editor workbook and selecting Edit Mode from the popup menu.

Data views can be removed by selecting a data view, right-clicking and selecting Item | Delete from the popup menu.

D AT A VI EW S

2 77

Selecting Item | Properties from the popup menu opens the Properties dialog. Border size defines the thickness of the data view border. Selecting Draw shadow draws a distinguishing shadow under the selected data view.

Adding Content to Manual Layout Editor Data Views


Measurement files and folders, parameters, BTS files and map files can be drag and dropped to the empty data views.

A measurement folder (or file) and a parameter are both selected in the Workspace and then drag and dropped (holding left mouse button down) on an empty data view in the Manual Layout Editor workbook.

Once you have finished editing the layout, right-click on the Manual Layout Editor window, and clear Edit Mode in the popup menu. This will lock the layout. If you later want to edit it some more, simply right-click on the Manual Layout Editor workbook and select Edit Mode from the popup menu.

QU ER IES

2 79

WORKBOOKS
Viewing measurement data in Nemo Analyze is extremely flexible and user-configurable. The various views are organized into workbooks, pages, and data views. A workbook is the main component that contains all the different pages and data views. There is a selection of ready-made workbooks but you can also create new workbooks. To open an empty workbook, select View | Add data view | Workbook. To open a ready-made workbook on a measurement file/folder, right-click the measurement file/folder and select [workbook folder] | [workbook] (e.g. UMTS | UMTS full details) from the popup menu.

To cancel a running query during workbook execution, click Cancel in the Executing Queries dialog. To cancel the execution of all queries within the workbook, select Cancel All.

ADDING DATA VIEWS


In order to view measurement data in the workbook you will need to create a data view. Right-click on the workbook and select Page | Add Data View | [data view type]. An empty data view will be opened in the workbook.

Now you can add data in the data view. You have two options. You can either drag a measurement file in the data view or right-click on the view and select Pick Parameter. In the latter case, first choose the graph type.

QU ER IES

2 81

In the Pick Parameter dialog select the measurement file in the Measurement field and pick the parameter to be viewed. In the Filter field you can type the name of the parameter you want to add in the data view to locate all the options, in this case all instances of BLER, more quickly from the list.

The selected view is opened.

You can create several pages and each page can contain several data views. To add more data views on the same page, right-click on the view and select Page | Add Data View | [data view type]. If you select Data View | Insert | [data view type], the existing data view is replaced with the new one.

QU ER IES

2 83

Right-click on the empty grid view, select Pick Parameter, select the parameter to be displayed and click OK. Now you have a workbook with one page that contains two data views. To add more pages to you workbook, see the next chapter.

ADDING PAGES
You can have several pages in a workbook each containing one or more data views. Right-click on a view and select Page | Add Page.

An empty page is added to the workbook and you can now open data views on the page. You can switch between the pages from the tabs at the bottom of the workbook.

QU ER IES

2 85

SAVING A WORKBOOK
After you have opened all the relevant views, right-click on a view and select Workbook | Save to save the workbook for later use.

The Save Workbook dialog opens.

Filename defines the name of the report template. Popup menu defines the menus (e.g. Workspace Device Menu) in which the workbook will be saved. For more information on menus, see p.494. Note: Workbooks containing bar or line graphs should always be saved only in the Device Menu. This is because the nature of these graph types requires the information to be ordered by time, which in turn can be done only within a single log file at a time.

Groups defines the group (e.g. User) in which the workbook will be saved to. For more information on groups, see p. 500.

After defining the filename and the destination popup menu and group, click OK.

QU ER IES

2 87

EXPORTING WORKBOOKS
An open workbook can be exported as a PDF or as a set of image files (image per page).

Exporting Workbooks as PDF Files


To export an open workbook as a PDF report, select File | Export | Workbook to PDF.

The Save As dialog opens.

To save the report as a PDF file, type in a file name in the File name field and click Save.

QU ER IES

2 89

Exporting Workbooks as Image Files


To export an open workbook as image files, select File | Export | Page Image to clipboard, and select export type from the drop-down menu.

Page image to clipboard saves the workbook page currently open on the clipboard. Page image to file saves the workbook page currently open as an image file. Workbook images to file saves all workbook pages as separate image files.

COPYING A WORKBOOK
After you have opened all the relevant views, right-click on a view and select Workbook | Create Copy for Measurement or click the Create copy of workbook button in the toolbar to make a copy of the workbook for another measurement file.

Select the measurement for which you want to open the workbook and click OK. The workbook is opened with the new data.

QU ER IES

2 91

WORKBOOK PROPERTIES
You can rename a workbook by right-clicking on the workbook and selecting Workbook | Properties. In the Workbook Properties dialog, type a new title for the workbook and click OK.

PAGE PROPERTIES
In the Page Properties dialog you can change the settings of each page. Right-click on a page and select Page | Properties.

On the Page page you can define a title for the page. The name is displayed at the bottom of a workbook. Here you can also define the page size.

If the Fit to window option is selected, all the layers are displayed in the visible area of the page. This means that when a new layer is added, the space for each layer gets smaller. With the Fixed size option you can define an exact size in pixels for the page. This will give more space for each layer. With the scroll bar you can scroll up and down the page to view all layers

EXAMPLE WORKBOOK
The example workbook, GSM, contains four pages, each of them displaying different GSM-related KPIs. The first page displays three different RF Parameters in line graphs stacked on top of each other in separate graphs. The Numerical data table on the right displays information in numerical form.

QU ER IES

2 93

Page two contains Serving and Neighbor cell data in line and bar graphs.

Page three contains L3 signaling data in a grid, and the decoded data in an info view.

The fourth page contains a map with the measurement route. The route is colored with the RX level color set.

QU ER IES

2 95

REPORTS
With Nemo Analyze, you can select the optimal approach for each statistical reporting task. Instead of being tied to the strengths and limitations of a single tool, you can employ a range of solutions, from Nemo Analyze workbooks and spreadsheets to Crystal Reports report templates and Microsoft Excel export templates.

WORKBOOKS
The fully customizable and automatable Nemo Analyze workbooks provide you with a completely new angle on statistical reporting. Each workbook can contain several pages and data views. All open workbooks and all data views within the workbooks are automatically synchronized in time. The comprehensive selection of data views includes spreadsheets, maps, grids, line graphs, bar graphs, pie charts, layer 3 messages, surface grids, and color grids. Nemo Analyze features a full set of readymade workbooks with pages and views for all the relevant KPIs. Nemo Analyze workbooks can be scheduled to be run automatically on predefined measurement file folders. Nemo Analyze paired with an automatic measurement solution such as Nemo Autonomous can indeed equal to automatic up-todate measurement reports popping up on your desktop as and when measurement sessions are completed. All custom settings, such as workbooks, queries, layouts, color sets, KPIs, etc., can be imported and exported to enable sharing between colleagues. For performing reporting using Nemo Analyze workbooks, see Workbooks.

SPREADSHEETS
Nemo Analyze spreadsheets offer a spreadsheet functionality that follows the current de facto standard in spreadsheet applications, fully integrated with Nemo Analyze's drive test data-specific post-processing features. You can drag and drop parameter aggregates and statistics on a single spreadsheet based on multiple measurement files and KPIs, and calculate e.g. success rates by employing the standard range of arithmetic operations and other capabilities of the spreadsheet approach. The resulting spreadsheet can be saved as a report template and subsequently run as a workbook on any measurement data. For performing reporting using Nemo Analyze spreadsheets, see Spreadsheets.

REPORTING WITH MICROSOFT EXCEL


With Nemo Analyzes Microsoft Excel export templates, you can export query result data sets from Nemo Analyze to a selected location within an existing MS Excel document. The exported parameters and the export locations within the MS Excel workbook are saved as a report template. The template can be subsequently run on any measurement data. The feature allows you to use the full data formatting functionality of MS Excel in reporting and make use of any existing, custom-made MS Excel spreadsheets. In the following example, Sheet1 of the existing target MS Excel sheet displays the final report output which is generated based on input data sets on Sheet2. In other words, as Sheet2 acts as a basis for the final output on Sheet1, the data sets should in this case be exported to the appropriate columns on Sheet2. Note: Before doing anything else, it is recommended to open the data set intended for export in grid data view to check the data sets number of columns and column order. The data set will be exported from Nemo Analyze in fixed table format and thus it may be necessary to make changes to the column order of the target MS Excel sheet before the export operation.

The target MS Excel workbook must be saved to Nemo Analyzes default Reports directory. To find out what the default folder for reports is, select View | Options from the Ribbonbar, and Environment | Default Paths | Reports in the Options dialog.

QU ER IES

2 97

To create an MS Excel export template, select View | Tools | Excel export.

The Excel Export Template dialog opens.

Excel workbook defines an existing MS Excel workbook as the target spreadsheet for the export. Excel template defines a name for the new MS Excel export template. Run macro on finish, when selected, runs the chosen macro in Excel once the export has been performed. Exported queries displays all queries selected for export. Add opens the Excel Template Row dialog, enabling you to define a new query. It is also possible to drag parameters from the parameter tree to the excel report creation template dialog.

Modify allows you to modify an existing query. Select the query in the Exported queries field and click the Modify button. Delete allows you to delete an existing query. Select the query in the Exported queries field and click the Delete button. Preview displays the columns of the query data set selected in the Exported queries field. Exported images displays all data view images selected for export. Add opens the Excel Export Template dialog, enabling you to define a new image. Modify allows you to modify an existing image. Select the image in the Exported images field and click the Modify button. Delete allows you to delete an existing image. Select the query in the Exported images field and click the Delete button. To browse for an existing Excel workbook (to be used as the target spreadsheet for the export), click the button.

QU ER IES

2 99

The Open dialog opens.

Select the Excel workbook you want to use as the target spreadsheet for the export and click Open. The path of the target spreadsheet is displayed in the Excel workbook field. To rename the export template, type a new name to the Excel template field.

To define a new query for export, click the Add button. The Excel Template Row dialog opens.

Sheet name defines the target sheet in the target MS Excel workbook for the exported parameter data set. Cell defines the target cell in the target MS Excel workbook for the exported parameter data set. In the present example, A2 is selected (instead of e.g. A1) because the first row of Sheet2 is reserved for column names (see screenshot on page 296).

Note: As the parameter data set is likely to contain more than one column, the target cell defined here will act as the upper left corner of the exported parameter data set. Thus, if the exported parameter data set contains three columns, and the target cell is defined as C2, the first-row cells of the second and the third columns will be D2 and E2 respectively. The Preview field in the Excel Export Template dialog displays the columns included in the selected parameter data set. Note: To make the exported data sets as versatile as possible in terms of the existing operations in the target MS Excel workbooks, all irrelevant data, including the Nemo Analyze column names, are removed from the exported data sets. Parameter defines the exported parameter data set. Click the button to browse for a parameter. With Modify you are able to modify the parameter data set that has been selected in the Parameter box. Right-clicking the Config button opens the Statistics Configuration dialog which enables you to modify the settings of statistics parameters. Select OK and the Analyze Wizard Start dialog opens.

Pick measurement parameter opens the Parameter Tree. The Parameter Tree enables the use of both ready-made parameters and previously saved custom KPIs (i.e. parameters created using the KPI workbench or SQL). By default, custom KPIs can be found in the Parameter Tree under the User item. For Generic query wizard, see page 337. For Manual query, see page 342. For Correlate parameters, see page 344. Selecting Pick measurement parameters in Analyze Wizard Start dialog opens the Analyze Wizard Measurement Parameters dialog.

QU ER IES

3 01

Select a parameter and click Finish. Previously saved custom KPIs can be found in the Parameter Tree under the User item. The Excel Template Row dialog opens.

Define Sheet name, Cell, Parameter, and Statistics. With Modify you are able to modify the parameter that has been selected in the Parameter box. Right-clicking the Config button opens the Statistics Configuration dialog which enables you to modify the settings of statistics parameters. Select OK. The defined query is now displayed as a row in the Exported queries field. The Preview field displays the columns the selected query contains and the column numbers they will take up in the target spreadsheet. If you want to add another query, click Add in the Excel Export Template and repeat the procedure above. To edit an existing query, select the query in the Exported queries field and click the Modify button.

To delete an existing query, select the query in the Exported queries field and click the Delete button. It is also possible to drag parameters from the parameter tree to the Excel report creation template dialog. In the Exported images section it is possible to define images to be exported into the Excel repot. Images can be exported from a file or from Analyze workbook. To define a new data view image for export, click Add in the Exported Images section. The Export Image dialog opens

File defines the name of an image file that will be exported to Excel. Workbook defines the workbook that will be exported to Excel as an image. Page, Image position, Image size, and Relative size define to which page of the workbook the image is to be exported, the desired position and the size of the image, and the size of the image in percentage. Sheet name defines the name of the Excel sheet that the image is exported to. Once Page name, Cell, and Parameter have all been defined, select OK. If you want to add another data view image, click Add in the Excel Export Template and repeat the procedure above. To edit an existing image, select the image in the Exported images field and click the Modify button. To delete an existing query, select the image in the Exported images field and click the Delete. Once all the intended queries and images have been added to the template, click OK. The Excel export template is saved to the default Report directory. To find out what the defined folder is, go to View | Options | Environment | Default Paths | Reports. To load the Excel export template to the report template database, first click on the bottom of the workspace panel. The Reports page of the Workspace opens. icon on the

QU ER IES

3 03

Next, double-click on the Excel export template in the Report Folder field. The Excel export template is added to the database and is now displayed in the Loaded field.

To run the Excel export template on a measurement file/folder, right-click on the measurement file/ folder and select Report|[Excel export template].axt from the popup menu.

The parameter data sets defined in the export template are exported to the defined page (Sheet2 in the present example) of the MS Excel workbook.

Microsoft Excel generates the final output page (Sheet1) based on the input page (Sheet2).

QU ER IES

3 05

REPORTING WITH CRYSTAL REPORTS


In Nemo Analyze you can run statistics in the form of reports over a single or multiple measurement files. The reports include essential network and application performance KPIs as defined by standards specifications, such as, ETSI, ANSI, etc. In addition, smaller-scale statistics can be run over a single parameter in the User Interface. Reporting in Nemo Analyze is based on Crystal Reports. The Crystal Report Viewer tool is embedded in Nemo Analyze. Reports can be exported to PDF, Excel, etc., as in standard Crystal Reports. Default report templates with all the most essential KPIs are provided in Nemo Analyze, but you can also create your own report templates. For this you will need the Crystal Reports Professional edition (independent third-party software).

Generating Reports
Generating reports is very easy. Right-click on the measurement file from which the report will be generated and select Report. Select one of the existing report templates, e.g., CDMA performance report.rpt.

QU ER IES

3 07

The report is generated and opened in a new workbook. Note that this may take a few minutes, depending on the processing power of the computer and on the complexity of the report.

The report contains several pages with all relevant network KPIs in different graphs and tables.

QU ER IES

3 09

Generating Reports from Multiple Files


You can also generate reports from multi-measurements or from folders containing several files. Rightclick on the multi-measurement file name, or on the folder, and select Report and the report template.

In the dialog you can define what kind of data will be included in the report. After making all the selections, click OK and the report is generated.

System and Band define from which cellular systems and frequency data will be included in the report.

Begin time and End time define the time period from which data is included in the report. By selecting the Area option, and clicking the Define Area.. tab, you can enter the Select Area dialog below.

Select Area defines the geographical area from which data is included in the report. Select a map in the Filename field, and select an area on the map with your mouse. Right-click on the map to change the tool, for example, to zoom in on the map.

Exporting Reports
After you have generated a report you can export it to another format. Click the Export Report button in the report workbook.

QU ER IES

3 11

In the Export dialog, select the output format and the destination for the exported file. Click OK.

In the Export Options dialog, select the report pages that will be exported and click OK.

The report is saved in the location and format defined.

Configuring Reports
The report contents and layout are user-configurable. Select the Reports page in the Workspace, right-click on the report template that you want to configure, and select Configure.

In the Report Configuration dialog you can edit the report contents. On the General page you can select which statistics are shown in the report. You can also type a Report description and a Report title that will be shown on the report front page. Show bins defines whether numerical data is shown in cumulation and density histograms.

QU ER IES

3 13

Under the other items you find a list of available parameters. For each parameter you can define some properties, such as, the lower and upper bound, threshold, and channel numbers. The selection depends on the parameter.

The changes will affect all reports that are generated using this template. For benchmarking reports you need to define how the benchmarking report is grouped. Open the Report Configuration dialog for a benchmarking report, click on General. In the Benchmark by field select the grouping factor and click OK.

CREATING CUSTOM QUERIES


NOTE: Using the easy-to-use graphical user interface of KPI Workbench, the user is able to perform all SQL-specific tasks and more without proficiency in the SQL language.

Nemo Analyze is based on an SQL database. The data is stored and retrieved from the database using SQL queries via the standard ODBC interface. The results of the SQL queries can be further processed using the KPI workbench, Nemo Analyzes graphical, flowchart-based scripting engine. The results of the SQL queries and the KPI Workbench scripts can be visualized using Nemo Analyzes comprehensive selection of data views, or any third party post-processing tool supporting ODBC connectivity. This section describes the guidelines for making custom KPIs with both SQL (see SQL Queries on page 315) and the KPI Workbench (see CUSTOM KPI Workbench on page 345). SQL queries represent a powerful tool for complex data filtering and data processing when the data is in scalar format. However, SQL has some downsides as well: it is impossible to create queries that track certain event sequences, such as changes in a particular parameter value from row to row. It is also impossible to merge data from more than two data sets into a single result set based on time. Moreover, the use of SQL queries requires knowledge of the Nemo Analyze database schema and proficiency in SQL. The KPI workbench enables the creation of complex queries, including time-based merging of multiple inputs, and the tracking of particular event sequences. The graphical user interface makes it also easier to use than SQL. To summarize the differences between the KPI Workbench and SQL, the KPI workbench is a more powerful tool for data customization, whereas SQL is sometimes more convenient when filtering or formatting raw data. For a user without previous experience with SQL, the KPI workbench is recommended as the primary data customization tool.

QU ER IES

3 15

SQL QUERIES
Getting started
As a prerequisite, the following documents and tools are recommended as a reference for developing SQL queries: 1. SQL editor. Queries can be written with the Nemo Analyze database browser (see 2. Database Browser on page 344) but there are also many specifically SQL-oriented third-party editors available. Good freeware tools are also available, such as http://gpoulose.home.att.net/Tools/QTODBC61.msi 3. Nemo File Format specification. The document describes the Nemo measurement file format on an event-by-event basis. The file format specification can be downloaded from the Nemo User Club, or accessed via Nemo Analyze by selecting Help | Documentation | Nemo file format from the Ribbonbar. 4. Open access SQL reference. The document describes the supported SQL syntax and can be downloaded from the Nemo User Club. In addition to the Open access SQL, Anite has added some proprietary scalar functions and stored procedures for common tasks needed in drive test data processing. The description of Anites proprietary scalar functions and stored procedures can be accessed by selecting Help | Help | Help topics from the Ribbonbar and searching the help with the keywords scalar functions or stored procedures. 5. Nemo Analyze database schema. Database schema can be accessed and queried by selecting Help | Documentation | Database schema from the Ribbonbar. The schema opens as a Crystal Reports document. Note that querying the schema may take several minutes. The schema can be exported as a PDF document.

Nemo Analyze Database Schema


Data in a relational database is stored as relations which are perceived by the user as tables. Each table represents an object or an event (e.g. employee, sales order, etc.). Each row in a table represents a unique instance of the object or event. Each column represents a different class of information defining the object or event, e.g. first name, last name, salary, customer name, order charge, etc. In the Nemo Analyze database, the table structure is based on the event structure of the Nemo log file format. That is, each event in the Nemo file format has a corresponding table in the database. For instance, a BLER event is always recorded when the mobiles Block Error Rate changes. When the log file is loaded into the Nemo Analyze database, each BLER event is parsed as a new row in the BLER table. When referring to a table in a query, the syntax is [schema name].[table name]. For example, "Nemo.UMTS"."BLER, where the schema is Nemo.UMTS and the table BLER.

Relations between tables


If rows in a given table can be associated with rows in another table, the tables are said to have a relationship between each other. Generally, these relationships can be of three different types: one-toone, one-to-many, and many-to-many. The Nemo Analyze database schema allows one-to-one and one-to-many relations.

One-to-one relation means that each row in a given table can have a relation with a single row in another table. In one-to-many relation, each row in a given table can have a relation with multiple rows in another table. These relations can be defined by two columns in a table: the columns primary key and foreign key. Each table by necessity has a primary key. Primary key is a column that uniquely identifies each row within a table. In Nemo Analyze database, the primary key column is always named as oid. If a table has relations with other table(s) it will also contain a foreign key column for each related table. The foreign key column contains the value of the primary key values (oid) of the associated row(s) in the other table. Foreign key column names always include the prefix the_, as with the following names: the_parent, the_event, and the_connection. For examples of how the relations are used in practice, please see the following two chapters that describe the mapping of static and dynamic events to database tables.

Mapping of Nemo log file events to database tables


The data in Nemo log files is written in event-based format. All data related to a same object (e.g. Block Error Rate) is written as a single event. Each different event type in Nemo file format has a corresponding table structure in the database. Events can be split into two categories. Static events always have the same amount of information elements, i.e. columns. Dynamic events have a varying number of information elements. A typical example of a static event is SHO. It reports soft handover events in UMTS. It contains always the same information elements, including SHO status, number of scrambling codes added to the active set, etc. An example of a dynamic event is ECNO. The ECNO event reports the RSCP and Ec/N0 of all measured active, monitored, detected, and undetected cells. The number of measured cells constantly varies based on the location of the measurement device in the network. The number of information elements, i.e. columns, in the ECNO event varies accordingly. The diagram below illustrates the mapping of SHO events to different database tables. A similar diagram applies to all static events. For each SHO event in the log file, there is one row written in the SHO table. All soft handover-specific information elements within the SHO event are written to the corresponding columns in the SHO table. Generic information elements common to all events are written in the Event table. Such information elements include time, latitude, longitude, GPS distance, etc. There is a one-to-one relation between the SHO table and the Event table. The column the_event in the SHO table is a foreign key, pointing to the oid column in the Event table. In other words, each unique row value in the column the_event in the SHO table corresponds with a unique row value in the column oid in the Event table. Based on this correspondence, the two tables can be correlated. The Device table contains a row for each log file loaded into the database. The only information column in this table is the log file extension. There is a one-to-many relation between the SHO table and the Device table. The SHO table contains a row for each SHO event from each log file loaded into the database. Each row of the SHO table belonging to same log file points to the same row in the Device table. That is, the value in the column the_device in the SHO table is the same on all rows that come from the same log file, corresponding with a single row value in the column oid in the Device table.

QU ER IES

3 17

The Measurement table contains a row for each measurement loaded into the database. It contains all measurement-specific information columns, including the log file title. It should be noted that the term measurement refers to one or multiple log files collected at the same time. If a Nemo Outdoor Multi measurement is made with 4 terminals, there will be a single measurement session and 4 log files. In this case, the measurement title is the same for all log files and the log files are differentiated from each other using different file extensions. When these files are loaded into the Nemo Analyze database, there will be 4 rows in the Device table, one for each log file, and one row in the Measurement table. So there is a one-to-many relation between the Measurement and Device tables. That is, the value in the column the_measurement in the Device table is the same on all rows that come from the same measurement session, corresponding with a single row value in the oid column of the Measurement table.

The diagram below illustrates the mapping of ECNO events into different database tables. The ECNO event in the log file contains Ec/N0, RSCP, channel number and scrambling code for each measured cell, and carrier RSSI for each measured channel. Because the number of information elements in each ECNO event varies, the data has been split to multiple different tables in order to achieve efficient data storage in the database. As with all dynamic size events, the data is split to multiple tables in the database. There is a single row written in the ECNO table for each ECNO event in the log file. The ECNO table consists of columns such as the number of active set cells, the number of monitored set cells, etc. The relations between the tables ECNO, Event, Device, and Measurement are similar to those of the SHO event, see previous example. There is a one-to-many relation between the tables ECNO and Channel. In the Channel table, there is a row for each carrier RSSI value reported in an ECN0 event. All the rows in the Channel table that belong to the same ECN0 event have the same value in the column the_parent. The value in the column the_parent equals the oid value of the matching row in the ECN0 table. There is also a one-to-many relation between the tables ECNO and Cell. For each measured cell in the ECNO event, there is a row in the Cell table. All the rows in the Cell table belonging to the same ECN0 event have the same value in the column the_parent. The value in the column the_parent equals the oid value of the matching row in the ECN0 table. As an example, consider an ECN0 event that includes measurement results from six different cells. Two of these cells are in channel x, and the rest in channel y. When parsed to the database, the event will produce one row in the ECN0 table, two rows in the Channel table, and six rows in the Cell table.

QU ER IES

3 19

The structures of all tables in the database follow the same principle as the SHO event and the ECNO event above. All tables in the database are structured based on the same principle with either the SHO event or the ECNO event (discussed above). Each column, table, and relation is described in detail in the Nemo Analyze database schema.

Views Automatically joined tables


As mentioned earlier, each information element is stored as a separate table based on the event structure of the Nemo log file format. General information, including time, lat, long, and log file name, are stored in the tables Measurement, Device and Event. These tables have relations with measurement data tables, as described in the dynamic and static event examples above. Typically, when a query is made, the data of interest, such as RX level, Ec/N0, etc. is retrieved together with corresponding general information which at the minimum includes time, latitude and longitude. This means that the tables Event, Device and Measurement must be joined with the table containing the data of interest.

The Nemo Analyze database schema contains views for each data table that results from joining the Event, Device and Measurement tables. In other words, a view is a virtual table that joins the tables Event, Device, and Measurement as one. When using a view in an SQL query, the tables are joined by the ODBC driver and not by the user. The views should always be used when there is need for viewing time, lat, long, etc. information together with the data of interest. This is for two reasons. Firstly, the views are easier to use because queries become simpler when there is no need for performing the join operation as part of the SQL query itself. Secondly, the query performance is also a little bit better when using views instead of query-based join operations. A + character at the end of the table name identifies a table as a view. For example, the view corresponding with the SHO table is titled SHO+. In the screenshot below, all the columns in the BLER+ table are listed. It can be seen that BLER, time, lat, long, measurement name, etc. are all available in the same table.

QU ER IES

3 21

The screenshot below demonstrates how views are used as part of queries. Both of the illustrated queries return the same information, i.e. time and bler. The upper query represents a situation where the join is performed as part of the SQL query, i.e. without using the relevant view. The query retrieves bler from the BLER table and time from the Event table. The Event and BLER tables are joined with the WHERE clause of the query. The lower query represents a situation where the relevant view is used, i.e. the query retrieves both time and the parameter bler from the BLER+ table.

Database timestamps
The Nemo Analyze database stores time in two different formats. The column sql_time in the Event table contains the timestamps of each measurement event in a standard, readable SQL timestamp format. The column time in the Event table contains the timestamp in proprietary binary format. The Nemo measurement file format is time-based, as opposed to sample-based. In other words, a new sample is created on a timeline only when changes occur in the monitored parameter and the temporal length of the samples depends entirely on how often changes occur in the parameter value. This also means that the samples are not equal in weight, as the duration of each sample affects the weight of the sample. In order to calculate correct statistics, each sample has to be weighted by its duration. Also, to ensure that line graphs and map plots are drawn accurately, the duration of each sample must be taken into account. The binary time stored in the column time (of the Event table and of all the views) contains both the timestamp and the duration of the sample. The column sql_time contains only the timestamp. When custom queries are used in Nemo Analyze, the binary timestamp time should always be used. Nemo Analyze converts the binary time to timestamp at runtime, and automatically utilizes the sample duration (embedded in the timestamp) in route colouring and line drawing.

QU ER IES

3 23

The screenshot above shows how the sample duration in milliseconds can be retrieved from the binary timestamp using the scalar function TI_INTERVAL(). It is also evident that the duration of some events is 0. These event parameters (such as dropped call, call attempt failure, and cell reselection) do not have duration, as they represent events that occur at a point in time. Therefore, the scalar function TI_INTERVAL(time) returns the value 0 ms. All other parameters, such as Rx level, Ec/N0, throughput, BLER, etc., contain a duration. When using Nemo Analyze queries with third-party tools, sql_time and TI_INTERVAL(time) must be used instead of the binary timestamp. Moreover, sql_time must always be used as part of the query when sorting the result data set by time or when using conditions (e.g. x.sql_time > y.sql_time).

Filtering log files


The queries are usually run on a subset of log files that is narrowed down from all the files in the database. The optimal query performance is achieved using Anites proprietary, non-SQL filters in the end of the queries. These are called hints. Because of the physical structure of the Nemo Analyze database, all irrelevant data is filtered out and not processed. The hints point directly to the relevant log file(s), enabling outstanding query performance regardless of the database size. Hints are added to the end of the query in the following format: /* MEAS({measurement_file_name_1}:{file_extension}|{measurement_file_name_2}:{file_extensio n}|....) */ The screenshot below shows an example of how the filter should be integrated into a query. An example of the same query in SQL format is also provided. It should be noted that log file filtering queries are unnecessary when creating custom SQL queries that are intended for use with Nemo Analyze only. This is because Nemo Analyze automatically adds the hints to the queries at runtime.

Value enumeration
All textual parameters with a discrete list of possible values are written to the database as numbers. In other words, each textual value has a corresponding numerical value. Examples of such parameters include system (GSM, UMTS, etc), handover type, call disconnect status, call failure cause, etc. The textual values of all parameters of this type are stored in a ValueEnum table (see screenshot below).

The two screenshots below show two versions of the same query. The first one returns values in numerical format. The second one uses the VAL_TO_STRING(<param_name>, <input column>) scalar function for retrieving the textual value for each numerical value. The scalar function VAL_TO_STRING retrieves the value enumeration automatically from the ValueEnum table.

It should be noted that the use of the scalar function VAL_TO_STRING is unnecessary when creating custom SQL queries that are intended for use with Nemo Analyze only. This is because Nemo Analyze displays the value enumeration automatically.

QU ER IES

3 25

Connections
The connection sessions measured during a drive test are stored as connection tables. The following connections and connection tables exist: Voice, Handover, Attach, PDPContextActivation, Data, RRC, DataTransfer, MMS, SMS, POC, LAU, RAU, and Ping. The connections often have a hierarchical structure. For example, the DataTransfer connection always has the Data Connection table as its parent connection. The Data Connection table has the PDPContextActivation table as its parent connection. Finally, the PDPContextActivation table has the Attach table as its parent connection. All data recorded during a measurement session has by necessity a relation (or relations) with the connection tables. Examples of such data include Tx power (recorded only during voice call or packet session), application throughput (recorded only during data transfer), and BLER (recorded only during packet session and voice call). As an example, the screenshot below illustrates the relations between the tables DAS and Attach. Attach is a connection table, containing a row for each Attach session. The DAS table contains a row for each uplink and downlink application throughput sample. Each row of the DAS table is related to the corresponding connection in the Attach table via the_connection foreign key. The Attach table has also relations with the GAA (Attach attempt), GAC (Attach connected), CAD (Attach disconnected), and CAF (Attach failure) tables. All the different connection tables follow the same principle. They contain all data specific to a particular session, including the data events that define the beginning and the end of the session.

The connection tables can be used for various purposes: Grouping data per session, e.g. average Data throughput per session:

SELECT the_connection, AVG("app_throughput_downlink") FROM "Nemo"."DAS+" WHERE "throughput_status"=1 GROUP BY the_connection

Filtering in only those throughput samples that were recorded during a data transfer:
Connection types: 0="Unknown" 1="Voice" 2="Handover" 3="Attach" 4="PDP contect activation" 5="Data" 6="RRC" 7="Data transfer" 8="MMS" 9="SMS" 10="POC" 11="LAU" 12="RAU" 13="Ping"

Filtering in only those throughput samples that were recorded when PDP context was active and Access point was Internet:

The two scalar functions utilized in the examples above can be defined as follows:

CONN_IS_SHARED(conn1.oid, conn2.oid). Checks if the two connections are actually the same, or if one of the connections is the parent of the other CONN_IS_TYPE(conn.oid, numeric_exp). Checks if the connection is of a particular type. The connection types are listed in the example above.

QU ER IES

3 27

Correlating tables based on time


Most tables do not have relations with each other. However, it is still often necessary to correlate data from these unrelated tables based on time. For example, the tables Rx level and Rx quality are not related, but the user may still want to correlate them based on time to find instances where bad Rx quality has coincided with good Rx level. Generally, for performing correlations of this type, it is recommendable to use Nemo Analyzes Correlate parameters functionality or the KPI Workbench. However, it is also possible to correlate the two tables using SQL and proprietary keywords, as explained in the two examples above.

Sample-based correlation
The procedure for correlating two data sets sample by sample based on time is as follows: 1. Define the tables to be joined using a FROM clause: FROM table x, table y 2. Define the x.time = y.time condition using a WHERE clause Note that the order in which the tables are entered to the FROM clause is extremely important. When the table x is the first one in the FROM clause: Each sample from the table y is retrieved and checked based on the following condition: if there is an x sample that has the same timestamp as the y sample OR if the x timestamp falls to the validity time interval of the y sample When the table y is the first one in the FROM clause, the comparison is performed vice versa Note that the comparison is not bi-directional. If there is a significant difference in the sample periods of the tables x and y, the table that has a smaller sample period should be selected as the last one (y) in order to get highest possible resolution 3. Define the condition x.the_device = y.the_device using a WHERE clause. This limits the timebased correlation into individual measurement files, improving the performance of the query and ensuring that the correlation is performed within a single log file. 4. Add the hint /* OPTIONS(USE_TIME_SCOPE) */ at the end of the SQL query. This enables the join based on time scope. The following screenshot shows an example correlation of the best active set cell Tx power and RSCP, filtered from periods where RSCP is < -95 dBm Tx power is in the table Nemo.UMTS.TXPC and RSCP is in the table Nemo.UMTS.ECNO.Cell. There is no established relation between the tables. All Tx power samples that fall within the validity time interval of an RSCP sample with the value <-95 are retrieved.

Time range correlation


Data can be correlated with a connection session or a time range if start and stop events can be retrieved from the database. A common example of time range correlation is a situation where data is tagged to a particular connection session and there is no established relation with the connection table. For instance, the parameters Ec/N0, RSCP, and RxLevel are always recorded, whether or not a voice call or other connection session is active. Although these parameters do not have established relations with the connection tables (see Connections chapter on page 325), the user may still want to correlate connection session information with them, e.g. to filter Ec/N0 from data transfers only, to calculate average Ec/N0 per data transfer, etc. Below you can find an example query that returns active set best RSCP from the time instants that coincide with active data transfer sessions. The time range correlation is performed based on simple time conditions as part of the WHERE clause of the query. The time condition should be defined using the column sql_time. The column time cannot be used because it is in a binary format understood by Nemo Analyze only. Also T_(time) should not be used although it is the actual timestamp. This is because the query performance with T_(time) is lower than with sql_time because T_(time) does not use indexing. Note also that the start and end points of the time range need to be joined (see the last condition in the WHERE clause). In practice this is possible only with tables that have relations with connection tables.

Guidelines for creating SQL queries


New custom SQL queries can be developed using an SQL editor. The Database Browser functionality, or one of the many available third-party query editors, can be used for the purpose. When the query is complete and tested, it can be copied to the Query Manager as a new custom query. To access the Query Manager, select View | Query manager from the Ribbonbar. Both the custom SQL queries and the ones created with the KPI Workbench can be found under the User branch of the Parameter Tree. To execute a query on a measurement file/folder, select a measurement file/folder in the Workspace and then double-click on a query. The process of creating a new query starts with the identification of relevant input data. The tables that contain the relevant data can be identified by examining the Nemo Analyze database schema (see Nemo Analyze Database Schema on page 315). The best way to find the data of interest is to export the schema to a PDF document and to search the document for the parameter name. The schema describes the database table by table, column by column, and relation by relation. Please note that the same parameter can exist in multiple tables, mainly because of the system split, e.g. the tables GSM.BLER and UMTS.BLER both contain the column bler. The design phase of the new query follows. The source data is appropriately processed based on the selected tables. The supported SQL syntax is discussed in the reference documents (see Getting started on page 315 for more information).

QU ER IES

3 29

With Nemo Analyzes extensive set of premade queries, it is most of the time sufficient to only modify some of the queries found in the Parameter Tree. The recommended starting point is to first open the query in a data view. Once the query has been run, the query can be captured in SQL format from the SQL History view. To access the SQL History view, click the SQL History tab in the bottom left corner of the Nemo analyze UI. If not visible, the SQL History view can be enabled by selecting View | Panels | SQL History from the Ribbonbar. The SQL history contains all executed queries in the order of execution (see screenshot below). To copy the contents of the SQL History view, select the contents, right-click on the selection, and select Copy to clipboard from the popup menu. Paste the query to the Nemo Analyze database browser.

If using the Nemo Analyze database browser for query development, the SQL of an existing query can also be captured using the query clipboard. To do this, open the query of interest in a data view, rightclick on the data view, and select Query | Copy from the popup menu. Paste the query to the Nemo Analyze database browser.

Unique features of Nemo Analyze SQL interface


Nemo Analyze database has some special behaviours and features that have been discussed in detail in the previous chapters. In many cases, the correct query creation procedure depends entirely on whether the query is to be used with Nemo Analyze or with a third-party ODBC data retrieval tool. The unique characteristics of the Nemo Analyze database are summarized below. Timestamp. The binary timestamp time should be retrieved if the query is to be used with Nemo Analyze. Nemo Analyze converts it automatically to a readable timestamp. Nemo Analyze can also make use of the validity time interval embedded within the timestamp when drawing line graphs and map plots, and when calculating statistics. The standard time stamp sql_time should be used with ORDER BY, WHERE and SELECT clauses if the query is to be used with a third-party tool. Measurement file filter. If the query is to be used with Nemo Analyze, the limiting of the query to defined log files must not be done as part of the SQL query, as Nemo Analyze limits the query automatically at runtime using proprietary hint-based filtering. To achieve optimal performance, it is recommended that the hints be used also with queries intended for use with third-party tools. Converting numeric values to corresponding textual values does not have to be done with the scalar function VALUE_ENUM if the query is to be used with Nemo Analyze. When necessary, Nemo Analyze automatically converts the numeric values to the corresponding textual values. Results have to be in chronological order. If a query result set contains the time column and the results are to be displayed in map, line graph, bar graph or numerical data format, the results must be ordered by time. Nemo Analyze produces the results automatically in time order, log file by log file. However, with e.g. the aggregates GROUP BY and UNION, time order is lost and the results must be rearranged using the clause ORDER BY sql_time.

SQL VS. KPI Workbench


For many purposes, SQL queries represent the optimal solution. However, the SQL language and the Nemo Analyze database schema set some limitations that can make for instance the following tasks easier to perform with the KPI Workbench. Correlating more than two tables based on time. Sample-based correlation (described above) enables two tables without an established relation to be joined based on time. However, with three or more tables the task becomes impossible. For example, correlating BLER, Ec/N0 and TX power based on time is impossible with SQL because all of the parameters are located in separate tables with no established relations between them. Creating queries that track certain event sequences, such as the change in a particular parameter value columns from row to row. SQL works well in scalar data processing where mathematical or logical operations are performed on a row-by-row basis, using different columns of each row as input. SQL sub queries can be utilized in creating an event to track situations where the system changes from GSM to UMTS (i.e. where a row contains the serving_system parameter value GSM and the subsequent row the value UMTS), but the method will cause the performance of the query to collapse. The state machine element of the KPI Workbench enables such queries to be created with ease and optimal query performance.

QU ER IES

3 31

SQL queries are at their best in scalar data processing. If all the required input data is located in the same table (i.e. time-based correlation based on multiple tables is not necessary), SQL is a good approach to complex scalar (row-by-row) data processing. Some examples of such tasks can be found below: Data filtering. Complex filters can be created using SQL scalar functions and logical operations. Data reformatting. Query result set columns can be modified and combined using various string and math functions. Conditional outputs can be created using the SQL statement CASE. As mentioned earlier, using SQL requires proficiency in SQL and knowledge of the Nemo Analyze database schema, and even then, the SQL language itself poses some limitations. Therefore, for a user without previous experience with SQL, KPI Workbench is recommended as the primary data customization tool. KPI Workbench avoids the limitations of SQL and offers full functionality also for data filtering and data reformatting purposes.

QUERY MANAGER
Queries are sets of rules based on which data is retrieved from the database, and then displayed in the various views in Nemo Analyze. There is a wide range of queries ready in Nemo Analyze that will assist you in your day-to-day data analysis but it is also possible to create new queries. You should not limit a query to certain measurement file(s), as Nemo Analyze limits the query automatically to the measurement files required. This way the query is not statically limited to some predefined measurement file or files, but instead, it can be run over any given measurement file or files in the database. Select View | Tools | Queries in the Ribbonbar to open the Query Manager.

In the Query Manager dialog, select the User item and click Add.

There are different types of queries: you can select a measurement parameter from a list (Pick measurement parameter), select query data from tables (Generic query wizard), or type the query string manually (Manual query).

QU ER IES

3 33

Pick Measurement Parameter


Pick Measurement Parameter is a quick and easy way to make new queries. The Analyze Wizard will guide you through the steps. Select Pick measurement parameter option from the list and click Next.

The Measurement Parameters page displays a list of available parameters. Select a parameter from the list and click Next. In the Filter field you can type the name or part of the name of the parameter you are looking for to locate it faster. This is especially useful when the parameter list is long.

In the Fill Parameters dialog you can further limit the query results by adding different filters. All parameters can be filtered, e.g. based on Area and Time. Filters can be applied to parameters that are part of the query. Right-click on top of the filter name or value to Add, Modify or Delete filters.

With numeric parameters (e.g. Ec/N0, but not for instance L3 signaling), the Analyze Wizard Statistics view opens. The view allows you to define statistics to be run over the selected parameter.

Statistics type defines the statistics type. Percentile value allows you to define the percentile value, e.g. 95th percentile. Group by allows you to define the basis for grouping the parameter aggregates, e.g. Ec/N0 values could be grouped according to scrambling code.

QU ER IES

3 35

Type a name for the new query and click Next. You can also type a description for the parameter in the Description field.

In the Column Aliases page you can define how the retrieved data is displayed in the various data views. This setting is optional. Click Finish.

The query is added to the Query Manager and it can also be found in the Analyze Wizard under the User item, and in the Parameter tree under the User item. In the Query Manager you can manually edit the query and the query filters, properties, statistics, and aliases by selecting the query from the list and clicking the Modify button. You are also able to modify correlation queries, however, note that filters cannot be modified for correlations queries. To make a copy from an existing query by selecting a query from the list and clicking the Copy button.

QU ER IES

3 37

Generic Query Wizard


Generic Query Wizard gives you more query options to choose from.

In the Select Tables page are displayed all the data tables available in the Nemo Analyze schema. Select the table(s) from which you want to retrieve the query data.

Choose the parameters for the query, and press the simple arrow for single parameters, and the double arrow to transfer all the parameters in the table.

In the Select Columns page, select the information that you are interested in. Click Next.

QU ER IES

3 39

In the Group Columns page you can define in which order the data will be displayed. It is recommended that you group columns based on time. If there is only one table selected to the query, the columns are by default ordered by time. This setting is optional.

In the Filters page, you can define filters for the query data. For example, you can define that the query is only applied to parameter values less than 0. You can also define a textual filter, for example, {?Threshold}. When the query is run, Nemo Analyze will open a dialog where you can enter a numerical value for the filter, for example, 10. This setting is optional.

In the Properties page, define a name for the query. In the Description field you can write a description for the parameter. Click Next.

If you defined a textual filter in the Filters page, define a numerical value in the Fill Parameters page and click Finish. Right-click on top of the filter name or value to Add, Modify, or Delete filters.

QU ER IES

3 41

In the Column Aliases page you can define how the retrieved data is displayed in the various data views. This setting is optional. Click Finish.

The query is added to the Query Manager and it can also be found in the Analyze Wizard under the User item, and in the Parameter tree under the User item. In the Query Manager you can manually edit the query by selecting it from the list and clicking the Modify button.

Manual Query
Manual Queries require knowledge of SQL and the Nemo Analyze database schema. Nemo Technical Support will provide you with the database schema.

Type a name for the query and select the Edit SQL manually option. In the Description field you can write a description for the parameter. Click Next.

QU ER IES

3 43

Type in the query string and click Finish.

The query is added to the Query Manager and it can also be found in the Analyze Wizard under the User item, and in the Parameter tree under the User item. In the Query Manager you can manually edit the query by selecting it from the list and clicking the Modify button.

Correlate Parameters
The Correlate Parameters query type is explained in more detail on page 120.

DATABASE BROWSER
You can enter the database browser via View | Tools | Database browser.

Database Browser can be used to display the table structure of the Nemo Analyze Database. The contents of each table or individual columns of a table can be viewed by double-clicking it. The corresponding query will be shown in the text window. Database Browser can also be used to write and test custom queries. The query can then be copied to the clipboard and saved to the UI with Query Manager.

QU ER IES

3 45

CUSTOM KPI WORKBENCH


Custom KPI Workbench makes it possible to create custom KPIs by dragging and dropping parameters, correlations, various operations, filters and sort elements to the Workbench, defining their properties, and connecting them into a logical flow chart. The KPI Workbench is part of the Troubleshooting Toolkit option. To open Custom KPI Workbench, select View | KPI Workbench in the Ribbonbar.

An empty Custom KPI Workbench view opens.

To clear the Workbench view, right-click on the Workbench background and select New Script from the popup menu. To run the KPI on a measurement file, right-click on the Workbench background and select Run Script from the popup menu. For further instructions, see p.397. To save the KPI, right-click on the Workbench background and select Save from the popup menu. For further instructions, see p.399.

Designing Custom KPIs and Adding Input Data Sets


The first step of creating custom KPIs should always be careful planning. First, determine what the KPI is supposed to do, i.e. what kind of data the output should consist of. Next, determine what kind of raw data parameters are needed to produce the right kind of output. Note: In some cases, it may be necessary to create a KPI that correlates two separate files that do not have any parameters in common (e.g. mobile and scanner files from the same measurement session). You may also wish to define value constants, i.e. values that can be referred to in scripts with a generic reference. If the use of either of these functionalities is necessary in terms of the KPI you intend to create, it is recommended to see the chapter Defining KPI Execution Method and Value Constants before continuing. To add a raw parameter data set (containing all samples within the data set), select a measurement file in Workspace | Folder Contents, and drag and drop the appropriate parameter to the Workbench from the Parameter Tree (located on the bottom left of the screen). To add a data set limited to pre-calculated statistics on a folder of measurement files, select a folder in Workspace | Folders, and drag and drop the appropriate parameter to the Workbench from the Parameter Tree.

You can also create a custom input parameter by dragging and dropping a Parameter element to the Workbench view from the menu structure on the right and defining the parameter using the Query Manager (see Parameter on page 348).

QU ER IES

3 47

To leave out irrelevant columns from the input and output data sets, columns in the parameter data set can be disabled. To do this, double-click on the parameter element in the Workbench view. Note that the columns can only be modified before the parameter element is connected to any subsequent KPI elements.

The Properties dialog opens.

To disable column(s), click the Result Columns tab, clear the column(s) and click OK.

Parameter
A parameter is a data set in tabular format that has a column containing parameter values and typically other columns such as time, coordinates, system, etc. To create custom parameters to form the basis for the new custom KPI, drag and drop the Parameter element from the menu structure to the Workbench view.

This will open the Query Manager and enable you to customize and optimize parameters by selecting only the relevant columns to be included in the parameter output.

Select user query allows you to select a custom query (e.g. one previously created and saved using KPI Workbench) as an input data set. For Pick measurement parameter, Generic query wizard and Manual query, see Query Manager on page 332.

QU ER IES

3 49

If there is no need to customize parameters, you can open default parameters on either entire raw parameter data sets or on data sets limited to pre-calculated statistics. To add a raw parameter data set (containing all samples within the data set), select a measurement file in Workspace | Folder Contents, and drag and drop the appropriate parameter to the Workbench from the Parameter Tree (located on the bottom left of the screen). To add a data set limited to pre-calculated statistics on a folder of measurement files, select a folder in Workspace | Folders, and drag and drop the appropriate parameter to the Workbench from the Parameter Tree.

Combining Input Data Sets


Parameter elements consist of measurement data in tabular format. Thus, the raw data parameter Rx Level is a table with the columns rx level, time, latitude, longitude, etc. If you need more than one raw data parameter, it is usually necessary to combine these data sets or tables into a single table before you can perform any arithmetic operations on or between them. Determine which method of combination would be the most useful in terms of the operations you are going to use. Note: Because different operations often have different requirements on the method of how the data sets are combined, it is important that the phases Combining Input Data Sets and Adding Operations are planned together. With operations that correlate parameters or search values from different parameters from the time range of a particular data set (namely the correlations Previous, Next and Current Value and All Values Within Time Range), it is important that the primary data set is placed on the extreme left compared to the other parameters or data sets connected to the same correlation, operation, etc. For instance, with the correlation Left Outer Join, all data entries from the data set on the left are included in the output, and from the secondary parameters or data sets, only the ones that match those of the primary data set. Similarly, the correlation All Values Within Time Range produces all values of the secondary parameters from the time range of the primary (leftmost) parameter or data set.

Secondary parameter Primary parameter

Correlation Connecting elements by dragging connector lines from output sockets to input sockets

In the example above, the Previous Value element is used to examine values of a secondary data set in relation to the primary data set, the parameter Call Attempt Failure. Thus, the values of the secondary parameter Ec/N0 Active Set are examined from a time range determined by both the correlation Previous Value and the primary data set Call Attempt Failure. As the primary data set Call Attempt Failure consists solely of call attempt failure events and the correlation Previous Value produces only those values of the secondary parameter Ec/N0 Active Set that immediately precede events in the primary parameter, the output will consist of Ec/N0 Active Set values preceding the start time of each Call Attempt Failure event. To connect the various KPI elements to each other, connect output sockets to input sockets by pressing the left mouse button down on the output socket of one element and holding it down while dragging the connector line to the input socket of the other element. To connect the same set of data to several different operations, simply draw as many connector lines as needed from the output socket of the data set (see picture below).

Once connected, the elements will turn green to signify that they are operational. If an element remains red, it requires further configuration. Right-click the element and select Properties from the popup menu. For more information on configuring individual elements, see the related topic below.

QU ER IES

3 51

Correlation: Previous/Current/Next Value


The Previous Value, Current Value and Next Value elements are used in examining values of secondary data sets from points in time determined by a primary data set. This makes it possible to examine what occurred in the network before, during, or after parameter events in the primary data set based on one to many secondary parameters. Previous Value produces only the previous value immediately before the start time of the parameter events in the primary data set, Current Value only the current value during the parameter events, and Next Value only the next value immediately after the events. To add a Previous/Current/Next Value element to the KPI, first make sure that the primary parameter or data set is situated the leftmost compared to the secondary parameters whose values are to be examined.

Drag and drop the Previous/Current/Next Value element to the Workbench. Connect the primary data set to the leftmost input socket of the Previous/Current/Next Value element.

Connect the secondary parameters (i.e. the parameters whose previous/current/next values are to be examined from points in time determined by the primary parameter) to the other input sockets of the Previous/Current/Next value element.

The Previous/Current/Next Value element changes color from red to green, signifying that the element is operational. Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu.

QU ER IES

3 53

The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

Correlation: Previous Or Current Value/Next Or Current Value


The Previous or Current Value and Next or Current Value elements are used in examining the parameter values in secondary data sets from points in time determined by a primary data set. These functions enable you to examine the parameter values in any number of secondary data sets before and during, or during and after, an event (such as a dropped call) has been logged in the primary data set. Previous or Current Value produces the previous value (immediately before the start time of the parameter event(s) in the primary data set), or the current value (during the parameter event(s) in the primary data set) if a current value is available. Next or Current Value produces the next value (immediately after the parameter event(s) in the primary data set), or the current value (during the parameter event(s) in the primary data set) if a current value is available. To add a Previous or Current Value/Next or Current Value element to the KPI, first make sure that the primary parameter or data set is situated the leftmost compared to the secondary parameters whose values are to be examined.

Drag and drop the Previous or Current Value/Next or Current Value element to the Workbench. Connect the primary data set to the leftmost input socket of the Previous or Current Value/Next or Current Value element.

Connect the secondary parameters (i.e. the parameters whose previous/current/next values are to be examined from points in time determined by the primary parameter) to the other input sockets of the Previous or Current Value/Next or Current Value element.

QU ER IES

3 55

The Previous or Current Value/Next or Current Value element changes color from red to green, signifying that the element is operational. Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu.

The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

Correlation: All Values Within Time Range


All Values Within Time Range, as compared to Previous, Current, and Next Value elements that produce only one value, produces all values from the secondary parameters from the time range of the primary (leftmost) parameter or data set.

Generally all RF parameters of the Nemo file format have a time range. This is because the Nemo measurement file format is time-based as opposed to sample-based. In other words, a new sample is created on a timeline only when changes occur in the monitored parameter and the temporal length of the samples depends entirely on how often changes occur in the parameter value. Event parameters (such as dropped call, call attempt failure, and cell reselection) do not have a time range as they represent events that occur at a point in time. A major difference between All Values Within Time Range and the other join elements (namely Inner Join, Left Outer Join, and Union) is that All Values Within Time Range combines the data based on time, whereas these other join elements do this based on matching values in the data sets. To add an All Values Within Time Range element to the KPI, first make sure that the primary parameter or data set is situated the leftmost compared to the secondary parameters whose values are to be examined from the time range determined by the primary element. Note: When deciding which input data set to use as the primary dataset, one should consider the fact that the output will be written only when there are valid values in the primary dataset. Therefore, if filtering out data from the secondary data set is not desirable, the parameter that contains valid samples for the most of the duration of the log file should be selected as the primary dataset. For example, selecting RX qual as the primary data set when correlating the parameters RX lev and RX qual, will produce results only from instances when a call was active. On the other hand, if RX level is selected as the primary data set, the results will include both idle periods and periods when a call was active.

If Rx level is selected as primary dataset, results are given both from idle, and call connected periods. Drag and drop the All Values Within Time Range element to the Workbench. Connect the primary parameter or data set to the leftmost input socket of the All Values Within Time Range element. Connect the secondary parameters or data sets (i.e. the parameters whose values are to be examined from a time range determined by the primary parameter) to the other input sockets of the element. The All Values Within Time Range element changes color from red to green, signifying that the element is operational.

QU ER IES

3 57

Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu. The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

Join: Inner Join


Inner Join can combine two sets of data that share a join value (such as identification code, channel number or scrambling code) defined by the user. Only the rows containing a matching join value will be included in the output. For instance, if two parameters both contain scrambling code information, the parameters can be joined using Inner Join by defining scrambling code as the join value. All rows with a scrambling code value that does not match with any of the scrambling code values in the other data set, and those containing a null value in the scrambling code column, will not be included in the output. To add an Inner Join element to the KPI, first open both of the data sets that are to be combined on a grid (for opening measurements on a grid, see Viewing Measurement Data on p.68) and consider which join values (such as identification code, channel number, scrambling code, etc.) you can use as a basis for the join, i.e. what values do the data sets have in common, what values would form a logical basis for the join in terms of the KPIs purpose, etc. In the example screenshots below, the two tables can be joined based on the scrambling code values they share.

Drag and drop the Inner Join element to the Workbench. Connect the data sets that are to be combined to the upper (i.e. input) sockets of the Inner Join element by pressing the left mouse button down on the output socket of each data set and holding down the mouse button while dragging a connector line from the data set output socket to the input socket of the Inner Join element. Right-click the Inner Join element and select Properties from the popup menu.

QU ER IES

3 59

From the Join tab of the Properties dialog, select the join values (i.e. the values that form the basis for the join) using the Left column and Right column drop-down menus. After you have defined the properties of the Inner Join element, click OK, and the element changes color from red to green, signifying that the element is operational. Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu.

The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

Join: Left Outer Join


Left Outer Join can combine two sets of data that share a join value (such as identification code, channel number or scrambling code) defined by the user. From the leftmost set of data, all rows are always included. From the secondary data set, only the rows containing join values that match rows in the leftmost set of data will be included in the output. For instance, if two parameters both contain scrambling code information, the parameters can be joined with Left Outer Join by defining scrambling code as the join value. All rows from the leftmost parameter are included in the output. From the secondary parameter, only the rows with scrambling codes that match with those of the left parameter are included. To add a Left Outer Join element to the KPI, first open both of the data sets that are to be combined on a grid and consider which join values (such as identification code, channel number or scrambling code) you can use as a basis for the join, i.e. what values do the data sets have in common, what values would form a logical basis for the join in terms of the KPIs purpose, etc. Drag and drop the Left Outer Join element to the Workbench.

QU ER IES

3 61

Connect the data sets that are to be combined to the upper (i.e. input) sockets of the Left Outer Join element by pressing the left mouse button down on the output socket of each data set and holding down the mouse button while dragging a connector line from the data set output socket to the input socket of the Left Outer Join element. Right-click the Left Outer Join element and select Properties from the popup menu.

Note: Depending on the input data sets, it may sometimes be necessary to group the input data sets based on the join value before using Left Outer Join to combine them. In the screenshot below, both input data sets have been grouped by scrambling code because scrambling code is going to be used as the Left Outer Join join value.

From the Join tab of the Properties dialog, select the join values (i.e. the values that form the basis for the join) using the Left column and Right column drop-down menus.

After you have defined the properties of the Left Outer Join element, click OK, and the element changes color from red to green, signifying that the element is operational. Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu.

The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

QU ER IES

3 63

Join: Union
Union joins two or several sets of data, with or without matching rows, into a single data set without removing any duplicate rows. Same-named columns containing data of the same type (e.g. both contain a string) are combined into a single column. Otherwise, each data set forms a new column into the resulting data set. Because the rows and columns in the resulting data set are in no particular order, it is often necessary to sort the resulting table based on, for instance, time in order to produce meaningful results. To add a Union element, drag and drop the Union element to the Workbench. Connect the data sets that are to be combined to the upper (i.e. input) sockets of the Union element by pressing the left mouse button down on the output socket of each data set and holding down the mouse button while dragging a connector line from the data set output socket to the input socket of the Union element.

The Union element changes color from red to green, signifying that the element is operational. Because the rows and columns in the unified data set are in no particular order, it is often necessary to sort the resulting table based on, for instance, time in order to produce meaningful results.

To add a sort element, drag and drop either Ascending or Descending sort element to the Workbench. Connect the Union element to the sort element. Define the sort elements properties by right-clicking the element and selecting Properties. For more information on defining the sort elements properties, see p.382.

Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu.

The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

QU ER IES

3 65

Join: Cartesian Product


Cartesian Product (or Cross Join) combines all rows from two or more data sets, producing a row for each possible combination. For the purposes of the KPI Workbench, Cartesian Product is most commonly used in combining two or more sets of data that are both known to contain only a single row of data (such as the results of Aggregate functions) into separate columns on a single row. To add a Cartesian Product element, drag and drop the Cartesian Product element to the Workbench. Connect the data sets that are to be combined to the upper (i.e. input) sockets of the Cartesian Product element by pressing the left mouse button down on the output socket of each data set and holding down the mouse button while dragging a connector line from the data set output socket to the input socket of the Cartesian Product element.

The element changes color from red to green, signifying that the element is operational. Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu.

The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

Adding Operations
Once you have combined all the necessary raw data sets into a single table, use the state machine, arithmetic operations, filters, time functions, etc. to define how the raw data is to be processed by the KPI. As the various KPI elements can be used together in countless different and complex combinations (with complex state machines, aggregate functions, Cartesian products, sort elements, etc.), you need to determine what correlations and operations are required and how they should interact in order to isolate only the relevant parts of the measurement data. For instance, whether a state machine needed, and if it is, what states would be relevant in terms of the data you are looking for, how should the transitions from state to state be triggered, what entries in the data would be the most logical or useful triggers for these transitions, etc. With the State Machine and other more complicated operations it is often necessary to first open a sample of the kind of measurement data the KPI is intended for with suitable parameters on a grid to determine how the data progresses and for instance what data entries should be used as triggers to obtain the most accurate results. For opening measurements on a grid, see Viewing Measurement Data on p.68. It is also worth noting that some operations may require a value (e.g. a transition trigger in state machine) to be defined when defining the operations properties. Entering a value to a field of this type can prove restricting in terms of the various different purposes the KPI could otherwise be used for. To avoid this problem and effectively create a more versatile KPI with values that can be entered later on based on the requirements of the KPIs each individual application, define each value as a variable by entering the string {?<variable name>} to the value field. This way, the program will prompt the user to define a value for each value defined as a variable every time the KPI is run. This makes it possible, for instance, to create a generic state machine to calculate the delay from any one L3 signaling message to any another. The state machine will prompt the user to define the two signaling messages each time the KPI is run, enabling the same KPI to be used for calculating delays between any number of different signaling message pairs.

QU ER IES

3 67

It should also be kept in mind that the Nemo file format is time-based as opposed to sample-based (see Correlation: All Values Within Time Range p. 355). Because of this, operations should always be weighted by time when applicable. If the results are not weighted by time, it will not be possible to know how long a particular parameter value has been valid and the results of operations such as Average and Count will not be accurate.

Operation: Case
With Case, the user can create rules that filter and divert different values from the data set to different output sockets, making it possible to direct these values to different operations or to examine them separately in the final output. Each of the output sockets filters out all the values that are not in accordance with the rule defining the socket. To add a Case element to the KPI, drag and drop the Case element to the Workbench. Connect the data set to the upper (i.e. input) sockets of the Case element by pressing the left mouse button down on the output socket of each data set and holding down the mouse button while dragging a connector line from the data set output socket to the input socket of the Case element. Right-click the Case element and select Properties from the popup menu. From the Flow Control tab in the Properties dialog, click the Add button to add a rule.

The Rule Properties dialog opens.

Name defines a name for the rule. Column defines the grid column (in the input data set) that contains the values that are compared to Value based on the rule defined in Operator.

Operator defines the rule that determines which values are filtered out. For instance, with Scrambling Code (Column) <= 2 (Value), all the values that do not fulfill the condition of being smaller than or equal to 2 are filtered out. Value defines the limit value that the Operator compares to the Column value. To define the value as a variable (i.e. the value is to be defined by the user each time the KPI is run), enter the string {?<variable name>} to the value field. Define Case properties and select OK. The Rule Properties dialog closes. To create another rule, click the Add button again in the Properties dialog and define properties for the rule. In the example screenshot below, two rules have been defined, each represented by an output socket of the Case element.

The element changes color from red to green, signifying that the element is operational. Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu. The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

Operation: Moving Average


Moving Average calculates value averages based on a user-defined number of samples preceding each value. In practice, this removes anomalies from, for instance, a graph, making it more stable and explicit. To add a Moving Average element to the KPI, drag and drop the Moving Average element to the Workbench.

QU ER IES

3 69

Connect the data set to the upper (i.e. input) sockets of the Moving Average element by pressing the left mouse button down on the output socket of each data set and holding down the mouse button while dragging a connector line from the data set output socket to the input socket of the Moving Average element. To define Moving Average properties, right-click the Moving Average element and select Properties from the popup menu. The Running Function tab of the Properties dialog opens.

Column defines the grid column (in the input data set) from which the moving average is to be calculated. Samples defines how many surrounding values will be sampled when calculating the average for each point in time. Define Moving Average properties and select OK. The element changes color from red to green, signifying that the element is operational. Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu. The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

Operation: Conversion
Conversion enables the user to convert value formats between integer, real (i.e. decimal) and string (i.e. text). If, for instance, a division is to be performed for data entries defined as integer values, it may be useful to convert these values to real format before the division to avoid rounded and thus inaccurate results. Also, if it is known that data entries have been falsely defined as being of string (i.e. text) data type but in reality consist of numbers, conversion can be used to convert the data type to numbers (i.e. either integer or real, depending on whether the entries contain integer or decimal values.) To add a Conversion element to the KPI, drag and drop the Conversion element to the Workbench. Connect the data set to the upper (i.e. input) sockets of the Conversion element by pressing the left mouse button down on the output socket of each data set and holding down the mouse button while dragging a connector line from the data set output socket to the input socket of the Conversion element. To define Conversion properties, right-click the Conversion element and select Properties from the popup menu. The Conversion tab of the Properties dialog opens.

Column defines the grid column (in the input data set) that contains the values that are to be converted. Target type defines the target data type to which the data within the column (defined in Column) will be converted. Possible target types include integer, real (i.e. decimal), and string (i.e. text). After you have defined the properties of the Conversion element, click OK, and the element changes color from red to green, signifying that the element is now operational. Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu. The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

QU ER IES

3 71

Operation: State Machine


State Machine makes it possible to define and isolate specific states and transitions between them from a single data set. It can be used, for instance, for examining the start and the end of particular events, the duration of such events, values of other parameters before, during and after these events, etc. For the state machine to work properly, it is recommended that the states and transitions are conceptually mapped (e.g. in the form of a flow chart, with references to events that trigger the transitions and define the boundaries of the states) beforehand based on grid data. Open the data set you are interested in on a grid (for opening measurements on a grid, see Viewing Measurement Data on p.68 ) and decide which value, signaling message name, event_ID, etc. is going to trigger which transition between which states. For instance, a transition could be triggered by an Ec/N0 value or an L3 signaling Message Name.

It is also very important that there is always a returning transition from each state in case the conditions of the transitions to the state are not fulfilled anymore. This ensures accurate results and the correct operation of the state machine.

BLER >= 20 OK Bad BLER

Ec/N0 difference <0 Missing handover

BLER < 20

Ec/N0 difference > 0

BLER < 20

After the operation of the State Machine has been conceptually mapped (see flow chart above), drag and drop State Machine to the Workbench. With State Machine it is important that the input data set is sorted based on time. Open the data set you are interested in on a grid and check whether it is sorted by time. If not, sort the input data set based on time by adding a sort element (i.e. Ascending or Descending, see p. 382) between the input data set and the State Machine. Connect the data set to the upper (i.e. input) sockets of the State Machine element by pressing the mouse button down on the output socket of each data set and holding down the mouse button while dragging a connector line from the data set output socket to the input socket of the State Machine element. Right-click the State Machine element and select Properties from the popup menu.

The State Machine tab of the Properties dialog opens.

First, create all the required states. To create a new state, click the Add button. The State dialog opens.

Enter a name for the state in the Name field and click OK. The State dialog closes. To create another state, click the Add button again in the Properties dialog. Note: Repeat these steps until you have created all the states you are going to need. You should have at least an idle state (i.e. a state that is not of interest to you) and an active state (i.e. a state that is in the focus of your analysis).

Next, select the idle state as the Initial State from the drop-down menu. To start defining transitions between states, select one of the states and click the Modify button.

QU ER IES

3 73

The State dialog opens.

Click the Add button. The Transition dialog opens.

Conditions displays the defined transition conditions. Time trigger defines a transition that is triggered if the condition is not fulfilled within the defined time period (in milliseconds).

Target defines the state to which the transition occurs. Output defines a title for the state that will be displayed in the final output data set. If the field is left blank, no output will be generated from this transition. For instance, with the idle state, leave this field blank. With the state(s) that interest you, be sure to define a title for the state. The output data set generated based on a state is defined as follows. When a transition occurs from the state x to state y, the point in time when the transition occurred from state x to state y (start_time), the point in time when the transition occurred from the state y to the next state, and the time in milliseconds that passed while in the state y (time_interval) are recorded in the output data set. Depending on the application of the state machine, different columns may be useful. When using state machine to create a custom event, the start_time is usually needed in the results to indicate the timestamp when the event occurred. State machine can be also used to calculate delay of certain procedure, for example radio bearer establishment in UMTS. In such case, state machine must be defined so that there is a state that is entered when radio bearer establishment begins, and the state is left when radio bearer is established. In this example, time_interval output column directly indicates the delay of radio bearer establishment in milliseconds. To define a condition, click Add. The Condition dialog opens.

Left Column defines the grid column (in the input data set) that contains the values that are compared to a value defined in Value, or another column defined in Right column, based on the rule defined in Operator. Operator defines the rule that determines what kind of relationship there has to exist between the values in Left Column and Right column/Value for the transition to be triggered (and for the present state to end). For instance, with the L3 Signaling message (Left Column) = CALL ATTEMPT (Value), the transition to the next state occurs when the value CALL ATTEMPT is found in the column L3 Signaling. Right column allows you to define a column to which the values in the Left Column are compared. If <Value> is selected, the Left Column value is compared to the value entered in the Value field. If <Previous value> is selected, and used with the Operator != (i.e. not equal to), the condition is fulfilled when the value in the Left Column changes. Value defines the value that triggers the transition, i.e. the end of one state and the beginning of a new one. To define the value as a variable (i.e. the value is to be defined by the user each time the KPI is run), enter the string {?<variable name>} to the value field. Note that each transition should have a unique variable name. After the condition has been defined, click OK.

QU ER IES

3 75

The condition is displayed in the Transition dialog.

To define another condition for the same transition, click the Add button again, and a new Condition dialog opens. Define condition properties in the Condition dialog and click OK. Both conditions are now displayed in the Transition dialog.

To define the AND/OR relationship between the conditions, right-click on the AND element.

Select AND or OR from the popup menu. When you have defined all the conditions for the transition, click OK to return to the State dialog. Repeat the steps to define transitions and conditions from each state you have created.

Finally, when all the transitions and transition conditions have been defined, click OK in the Properties dialog. The State Machine element changes color from red to green, signifying that the element is operational. Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu. The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397. For a practical example, see Use Case22.: Creating a KPI for Dropped Calls Resulting from a Missing Handoveron page 412.

Operation: Group By/Binning


The Group By element enables you to select multiple parameters from an input data set and group or bin the rest of the data set based on these parameters. The Group By element also enables you to simultaneously run multiple different aggregates based on any number of parameters contained within the input data set. Thus, the Group By element can be used for both grouping input data and producing parameter aggregates based on the input data. The available aggregate functions include Minimum, Maximum, Average, Standard Deviation, Variance, Sum, Count, Mode, Median, Percentile, First, and Last. For more information on aggregates, see Aggregate Functions on page 380. First, combine the parameter data sets into a single data set by dragging and dropping a relevant correlation element (e.g. All Values Within Time Range) to the Workbench. For more information on correlation elements, see Combining Input Data Sets on page 349. With All Values Within Time Range, connect the primary parameter or data set to the leftmost input socket of the All Values Within Time Range element. Connect the secondary parameters or data sets (i.e. the parameters whose values are to be examined from a time range determined by the primary parameter) to the other input sockets of the element. Next, add a Group By element to the KPI by dragging and dropping the element to the Workbench. Connect the correlation element (i.e. the data set that combines all the individual parameter data sets) to the Group By element by pressing the left mouse button down on the output socket of the correlation element and holding down the mouse button while dragging a connector line from the correlation element output socket to the input socket of the Group By element.

QU ER IES

3 77

Right-click the Group By element and select Properties from the popup menu.

The Group By tab of the Properties dialog opens.

Input displays all parameters within the input data set. From these parameters, you can select parameters to the Group By and Aggregates categories. Group by defines how the parameter aggregates defined in Aggregates are to be grouped. For instance, when calculating an Ec/N0 average per scrambling code, the aggregate function Average would have the following settings: Group by: scrambling_code and Aggregates: ec/no (with the aggregate function Average selected from the Function drop-down menu). To define a parameter as a basis for grouping, select a parameter in the Input field and press the button. To remove a

grouping parameter, select the parameter in the Group by field and press the button. There can be multiple grouping parameters. The order in which the grouping parameters are entered is also important, as every grouping parameter hierarchically determines also the grouping of the subsequent grouping parameters. Aggregates defines the aggregate parameters. To add a parameter to Aggregates, select a parameter in the Input field and press the button. To define parameter aggregate properties, select the parameter in the Aggregates field and define aggregate properties using Function, Weight by, and Result title. Once the properties of a single parameter aggregate have been defined, start defining the next one by simply selecting another parameter from the Input field and clicking the button. To remove a parameter aggregate, select the parameter in the Aggregates field and press the button. For more information on defining the properties of aggregate elements, see Aggregate Functions on page 380. Function defines the aggregate function applied to the selected parameter in the Aggregates field. Function properties are defined using Weight by and Result title. Weight by defines the how the calculation is to be weighted, i.e. by time, distance (based on GPS coordinates), etc. Weight by is usually used only with the aggregate functions Average and Count.

QU ER IES

3 79

Note: Because the Nemo measurement file format is time-based as opposed to samplebased (i.e. a sample is created on a timeline only when changes occur in the monitored parameters), the aggregate functions Average and Count should be weighted by time in order to obtain accurate results. Result title defines a title for the result. Without a title, the parameter aggregate will not be operational. In the example below, three ec/n0-based aggregates (ec/no minimum, maximum, and average) and three tx_power -based aggregates (tx power minimum, maximum, and average) have been defined. These have been grouped based on two parameters: bts_site_name and bts_cell_name. As the parameter bts_site_name is the uppermost, all other parameters, i.e. the grouping parameter bts_cell_name and the six parameter aggregates, will be grouped based on bts_site_name in the resulting data set. In addition to this, the parameter aggregates will also be grouped based on the parameter bts_cell_name, as bts_cell_name is a grouping parameter.

With the settings above, the resulting data set will look like this.

Once the properties of the Group By element have been defined, select OK. The Group By element changes color from red to green, which means that it is operational. Test the operation of the KPI this far by connecting it to the Output element (even if you are still going to add more elements), right-clicking on the Workbench background and selecting Run Script from the popup menu.

The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

Aggregate Functions
Aggregate functions include Minimum, Maximum, Average, Standard Deviation, Variance, Sum, Count, Mode and Median. Aggregates are simple functions used for producing minimum and maximum values, averages, sums, and counts (number of samples, values, etc.) from data sets. The same steps are taken in defining all of them. To simultaneously produce multiple different aggregates based on multiple parameter data sets, use the Group By element (for more information on Operation: Group By, see page 376). To add an aggregate element to the KPI, drag and drop the element to the Workbench. Connect the data set(s) on which the calculation is to be performed to the upper (i.e. input) sockets of the aggregate element by pressing the left mouse button down on the output socket of each data set and holding down the mouse button while dragging a connector line from the data set output socket to the input socket of the aggregate element.

QU ER IES

3 81

Right-click the aggregate element and select Properties from the popup menu. The Aggregate tab of the Properties dialog opens.

Column defines the column on which the calculation is going to be performed. In most cases, the parameter present in the list (such as bler, ec/no, etc.) is selected. The symbol * is selected only with the aggregate function Count when null values are to be included in the count. If null values are to be left out of the count, select the parameter. Group by defines how the value type defined in Column is to be grouped. For instance, when calculating an Ec/N0 average per scrambling code, the aggregate function Average would have the following settings: Column: ec/no, Group by: scrambling_code, and Weight by: time. Weight by defines the how the calculation is to be weighted, i.e. by time, distance (based on GPS coordinates), etc. Weight by is usually used only with the aggregate functions Average and Count.

Note: Because the Nemo measurement file format is time-based as opposed to samplebased (i.e. a sample is created on a timeline only when changes occur in the monitored parameters), the aggregate functions Average and Count should be weighted by time in order to obtain accurate results. Result title defines a title for the result. Without a title, the aggregate element will not be operational. Define the properties of the aggregate function and click OK. The element changes color from red to green, signifying that the element is operational. Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu.

The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

Sort Elements
The sort elements Ascending and Descending respectively change the order of values in a data set to ascending and descending based on a user-defined value (e.g. time). To add a sort element to the KPI, drag and drop the sort element (either Ascending or Descending) to the Workbench. Connect the data set to the upper (i.e. input) sockets of the Ascending element by pressing the left mouse button down on the output socket of each data set and holding down the mouse button while dragging a connector line from the data set output socket to the input socket of the Ascending element. To define a sorting value, right-click the Ascending sort element and select Properties from the popup menu.

QU ER IES

3 83

The Sort tab of the Properties dialog opens.

Select the sorting value from the drop-down menu and click OK. The element changes color from red to green, signifying that the element is operational. Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu. The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

Filters
Filter makes it possible to filter out all the values from a data set that do not fulfill the defined condition(s). For instance, a filter for Scrambling Code <= 2 would AND Scrambling Code > 0 would filter out all other scrambling codes than 1 and 2. To add a Filter element to the KPI, drag and drop the Filter element to the Workbench. Connect the data set to the upper (i.e. input) sockets of the Filter element by pressing the left mouse button down on the output socket of each data set and holding down the mouse button while dragging a connector line from the data set output socket to the input socket of the Filter element. Right-click the Filter element and select Properties from the popup menu.

The Filter tab of the Properties dialog opens.

Click the Add button to add a filter.

QU ER IES

3 85

The Condition dialog opens.

Left column defines the grid column (in the input data set) that contains the values that are compared to Value (or to another column if a column name is selected in the Right column drop-down menu instead <Value>) based on the rule defined in Operator. Operator defines the condition that determines which values are filtered out. For instance, with Scrambling Code (Column) <= 21 (Value), all the values that do not fulfill the condition of being smaller than or equal to 21 are filtered out. Right column defines the data to which the data defined in the Left column is compared. <Previous value> defines the filter condition based on a comparison between the value in the Left column and the value preceding it (also in the Left column). <Value> enables you to enter a numerical limit value in the Value field (see below). The drop-down menu displays also parameter names based on the parameter data available in the input data set. Selecting one of the parameter names defines the comparison to be performed between the data selected in the Left column and the data selected in the Right column. Value defines the limit value that the Operator compares to the Left Column value. To define the value as a variable (i.e. the value is to be defined by the user each time the KPI is run), enter the string {?<variable name>} to the Value field. Define conditions and select OK. The Properties dialog now displays the defined condition.

To add another condition, click Add.

Alternatively, if no more conditions are required, click OK to return to the Workbench view. The logic of the filter element follows that of a binary tree. Thus, one node can always have only two child nodes.

If a third child node, i.e. condition, is added, another level is automatically added to the tree.

Individual nodes can be dragged and dropped from one level or node to another.

QU ER IES

3 87

To define AND/OR relationships between the nodes, right-click on a logical operator, i.e. AND/OR, and select AND or OR from the popup menu.

You are also able to select several filters and then group them by right-clicking the Group option from the context menu. A line will appear in the left-hand corner to signify grouping. You can also ungroup filters by right-clicking Ungroup from the context menu. Once all the required conditions have been defined, click OK to return to the Workbench view. The Filter element changes color from red to green, which means that it is operational.

Test the operation of the KPI this far by connecting it to the Output element (even if you are still going to add more elements), right-clicking on the Workbench background and selecting Run Script from the popup menu. The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397. For a practical example, see Use Case 21.: Creating Complex Filters Using Multiple Conditions on page 404.

Filters: Top-N, Bottom-N, Nth Best, and Nth Worst

Top-N and Bottom-N list a defined number of best/worst entries in a predefined column of a data set, filtering out all other data entries. Nth Best and Nth Worst produce only the Nth best/Nth worst value entry from a defined column in a data set. To add a filter element to the KPI, drag and drop the element to the Workbench.

QU ER IES

3 89

Connect the data set(s) on which the calculation is to be performed to the upper (i.e. input) sockets of the filter element by pressing the left mouse button down on the output socket of each data set and holding down the mouse button while dragging a connector line from the data set output socket to the input socket of the filter element.

Right-click the filter element and select Properties from the popup menu. The Nth tab of the Properties dialog opens.

N enables you, depending on the filter, either to define the number of best/worst data entries to be listed from a column in a data set, or to specify the position on such list that is of interest to you (i.e. the Nth best/worst value of a column in a data set). Column defines the column from which the filter will, depending on the filter, produce either an N number of best or worst value entries (i.e. Top-N, Bottom-N) or the Nth best/worst value (i.e. Nth Best, Nth Worst). In most cases, the parameter present in the list (such as bler, ec/no, etc.) is selected.

Group by defines how the value type defined in Column is to be grouped. For instance, to produce the best two Ec/N0 values per each scrambling code, the Top-N filter would have the following settings: N: 2, Column: ec/no, Group by: scrambling_code. Define the properties of the filter and click OK. The element changes color from red to green, signifying that the element is operational. Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu.

The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

Mathematical Functions
Mathematical functions enable the performing of various arithmetic operations. To add a mathematical function to the KPI, drag and drop a math element to the Workbench. Connect the data set on which the calculation is to be performed to the upper (i.e. input) sockets of the math element by pressing the left mouse button down on the output socket of each data set and holding down the mouse button while dragging a connector line from the data set output socket to the input socket of the math element. Right-click the math element and select Properties from the popup menu. The Operator tab of the Properties dialog opens.

QU ER IES

3 91

Left column defines the parameter or data set on which the calculation will be performed. Right column defines the value by which the Left column value is multiplied, divided, etc. Select <Value> to manually define the value in the Right value field. Otherwise, select a value from the drop-down menu. Right value defines the value by which the Left column value is multiplied, divided, etc. when <Value> is selected in Right column. To define the value as a variable (i.e. the value is defined by the user each time the KPI is run), enter the string {?<variable name>} to the value field. Result title defines a title for the column displaying the results in the output data set. The method of defining each of the mathematical operations is discussed below. Addition (+),subtraction (-),multiplication (*),division (/), and modulation (%) (i.e. calculating the remainder of a division of one number by another) performs the selected operation on the data set defined as the Left column in the element properties. Right column should be set to <value>. The value entered into the Right value field defines the value which is added or subtracted from the Left column values, or by which the Left column values are multiplied, divided, etc. For example, the Right value defined as 3 in a division (/) element would divide each of the values in the column (defined as the Left column) by 3. Left (<<) and right (>>) bit shifting performs bit shift on the data set defined as the Left column in the element properties. Right column should be set to <value>. The value entered into the Right value field defines the number of digits shifted. For example, the Right value defined as 3 in a right (>>) bit shift element would perform in the input data set a three-digit shift to the right. The Ceiling function rounds a real number to the next larger integer based on each of the values in the column defined as the Left column in the element properties. Right column should be set to <value>. The value entered into the Right value field defines the rounding method, i.e. 1 defines rounding to the next integer, 10 defines rounding based on ten, etc. The Floor rounds a real number to the next smaller integer based on each of the values in the column defined as the Left column in the element properties. Right column should be set to <value>. The value entered into the Right value field defines the rounding method, i.e. 1 defines rounding to the next integer, 10 defines rounding based on ten, etc. The Log (logarithm) function calculates the logarithm of each of the values in the column defined as the Left column in the element properties. Right column should be set to <value>. The value entered into the Right value field defines the base for the logarithm.

Exponentiation performs exponentiation on each of the values in the column defined as the Left column in the element properties. Right column should be set to <value>. The value entered into the Right value field defines the exponent. Root performs an Nth root (square root, cube root, etc.) operation on each of the values in the column defined as the Left column in the element properties. Right column should be set to <value>. The value entered into the Right value field defines the root, e.g. 2 defines the operation as square root. The Round function performs rounding of a real number to the nearest integer based on each of the values in the column defined as the Left column in the element properties. If the fraction part of the number is <0.5, the number is rounded to the next smaller integer. If the fraction part is >=0.5, the number is rounded to the next larger integer. Right column should be set to <value>. The value entered into the Right value field defines the rounding method, i.e. 1 defines rounding to the next integer, 10 defines rounding based on ten, etc. Define the operation properties and click OK. The element changes color from red to green, signifying that the element is operational. Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu. The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

QU ER IES

3 93

Time Functions: Resample


As Nemo file format is event-based, samples with varying time ranges are logged only when the parameter value changes. This means the sample interval is not constant. Resample allows the data to be resampled at a periodic, user-definable time interval. Thus for instance 0.5-second and 1.3second RSCP samples with values -86 and -87 respectively, would constitute five samples of value 86 and 13 samples with value -87 when resampled at a 100-millisecond interval. Note: Resampling data using a sample period that is longer than the original sample period can reduce the accuracy of the data and may cause some data to be lost.

To add a Resample element to the KPI, drag and drop the element to the Workbench. Connect the data set to be resampled to the upper (i.e. input) sockets of the Resample element by pressing the left mouse button down on the output socket of the data set and holding down the mouse button while dragging a connector line from the data set output socket to the input socket of the Resample element.

Right-click the Resample element and select Properties from the popup menu. The Resample tab of the Properties dialog opens.

Interval defines the interval at which the data set is to be resampled. The drop-down menu allows you to define the unit of time as millisecond (ms) or second (s). Define resampling interval and click OK. The element changes color from red to green, signifying that the element is operational. Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu.

The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

QU ER IES

3 95

Time Functions: Time Shift


Time Shift allows you to modify the timestamp and time range of the input dataset. For example, time shift can be used to create a time range for event-type input that is logged as a point in time, such as Dropped call. Defining a time range surrounding this point in time allows you to correlate the time range surrounding the event with other parameters (e.g. ec/n0 best active set) to find out what the values of the correlated parameters were not only at the time of the event but also before and after the event. To add a Time Shift element to the KPI, drag and drop the element to the Workbench. Connect the data set(s) on which the time shift is to be performed to the upper (i.e. input) sockets of the Time Shift element by pressing the left mouse button down on the output socket of each data set and holding down the mouse button while dragging a connector line from the data set output socket to the input socket of the Time Shift element.

Right-click the Time Shift element and select Properties from the popup menu. The Time Shift tab of the Properties dialog opens.

Time offset enables you to define an offset in relation to the point in time the event occurred. Selecting backward from the drop-down menu defines the offset point before the event. Selecting forward from the drop-down menu defines the offset point after the event. Define the distance of the offset point from the event by entering an amount of time in the time field and by setting unit of time as either seconds (s) or milliseconds (ms).

Duration allows you to define the direction and length of the time range from the offset point defined in Time offset. Selecting backward from the drop-down menu defines the time range as ending to the offset point defined in Time offset. Selecting forward from the drop-down menu defines the time range as beginning from the offset point (defined in Time offset). Define the length of the time range from the offset point (defined in Time offset) by entering an amount of time in the time field and by setting unit of time as either seconds (s) or milliseconds (ms). For example, to define a time range that begins 10 seconds before the event and ends 10 seconds after the event, define Time offset as backward 10 s and Duration as forward 20 s. Define the properties of the Time Shift and click OK. The element changes color from red to green, signifying that the element is operational. Even if you are still going to add more elements to the KPI, it is recommended to test the operation of the KPI at this point by connecting it to the Output element, right-clicking on the Workbench background, and selecting Run Script from the popup menu.

The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

QU ER IES

3 97

Running and Testing KPIs


To test the operation of the KPI, make sure all elements are of green color (operational), connect the last element to the Output element, right-click on the Workbench background, and select Run Script from the popup menu. The Select Measurement dialog opens.

Folder defines the measurement folder whose contents will be displayed as a list in the field below. Filter defines a filter for searching measurement files. Type the measurement name in the field and only the defined measurements will be displayed in the list below. Select a measurement from the list and click OK. Once the script has been successfully run, there is a good opportunity for fine-tuning the KPI based on the output, for instance, by adding a sort element to the KPI to make the results more explicit, defining even more optimal states and triggers if using the state machine, or further filtering the data if the output still contains irrelevant information. You can also hide irrelevant columns in the output data set by double-clicking on the Output element and clearing the irrelevant columns on the Results tab of the Properties dialog. Ideally, the KPI should produce only the target data without any irrelevant information and in as explicit form as possible.

Saving KPIs as Components


To save the KPI as a component that can later be reused as part of more complex KPIs, right-click on the Workbench background and select Save Component from the popup menu. Note: In order for the component to work properly as part of a KPI, it must be operational and complete with Parameter and Output elements when saved.

The saved KPI component can be found from the menu structure on the right, under the menu item Components.

To add a component to a KPI, drag and drop the component to the Workbench. The Component Type dialog opens.

To open the component as a single element, select Single component and click OK.

QU ER IES

3 99

To open the component in multiple-element format that allows you to edit the properties of individual parts of the component, select Multiple nodes (will reset model) and click OK.

For editing the properties of individual KPI elements, see the topic pertaining to the relevant KPI element above.

Saving Custom KPIs


To save a KPI, right-click on the Workbench background and select Save from the popup menu. If the script is not operational, the Save option will not be displayed in the popup menu.

The Analyze Wizard Properties dialog opens.

Name defines a name for the KPI. Name is displayed in the Parameter Tree under the User item.

Title defines a title for the KPI. Title is displayed when the KPI has been dragged and dropped to the Workbench Description defines a description for the KPI. The description of a KPI can be accessed by rightclicking the KPI in the Parameter Tree and selecting Description from the popup menu. Select Next. The Analyze Wizard Column Aliases dialog opens.

In the Column Aliases dialog you can define how the retrieved data is displayed in the various data views. These settings are optional. Click Finish. The saved custom KPI can be found in the Parameter Tree under the User item.

Reopening Custom KPIs for Editing


Go to the User item in the Parameter Tree. Select View | KPI Workbench in the Ribbonbar to open the KPI Workbench. Select the KPI you want to open for editing and drag and drop it to the Workbench. The Component Type dialog opens.

To open the component in a format that allows editing, select Multiple nodes (will reset model) and click OK.

QU ER IES

4 01

For editing the properties of individual KPI elements, see the topic pertaining to the relevant KPI element above.

Defining KPI Execution Method and Value Constants


The Properties control in the bottom right corner of the screen allows you to define the KPI execution method and value constants. If the intended purpose of the new KPI calls for switching the KPI execution method, or if you wish to define value constants for use during the KPI creation process, these both should be done before beginning to define the actual KPI.

KPI execution method


Note: Unless accomplishing a particular operation specifically requires executing the query simultaneously on all measurement files, always use the KPI execution method Execute Per Measurement=True. In some special cases, it may be necessary to create a KPI that correlates data from separate files (e.g. mobile and scanner files from the same measurement session). This kind of correlation is not possible when Execute Per Measurement = true because the files are processed one at a time. To accommodate these situations, Nemo Analyze allows the KPI execution method Execute Per Measurement to be disabled. To disable the method Execute Per Measurement, select False from the Execute Per Measurement drop-down menu.

Defining value constants


To define values that can be referred to in scripts with a reference, right-click on the empty Constants field in the bottom right corner of the screen and select Add Constant from the popup menu.

QU ER IES

4 03

The Add Constant dialog opens.

Name defines the reference string which is replaced by the constant value (e.g. 10000, see screenshot below) in a script. In script, the usage of the reference string example is {$example}. Click OK. The Constants field now displays a row with the reference string example.

Click the content field (on the right of the reference string), enter the constant value (e.g. 10000), and press Enter. The constant example can now be referred to with the string {$example} when defining the properties of various KPI elements. For example, when defining a filter element, the reference string can be used instead of the actual value 10000.

Use Case 21.: Creating Complex Filters Using Multiple Conditions


Filter makes it possible to filter out all the values in a data set that do not fulfill a defined condition. For instance, a filter for Scrambling Code <= 2 would AND Scrambling Code > 0 would filter out all other scrambling codes than 1 and 2. The following example creates a filter that filters out all scrambling codes other than those belonging to the ranges 12-21, 29-30, and 74-88. As a formula, the logic can be presented in the following format: (scr. code<=21 AND scr. code > 11) OR (scr. code<=30 AND scr. code >= 29) OR (scr. code <=88 AND scr. code >= 74). To add a Filter element to the KPI, drag and drop the Filter element to the Workbench. Connect the data set to the upper (i.e. input) sockets of the Filter element by pressing the left mouse button down on the output socket of each data set and holding down the mouse button while dragging a connector line from the data set output socket to the input socket of the Filter element. Right-click the Filter element and select Properties from the popup menu.

QU ER IES

4 05

The Filter tab of the Properties dialog opens.

Click the Add button to add a filter. The Condition dialog opens.

Define the first condition, e.g. scr. code <= 21, and select OK. The Properties dialog now displays the defined condition.

To add another condition, click Add.

The Condition dialog opens.

Define the second condition, e.g. scr. code > 11, and select OK. The Properties dialog now displays the two conditions.

The first of the three value ranges, i.e. 12-21, has now been defined. The logic of the filter element follows that of a binary tree, and thus one node can always have only two child nodes. In other words, as the first already defined condition pair takes up one of the two possible nodes on the highest level, there can be only one available node left on that level. However, in order for us to be able to create a filter that can accommodate also situations where there are values from only one of the ranges, three nodes of equal importance are needed, one for each value range. We can accomplish this by adding the two remaining condition pairs, i.e. (scr. code<=30 AND scr. code >= 29) and (scr. code <=88 AND scr. code >= 74) as children to the remaining free higher-level node. In practice, the original formula is restructured in the following manner: (scr. code<=21 AND scr. code > 11) OR ((scr. code<=30 AND scr. code >= 29) OR (scr. code <=88 AND scr. code >= 74)).

QU ER IES

4 07

Next, the second higher-level node is created as a parent for the remaining two condition pairs. To create the second higher level node, select one of the two existing conditions and click Add.

Define the condition, e.g. scr. code <= 30, in the Condition dialog and click OK. The new condition is automatically added as the second higher-level node. (This is because the node selected before clicking the Add button in the previous phase already contained the maximum of two child nodes.)

As the filter should be able to accommodate also situations where there are values from only one of the ranges, the logical operator between the two higher-level nodes must be changed to OR.

To change the logical operator, right-click on the operator and select the intended operator from the popup menu. The operator is changed to OR.

Next, to complete the second condition pair, select the previously created condition, i.e. scr. code <= 30, and click Add. Define the condition, i.e. scr. code >= 29, in the Condition dialog and click OK.

QU ER IES

4 09

The second condition pair is completed.

Again, as the filter should be able to accommodate also situations where there are values from only one of the ranges, the logical operator must be changed to OR. To create the remaining condition pair, i.e. 74-88, select one of the conditions in the latest pair and click Add.

Define the condition, i.e. scr. code >= 74, in the Condition dialog and click OK.

A new level is added automatically to the binary tree.

Next, to complete the third condition pair, select the previously created condition, i.e. scr. code >= 74, and click Add. Define the condition, i.e. scr. code <=88, in the Condition dialog and click OK. The properties of the filter element are now complete.

Click OK to return to the Workbench view.

QU ER IES

4 11

The Filter element changes color from red to green, which means that it is operational.

Test the operation of the KPI this far by connecting it to the Output element (even if you are still going to add more elements), right-clicking on the Workbench background and selecting Run Script from the popup menu. The Select Measurement dialog opens. Select a measurement from the list and click OK. For more information on the Select Measurement dialog, see Running and Testing KPIs on p. 397.

Use Case22.: Creating a KPI for Dropped Calls Resulting from a Missing Handover
This use case describes the steps for creating a KPI for dropped calls resulting from a missing handover. To open the KPI Workbench, select View | KPI Workbench from the Ribbonbar.

Step 1: Planning and Adding Parameters


First you need to consider what parameters would be the best indicators of a missing handover. This sort of situations can occur for instance in downtown areas where a corner of a building may instantly block the coverage of the serving cell, giving it no time to signal the next cell of a handover. Typically, when a missing handover occurs, two parameters are affected almost simultaneously: block error ratio value rises and the Ec/N0 value of the active set falls below that of the monitored set. With the latter case, you need to consider what kind of calculation and which parameters are needed to produce this sort of value. Thus, the relevant parameters are BLER and, to be able to compare the Ec/N0 value of the active set to the absolute best Ec/N0 value of both the active and monitored sets, both Ec/N0 best active set and Ec/N0 1. best. In other words, if Ec/N0 1. best is better than Ec/N0 best active set, the handover has not occurred. The value can be produced by subtracting Ec/N0 1. best from Ec/N0 active set. In other words, if the result of the subtraction of Ec/N0 1. best from Ec/N0 best active set is smaller than zero, a better Ec/N0 value exists than that of the active set. Isolating missing handovers from the data will be necessary in order for it to be possible to correlate missing handovers with dropped calls. This is best achieved by using the state machine. Missing handover will be defined as a state and thus isolated from the data. In addition to this state, the operation of the state machine requires also an idle state to make the transitions from state to state possible. As Bad BLER values by necessity precede missing handovers, a state for Bad BLER might also be useful. Thus, the state machine will use three states: OK, Bad BLER, and Missing handover. For the transitions to occur from state to state, you have to decide what values will trigger them. In other words, what value is considered good, bad, etc. and which transition it will trigger. For the purposes of this use case, a BLER value that is greater or equal to 20 is considered bad and an Ec/N0 difference that is smaller than zero is considered to signify a missing handover. Thus, the transition from the normal state OK to the missing handover prerequisite state Bad BLER should be triggered by a bad BLER value (i.e. when BLER is greater or equal to 20). For the transition to occur from Bad BLER to the actual Missing handover, also the Ec/N0 difference will have to be smaller than zero. If however either one of the trigger values improves, the state machine should return to the relevant state. Thus three more transitions are needed, one for returning from the state Bad BLER if BLER improves, one for returning from the state Missing Handover to Bad BLER if the Ec/N0 difference value improves but the BLER value remains bad, and one to return from the state Missing handover straight to the state OK if BLER improves.

QU ER IES

4 13

It is very important that there is always a returning transition from each state in case the conditions of the transitions to the state are not fulfilled anymore. This ensures accurate results and the correct operation of the state machine. It is also recommended to always plan the operation of the state machine in the form of a flow chart, complete with trigger values. See an example below.

BLER >= 20 OK Bad BLER

Ec/N0 difference <0 Missing handover

BLER < 20

Ec/N0 difference > 0

BLER < 20

Once the missing handovers have been isolated from the data using the state machine, it will be necessary to correlate the missing handover events with the parameter Call dropped to find out if any of the dropped calls in the data occurred while a handover was missing. From the Parameter Tree, drag and drop the parameters BLER, Ec/N0 best active and Ec/N0 Nth best to the Workbench.

When you drag and drop the parameter Ec/N0 Nth best to the Workbench, the Analyze Wizard Filters dialog opens.

Make sure the Value field contains the value 1 and select Finish.

Step 2: Combining raw data sets


Next, as the mathematical function Subtraction requires the two values between which the operation takes place to be found within the same data set, to calculate the Ec/N0 difference, the two parameter data sets have to be combined into a single data set. As we are interested in two Ec/N0 values that occur simultaneously, we can use the correlation All Values Within Time Range to combine the two data sets.

QU ER IES

4 15

From the Correlations menu, drag and drop All Values Within Time Range to the Workbench and drag connector lines from the output sockets of the parameters Ec/N0 best active set and Ec/N0 Nth best to its input sockets by holding the left mouse button down.

Step 3: Adding Operations


From the Math menu, drag and drop Subtraction (-) to the Workbench and connect All Values Within Time Range to it.

Right-click on the Subtraction element and select Properties from the popup menu. The Properties dialog opens.

From the Left column drop-down menu, select ec/no. From the Right column drop-down menu, select 1. best Ec/N0. Leave the Right value field empty. Type Ec/N0 difference to the Result title field. At this point, there are two input data sets: BLER and Ec/N0 difference. To combine these data sets into a single data set, you need to select a correlation method that does not remove any data from either of the sets, namely Union.

QU ER IES

4 17

From the Correlations menu, drag and drop Union to the Workbench and connect BLER and Ec/N0 difference to it.

To test the output at this point, connect the Union element to the Output element, right-click on the Workbench background and select Run Script from the popup menu. The Select Measurement dialog opens. Select a measurement from the list and click OK. The output grid opens as a new workbook.

From the output grid, it is evident that although the two tables and their time columns have been merged, the rows are not ordered by time. As most operations require the input data to be ordered by time, you need to sort the data set before performing any further operations on it. From the Sorting menu, drag and drop the Ascending sort element to the Workbench and connect the Union element to it.

QU ER IES

4 19

Right-click on the Ascending sort element and select Properties from the drop-down menu. The Properties dialog opens.

Select time from the Column drop-down menu and click OK. The Ascending sort element is now operational.

QU ER IES

4 21

To test the output again at this point, connect the Ascending sort element to the Output element, rightclick on the Workbench background and select Run Script from the popup menu. The Select Measurement dialog opens. Select a measurement from the list and click OK.

The output grid is now ordered according to time.

From the Operations menu, drag and drop State Machine to the Workbench and connect the Ascending sort element to it.

Right-click on the State Machine element and select Properties from the drop-down menu.

QU ER IES

4 23

The Properties dialog opens.

Select Add. The State dialog opens.

To create a state called OK, type OK to the Name field and click OK. In the Properties dialog, click Add again. In the State dialog, type Bad BLER to the Name field and click OK.

Again, in the Properties dialog, click Add. In the State dialog, type Missing handover to the Name field and click OK.

All states have now been created. As the state OK serves the function of the idle state in this use case, select the state OK from the Initial state drop-down menu. This defines OK as the starting point for the operation of the state machine. Next, define a condition that has to be fulfilled in order for the state machine to move from the Initial State (i.e. the state OK) to the state Bad BLER. To define transitions from the state OK, select OK from the States field, and click Modify. An empty State dialog opens. The name of the state, i.e. OK, is displayed in the Name field.

To create a transition from the state OK to another state, click Add.

QU ER IES

4 25

The Transition dialog opens.

Next, define the target state (i.e. the state to which the transition occurs) by selecting Bad BLER from the Target drop-down menu. Because the only relevant state in terms of the KPI is Missing handover and the output should not include any data from the state OK, leave the Output field empty. The Condition dialog opens.

Next, define the condition that causes the transition to occur from the state OK to the state Bad BLER. For the purposes of this use case, it has been decided that for a BLER value to be considered bad, it has to be 20 or more. To define this as a transition condition, select bler from the Column drop-down menu, select >= (i.e. greater or equal) from the Operator drop-down menu, and type the value 20 to the Value field. When the condition has been defined, click OK to return to the Transition dialog.

The condition bler >= 20 is now displayed in the Conditions field.

As there is no need to define any more conditions for the transition from the state OK to the state Bad BLER, click OK in the Transition dialog to return to the State dialog. In the State dialog, the transition to Bad BLER is now displayed in the Transitions field.

As there is no need to define any more transitions from the state OK, click OK in the State dialog to return to the Properties dialog.

QU ER IES

4 27

To define transitions from the state Bad BLER, select Bad BLER in the Properties dialog, and click the Modify button.

Again, an empty State dialog opens. The name of the state, i.e. Bad BLER, is displayed in the Name field.

Two transitions are needed from the state Bad BLER: one that returns to the initial state OK in case the BLER value is less than 20 again, and another to move from the state Bad BLER to the state Missing handover if the Ec/N0 difference between the best active set and the absolute best set is less than 0. To create a new transition, click Add in the State dialog.

The Transition dialog opens.

Next, define the target state (i.e. the state to which the transition occurs) by selecting Missing handover from the Target drop-down menu. Because the only relevant state in terms of the KPI is Missing handover and the output should not include any data from the state Bad BLER, leave the Output field empty. Click the Add button. The Condition dialog opens.

Define the condition for the transition from the state Bad BLER to the state OK. Select bler from the Column drop-down menu, select < (i.e. smaller than) from the Operator drop-down menu, and type the value 20 to the Value field. When the condition has been defined, click OK to return to the Transition dialog. The condition bler < 20 is now displayed in the Conditions field.

Click OK to return to the State dialog.

QU ER IES

4 29

The transition from Bad BLER to OK is displayed in the Transitions field of the State dialog.

Click Add to add a transition from Bad BLER to Missing handover. The Transition dialog opens.

Next, define the target state (i.e. the state to which the transition occurs) by selecting Missing handover from the Target drop-down menu. Because the only relevant state in terms of the KPI is Missing handover and the output should not include any data from the state Bad BLER, leave the Output field empty. Click the Add button. The Condition dialog opens.

Define the condition for the transition from the state Bad BLER to the state Missing handover. Select Ec/N0 difference from the Column drop-down menu, select < (i.e. smaller than) from the Operator drop-down menu, and type the value 0 to the Value field. When the condition has been defined, click OK to return to the Transition dialog.

The Transition dialog opens with the condition Ec/N0 difference displayed in the Conditions field of the Transition dialog.

As there is no need to define any more conditions for transitions from the state Bad BLER, click OK in the Transition dialog to return to the State dialog. In the State dialog, the transitions to the states Bad BLER and Missing handover are now displayed in the Transitions field.

As there is no need to define any more transitions from the state Bad BLER, click OK in the State dialog to return to the Properties dialog.

QU ER IES

4 31

To define transitions from the state Missing handover, select Missing handover in the Properties dialog and click the Modify button.

The State dialog opens.

Two transitions are needed from the state Missing handover: one that returns to the initial state OK in case the BLER value is less than 20 again, and another to move from the state Missing handover to the state Bad BLER if the Ec/N0 difference between the best active set and the absolute best set is greater or equal to 0. To create a new transition, click Add.

The Transition dialog opens.

Define the target state (i.e. the state to which the transition occurs) by selecting OK from the Target drop-down menu. As the output should include the data from the state Missing handover, enter the name Missing handover to the Output field. Once the target state has been defined, click Add to add a condition for the transition. The Condition dialog opens.

Define the transition from the state Missing handover to the state OK. Select bler from the Column drop-down menu, select < (i.e. smaller than) from the Operator drop-down menu, and type the value 20 to the Value field. Once the condition has been defined, click OK to return to the State dialog.

QU ER IES

4 33

The condition bler < 20 is now displayed in the Conditions field of the Transition dialog.

When the transition has been defined, click OK to return to the State dialog. The transition to OK is now displayed in the Transitions field of the State dialog.

As one more transition is needed from the state Missing handover, click the Add button in the State dialog. The Transition dialog opens.

Define the target state (i.e. the state to which the transition occurs) by selecting Bad BLER from the Target drop-down menu. As the output should include the data from the state Missing handover, enter the name Missing handover to the Output field. Next, click Add to define the condition for the transition from the state Missing handover to the state Bad BLER. The Condition dialog opens.

Select Ec/N0 difference from the Column drop-down menu, select >= (i.e. greater or equal) from the Operator drop-down menu, and enter the value 0 to the Value field. When the condition has been defined, click OK to return to the Transition dialog.

Click OK to return to the State dialog.

QU ER IES

4 35

The transition to Bad BLER is now displayed in the Transitions field of the State dialog.

Click the OK button to return to the Properties dialog. Click the OK button in the Properties dialog. The State Machine element is now operational. The states where the BLER value rises and a handover is missing have now been isolated. To find the events where missing handovers lead to dropped calls, we need to add the parameter Call dropped to the KPI. Drag and drop the parameter Call dropped from the Parameter Tree to the KPI Workbench.

As the only relevant dropped calls in terms of the KPI are the ones that occur during the state Missing handover, the most optimal correlation for combining these two data sets is All Values Within Time Range. Drag and drop an All Values Within Time Range element from the Correlations menu to the KPI Workbench. For the output data set to include all Missing handover time ranges and only those dropped call events that occur during these time ranges, the state machine has to be the primary data set and the parameter Call dropped the secondary one. Connect the state machine to the leftmost input socket of the All Values Within Time Range element. Connect the parameter Call dropped to the next socket. Finally, connect the All Values Within Time Range element to the Output element.

Step 4: Running and Testing the KPI


To run the KPI, right-click on the Workbench background and select Run Script from the popup menu. The Select Measurement dialog opens. Select the measurement file you want to run the KPI on and click OK.

QU ER IES

4 37

The final output includes only the rows with Missing handover events from the measurement data, and if there are Call dropped events within the time range of the Missing handover events, these will be displayed as well.

To save the custom KPI for later use, right-click on the Workbench background and select Save. The Analyze Wizard Properties dialog opens.

Enter Dropped calls resulting from missing handover to the Name and Title fields and click Next. The Analyze Wizard - Column Aliases dialog opens. Click Finish. The KPI can now be found in the in the Parameter Tree under the User item.

ANALYTICS
Analytics feature is part of the Nemo Analyze Professional feature package. Analytics consists of new summary tables in the Analyze database for voice/video call, data, MMS, SMS, and handover events. The information in the Analytics tables is essentially the same as in the standard database tables. The key difference is that Analytics tables do not have the physical per measurement file federation as the standard tables. This makes Analytics queries to execute significantly faster over large (>100 log files) datasets when comparing to corresponding standard queries. Analytics tables are populated with queries executed over the standard tables. Thus the process has two steps, first the log file(s) are loaded into the Nemo Analyze database as before. Then, AnalyzeAnalytics.exe, a process running on the background, will automatically detect new files added to the database, and run a set of queries over the files and save the query results into the Analytics tables. Analytics can be enabled from the View | Options | Database | Queries | Use Analytics to preprocess measurement data. Analytics tables are visible for the end user via new set of queries in the parameter tree, under Analytics branch. The queries can be used to plot events on the map, calculated statics, etc. in a similar manner as standard queries. Analytics queries can be also used in custom workbooks and report templates. Note that the file loaded to the database cannot be queried with Analytics queries immediately after it appear in the database. There is an additional delay to process the files, that is, to run the queries to populate Analytics tables. The extra delay is approximately 4min per 1 hour of raw log file data. When using Nemo Analyze Enterprise Analytics starts automatically in the database server.

QU ER IES

4 39

(S)FTP Log File Autoload


Nemo Analyze enables loading log files and autonomous rep files to the database automatically over passive FTP or SFTP connection. Active FTP mode is not supported. The feature is mainly targeted for Nemo Analyze Enterprise, but it can be enabled also in Nemo Analyze Standalone. Loading of the files over FTP is two-step process. First the files are retrieved from the FTP server and saved to a local folder. Then the files are loaded from the local folder into the database.

Please note that the measurements files are removed from the FTP servers autoload folder after loading.
Following registry keys have to be defined to enable the FTP autoload: - db.autoload.directory (STRING). Local folder path where files received from FTP are saved and from where AnalyzeSocketServer automatically loads them to database db.autoload.extensions (STRING). Comma-separated list of file extensions to be loaded. All Nemo logfile extensions are supported: nmf, nbl, rep, gz,dt1,dt2,fs1,fs2, db.autoload.ftp (STRING) where the format is same as with using URL to logon to an ftp server. For example sftp://username:password@server.com:22/files. Login credentials can be empty as well as port and remote path in the end. E.g. ftp://server.com Optional registry key values: - db.autoload.ftp.recursive (DWORD) 0 or 1 (default 0). Defines whether files under subfolders on the FTP server are loaded or not db.autoload.ftp.interval (DWORD) 0 - INT_MAX (default 360) FTP file checking interval in seconds.

Registry keys have to be set under HKEY_LOCAL_MACHINE\SOFTWARE\Anite\Nemo Analyze\ServerSettings in 32bit Windows and Nemo Analyze Enterprise server. In 64 bit Windows running Nemo Analyze standalone the registry keys have to be set under HKEY_LOCAL_MACHINE\SOFTWARE\Wow6432Node\Anite\Nemo Analyze\ServerSettings

QU ER IES

4 41

Registry settings needed for automatic file upload to database over FTP.

OTHER TASKS
EDITING COLOR SETS
Color sets are used to color the measurement route and BTS icons on a map based on parameter values. This makes it easier to spot problem areas on a map. Color sets can also be used in grid and numerical data views to highlight certain data. To edit and create color sets, select View | Tools | Color set editor.

In the Color Set Editor dialog you will see a list of all existing color sets. You can sort the color sets by selecting the Type and Group. You can also search for a color set by its name. To edit an existing color set, double-click on the color set.

O TH ER T ASKS

4 43

In the Color Set Properties dialog you can edit the settings for each color set. Click Modify.

Name and Short Name identify the color set. Select from the Groups list the items with which the color set will be associated. Type defines what kind of data the color set is used with, numerical, gradient, or string. The Groups and Type settings can be used to sort the color sets in the Color Set Editor dialog. Values table displays the different value ranges and the colors associated with them.

Importing Color Sets


It is possible to import color sets from Nemo Analyze 4.0 and Nemo Outdoor. Select File | Import | Old Color Sets. In the Open dialog, browse the color set file (.csf) you want to import and click Open. In the Import Color Sets dialog, select the color sets that you would like to import and finally click OK.

The imported color sets are added in the Color Set Editor.

Automatic Generation of Color Sets


With Nemo Analyze it is also possible to generate color sets automatically from parameter values that appear in the measurement file. First go to the Maps page in the Workspace by clicking the icon, and open a blank map.

On the Measurements page in the Workspace, click on a measurement file. From the parameter tree, select a parameter that you want to plot on the map, for example, RX level full, and drag it on the map. A route is drawn on the map and it is colored by default by using color rotation.

O TH ER T ASKS

4 45

Right-click on the map and select Generate Color Set from the popup menu.

In the Color Set Wizard, select the parameter that you dragged on the map, e.g., RX level full. In the Column field, select the parameter again. Click Next.

Type a name and a short name for the new color set. Also define in which group the color set will be added. Finally, click Finish. The new color set is generated.

O TH ER T ASKS

4 47

To apply the new color set on the route, right-click on the route and select Properties. In the Color page, select Based on value as Mode and then the color set you created (e.g., RX level full 2(dBm)) and the corresponding parameter in the Scheme field (e.g., RX level full). Click OK.

The color set is applied to the route. The color legend on the right displays the colors and the corresponding values.

Use Case 23.: Automatic Generation of Color Set for a Value Range
This use case explains how to generate a color set automatically for a value range. This functionality is particularly useful when creating color sets for, e.g. scrambling code, cell id, and BSIC. First open a blank map from the Maps page in the Workspace by double-clicking on a map in the Loaded view. It is assumed that the map has been loaded to the Workspace from the Map Folder.

O TH ER T ASKS

4 49

Next, select a measurement file from the Workspace. From the Parameter Tree, select a parameter that you want to plot on the map, for example, Cell identification, and drag it on the map.

A route is drawn on the map. The route is colored by default by using color rotation. If the route appears as symbols, right-click on the route, select Properties, and on the Route page select Line in the Draw Mode field. Click OK.

Next, right- click on the Color Legend in the side panel and click on Edit Selected. In the Color Set Properties dialogue click Add Range.

O TH ER T ASKS

4 51

In the Add Range dialogue, you can define a value range for a new color set, and the interval between the values. Here the value range is from 116731 to 117419 with an interval of 1. Click OK.

The new value range for the color set has been created. Now the various parameter values show more distinctly on the map due to the user-defined value range.

Use Case 24.: Creating a Color Set


This use case describes the steps for creating a new color set. Select View | Tools | Color set editor. In the Color Set Editor, click the Add button.

An empty Color Set Properties dialog opens up. Define first the settings at the top of the dialog. Next start defining the value ranges and colors. Click the Add button.

O TH ER T ASKS

4 53

In the Range Properties dialog, select the color and then the upper and lower limits of the value range. You can also write a description for the value range. Finally click OK.

Repeat the step until you have a complete set of values. Click OK and the new color set is added to the Color Set Editor dialog.

Use Case 25.: Creating and Applying a Color Set on Map


This use case describes the steps for creating and applying a color set on a map. Select View | Tools | Color set editor. In the Color Set Editor, click the Add button. An empty Color Set Properties dialog opens up. Type a name for the color set, and select in which group it is added. Select gradient as the type. Click the Add button.

Next define the string that will be colored. In the Value Properties dialog, select a color and then the minimum value for the parameter. You can also write a description for the value. Finally click OK. Repeat the step for the maximum value.

O TH ER T ASKS

4 55

Note that you should always define the values from lowest to highest. Otherwise the color set will not work properly.

The defined values and colors are displayed in the Color Set Properties dialog.

Click OK and the new color set is added to the Color Set Editor dialog. Click Close

Open a blank map, e.g. from the Maps page in the Workspace, and on the Measurements page of the Workspace, select a measurement in the workspace. Type in Ec/N0 best active set in the Parameter tree filter field, and drag the parameter on the map. The measurement route appears on the map if the file contains location data.

Right-click on the route and select Properties.

O TH ER T ASKS

4 57

In the Properties dialog, go to the Color page and select Based on value as the Mode. In the Scheme field, select Ec/N0 best active set and in the Color set field, select the EcNo gradient color set that was created earlier. Click OK.

The route will be colored based on the set of rules defined in the new color set.

Use Case 26.: Creating and Applying a Color Set in Grid


This use case describes the steps for creating and applying a color set in a grid. Select View | Tools | Color set editor. In the Color Set Editor, click the Add button. An empty Color Set Properties dialog opens up. Type a name for the color set, and select in which group it is added. Make sure that the name of the new color set is not already in use. Select string as the Type. Click the Add button.

Next define the string that will be colored. In the String Properties dialog, type in the string of text that will be colored and select the color used. You can also write a description for the string.

Click OK.

O TH ER T ASKS

4 59

The new string is displayed in the Values field of the Color Set Properties dialog.

Click OK and the new color set is added to the Color Set Editor dialog.

Open a Layer3 signaling grid. Select a device in the Workspace, and type L3 in the Parameter tree filter field. Right-click the L3 signaling parameter, and select Open in | Grid. A workbook containing L3 signaling data in a grid is opened.

O TH ER T ASKS

4 61

Right-click on the grid, and select Properties. From the Color Sets tab of the Properties dialog, select the column Message Name and from the drop-down menu the previously created color set and click OK.

The signaling messages defined in the color set are colored.

SCHEDULING EVENTS
The Event Scheduler makes it possible to schedule workbooks and reports to be run and measurement files to be loaded to the database automatically from a predefined folder. To open Event Scheduler, select View | Event Scheduler from the Ribbonbar.

The Scheduler view opens.

O TH ER T ASKS

4 63

To add an event, select a timeslot from the calendar, right-click over it, and select Add Event from the popup menu.

The Schedule Event Batch dialog opens.

One event batch can contain several events that are being run successively. Batch name defines a name for the batch. Start time defines a start time for the batch. To define a recurring batch, click the Recurrence button. For more information on recurring events, see Defining Recurring Events on page 473.

Add enables you to add events through the Schedule Event dialog.

Event name defines a name for the event. Event type defines the event as a report event, a workbook event, a load folder event, or as a KPI threshold alarm event. To create and define an event of the selected type, click the Configure button.

Delete enables you to delete an event. Up enables you to move an event up in the batch. Down enables you to move an event down in the batch.

Report Events
Selecting Run Report as the Event Type in the Schedule Event dialog opens the Configure Report dialog.

O TH ER T ASKS

4 65

Measurement folder defines the folder on which the report will be run. Update folder before report execution, if selected, checks whether there are new measurement files in the folder before running the report. Report per measurement, if selected, runs the report on each measurement file separately. Report file defines the report that is to be run on the measurement files within the folder defined in Measurement folder. To browse for a report file, click the browse button, select a report file and click OK. Export, if selected, enables automatic export of scheduled reports to PDF, MS Excel, MS Word, Rich text, text, and HTML. Export folder defines a destination folder for the exported report. Autogenerate name generates a filename for the exported file automatically. Additional Report Scope Settings opens the Report Scope dialog. Close workbook when report ready, if selected, automatically closes the workbook after export. Send email when report ready defines an automatic notification email to be sent upon scheduled report execution. Attach report, if selected, automatically attaches scheduled reports to event reporting emails. It is recommended that the attached reports are zipped first to save bandwidth. Configure email allows you to define notification email settings. Clicking the Configure email button in the Configure Report dialog opens the Email dialog.

To defines the recipient email address. Subject displays the default subject line of the notification email. Message allows you to edit the automatic notification email.

Clicking the Additional Report Scope Settings button in the Configure Report dialog opens the System tab of the Report Scope dialog.

System defines the systems that will be included in the report. Band defines the bandwidths that will be included in the report. To define the scope of the report based on time or area, select the Time & Area tab.

O TH ER T ASKS

4 67

Begin time defines the start time for the time range from which the measurement files will be processed. End time defines the end time for the time range from which the measurement files will be processed. Area, if selected, includes enables the limiting of the report scope based on area. Define Area opens the Select Polygon Area dialog. Clicking the Define Area button on the Time & Area tab of the Report Scope dialog opens the Select Polygon Area dialog.

Select a map in the Filename field, or load one through the browse button. Next, define an area on the map using the left mouse button. The Save As button enables you to save the map with the selected area for later use. Click OK.

Workbook Events
Selecting Run Workbook as the Event Type in the Schedule Event dialog opens the Configure Workbook dialog.

Measurement folder defines the folder on which the report will be run. Update folder before report execution, if selected, checks whether there are new measurement files in the folder before running the report. Report per measurement, if selected, generates a workbook based on each measurement file separately. Workbook defines the workbook that is to be run on the measurement files within the folder defined in Measurement folder. To browse for a workbook file, click the browse button, select a workbook file and click OK. Export to defines whether the workbook is to be exported as image files (image per page) or as a PDF document. Export folder defines a destination folder for the exported workbook. Autogenerate name, if selected, generates a filename for the exported file automatically. Export name defines the name of the exported report. Close workbook when report ready, if selected, automatically closes the workbook after export. Send email when workbook ready defines an automatic notification email to be sent upon scheduled workbook execution.

O TH ER T ASKS

4 69

Configure email allows you to define notification email settings. Clicking the Configure email button in the Configure Workbook dialog opens the Email dialog.

To defines the recipient email address. Subject displays the default subject line of the automatic notification email. Message allows you to edit the automatic notification email.

Load Folder Events


Selecting Load Folder Event as the Event Type in the Schedule Event dialog opens the Configure Folder dialog.

Load folder defines the folder from which all measurement, map, BTS, report, etc. files are loaded. Include subfolders, if selected, causes all measurement, map, BTS, report, etc. files to be loaded from all subdirectories of the folder defined in Load folder.

KPI Threshold Alarm Events


Selecting KPI Threshold Alarm as the Event Type in the Schedule Event dialog opens the Configure KPI Threshold Alarm dialog.

Measurement folder defines the measurement folder on which the KPI is run. Select KPI button allows you to define the KPI based on which the threshold condition is defined. Click the Select KPI button to open the Analyze Wizard Measurement Parameters dialog (see below). KPI threshold controls allow you to define the threshold condition. The leftmost drop-down menu allows you to define the column from the KPI dataset that was selected as the basis for the condition with the Select KPI control. The next drop-down menu defines the logical operator of the condition, e.g. larger or equal (>=), smaller than (<), etc. The rightmost field allows you to enter the threshold value to which the values in the defined KPI column are compared based on the logical operator. Run script, when selected, allows you to set any script (.bat) to be run when the threshold condition is fulfilled. Send email, when selected, allows you to configure an email notification to be sent when the threshold condition is fulfilled. Clicking the Add below the KPI thresholds box in the Configure KPI Threshold Alarm dialog opens the Analyze Wizard Measurement Parameters dialog.

O TH ER T ASKS

4 71

Select a parameter from the parameter tree and click Next. The Analyze Wizard Filters dialog opens.

The dialog allows you to define optional filters by clicking the Add button. It is also possible to set multiple KPI triggers with AND/OR elements. After defining optional filters, click Finish.

Clicking the Configure email button in the Configure KPI Threshold Alarm dialog opens the Email dialog.

To defines the recipient email address. Subject displays the default subject line of the automatic notification email. Message allows you to edit the automatic notification email.

O TH ER T ASKS

4 73

Defining Recurring Events


Clicking the Recurrence button in the Schedule Event Batch dialog opens the Event Recurrence dialog.

Event time allows you to define a Start and End time (i.e. duration) for the event. Recurrence pattern allows you to define how often the event recurs, i.e. Daily, Weekly, Monthly, or every N week(s) on [weekday(s)]. Range of recurrence allows you to define duration for the recurrence pattern, either based on Start and End by dates, or based on a user-defined number of occurrences using the End after N occurrences control.

TASK MANAGER AND PROBLEM REPOSITORY


The Task Manager and the Problem Repository allow the users of Nemo Analyze Enterprise Edition to store reports on detected network problems into the server database. A problem report may contain screenshots and free-form description of the problem, as well as references to the log file(s) displaying the problem. Other users of the client/server system can view the problem report and drill down to the relevant log files to analyze the problem. To add a problem to the repository as an open task, right-click in the data view displaying the problem at problem spot (i.e. the point in time where the problem is apparent) and select Add task from the popup menu. The Task Configuration dialog of the Task Manager opens (see Configuring and Editing a Task on page 475).

To access the Task Manager without adding a task (e.g. to manage, edit, and delete existing tasks), move your mouse over the Task Manager tab on the right side of the screen to open the Task Manager view. If the Task Manager tab is not visible, enable the Task Manager by selecting View | Panels | Task Manager. The Task Manager view displays a control that allows you to open four different task views: tasks assigned to you (My tasks), all existing tasks (All tasks), open tasks assigned to you (My open tasks), and closed tasks that have been assigned to you (My closed tasks). Select and open a view by double-clicking on the view type (see Managing Tasks on page 485).

O TH ER T ASKS

4 75

Configuring and Editing a Task


When a new task is created, or an existing one is opened for editing, the Task Configuration dialog opens.

Title defines the task name. Assigned to allows the task to be assigned to a specific user. Status defines the status of the task as Open or Closed, or as any user-defined status entered into the field. Priority defines the priority of the task as 1 (high), 2 or 3 (low). Due date defines a due date for finishing the task. Created by displays the name of the user who created the task. Task id displays the automatically created unique task identification number. Category allows the task to be assigned with a user-defined category (e.g. classification according to the problem type). Timestamp displays the point in time when the problem occurred. Longitude displays the geographical location of the problems occurrence in longitude. Latitude displays the geographical location of the problems occurrence in latitude. Cell name allows the cell that was active at the time of the problem to be selected.

Area allows the user to enter the placename or a textual description of the area where the problem occurred. Details allows the user to enter a textual description of the problem.

The Attachments tab of the Task Configuration dialog allows you to add attachments to the task.

Add allows you to add an attachment to the task. Delete allows you to delete existing attachments. Save allows you to save the selected attachment to the hard drive. Execute allows you to open or run the selected attachment.

O TH ER T ASKS

4 77

The Location tab of the Task Configuration dialog opens

The Location tab displays the reported location of the problems occurrence on map based on the Longitude and Latitude defined on the Problem Details tab. The map can be selected using the drop-down menu immediately above the map.

The Preview tab displays a screen capture of the data view when the task was created.

O TH ER T ASKS

4 79

MANAGING USERS AND USER GROUPS


The Database Manager enables you to manage users and user groups (Nemo Analyze Enterprise Edition). To open Database Manager, select View | Database Manager from the Ribbonbar.

To create a new user, select the folder Users in Database Management | Users and Groups from the directory tree, right-click on the empty background of the Name field on the right of the directory tree, and select New User from the popup menu.

The New User dialog opens.

Name defines the user name. Password defines the password for the user account.

To set a new password for a user, right-click on the user in the Name field, and select Set Password from the popup menu.

The Set Password dialog opens.

Type a new password in the Password field and click OK. To delete a user, right-click on the user in the Name field, and select Delete from the popup menu.

O TH ER T ASKS

4 81

User Groups
The available user groups include Administrators, Users, Power Users, Autonomous Users, and Autonomous Managers. Administrators have unrestricted rights, including access to the cleanup function. Users do not have delete rights and are only allowed to make queries and load files into the database. Users are allowed to view all data and alter minor elements of a file, e.g. configure notifications. Power Users have delete rights, i.e. they are allowed to delete measurement and BTS files. Otherwise the rights of the Power Users are identical to those of the Users. Autonomous users have right to view all history and measurements of Wanderer units loaded into the system, Autonomous users do not have right to issue configurations, that is, to order Wanderers to do measurements. Automomous managers have full rights to all Wanderer units loaded to the system, including right to view history and logfiles produced by the units, as well as managing the measurements to be done. Wanderer unit groups created from the Fleet Manager are also shown in the User groups. Dallas units in the picture below is an example of such a group. Users added as members of that group get right to tissue configurations (order measurements) to the Wanderer units allocated to that group. As a prerequisite, such users have to be also members of Autonomous users group. This means such users will be able to see all the Wanderer units in the system, but can only issue configurations for selected subset of the Wanderers as defined by the Wanderer unit group.

To add users to a group, select the folder Groups in Database Management | Users and Groups from the directory tree, right-click on a user group on the right of the directory tree (e.g. Power Users), and select Add to Group from the popup menu.

The Group Properties dialog opens.

To add a user to the group, click Add. The Select Users dialog opens.

O TH ER T ASKS

4 83

Select the user(s) you want to add. To select multiple users, hold the Control key down while selecting the users. Click OK. The Group Properties dialog opens, displaying the group members.

To remove users from the group, select a user and click the Remove button.

VIEWING SERVER STATISTICS


To access server statistics (Nemo Analyze Enterprise Edition), select View | Tools | Database Manager from the Ribbonbar.

The Database Manager view opens. Select Server Administration | Server Statistics.

Server statistics are displayed. To refresh the statistics, right-click on the statistics view and select Refresh from the popup menu. To enable automatic statistics updates (updated every 5 seconds), right-click on the statistics view and select Poll from the popup menu.

O TH ER T ASKS

4 85

Managing Tasks
To access the Task Manager (e.g. to manage, edit, and delete existing tasks), move your mouse over the Task Manager tab on the right side of the screen to open the Task Manager view. If the Task Manager tab is not visible, enable the Task Manager by selecting View | Panels | Task Manager.

The Task Manager view displays a control that allows you to open four different task views: tasks assigned to you (My tasks), all existing tasks (All tasks), open tasks assigned to you (My open tasks), and closed tasks that have been assigned to you (My closed tasks). Select and open a view by double-clicking on the view type.

The selected task view type (e.g. My open tasks) opens with all tasks of the selected type listed.

The map view below the task list displays all tasks of the selected type, e.g. All tasks, My open tasks, etc. The displayed map may be selected using the drop-down menu immediately above the map.

Right-clicking on the task list opens a popup menu.

Configure Task allows you to edit the selected task. Add Task allows you to create a new task. Delete Task allows you to delete the selected task. Refresh allows you to refresh the task list.

O TH ER T ASKS

4 87

EDITING OPERATOR SETTINGS


The Database Manager enables you to manage operator settings. These settings are used as the basis of operations such as statistics per operator. To open Database Manager, select View | Database Manager from the Ribbonbar.

The Database Manager view opens. Select Operators.

The Operator list is displayed. To edit the properties of an operator, right-click on the operator, and select Modify from the popup menu.

The Operator Configuration dialog opens.

Operator name defines the operator name. Mobile country code defines the operators MCC. Mobile network code defines the operators MNC. To add a new operator to the list, right-click on the Operators list, and select Add from the popup menu.

The Operator Configuration dialog opens.

Operator name defines the operator name. Mobile country code defines the operators MCC. Mobile network code defines the operators MNC.

O TH ER T ASKS

4 89

EDITING CDMA OPERATOR SETTINGS


The Database Manager enables you to manage CDMA operator settings. These settings are used as the basis of operations such as statistics per operator. To open Database Manager, select View | Database Manager from the Ribbonbar.

The Database Manager view opens. Select CDMA Operators.

The CDMA Operator list is displayed. To edit the properties of an operator, right-click on the operator, and select Modify from the popup menu.

The Operator Configuration dialog opens.

Operator name defines the operator name. Sid defines the system identification number. To add a new operator to the list, right-click on the CDMA Operators list, and select Add from the popup menu.

The Operator Configuration dialog opens.

Operator name defines the operator name. Sid defines the system identification number.

O TH ER T ASKS

4 91

CONFIGURING NOTIFICATION ICONS


Nemo Analyze allows you to configure notification icons for all measurement events. These icons can be displayed in graphs and maps to alert you when something interesting happens. See viewing notification icons in maps and viewing notification icons in graphs. The notification icons can be configured using the Parameter Tree. Alternatively, notifications can also be configured using the Notification Configuration functionality.

Configuring Notifications Using the Parameter Tree


The parameter-specific notification icons are displayed in connection with the relevant parameter (e.g. with the parameter Call connected) in the Parameter Tree. The parameters that have not been assigned with a notification icon are indicated by the icon. To change the default icon of a parameter, right-click on the parameter and select Change Defaults from the popup menu.

The Change Defaults dialog opens.

To select a new notification icon, click the browse () button. The Open dialog opens. Select the new notification icon and click the Open button. The Image field of the Change Defaults dialog now displays the directory path of the selected icon.

Click OK to set the selected icon as the default notification icon for the parameter. The icon is displayed in the Parameter Tree in connection with the parameter.

O TH ER T ASKS

4 93

Notification Configuration
You can access the Notification Configuration tool by selecting View | Tools | Notification Configuration.

In the Notification Configuration dialog you can define icons for all measurement events. These icons can be displayed on graphs and maps to alert you when something interesting happens. See viewing notification icons on maps and viewing notification icons on graphs. There are a number of default notifications for you to choose from but you can also modify the default notifications or make new ones. Select a notification from the list and click the Modify button.

In the Notification Properties dialog, browse a new bitmap in the Icon field and click OK.

EDITING MENUS
With the Menu Editor you can add workbooks to the Workspace menus. You can open the Menu Editor by selecting View | Tools | Edit | Menu editor, or by right-clicking on a folder in the Workspace and selecting Customize Menu.

You can add workbooks to three different menus in the Workspace: Device, Folder, and Measurement menus.

O TH ER T ASKS

4 95

The Device menu is the popup menu that opens when you rightclick on a device. It contains items, such as, Correlate Parameters and Network Parameters.

The Folder menu is the popup menu that opens when you rightclick on a file folder under the Measurements folder.

O TH ER T ASKS

4 97

The Measurement menu is the popup menu that opens when you right-click on a measurement file. It contains items, such as, Timeline and Adjust Coordinates.

Select the menu where you want to add a workbook and click Add. In the Workbook Command Properties dialog select the workbook that will be added to the menu and define in which group it will be added. Finally click OK. Workbook defines the name for the workbook. Groups defines in which group the query will be added.

The workbook will appear in the Menu Editor and in the selected menu.

Rename workbooks by right-clicking on a workbook in the Menu Editor and selecting Rename workbook.

O TH ER T ASKS

4 99

EDITING QUERY AND WORKBOOK GROUPS


With the Group Editor you can edit the groups in which queries and workbooks are organized. These groups are visible, for example, in the workspace when you right-click on a device. You can open the Group Editor by selecting Tools | Group editor.

On the left-hand side in the Group Editor dialog you see a list of groups. These are the group labels visible in the user interface. For each group you can select the cellular systems for which the particular group is available. For example, if you have a file measured in the TETRA network you do not necessarily need to see the UMTS or CDMA groups.

In the Items belonging to group table you see a list of color sets and workbooks that are available for that group. You can select and clear each item. If you add workbooks in menus (see chapter Edit on page 494), they will appear here as well.

O TH ER T ASKS

5 01

CONFIGURING SYSTEM SETTINGS


In the Options dialog you can make some general system settings. Select View | Tools | Options to open the dialog.

Options Environment

Hide advanced UI features hides and displays advanced features that are not needed in basic use. Hide advanced SQL functionality hides and displays advanced SQL features that are not needed in basic use. When the Maintain workspace state when application is closed and restarted option is selected, also items in the workspace that are not in the database, i.e., maps and reports, are loaded when you start Nemo Analyze. When the Clear query clipboard when application is closed option is selected, the Query Clipboard is emptied when you exit Nemo Analyze. When a number of minutes is entered into the Auto-save default workspace every X minutes field, all open workbooks are auto-saved as the default workspace on regular intervals. This prevents loss of work, e.g. in the event of a system crash, as the latest auto-saved workbooks are always automatically reopened in the program startup. To prevent the default workspace from opening during startup, hold down the Shift key.

JPEG compression quality defines the quality of images exported from Nemo Analyze. You can save workbooks as images by selecting Edit | Send as Image to. Image to clipboard fixed size enables you to define a fixed size for pictures placed on the Clipboard.

Default paths defines the default locations of different types of files. Nemo Analyze will automatically look for the defined file types in these folders.

O TH ER T ASKS

5 03

Throughput defines the unit of measure for throughputs Base station identification code defines the number format for base station identification codes. Cell identification defines in which mode the parameters are presented. Distance defines the unit of measure for distance. Velocity defines the unit of measure for velocity. Hide date from timestamps option hides and displays the date in timestamps. Limit number of decimal digits in UI to defines the maximum number of decimal digits displayed.

Statistics enables you to specify a default basis for statistics calculations (i.e. whether the calculations are based on Time, Distance, or Sample), whether Nemo Analyze should ask you to select the basis each time a calculation is performed and whether the dB-parameter statistics are calculated using linear math.

O TH ER T ASKS

5 05

Header title is displayed at the top of the page when you print views from Nemo Analyze.

The Palette is used, for example, to determine colors for neighbor-style graphs and routes.

Server defines the SMTP server IP address. Port defines the SMTP port. Username defines the SMTP server username. Password defines the SMTP server password. From defines the sender email address. To defines the recipient email address for notification emails sent by the event scheduler.

O TH ER T ASKS

5 07

Options Database

The Server control enables you to switch between different database configurations and databases (e.g. between the standalone database and the database server). It also enables you to Add, Delete, and Modify configurations. Analyze-Local is the default standalone database of Nemo Analyze. Modify opens the Add Server dialog with existing database settings, enabling you to modify an existing database configuration. Add opens the Add Server dialog, enabling you to define a new database configuration. The Additional data sources control enables you to configure connections to other ODBC databases on your computer. It enables you to Add, Delete, and Modify configurations.

Title defines the title of the server. Server host defines the servers IP address. Server name defines the server name. ODBC port defines the ODBC port. Server port defines the server port. By default, Nemo Analyze uses the server port 12001. User defines the username. Password defines the password.

Name selects the data source. User defines the username. Password defines the password.

O TH ER T ASKS

5 09

Auto-hide columns from queries option hides query data that is not relevant. When the Enable query memory usage warning option is selected, Nemo Analyze will warn you when the query memory usage exceeds the Threshold value limit. When selected, Allow queries to be run in parallel allows you to enable queries to be run in parallel and select the maximum amount of parallel queries. Use Analytics to preprocess measurement data, when selected, enables Nemo Analyze to preprocess the measurement data, allowing instant event queries from large datasets.

Options Report

When the Open HTTP links in external browser option is selected, web pages are opened in a browser window. Otherwise they are opened in Nemo Analyze.

Options Graph

When the Smooth scrolling when synchronizing option is selected, the graph view will scroll smoothly instead of displaying a part of the measurement and then jumping forward as the synchronization marker moves out of view. Palette defines the color palette style for graphs. Sort bar graph by parameter enables you to group bars in bar graphs by parameter rather than by x axis values. For more information, see Group Values.

O TH ER T ASKS

5 11

Options Map

Default map type defines the map that is opened by default when no other map file is available. Default map defines the map that is opened by default. When the Automatically add default map to empty map control option is selected, the default map is opened automatically if you open an empty map view. When the Use default map in workbooks option is selected, all workbooks use the map defined in Default map instead of their own default maps. When the Automatically zoom to first added layer option is selected, Nemo Analyze will zoom in on the map layer that is added first. Print ratio defined the print ratio in percentage.

When the Automatically offset simultaneous measurement routes on the map option is selected, an offset is automatically added to measurement routes that overlap each other, as is the case, for example, with multi measurements. You can define the default offset in x and y directions. When the Automatically add default BTS for measurement route option is selected, Nemo Analyze will check from the file header what BTS file was used when the file was recorded and open the same file if it is loaded in the database. Default route thickness enables you to define the default thicknesses of high band and low band routes.

O TH ER T ASKS

5 13

The Color Legend dialog enables you to set statistics to be calculated based on time, distance, or samples.

Default map provider defines which map provider is used by default (Google Maps/OpenStreetMap) Default location defines the default location when using the live map functionality.

Options BTS

BTS texts options allows you to define the size and style of BTS text. Display BTS overlay on top of other layers sets BTS layers to be displayed over all other layers. Default BTS filter and Carrier allows you to define the default BTS filter and carrier.

Icon size defines the size of the BTS icon in pixels. Fixed defines the fixed size of BTS icons in pixels. Dynamic, max size defines a maximum size for dynamic BTS icons (i.e. icons which change size based on zoom level). Show site names option hides and displays the site names. Draw frame hides and displays the outlines of BTS icons. Site transparency defines the transparency of the BTS icons.

O TH ER T ASKS

5 15

Show cell information option hides and displays the selected cell information. Hide cell texts on low zoom levels (performance optimization) option hides partly or completely the cell texts when the map is zoomed out so that the map can be drawn faster and with less visual distractions. The level is adjustable with a slider. Zoom levels are specific to used maps. Use cell beam range from BTS file allows you to enable/disable the display of cell beam range based on cell beam range data in the BTS file. Default beam defines the default beam range in meters. Default beam angle defines the default beam angle in degrees. Beam transparency defines the transparency of the beam range layer.

Default color defines the default color of the BTS icons. Default beam color defines the default beam color.

Default settings for BTS parameter allow you to define default color sets for BTS parameters.

BTS Technology Based Settings enable you to choose the color and the size of the base station depending on which networks base station it is, or based on the carrier number. When the Use technology based settings (overrides icon size settings) option is selected, you are able to open the Set Technology Setting view by clicking Add. By clicking the Icon color option, the size and the color of the base stations changes on the map. If this option is not selected, only the size of the base station changes.

IMPORTING CUSTOM SETTINGS


It is possible to import custom settings (such as custom parameters, workbooks, queries, color sets, KPIs, etc.) from other users. Select File | Import | Settings from the Ribbonbar.

The Open dialog appears.

O TH ER T ASKS

5 17

Browse for the settings file (.aex) to be imported and select Open.

EXPORTING CUSTOM SETTINGS


It is possible to export custom settings (such as parameters, workbooks, queries, color sets, KPIs) to other users. All custom settings are saved in a single .aex file. Select File | Export | Settings from the Ribbonbar.

The Save As dialog appears.

Select a destination for the settings file (.aex) and select Save. The Export Settings dialog opens.

Select the settings you want to export and click OK. You can export e.g. database settings, custom queries, color sets, etc. It is also possible to do multiple selections and select/deselect groups, e.g. all workbooks, all queries, at once. The settings file (.aex) is saved in the selected location and can be imported to any other Nemo Analyze Standalone/Client using custom settings import (see Importing Custom Settings on page 516).

O TH ER T ASKS

5 19

IMPORTING MAPINFO POLYGONS


MapInfo polygons can be used e.g. when defining area-based search folders (see Search Folders on page 52) and when limiting queries to measurement files from particular area(s) (see Filtering Based on Polygon Area on page 110).

To import MapInfo polygons, click the polygons icon at the bottom of the Workspace view, rightclick on the Workspace | Polygons | Regions background, and select Import polygon from the popup menu.

The Polygon Import dialog opens.

File defines the MapInfo polygon (.TAB) file to be imported. Click the browse button a MapInfo polygon file.

to browse for

Available features allows you to select the polygon file features you want to import. Clear a feature to prevent it from being imported. Selected features allows you to select the polygon file features you want to import. Clear a feature to prevent it from being imported. To select a feature or a set of features, select it in the Available features field and click the button. To select all available features, select it in the Available features field and click the The selected feature(s) are displayed in the Selected features field. button.

Import name allows you to define a new name for the imported polygon. To clear a feature or a set of features, select it in the Selected features field and click the To select all available features, select it in the Selected features field and click the Once all polygons to be imported have been selected, click OK. button.

button.

O TH ER T ASKS

5 21

VIEWING LOG ON SYSTEM PERFORMANCE AND SQL FUNCTIONS


To access the Log window, select View | Panels | Log window from the Ribbonbar.

The Log window is displayed on the bottom of the screen.

Show output from allows you to define the functions you wish to see the log on. Log, if selected, displays log on general system performance. SQL, if selected, displays log on SQL queries and database functions.

The Clear all

button clears the currently displayed log.

U SER INTERF AC E

5 23

USER INTERFACE
MENUS
This chapter describes the contents of the menu bar in Nemo Analyze.

Analyze Menu
You can access the Analyze menu by clicking on the Analyze main window. logo at the top left-hand corner of the

Open enables you to open measurement files, maps, BTS files, report templates and data source files.

Workspace enables you to open, save and close workspace items.

Workbook enables you to save active workbooks and floating windows.

Export to PDF enables you to export workbooks to PDF format. This item is displayed only if you have a workbook open. Print enables you to preview and print active documents, and change printer and printing options. This item is displayed only if you have a workbook open.

Recent Workspaces enables you to open saved workspaces that you have used recently. You can close Nemo Analyze by clicking on Exit Application.

File Menu

Open | Measurement enables you to load measurement files, maps, BTS files, report templates, and data source files to the database. Open | Workspace enables you to open saved workspaces. Open | Folder enables you to browse for a folder, the contents of which are to be loaded to the database. Save | Workspace enables you to save workspace files (.aws). Save | Workbook enables you to save workbooks (p. 285) and add workbooks as templates to the Workspace menus. Import enables you to import image files to be used as maps, custom settings files (.aex) , and color set files. Export enables you to export workbooks in PDF format and custom settings files in .aex format. Image to clipboard enables you to save an active workbook or map as an image. You can either save the image on clipboard to enable pasting to other applications, or save the Image to file in .jpg format. You can also set an Export Ratio (or an export size) for the exported image in percentages or in fixed pixel size. Print enables you to preview and print active documents, and change printer and printing options. You can exit the File menu by double-clicking on it.

View Menu

Add Data View enables you to create empty workbooks, workbooks with selected empty data views, and floating windows. For adding parameters on empty data views, see Adding Data Views. View | Organize Measurements enables you to organize measurement files loaded in the database. See page 48 for more information on organizing measurements.

U SER INTERF AC E

5 25

Tools | Database Browser enables you to view all the tables in the schema. Menu editor. Group editor, Color set editor, Notification configuration, Query manager, Options, Event scheduler and User Manager enable you to access several editors and dialogs. In this manual, these dialogs and editors are described in their own chapters. Panels enables you to display and hide the Workspace, Output, Properties, Window browser and Query clipboard views. You can exit the View menu by double-clicking on it.

Tools Menu

Reconnect enables you to reconnect to the database if the connection is lost. Synchronization enables you to play back measurement files. Reset Source button resets the synchronization source. The field after the button displays the current synchronization source. Go to beginning Step backwards Play backwards Stop button jumps to the beginning of the file. button steps one step backwards in the file. button plays the file from end to beginning.

button ends the playback. button plays the file from beginning to end. button steps one step forward in the file.

Play forward Step forward

Go to timestamp button jumps to a particular point in the file if timestamps were inserted when the file was recorded. Counter displays the current point in the file. Playback speed defines how fast the file is played back. Area Binning enables you to define the bin size in horizontal X steps and vertical Y steps. To set the bin in meters, select Bins in meters. Zoom enables you to set zoom to occur synchronously in all graphs and maps. Drill Down enables you to set a time range from which log entries are to be included in a drill-down Before and After a selected event in the measurement data. Global Filters enables you to add and remove global filters applied to all operations performed with Nemo Analyze. For instance, if you are interested only in measurement data from a certain restricted area, you can select this area as a filter using polygon area selection.

Fleet Menu
Fleet Manager is an optional component of Nemo Analyze intended for the management of Nemo Autonomous fleet. It will be discussed in further detail in the Nemo Autonomous user manual.

Layout Menu
Note: The layout menu is only displayed when a workbook is open in Nemo Analyze. When you close all workbooks, also the layout menu is closed automatically.

Add | Page enables you to add a new, empty page in a workbook. By clicking Graph, Map, Grid, etc. you can add data views in the active page. For adding parameters on empty data views, see Adding Data Views on page 280. Remove | Page enables you to remove an active page from a workbook. Copy | Workbook enables you to create a copy of the active workbook for a selected measurement. Organize item enables you to close all workbooks and floating windows, and organize various workbooks and views in the Nemo Analyze main view.

Help Menu

In the Help menu, you will find access to the online help and also shortcuts to some important documents, such as, a description of the Nemo file format.

SHOR TCUT S

5 27

SHORTCUT KEYS
Shortcut keys available in Nemo Analyze:

Command Close Close Window New Workbook Next Print

Shortcut Key
ALT+F4 CTRL+F4 CTRL+N CTRL+F6 CTRL+P

Description Closes the program. Closes the active window (exc. main window). Opens an empty workbook. Activates the next open workbook (in the order of last appearance). Prints the current view.

Analyze Tools menu shortcut keys: Command Go to the beginning Step backwards Play backwards Stop Play forwards Step forwards Shortcut Key
CTRL+SHIFT+ LEFT ARROW

Jumps to the beginning of the file.

CTRL+LEFT ARROW SHIFT+F5 F6 F5 CTRL+RIGHT ARROW

Steps one step backwards in the file. Plays the file from end to beginning. Ends the playback. Plays the file from beginning to end. Steps one step forward in the file.

FAQ
Q: How soon are changes in the Nemo File Format implemented to Nemo Analyze? A: Changes in the Nemo File Format are implemented very quickly to Nemo Analyze and there is no need to reload any data in the database. Q: Which database is used in Nemo Analyze? A: Nemo Analyze uses the Object Store database. Q: Which reporting tool is used in Nemo Analyze? A: The Viewer version of Crystal Reports reporting tool is included in the Nemo Analyze package but you can also use other reporting tools. To create new report templates, you will need the Crystal Reports Designer tool or some other SQL-compatible designing tool. Q: I opened a parameter on a map but the samples are shown without color coding? A: If you opened data on the map through the Analyze Wizard (right-click on a device in workspace | Pick Parameter) and selected a map instead of the default view, only the selected parameters or events are displayed. You should first open a map and then drag and drop a measurement file on the map. Then use color sets to highlight parameter values on the route.

N EM O AN AL YZE SU PPOR T

5 29

NEMO ANALYZE SUPPORT


If you have questions on or beyond this documentation, please contact Nemo Analyze technical support. Note that for full support you need to have a valid Maintenance Agreement.

INTERNET SUPPORT AND NEMO USER CLUB


The Nemo User Club offers several new ways to benefit from Nemo products. You can find it at our WWW site http://www.anite.com/nemo. Access to the Nemo User Club is restricted to customers with SW maintenance and Technical Support agreements or partner agreements. If you are an existing Nemo customer with SW maintenance and Technical Support agreements, but you are without access to the user club, please complete an on-line registration form. After submitting the requested information, you will receive a personal access key and password by email in a few days time. Once you are registered with our User Club, you will automatically receive e-mails letting you know each time a new version of software for your Nemo Product is released. It is fast, easy, and it is available to you seven days a week, 24 hours a day. If you have questions relating to the Nemo User Club, please e-mail us at nemo.helpdesk@anite.com.

PHONE, EMAIL, AND FAX SUPPORT


During the warranty period, the phone and fax support related to potential software errors is free of additional charge. Registered users with a valid Maintenance Agreement are entitled to full support. When faxing or emailing, please let us know the number of your Software Maintenance and Support Agreement. Please contact us at the following locations:
Global Email Tel. Fax Address nemo.helpdesk@anite.com +358 50 395 7800 +358 8 551 6182 Anite Finland Ltd Kiviharjunlenkki 1 D 90220 Oulu Finland

Americas Email Mobile Fax Address

nemo.helpdesk@anite.com +1 469 951 9106 +1 972 929 9898 Anite Telecoms Inc. 6333 N. State Hwy 161, Suite 110 Irving, Texas 75038

APAC Email Tel. Fax Address

nemo.helpdesk@anite.com +65 6254 9003 +65 6254 9885 Anite Singapore Pte Ltd 101 Thomson Road, #18-04 United Square Singapore 307591

P.R. China Email Tel. Fax Address

nemo.helpdesk@anite.com +86 10 6567 8528 +86 10 6567 8521 Anite Wireless Trading (Beijing) Ltd. JB11-0-092, Jianwai Diplomatic Compound Jianguomenwai Street, Chaoyang District, Beijing 100600, CHINA

When contacting us for problem reporting, please supply us with the following information: your Windows version number (e.g., Windows XP), see Windows Control Panel | System your computer brand model (e.g., Dell) your Nemo Analyze version (e.g., Version Nemo Analyze 6.20), see Help | About Analyze your description of the problem whether you were able to repeat the problem the corresponding measurement results file (as email attachment or on diskette)

APPENDIX 1
MAKING MAPINFO MAPS
Using raster image files, you can bring paper maps, photographs, and other graphic images into MapInfo. You can scan paper maps and then use paper maps as the foundation for the maps you create in MapInfo. After scanning the paper map into a raster image file, you can display it in a map window. Nemo Outdoor and Nemo Analyze support only .TIF, .GIF, and .JPG formats. MapInfo can read the following types of raster images: - filename.TIF - filename.GIF - filename.JPG - filename.PCX - filename.BMP - filename.BIL Color options are: Monochrome images: each pixel in map image can be black or white. Gray scale images: each pixel in gray scale image can be black, white, or a shade of gray. Color images: each pixel can be of any color from a palette of available colors. MapInfo supports 256 colors. When using Nemo Outdoor and Nemo Analyze, we recommend using gray scale images. Doing this, measurement route can be discerned better than with color map images.

Registering a Raster Map


If you want to overlay vector data on top of a raster image, then you must register the raster map image so that MapInfo can position it properly in a Map window. You do this in the Image Registration dialog box. You must identify control point coordinates and projection of the raster image map. It is important to provide accurate control point information when registering a raster map image. Choose control points that can be easily identified and selected, such as street intersections or use the coordinates (latitude and longitude grid) for the selected point from a paper map.

To register a raster map with MapInfo SW:


1. Choose File | Open Table and Raster Image File Format. Select your raster map image file (.TIF, .GIF, .JPG) and open it. The Image Registration dialog box will be displayed. A preview of the raster map image appears on the screen. 2. Select the Projection button to specify the projection of the raster image map. If you do not know the right projection, the default value is Longitude/ Latitude.

3. Start adding control points. Click on a location in the preview (raster image map). The Add Control Point dialog box will be displayed showing the location of the point in pixels. Add Map X and Map Y coordinates in decimal formats. For example, to specify the coordinates 65 degrees, 30 minutes, enter 65.5 degrees. Use negative numbers when specifying west and south coordinates. If you want to register a raster map, which does not use longitude and latitude coordinates, specify the appropriate projection in the Image Registration dialog box. Enter your coordinates in the native units of the coordinate system. For example, if you are registering a UTM map image, enter coordinates in meters. Using MapBasic program, you can convert d/m/s coordinates into decimal format. 4. You must choose at least three control points. Select points that can be easily identified and selected in the map window. 5. After all control points have been defined, click OK in the Image Registration dialog box. The raster map will be displayed in the map window. 6. Move filename.tab and filename.tif/gif/jpg files to the map directory of Nemo Outdoor or Nemo Analyze. 7. Open filename.tab in Nemo Outdoor or Nemo Analyze.

IND EX

5 33

APPENDIX 2
END-USER LICENSE AGREEMENT
IMPORTANT - READ CAREFULLY: This End-User License Agreement ("EULA") is a legal agreement between you (either an individual or a single entity) and Anite Finland Ltd for the Anite Finland Ltd software product(s), in whatever form, identified above ("Software Product" or "Software"). The Software Product includes computer software, the associated media, any printed materials, and any "online" or electronic documentation. By installing, copying or otherwise using the Software Product, you agree to be bound by the terms of this EULA. If you do not agree to the terms of this EULA, Anite Finland Ltd is unwilling to license the Software Product to you. In such event, you may not use or copy the Software Product, and you should promptly return the unused product(s) in their original packaging to the place of purchase within thirty days of the date of original purchase. SOFTWARE PRODUCT LICENSE The Software Product is protected by copyright laws and international copyright treaties, as well as other intellectual property laws and treaties. The Software Product is licensed, not sold. 1. GRANT OF LICENSE. This EULA grants you the following rights: Software. You agree that the Software product is licensed to you to be used with the mobile phone(s) and software protection key designated by Anite Finland Ltd for such use ("Designated Hardware"). You also agree that you shall not have more than one copy of the Software Product in use at a time for any Designated Hardware in which the Software Product is to be used or to be distributed except as expressly provided for in this EULA. You may use one copy of the Software Product in binary code form only on one computer at a time in order to conduct system measurements using the Designated Hardware at various locations to the extent necessary for the permitted operation of the application software but may not duplicate the Software Product. Storage/Network/Single Use. You may also store or install a copy of the Software Product in binary code form only over an internal network, and distribute the Software Product to your other computers over an internal network. However, you must acquire and dedicate a license for the Software Product for each Designated Hardware on which the Software Product is used or to which it is distributed. A license for the Software Product may not be shared or used concurrently on different computers. 2. DESCRIPTION OF OTHER RIGHTS AND LIMITATIONS. Limitations on Reverse Engineering, Decompilation and Disassembly. Removal, emulation, or reverse engineering of all or any part of this product or its protection constitutes an unauthorized modification to the product and is specifically prohibited. Nothing in this license statement permits you to derive the source or assembly code of files provided to you in executable or object formats. Limitations on Modifications, Adaptations and Other Changes. You may not modify, adapt or otherwise make any changes to the Software Product except and only to the extent you are expressly permitted to do so under the applicable mandatory law notwithstanding this limitation or by Anite Finland Ltd's prior written consent. Separation of Components. The Software Product is licensed as a single product. Its component parts may not be separated for use on more than one Designated Hardware and computer as designated above.

Not for Resale. This Software Product is Not for Resale" or "NFR" and your license only permits personal use and you may not sell or otherwise transfer the Software Product for value except as specified below. Rental. You may not rent, lease or lend the Software Product to any person or entity. Software Transfer. You may not otherwise give or permanently transfer any or all of your rights under this EULA without the express prior written consent of Anite Finland Ltd which such consent shall not be unreasonably withheld. Duration and Termination. This EULA is effective from the day you open the sealed media package or download and continues until termination. You may terminate this EULA at any time. Without prejudice to any other rights, Anite Finland Ltd may terminate this EULA if you fail to comply with the terms and conditions of this EULA. In either event, you must destroy all copies of the Software Product and all associated media, printed materials, updates and component parts and verify to Anite Finland Ltd in writing that such has been done. 3. SUPPORT SERVICES. Anite Finland Ltd may provide you with support services related to the Software Product ("Support Services"). Any supplemental software code provided to you as part of the Support Services shall be considered part of the Software Product and subject to the terms and conditions of this EULA. With respect to technical information you provide to Anite Finland Ltd as part of the Support Services, Anite Finland Ltd may use such information for its business purpose, including for product updates and development. 4. COPYRIGHT AND TRADEMARKS. All title and copyrights in and to the Software Product (including but not limited to any images, photographs, animations, video, audio, music, text and "applets", incorporated into the Software Product), the accompanying printed materials, and any copies of the Software Product, are owned by Anite Finland Ltd or its suppliers. The Software Product is protected by copyright laws and international treaty provisions. You may not copy the printed materials accompanying the Software Product. You may not remove, modify or alter any Anite Finland Ltd copyright or trademark from any part of the Software Product, including but not limited to any such notices contained in the physical and/or electronic media or documentation, in the Anite Finland Ltd Setup Wizard dialogue or 'about' boxes, in any of the runtime resources and/or in any web-presence or web-enabled notices, code or other embodiments originally contained in or dynamically or otherwise created by the Software Product. Portions based on WinWAP technology. Copyright Slob-Trot Software Oy Ab 2004. http://www.winwap.com Google Earth is a registered Trademark of Google Inc. All screenshots of Google Earth included in this document are the property of Google Inc. Perceptual Evaluation of Speech Quality (PESQ) measurement technology included in this product is protected by copyright and by European, US, and other patents and is provided under license from OPTICOM Dipl. -Ing. M. Keyhl GmbH, Erlangen, Germany, 2004, http://www.opticom.de. For further information, please refer to http://www.pesq.org. 5. DUAL-MEDIA SOFTWARE. You may receive the Software Product in more than one medium. Regardless of the type or size of medium you receive, you may use only that one medium that is appropriate for the systems you use to operate the Software Product. You may not use or install the other medium on any system. You may not loan, rent, lease or otherwise transfer the other medium to another user, except as part of the permanent transfer (as provided above) of the Software Product.

IND EX

5 35

6. CONFIDENTIALITY AND NON-DISCLOSURE. You agree to keep in confidence and not to disclose in any manner or form to any third party nor publish any such data or information related to the Software Product. You agree to protect all information and data provided to you by Anite Finland Ltd related to the Software Product with the same degree of care that you use to protect your own like data and information, but with not less than that due care exercised by a reasonably prudent person under the same or similar circumstances. 7. U.S. GOVERNMENTAL RESTRICTED RIGHTS. The Software Product and documentation are provided with Restricted Rights. Use, duplication, or disclosure by the Government is subject to restrictions as set forth in subparagraph (c)(1)(ii) of the Rights in Technical Data and Computer Software clause at DFARS 252.277-7013 or subparagraphs (c)(1) and (2) of the Commercial Computer Software - Restricted Rights at 48 CFR 52-227-19, as applicable. 8. LIMITED WARRANTY. Anite Finland Ltd warrants that (a) the Software Product will perform substantially in accordance with the accompanying written materials for a period of three (3) months from the date of receipt, and (b) any third party hardware accompanying the Software Products will be free from defects in materials and workmanship under normal use and service for a period of one (1) year from the date of receipt. To the extent that implied warranties are disclaimable, they are disclaimed in paragraph 10. Some states/jurisdictions do not allow disclaimers of or limitations on duration of an implied warranty, so the above limitation may not apply to you. Any implied warranties on the Software Product and, as applicable, third party hardware that is not disclaimable are limited to three (3) months and one (1) year respectively. 9. CUSTOMER REMEDIES. Anite Finland Ltd and its suppliers' entire liability and your exclusive remedy shall be, at Anite Finland Ltd option, either (a) return of the price paid, or (b) repair or replacement of the software or hardware that does not meet the Limited Warranty contained in this EULA and which is returned to Anite Finland Ltd. The Limited Warranty contained herein is void if failure of the software or third party hardware has resulted from accident, abuse, misapplication or events beyond the control of Anite Finland Ltd including, but not limited to Acts of God, acts of nature or acts of war. Any replacement Software Products or hardware will be warranted for the remainder of the original warranty period or thirty (30) days, whichever is longer. Anite Finland Ltd is not responsible for ensuring adequate backup and storage for your data or records. 10. NO OTHER WARRANTIES. To the maximum extent permitted by applicable law, Anite Finland Ltd disclaims all other warranties, either express or implied, including but not limited to implied warranties of merchantability, fitness for a particular purpose, and non-infringement of third-party rights, with regard to the Software Product, the accompanying written materials, and any accompanying hardware. This limited warranty gives you specific legal rights. You may have others which vary from state/jurisdiction to state/jurisdiction.

11. NO LIABILITY FOR CONSEQUENTIAL DAMAGES. To the maximum extent permitted by applicable law, in no event shall Anite Finland Ltd or its suppliers be liable for any damages whatsoever (including without limitation, special, incidental, consequential, punitive or indirect damages for personal injury, loss of business profits, business interruption, loss of business information, or any other pecuniary loss) arising out of the use of or inability to use this Software Product, even if Anite Finland Ltd has been advised of the possibility of such damages. In any case, Anite Finland Ltd and its suppliers' entire liability under any provision of this EULA shall be limited to the amount actually paid by you for the Software Product and/or third party hardware accompanying the Software Product. The Limitations of Liability contained in this Section apply to claims of every kind, whether based on contract law, tort law or otherwise, even if these limitations have the effect of exculpating a party from responsibility for its negligence, fault or other conduct. Because some states/jurisdictions do not allow the exclusion or limitation of liability for consequential or incidental damages, the above limitation may not apply to you.

Das könnte Ihnen auch gefallen