<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="9788793609846.xsl"?>
<book id="home" xmlns:xlink="http://www.w3.org/1999/xlink">
<bookinfo>
<title>The MANTIS Book Cyber Physical System Based Proactive Collaborative Maintenance</title>
<affiliation><emphasis role="strong">Editors</emphasis></affiliation>
<authorgroup>
<author><firstname>Michele</firstname>
<surname>Albano</surname>
</author>
</authorgroup>
<affiliation>ISEP, Polytechnic Institute of Porto, Portugal</affiliation>
<authorgroup>
<author><firstname>Erkki</firstname>
<surname>Jantunen</surname>
</author>
</authorgroup>
<affiliation>VTT Technical Research Centre of Finland Ltd., Finland</affiliation>
<authorgroup>
<author><firstname>Gregor</firstname>
<surname>Papa</surname>
</author>
</authorgroup>
<affiliation>Jo&#x0017E;ef Stefan Institute, Slovenia</affiliation>
<authorgroup>
<author><firstname>Urko</firstname>
<surname>Zurutuza</surname>
</author>
</authorgroup>
<affiliation>Mondragon Unibertsitatea, Spain</affiliation>
<publisher>
<publishername>River Publishers</publishername>
</publisher>
<isbn>9788793609846</isbn>
</bookinfo>
<preface class="preface" id="preface01">
<title>RIVER PUBLISHERS SERIES IN AUTOMATION, CONTROL AND ROBOTICS</title>
<para><emphasis>Series Editors:</emphasis></para>
<para><emphasis role="strong">ISHWAR K. SETHI</emphasis><?lb?><emphasis>Oakland University</emphasis><?lb?><emphasis>USA</emphasis></para>
<para><emphasis role="strong">TAREK SOBH</emphasis><?lb?><emphasis>University of Bridgeport</emphasis><?lb?><emphasis>USA</emphasis></para>
<para><emphasis role="strong">QUAN MIN ZHU</emphasis><?lb?><emphasis>University of the West of England</emphasis><?lb?><emphasis>UK</emphasis></para>
<para>Indexing: All books published in this series are submitted to the Web of Science Book Citation Index (BkCI), to CrossRef and to Google Scholar.</para>
<para>The &#x0201C;River Publishers Series in Automation, Control and Robotics&#x0201D; is a series of comprehensive academic and professional books which focus on the theory and applications of automation, control and robotics. The series focuses on topics ranging from the theory and use of control systems, automation engineering, robotics and intelligent machines.</para>
<para>Books published in the series include research monographs, edited volumes, handbooks and textbooks. The books provide professionals, researchers, educators, and advanced students in the field with an invaluable insight into the latest research and developments.</para>
<para>Topics covered in the series include, but are by no means restricted to the following:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Robots and Intelligent Machines</para></listitem>
<listitem>
<para>Robotics</para></listitem>
<listitem>
<para>Control Systems</para></listitem>
<listitem>
<para>Control Theory</para></listitem>
<listitem>
<para>Automation Engineering</para></listitem>
</itemizedlist>
<para>For a list of other books in this series, visit www.riverpublishers.com</para>
</preface>
<preface class="preface" id="preface02">
<title>Foreword</title>
<para>This book aims to highlight the fundamental underpinnings of Condition-Based Maintenance and related conceptual ideas. It also presents an overall view of i) proactive maintenance, and ii) condition-based maintenance, including its potential economic impact and latest technical solutions.</para>
<para>The core content of this book describes the outcome of the MANTIS project. The ambition was to support the creation of a platform to enable novel kinds of maintenance strategies for industrial machinery. The key enabler was considered to be the carefully crafted symbiotic combination of data collection through Cyber-Physical Systems, and the use of machine learning techniques and advanced visualization for the enhanced monitoring and prognosis of the machines.</para>
<para>After more than 3 years of hard work, the MANTIS consortium has spent uncountable hours on the MANTIS platform, working on its implementation and its application to a wide range of industrial use cases.</para>
<para>It is our sincere hope that others will, with the help of this book and on-line resources, benefit from the MANTIS reference architecture, and the techniques employed in it. Hopefully we can build an open community around this technology so that the maintenance industry can take a step into the future of massive, agile yet affordable monitoring systems.</para>
<para>The book is therefore suitable for industrial and maintenance managers that want to implement a new, or enhance an existing, maintenance system for their companies. This book should give them a basic idea of the first required steps in implementing advanced monitoring and analytics systems.</para>
<para><link linkend="ch01">Chapter <xref linkend="ch01" remap="1"/></link> describes the fundamentals of advanced monitoring practices and strategies in industry, and sets the context for the rest of the book.</para>
<para><link linkend="ch02">Chapter <xref linkend="ch02" remap="2"/></link> delves into the economical pillars behind advanced monitoring and analytics that lead to the importance of both technological advances and new business models, which promote the acceptance of these technologies.</para>
<para><link linkend="ch03">Chapter <xref linkend="ch03" remap="3"/></link> describes the architecture envisioned for a collaborative platform for proactive maintenance and the functional requirements it must satisfy.</para>
<para><link linkend="ch04">Chapter <xref linkend="ch04" remap="4"/></link> focuses on the data collection work that is the basic requirement for advanced data analysis techniques to succeed. Two main problems emerged with regards to the huge amount of data that were collected in this context: how to process the data, and how to visualize them.</para>
<para><link linkend="ch05">Chapter <xref linkend="ch05" remap="5"/></link> introduces a number of data analysis and machine learning techniques that can be &#x02013; and were &#x02013; applied to support advanced maintenance strategies.</para>
<para><link linkend="ch06">Chapter <xref linkend="ch06" remap="6"/></link> provides insights into the techniques used to visualize the data, and how the data can be manipulated and inspected.</para>
<para><link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link> describes a number of real-life pilots that were implemented based on the MANTIS platform.</para>
<para><link linkend="ch08">Chapter <xref linkend="ch08" remap="8"/></link> discusses novel business models and strategies, proposes their application to modern maintenance techniques, and provides a quantitative example based on a real-life scenario.</para>
<para><link linkend="ch09">Chapter <xref linkend="ch09" remap="9"/></link> concludes the book by considering open research questions that arose during the execution of the MANTIS project.</para>
</preface>
<preface class="preface" id="preface03">
<title>List of Contributors</title>
<para>Achim Woyte, 3e, Belgium</para>
<para>Aitzol Iturrospe, Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Alberto Sillitti, Innopolis University, Russian Federation</para>
<para>Alejandro Veiga, Acciona Construcci&#x000F3;n S.A., Spain</para>
<para>Ale&#x00161; Kancilija, XLAB, Ljubljana, Slovenia</para>
<para>Alp Ak&#x000E7;ay, Technische Universiteit Eindhoven, The Netherlands</para>
<para>Ander Etxabe, IK4-Ikerlan, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Andoitz Aranburu, FAGOR ARRASATE, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Andriy Zubaliy, Sirris, Belgium</para>
<para>Anja Vidmar, XLAB, Ljubljana, Slovenia</para>
<para>Ansgar Bergmann, STILL GmbH, Germany</para>
<para>Antonio Ruggieri, Ansaldo STS, Italy</para>
<para>Antti Niemel&#x000E4;, Lapland University of Applied Sciences Ltd., Finland</para>
<para>Arja Kotkansalo, Lapland University of Applied Sciences Ltd., Finland</para>
<para>Arvind Kumar Shekar, Bosch, Germany</para>
<para>Asif Mohammed, ADIRA Metal-Forming Solutions SA, Portugal</para>
<para>Attila Frank&#x000F3;, AITIA International Inc., Hungary</para>
<para>Babacar Sarr, 3e, Belgium</para>
<para>Bas Tijsma, Philips Consumer Lifestyle B.V., The Netherlands</para>
<para>Carlos Soares, Instituto de Engenharia de Sistemas e Computadores do Porto, Portugal</para>
<para>Carolina Mej&#x000ED;a Ni&#x000F1;o, MONDRAGON Corporation, Spain</para>
<para>Catarina F&#x000E9;lix, Instituto de Engenharia de Sistemas e Computadores do Porto, Portugal</para>
<para>&#x000C7;i&#x000E7;ek G&#x000FC;ven, Technische Universiteit Eindhoven, The Netherlands</para>
<para>Cl&#x000E1;udio Rebelo de S&#x000E1;, Instituto de Engenharia de Sistemas e Computadores do Porto, Portugal</para>
<para>Csaba Heged&#x00171;s, AITIA International Inc., Hungary</para>
<para>Daan Terwee, Philips Consumer Lifestyle B.V., The Netherlands</para>
<para>Daniel Reguera, Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>David Chico, FAGOR ARRASATE, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Diogo Martinho, ISEP, Polytechnic Institute of Porto, Porto, Portugal</para>
<para>Eduardo Saiz, IK4-Ikerlan, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Ekhi Zugasti, Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Emmanuel Okafor, University of Groningen, The Netherlands</para>
<para>Eneko S&#x000E1;enz de Argando&#x00F1;a, Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Erkki Jantunen, VTT Technical Research Centre of Finland Ltd, Finland</para>
<para>Eva Mart&#x000ED;nez, Acciona Construcci&#x000F3;n S.A., Spain</para>
<para>F&#x000E9;lix Larrinaga, Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Franc Novak, Jo&#x0017E;ef Stefan Institute, Slovenia</para>
<para>Giovanni Di Orio, FCT-UNL, UNINOVA-CTS, Caparica, Portugal</para>
<para>Godfried Webers, Philips Medical Systems Nederland B.V., The Netherlands</para>
<para>Goreti Marreiros, ISEP, Polytechnic Institute of Porto, Porto, Portugal</para>
<para>Gorka Urchegui, Mondragon Sistemas De Informacion, Spain</para>
<para>Gregor Papa, Jo&#x0017E;ef Stefan Institute, Slovenia</para>
<para>Hans Sprong, Philips Medical Systems Nederland B.V.,</para>
<para> The Netherlands</para>
<para>Hugo Ferreira, Instituto de Engenharia de Sistemas e Computadores do Porto, Portugal</para>
<para>Iban Barrutia Inza, Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>I&#x00F1;aki Garitano, Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>I&#x000F1;igo Lecuona, Mondragon Sistemas De Informacion, Spain</para>
<para>Iosu Gabilondo, IK4-Ikerlan, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Isabel Pra&#x000E7;a, ISEP, Polytechnic Institute of Porto, Porto, Portugal</para>
<para>Istv&#x000E1;n Moldov&#x000E1;n, Budapest University of Technology and Economics, Hungary</para>
<para>Jaap van der Voet, Philips Medical Systems Nederland B.V., The Netherlands</para>
<para>Jaime Campos, Department of Informatics, Linnaeus University, Sweden</para>
<para>Jan-Otto Hooghoudt, Aalborg University, Denmark</para>
<para>Javier Fernandez Anakabe, Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Jeroen Gijsbers, Philips Medical Systems Nederland B.V., The Netherlands</para>
<para>Jon Basurko, IK4-Ikerlan, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Jon Olaizola, Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Jon Rodriguez, KONIKER, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Jos&#x000E9; Luis Flores, IK4-Ikerlan, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Jos&#x000E9; Manuel Abete, Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Juha Valtonen, Lapland University of Applied Sciences Ltd, Finland</para>
<para>Karel De Brabandere, 3e, Belgium</para>
<para>Karel Eerland, Philips Medical Systems Nederland B.V., The Netherlands</para>
<para>Karmele Intxausti, IK4-Ikerlan, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Kees Wouters, Philips Electronics Nederland B.V., The Netherlands</para>
<para>Lambert Schomaker, University of Groningen, The Netherlands</para>
<para>Lauri V&#x000E4;limaa, Wapice, Finland</para>
<para>Luis Lino Ferreira, ISEP, Polytechnic Institute of Porto, Porto, Portugal</para>
<para>Manja Gorenc Novak, XLAB, Ljubljana, Slovenia</para>
<para>Marcel Boosten, Philips Medical Systems Nederland B.V., The Netherlands</para>
<para>Mar&#x000ED;a Aguirregabiria, IK4-Ikerlan, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Mario Riccardi, Ansaldo STS, Italy</para>
<para>Marjan &#x00160;terk, XLAB, Ljubljana, Slovenia</para>
<para>Mathias Gr&#x000E4;dler, Wapice, Finland</para>
<para>Matti Kaija, Fortum Power and Heat Oy, Finland</para>
<para>Mauro Barbieri, Philips Electronics Nederland B.V., The Netherlands</para>
<para>Michele Albano, ISEP, Polytechnic Institute of Porto, Porto, Portugal</para>
<para>Michel I&#x000F1;igo Ulloa, MONDRAGON Corporation, Spain</para>
<para>Mike Holenderski, Technische Universiteit Eindhoven, The Netherlands</para>
<para>Mikel Anasagasti, Goizper, Spain</para>
<para>Mikel Mondragon, Goizper, Spain</para>
<para>Mikel Viguera, KONIKER, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>M&#x000F3;nica S&#x000E1;nchez, Acciona Construcci&#x000F3;n S.A., Spain</para>
<para>Nayra Uranga, Acciona Construcci&#x000F3;n S.A., Spain</para>
<para>Oier Sarasua, FAGOR ARRASATE, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>P&#x000E1;l Varga, AITIA International Inc., Hungary</para>
<para>Pankaj Sharma, Mechanical Engineering Department, IIT Delhi, India</para>
<para>Paolo Ciancarini, Consorzio Interuniversitario Nazionale per l&#x02019; Informatica, Italy</para>
<para>Paolo Sannino, Ansaldo STS, Italy</para>
<para>Patricia Dominguez Arroyo, University of Groningen, The Netherlands</para>
<para>Paulien Dam, Philips Consumer Lifestyle B.V., The Netherlands</para>
<para>Pedro Mal&#x000F3;, FCT-UNL, UNINOVA-CTS, Caparica, Portugal</para>
<para>Peter Craamer, Mondragon Sistemas De Informacion, Spain</para>
<para>Petri Helo, Wapice, Finland</para>
<para>Quinten Marcelis, Ilias Solutions, Belgium</para>
<para>Rafael Rocha, ISEP, Polytechnic Institute of Porto, Porto, Portugal</para>
<para>Rafael Socorro, Acciona Construcci&#x000F3;n S.A., Spain</para>
<para>Raquel Garc&#x000ED;a, Acciona Construcci&#x000F3;n S.A., Spain</para>
<para>Ra&#x000FA;l Torrego, IK4-Ikerlan, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Ricardo Romero, Tekniker, Spain</para>
<para>Riku Salokangas, VTT Technical Research Centre of Finland Ltd, Finland</para>
<para>Roberto Gonz&#x000E1;lez, Tekniker, Spain</para>
<para>Roberto Uribeetxeberria, Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Rui Casais, ADIRA Metal-Forming Solutions SA, Portugal</para>
<para>Salvatore Esposito, Ansaldo STS, Italy</para>
<para>Sheng He, University of Groningen, The Netherlands</para>
<para>Silvia Hern&#x000E1;ndez, Acciona Construcci&#x000F3;n S.A., Spain</para>
<para>S&#x000F6;ren Schneickert, Fraunhofer IESE, Germany</para>
<para>S&#x000F8;ren Theodorsen, Ilias Solutions, Belgium</para>
<para>Stefano Primi, Acciona Construcci&#x000F3;n S.A., Spain</para>
<para>Susana Ferreiro, Tekniker, Spain</para>
<para>&#x00160;pela Poklukar, Jo&#x0017E;ef Stefan Institute, Slovenia</para>
<para>Tom Tourw&#x000E9;, Sirris, Belgium</para>
<para>Urko Leturiondo, IK4-Ikerlan, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Urko Zurutuza, Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Veli-Pekka Salo, Wapice, Finland</para>
<para>Verus Pronk, Philips Electronics Nederland B.V., The Netherlands</para>
<para>Ville Rauhala, Lapland University of Applied Sciences Ltd, Finland</para>
<para>Vito &#x0010C;u&#x0010D;ek, XLAB, Ljubljana, Slovenia</para>
<para>Xabier Eguiluz, IK4-Ikerlan, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para>Yi Yang, Aalborg University, Denmark</para>
</preface>
<preface class="preface" id="preface04">
<title>List of Figures</title>
<table-wrap position="float">
<table cellspacing="5" cellpadding="5" frame="none" rules="none">
<tbody>
<tr><td valign="top" width="15%"><emphasis role="strong"><link linkend="F1-1">Figure 1.1</link></emphasis></td><td valign="top" align="left">Overall concept idea.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F2-1">Figure 2.1</link></emphasis></td><td valign="top" align="left">The business strategy main elements.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F2-2">Figure 2.2</link></emphasis></td><td valign="top" align="left">Production &#x02013; maintenance relationship.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F2-3">Figure 2.3</link></emphasis></td><td valign="top" align="left">Implementing CBM in support of business drivers.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F2-4">Figure 2.4</link></emphasis></td><td valign="top" align="left">Classification of business drivers.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F2-5">Figure 2.5</link></emphasis></td><td valign="top" align="left">Workflow in building maintenance optimisation models.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F2-6">Figure 2.6</link></emphasis></td><td valign="top" align="left">Maintenance decision making process.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F2-7">Figure 2.7</link></emphasis></td><td valign="top" align="left">Decision framework for predictive maintenance.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-1">Figure 3.1</link></emphasis></td><td valign="top" align="left">Mantis architecture construction approach.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-2">Figure 3.2</link></emphasis></td><td valign="top" align="left">A generalized IoT framework.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-3">Figure 3.3</link></emphasis></td><td valign="top" align="left">The industrial internet of things reference architecture.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-4">Figure 3.4</link></emphasis></td><td valign="top" align="left">The Lambda data processing architecture.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-5">Figure 3.5</link></emphasis></td><td valign="top" align="left">MIMOSA data model diagram.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-6">Figure 3.6</link></emphasis></td><td valign="top" align="left">A three-part CBM architecture in the light of the OSA-CBM.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-7">Figure 3.7</link></emphasis></td><td valign="top" align="left">Overview of the MANTIS reference architecture.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-8">Figure 3.8</link></emphasis></td><td valign="top" align="left">MANTIS interoperability proposed approach.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-9">Figure 3.9</link></emphasis></td><td valign="top" align="left">Model for conceptual integration.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-10">Figure 3.10</link></emphasis></td><td valign="top" align="left">Domain model for MANTIS CPS.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-11">Figure 3.11</link></emphasis></td><td valign="top" align="left">Semantic Data representation and information exchange model.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-12">Figure 3.12</link></emphasis></td><td valign="top" align="left">Value concept specification.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-13">Figure 3.13</link></emphasis></td><td valign="top" align="left">MANTIS event model (based on the IoT-A event information model).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-14">Figure 3.14</link></emphasis></td><td valign="top" align="left">Model for technical integration.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-15">Figure 3.15</link></emphasis></td><td valign="top" align="left">Elements having their digital identity (certificate).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-16">Figure 3.16</link></emphasis></td><td valign="top" align="left">Elements of the BLP model.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-17">Figure 3.17</link></emphasis></td><td valign="top" align="left">Policy hierarchy.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F3-18">Figure 3.18</link></emphasis></td><td valign="top" align="left">Architecture evaluation with Fraunhofer RATE.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-1">Figure 4.1</link></emphasis></td><td valign="top" align="left">SAW device&#x02019;s RLC equivalent circuit.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-2">Figure 4.2</link></emphasis></td><td valign="top" align="left">Signal processing scheme.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-3">Figure 4.3</link></emphasis></td><td valign="top" align="left">First prototype of a SAW sensor.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-4">Figure 4.4</link></emphasis></td><td valign="top" align="left">Second prototype of a SAW sensor.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-5">Figure 4.5</link></emphasis></td><td valign="top" align="left">Standalone SDR platform, antennas and sensor.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-6">Figure 4.6</link></emphasis></td><td valign="top" align="left">Developed GUI showing real time temperature measurements.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-7">Figure 4.7</link></emphasis></td><td valign="top" align="left">Ten year market forecast for printed sensors made in 2014 (in $ million).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-8">Figure 4.8</link></emphasis></td><td valign="top" align="left">Error codes over time, colored based on the tool IDs.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-9">Figure 4.9</link></emphasis></td><td valign="top" align="left">Error codes over time, and corresponding trust values.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-10">Figure 4.10</link></emphasis></td><td valign="top" align="left">BW optimisation by KPI extraction.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-11">Figure 4.11</link></emphasis></td><td valign="top" align="left">Linearly polarized quasi &#x02013; Yagi antennas: Maximum alignment (left) and maximum misalignment (right).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-12">Figure 4.12</link></emphasis></td><td valign="top" align="left">Interferences from different processes and devices in industry.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-13">Figure 4.13</link></emphasis></td><td valign="top" align="left">Example of a 3D model.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-14">Figure 4.14</link></emphasis></td><td valign="top" align="left">Channel emulator scheme.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-15">Figure 4.15</link></emphasis></td><td valign="top" align="left">Schematic outline of a process measured by a sensor with sensing elements A and B.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-16">Figure 4.16</link></emphasis></td><td valign="top" align="left">Sensing Physical Domains.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-17">Figure 4.17</link></emphasis></td><td valign="top" align="left">Categories table header.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-18">Figure 4.18</link></emphasis></td><td valign="top" align="left">Torque sensor.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-19">Figure 4.19</link></emphasis></td><td valign="top" align="left">Filament heat/emission correction.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-20">Figure 4.20</link></emphasis></td><td valign="top" align="left">Sphere Avisense.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-21">Figure 4.21</link></emphasis></td><td valign="top" align="left">Oil condition sensor.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-22">Figure 4.22</link></emphasis></td><td valign="top" align="left">Air/water Flow and temp: SA5000, IFM electronic.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-23">Figure 4.23</link></emphasis></td><td valign="top" align="left">Air/water/temperature sensor: Error signals for self-test result indications.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-24">Figure 4.24</link></emphasis></td><td valign="top" align="left">ZED-THL-M ZigBee sensor for temperature, humidity and light.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-25">Figure 4.25</link></emphasis></td><td valign="top" align="left">Window function (Hysteresis) indications.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-26">Figure 4.26</link></emphasis></td><td valign="top" align="left">Parametric model used to estimate the relationship between the air velocity and the pressure drop.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-27">Figure 4.27</link></emphasis></td><td valign="top" align="left">Kalman filter based process for sensor data fusion.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-28">Figure 4.28</link></emphasis></td><td valign="top" align="left">e-Alert controller to monitor environmental conditions in a medical device.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F4-29">Figure 4.29</link></emphasis></td><td valign="top" align="left">e-Alert control sensor context diagram.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-1">Figure 5.1</link></emphasis></td><td valign="top" align="left">The collaboration concept.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-2">Figure 5.2</link></emphasis></td><td valign="top" align="left">Overview for RCA.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-3">Figure 5.3</link></emphasis></td><td valign="top" align="left">Press machine head critical zones and measured forces at the rods.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-4">Figure 5.4</link></emphasis></td><td valign="top" align="left">RUL calculation at 4.5 104 cycles.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-5">Figure 5.5</link></emphasis></td><td valign="top" align="left">Illustration of Updated pdf of RUL for a REB.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-6">Figure 5.6</link></emphasis></td><td valign="top" align="left">ln(c-factor) vs. linear wear from lab experiments.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-7">Figure 5.7</link></emphasis></td><td valign="top" align="left">Schema of an artificial neural network.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-8">Figure 5.8</link></emphasis></td><td valign="top" align="left">Deep Neural Network with 3 hidden layers.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-9">Figure 5.9</link></emphasis></td><td valign="top" align="left">RUL error vs remaining days to clogging situation.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-10">Figure 5.10</link></emphasis></td><td valign="top" align="left">Overview for Alerting and prediction of failures.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-11">Figure 5.11</link></emphasis></td><td valign="top" align="left">Linearly separable point can be separated in an infinite number of ways.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-12">Figure 5.12</link></emphasis></td><td valign="top" align="left">Optimal hyperplane in a 2D feature vector obtained by maximizing the margin.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-13">Figure 5.13</link></emphasis></td><td valign="top" align="left">Slack variables introduced for not linearly separable problem.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-14">Figure 5.14</link></emphasis></td><td valign="top" align="left">SVM kernel transformation to gain linearly separation.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-15">Figure 5.15</link></emphasis></td><td valign="top" align="left">Principal components of a 2-dimensional data.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-16">Figure 5.16</link></emphasis></td><td valign="top" align="left">Detection of the presence or absence of a 3 &#x000D7; 3 square in the input image by a Convolutional Neural Network.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-17">Figure 5.17</link></emphasis></td><td valign="top" align="left">Illustration of the components of a LSTM cell.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-18">Figure 5.18</link></emphasis></td><td valign="top" align="left">Filter that map the relation between inputs u<emphasis><subscript>t</subscript></emphasis> and output y<emphasis><subscript>t</subscript></emphasis> and generate residuals &#x1D700;<subscript><emphasis>t</emphasis></subscript>.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-19">Figure 5.19</link></emphasis></td><td valign="top" align="left">Error curve for clustering results as a function of the expected number of clusters k.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-20">Figure 5.20</link></emphasis></td><td valign="top" align="left">Average p-significance levels as a function of number of groups for interactions (clusters &#x000D7; customers) on the left, and (clusters &#x000D7; servicers) on the right.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-21">Figure 5.21</link></emphasis></td><td valign="top" align="left">Block diagram describing the predictive recurrent neural network using an Integrated Dimensionality LSTM architecture applied on Litronic message codes.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-22">Figure 5.22</link></emphasis></td><td valign="top" align="left">Result on small pilot test 10K time samples, 64 possible message codes.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-23">Figure 5.23</link></emphasis></td><td valign="top" align="left">ID-LSTM prediction accuracy for both training and testing phases while training the network on subset 9 for 100 epochs.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-24">Figure 5.24</link></emphasis></td><td valign="top" align="left">Example of images in the NEU data set (left) and the newly collected MTEXTURE data set (right).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F5-25">Figure 5.25</link></emphasis></td><td valign="top" align="left">Performance of VGGnet and AlexNet on NEU</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-1">Figure 6.1</link></emphasis></td><td valign="top" align="left">MANTIS human-machine interaction.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-2">Figure 6.2</link></emphasis></td><td valign="top" align="left">MANTIS HMI model.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-3">Figure 6.3</link></emphasis></td><td valign="top" align="left">Categories of Context (Perera, et al., 2014).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-4">Figure 6.4</link></emphasis></td><td valign="top" align="left">Research Fields Influencing the Development of Intelligent User Interfaces.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-5">Figure 6.5</link></emphasis></td><td valign="top" align="left">Context-Dependent Data Integration in Maintenance Scenarios (POWER-OM, Lulea Technology Univ., 2015).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-6">Figure 6.6</link></emphasis></td><td valign="top" align="left">Interaction driven approach for visualization recommender.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-7">Figure 6.7</link></emphasis></td><td valign="top" align="left">Navigation parse and index on ElasticSearch.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-8">Figure 6.8</link></emphasis></td><td valign="top" align="left">Interaction capture and store in ElasticSearch index.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-9">Figure 6.9</link></emphasis></td><td valign="top" align="left">Index structure created in ElasticSearch to capture the interaction of the user with the interface.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-10">Figure 6.10</link></emphasis></td><td valign="top" align="left">Data analysis and visualization workflow in Orange.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-11">Figure 6.11</link></emphasis></td><td valign="top" align="left">MANTIS data visualization in Kibana.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-12">Figure 6.12</link></emphasis></td><td valign="top" align="left">A Kibana graph embedded in the MANTIS HMI prototype.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-13">Figure 6.13</link></emphasis></td><td valign="top" align="left">Machine Data Visualization (Table).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-14">Figure 6.14</link></emphasis></td><td valign="top" align="left">Graphic Visualization of Machine Data.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-15">Figure 6.15</link></emphasis></td><td valign="top" align="left">Maintenance Sensor Data Visualization.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-163">Figure 6.163</link></emphasis></td><td valign="top" align="left">D model of flue gas blower placed at meeting room table.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-17">Figure 6.17</link></emphasis></td><td valign="top" align="left">Marker object representing data from sensor.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-18">Figure 6.18</link></emphasis></td><td valign="top" align="left">Screen capture of the VR demonstrator displaying real measurement data.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-19">Figure 6.19</link></emphasis></td><td valign="top" align="left">FFT and raw vibration data opened for a measurement.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-20">Figure 6.20</link></emphasis></td><td valign="top" align="left">FFT data inspection with Leap Motion controls.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-21">Figure 6.21</link></emphasis></td><td valign="top" align="left">Screenshot showing a super-imposed alarm and the synthetic data that has caused it.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-22">Figure 6.22</link></emphasis></td><td valign="top" align="left">An example of a filled Usability Validation Plan document page 1.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F6-23">Figure 6.23</link></emphasis></td><td valign="top" align="left">An example of a filled Usability Validation Plan document page 2.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-1">Figure 7.1</link></emphasis></td><td valign="top" align="left">Maturity Level before and after the MANTIS project for the techniques in the use cases.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-2">Figure 7.2</link></emphasis></td><td valign="top" align="left">Goals and vision for the shaver manufacturing site.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-3">Figure 7.3</link></emphasis></td><td valign="top" align="left">Basic interactions.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-4">Figure 7.4</link></emphasis></td><td valign="top" align="left">Cumulative explained variance of the principal components.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-5">Figure 7.5</link></emphasis></td><td valign="top" align="left">Control chart showing the Hotelling&#x02019;s T<superscript>2</superscript> score and tolerances to identify trends and outliers in the multivariate data.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-6">Figure 7.6</link></emphasis></td><td valign="top" align="left">Graph of the predicted and the observed quality with PLS regression.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-7">Figure 7.7</link></emphasis></td><td valign="top" align="left">Error codes over time, and corresponding trust values.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-8">Figure 7.8</link></emphasis></td><td valign="top" align="left">Main screen for production operators, showing process performance for eight similar machines. The green circles indicate &#x02018;good&#x02019; behaviour.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-9">Figure 7.9</link></emphasis></td><td valign="top" align="left">Analysis screen of operator dashboard.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-10">Figure 7.10</link></emphasis></td><td valign="top" align="left">Time to failure (in amount of products) for the three main tool types.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-11">Figure 7.11</link></emphasis></td><td valign="top" align="left">Waterproofing the Pajares tunnel in Asturias, Spain.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-12">Figure 7.12</link></emphasis></td><td valign="top" align="left">Valencia lightouse (Spain).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-13">Figure 7.13</link></emphasis></td><td valign="top" align="left">Pedestrian bridge in Madrid (Spain).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-14">Figure 7.14</link></emphasis></td><td valign="top" align="left">Pultrusion machine at ACCIONA workshop.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-15">Figure 7.15</link></emphasis></td><td valign="top" align="left">Monitoring system.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-16">Figure 7.16</link></emphasis></td><td valign="top" align="left">Local HMI. PC, Tablets, mobile devices.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-17">Figure 7.17</link></emphasis></td><td valign="top" align="left">Remote HMI. Home menu.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-18">Figure 7.18</link></emphasis></td><td valign="top" align="left">Remote HMI. Temperature, Humidity &#x00026; Light (THL) menu.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-19">Figure 7.19</link></emphasis></td><td valign="top" align="left">Mixing area at workshop.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-20">Figure 7.20</link></emphasis></td><td valign="top" align="left">Temperature control in the Mixing Area. (Monitoring period 21/02/2018 -06/03/2018).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-21">Figure 7.21</link></emphasis></td><td valign="top" align="left">Humidity control in the Mixing Area. (Monitoring period 21/02/2018 -06/03/2018).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-22">Figure 7.22</link></emphasis></td><td valign="top" align="left">Cooling system from injection chamber.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-23">Figure 7.23</link></emphasis></td><td valign="top" align="left">Schema of the chiller cooling system from injection chamber.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-24">Figure 7.24</link></emphasis></td><td valign="top" align="left">Schema of the injection chambers cooling circuit.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-25">Figure 7.25</link></emphasis></td><td valign="top" align="left">Sensors installed in the refrigeration circuit.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-26">Figure 7.26</link></emphasis></td><td valign="top" align="left">Room temperature vs. liquid inlet temperature control chart.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-27">Figure 7.27</link></emphasis></td><td valign="top" align="left">Refrigeration circuit inlet temperature control chart.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-28">Figure 7.28</link></emphasis></td><td valign="top" align="left">Refrigeration liquid tank level control chart.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-29">Figure 7.29</link></emphasis></td><td valign="top" align="left">Detail of the main compressor that supplies compressed air to the resin injection system.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-30">Figure 7.30</link></emphasis></td><td valign="top" align="left">(a) Schema of the injection chamber resin injection circuit. (b) Detail of the resin injection circuit.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-31">Figure 7.31</link></emphasis></td><td valign="top" align="left">Data record of the evolution of the air pressure obtained from the Remote HMI interface.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-32">Figure 7.32</link></emphasis></td><td valign="top" align="left">Detail of the data record of the evolution of the air pressure obtained from the Remote HMI interface.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-33">Figure 7.33</link></emphasis></td><td valign="top" align="left">Data record of the evolution of the temperature obtained from the Remote HMI interface.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-34">Figure 7.34</link></emphasis></td><td valign="top" align="left">Pressure and temperature sensor installed in the compressed air circuit.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-35">Figure 7.35</link></emphasis></td><td valign="top" align="left">Fagor Arrasate mechanical and servo driven presses.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-36">Figure 7.36</link></emphasis></td><td valign="top" align="left">Clutch Brake (left) and gearboxes (center) by Goizper, used in Fagor&#x02019;s mechanical press machines (right).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-37">Figure 7.37</link></emphasis></td><td valign="top" align="left">Predictive maintenance HMI platform.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-38">Figure 7.38</link></emphasis></td><td valign="top" align="left">Location of data sources.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-39">Figure 7.39</link></emphasis></td><td valign="top" align="left">GOIZPER pneumatic Clutch Brake outside and inside.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-40">Figure 7.40</link></emphasis></td><td valign="top" align="left">MGEP press machine, demonstrator 1.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-41">Figure 7.41</link></emphasis></td><td valign="top" align="left">GOIZPER clutch brake in MGEP demonstrator before MANTIS.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-42">Figure 7.42</link></emphasis></td><td valign="top" align="left">Clutch Brake in MGEP demonstrator with the sensors installed.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-43">Figure 7.43</link></emphasis></td><td valign="top" align="left">GOIZPER Clutch Brake test bench, demonstrator 2.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-44">Figure 7.44</link></emphasis></td><td valign="top" align="left">Local data logger for data capturing and processing.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-45">Figure 7.45</link></emphasis></td><td valign="top" align="left">Architecture design.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-46">Figure 7.46</link></emphasis></td><td valign="top" align="left">FALINK Cloud Platform.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-47">Figure 7.47</link></emphasis></td><td valign="top" align="left">FAGOR ARRASATE future control room.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-48">Figure 7.48</link></emphasis></td><td valign="top" align="left">Static testbed scenario.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-49">Figure 7.49</link></emphasis></td><td valign="top" align="left">Dynamic testbed scenario.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-50">Figure 7.50</link></emphasis></td><td valign="top" align="left">Measurement setup in the press machine.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-51">Figure 7.51</link></emphasis></td><td valign="top" align="left">Wireless sensor node.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-52">Figure 7.52</link></emphasis></td><td valign="top" align="left">Torque measures.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-53">Figure 7.53</link></emphasis></td><td valign="top" align="left">Comparison between estimated and measured torque values (87% of the nominal torque, 100% of the nominal speed and gain 1000).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-54">Figure 7.54</link></emphasis></td><td valign="top" align="left">Selected structural detail from press head.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-55">Figure 7.55</link></emphasis></td><td valign="top" align="left">Fully sensorised welded specimen in the testbed.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-56">Figure 7.56</link></emphasis></td><td valign="top" align="left">Stress plot in the testbed due to a unitary load.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-57">Figure 7.57</link></emphasis></td><td valign="top" align="left">Mechanical test specimen with two sensors based on silver conductive adhesive deposited directly on top of the surface.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-58">Figure 7.58</link></emphasis></td><td valign="top" align="left">(a) Two sensors based on conductive inks placed in one side of the specimen; (b) Commercial crack detection gauge placed on the opposite side of the specimen; (c) Commercial extensometer placed in the crack, during the fatigue test.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-59">Figure 7.59</link></emphasis></td><td valign="top" align="left">Picture of experimental set-up.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-60">Figure 7.60</link></emphasis></td><td valign="top" align="left">Ink signal during the second phase of fatigue test and ink pictures corresponding to the two points indicated in the graph. Pictures of both sides of the test specimen, at the end of the fatigue test.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-61">Figure 7.61</link></emphasis></td><td valign="top" align="left">Bushing seizure model.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-62">Figure 7.62</link></emphasis></td><td valign="top" align="left">OilWear sensor.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-63">Figure 7.63</link></emphasis></td><td valign="top" align="left">Servo driven press machine model diagram.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-64">Figure 7.64</link></emphasis></td><td valign="top" align="left">Estimated and actual angular positions.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-65">Figure 7.65</link></emphasis></td><td valign="top" align="left">Step by step Bayesian soft sensor estimation performance diagram.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-66">Figure 7.66</link></emphasis></td><td valign="top" align="left">Estimated electric torque for different tonnage tests.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-67">Figure 7.67</link></emphasis></td><td valign="top" align="left">Calculated active power for different tonnage tests.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-68">Figure 7.68</link></emphasis></td><td valign="top" align="left">Estimated electric torque for 300mm and 500mm wide tough metal sheet cutting process.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-69">Figure 7.69</link></emphasis></td><td valign="top" align="left">Calculated active power for 300mm and 500mm wide tough metal sheet cutting process.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-70">Figure 7.70</link></emphasis></td><td valign="top" align="left">Calculated active power of centred and off-centred tests.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-71">Figure 7.71</link></emphasis></td><td valign="top" align="left">Estimated electric torque for a centred and off-centred tests.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-72">Figure 7.72</link></emphasis></td><td valign="top" align="left">Edge Tier.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-73">Figure 7.73</link></emphasis></td><td valign="top" align="left">Platform Tier.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-74">Figure 7.74</link></emphasis></td><td valign="top" align="left">Evolution of the Normality Factor over work cycles.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-75">Figure 7.75</link></emphasis></td><td valign="top" align="left">Azure-based Maintenance Cloud Platform.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-76">Figure 7.76</link></emphasis></td><td valign="top" align="left">Clutching and braking velocity and acceleration without slippage.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-77">Figure 7.77</link></emphasis></td><td valign="top" align="left">Clutching and braking velocity and acceleration with slippage at clutching.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-78">Figure 7.78</link></emphasis></td><td valign="top" align="left">Prototyping clutch brake set-up.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-79">Figure 7.79</link></emphasis></td><td valign="top" align="left">Estimated vs actual inner chamber pressure and brake springs stiffness.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-80">Figure 7.80</link></emphasis></td><td valign="top" align="left">Air pressure vs air mass representation.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-81">Figure 7.81</link></emphasis></td><td valign="top" align="left">Left side, brake friction material related curve section. Right side, clutch friction material related curve section.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-82">Figure 7.82</link></emphasis></td><td valign="top" align="left">Air mass flow profile during a single stroke cycle.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-83">Figure 7.83</link></emphasis></td><td valign="top" align="left">Bending process.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-84">Figure 7.84</link></emphasis></td><td valign="top" align="left">Sheet metal machine.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-85">Figure 7.85</link></emphasis></td><td valign="top" align="left">Design of the Proactive Maintenance solution.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-86">Figure 7.86</link></emphasis></td><td valign="top" align="left">Sensor component hardware</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-87">Figure 7.87</link></emphasis></td><td valign="top" align="left">Consumer Remote procedure call (RPC).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-88">Figure 7.88</link></emphasis></td><td valign="top" align="left">Checking the &#x0201C;Bend&#x0201D; variable on a GreenBender machine.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-89">Figure 7.89</link></emphasis></td><td valign="top" align="left">Example data from the controller.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-90">Figure 7.90</link></emphasis></td><td valign="top" align="left">Smart service landscape as a possible result of the Mantis basis work.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-91">Figure 7.91</link></emphasis></td><td valign="top" align="left">Overall scenario of a resulting field configuration and interaction.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-92">Figure 7.92</link></emphasis></td><td valign="top" align="left">Data sources for downstream analytics.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-93">Figure 7.93</link></emphasis></td><td valign="top" align="left">MS Azure demonstrator architecture approach.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-94">Figure 7.94</link></emphasis></td><td valign="top" align="left">Raspberry Pi based sensor based.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-95">Figure 7.95</link></emphasis></td><td valign="top" align="left">Data Logging on the internal CAN bus.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-96">Figure 7.96</link></emphasis></td><td valign="top" align="left">Example for graphical pattern recognition by Spotfire.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-97">Figure 7.97</link></emphasis></td><td valign="top" align="left">Tire changed due to wear and damage.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-98">Figure 7.98</link></emphasis></td><td valign="top" align="left">Agility heat map of analysed forklift trucks.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-99">Figure 7.99</link></emphasis></td><td valign="top" align="left">Second step in architecture to improve process results.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-100">Figure 7.100</link></emphasis></td><td valign="top" align="left">Simple turnout schematic. Moving switch blades are shown in red.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-101">Figure 7.101</link></emphasis></td><td valign="top" align="left">Switch Data &#x02013; Behaviour 1.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-102">Figure 7.102</link></emphasis></td><td valign="top" align="left">Switch Data &#x02013; Behaviour 2.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-103">Figure 7.103</link></emphasis></td><td valign="top" align="left">Switch Data &#x02013; Behaviour 3.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-104">Figure 7.104</link></emphasis></td><td valign="top" align="left">Switch Data &#x02013; Behaviour 4.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-105">Figure 7.105</link></emphasis></td><td valign="top" align="left">Switch Data &#x02013; Behaviour 5.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-106">Figure 7.106</link></emphasis></td><td valign="top" align="left">Identification of bands for quartiles and outliers.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-107">Figure 7.107</link></emphasis></td><td valign="top" align="left">Measurement system setup.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-108">Figure 7.108</link></emphasis></td><td valign="top" align="left">Switch&#x02019;s point blades displacements.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-109">Figure 7.109</link></emphasis></td><td valign="top" align="left">Measurement state diagram.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-110">Figure 7.110</link></emphasis></td><td valign="top" align="left">Graphics, displaying the state and the position of the switch (left) and the instant values of the environmental parameters (right).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-111">Figure 7.111</link></emphasis></td><td valign="top" align="left">Graph displaying blade displacement sensor measurements.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-112">Figure 7.112</link></emphasis></td><td valign="top" align="left">Energy flow in a grid-connected photovoltaic system.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-113">Figure 7.113</link></emphasis></td><td valign="top" align="left">Example of limit checking results for the energy conversion process in a PV plant; performance loss ratios per conversion step are compared to the model for each conversion step.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-114">Figure 7.114</link></emphasis></td><td valign="top" align="left">Flue gas recirculation blower monitored in the pilot project.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-115">Figure 7.115</link></emphasis></td><td valign="top" align="left">Conventional energy production use case platform.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-116">Figure 7.116</link></emphasis></td><td valign="top" align="left">RaspberryPi based measuring system with a mems accelerometer.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-117">Figure 7.117</link></emphasis></td><td valign="top" align="left">IoT-Ticket dashboard utilizing the 3D model built by LapinAMK. Live values are fused into 3D model.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-118">Figure 7.118</link></emphasis></td><td valign="top" align="left">nmas Analysator UI basic view.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-119">Figure 7.119</link></emphasis></td><td valign="top" align="left">nmas Easy condition monitoring device.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-120">Figure 7.120</link></emphasis></td><td valign="top" align="left">REST interface for this use case.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-121">Figure 7.121</link></emphasis></td><td valign="top" align="left">Unplanned system shutdown has a large impact on patients and hospital staff.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-122">Figure 7.122</link></emphasis></td><td valign="top" align="left">Graphic depiction of the healthcare use-case.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-123">Figure 7.123</link></emphasis></td><td valign="top" align="left">Magnetic Resonance Unit.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-124">Figure 7.124</link></emphasis></td><td valign="top" align="left">Image Guided Therapy Unit.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-125">Figure 7.125</link></emphasis></td><td valign="top" align="left">High-level overview of the data architecture.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-126">Figure 7.126</link></emphasis></td><td valign="top" align="left">e-Alert sensor context diagram.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-127">Figure 7.127</link></emphasis></td><td valign="top" align="left">Overview of the main data sources used.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-128">Figure 7.128</link></emphasis></td><td valign="top" align="left">Timeline of gathered categories of data from a single system.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-129">Figure 7.129</link></emphasis></td><td valign="top" align="left">Conversion of features into arrays.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-130">Figure 7.130</link></emphasis></td><td valign="top" align="left">ANN Architecture.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-131">Figure 7.131</link></emphasis></td><td valign="top" align="left">Learning curves.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-132">Figure 7.132</link></emphasis></td><td valign="top" align="left">Example DLT plot.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-133">Figure 7.133</link></emphasis></td><td valign="top" align="left">Segmenting the usage curve.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-134">Figure 7.134</link></emphasis></td><td valign="top" align="left">Timeline of Events (1: failure occurs before maintenance, 2: maintenance occurs before failure).</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-135">Figure 7.135</link></emphasis></td><td valign="top" align="left">Model output default values for flat detector with P=0.80.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-136">Figure 7.136</link></emphasis></td><td valign="top" align="left">Model output default values for flat detector with P=0.15.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-137">Figure 7.137</link></emphasis></td><td valign="top" align="left">Influence of P on expected costs in flat detector case.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-138">Figure 7.138</link></emphasis></td><td valign="top" align="left">Influence of P on expected downtime in flat detector case.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-139">Figure 7.139</link></emphasis></td><td valign="top" align="left">Influence of P on expected costs and downtime for different customers.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-140">Figure 7.140</link></emphasis></td><td valign="top" align="left">Portal to gain access to e-Alert sensor data.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-141">Figure 7.141</link></emphasis></td><td valign="top" align="left">Remote monitoring dashboard high level overview.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F7-142">Figure 7.142</link></emphasis></td><td valign="top" align="left">Remote monitoring dashboard detailed report.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-1">Figure 8.1</link></emphasis></td><td valign="top" align="left">Main maintenance objectives evolution.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-2">Figure 8.2</link></emphasis></td><td valign="top" align="left">Evolution of maintenance techniques.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-3">Figure 8.3</link></emphasis></td><td valign="top" align="left">Bathtub Curve used in reliability engineering.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-4">Figure 8.4</link></emphasis></td><td valign="top" align="left">Aspects of operational reliability.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-5">Figure 8.5</link></emphasis></td><td valign="top" align="left">Different models of collaboration between vender and client.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-6">Figure 8.6</link></emphasis></td><td valign="top" align="left">Margin in product-driven model vs margin in service models.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-7">Figure 8.7</link></emphasis></td><td valign="top" align="left">Degree of platforms automation among organizations.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-8">Figure 8.8</link></emphasis></td><td valign="top" align="left">Customized vertical where solutions based on standard platform.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-9">Figure 8.9</link></emphasis></td><td valign="top" align="left">Top-down approach, where the product has all the information in every step of the value chain.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-10">Figure 8.10</link></emphasis></td><td valign="top" align="left">Smart product as CPS providing new features and functions based on connectivity.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-11">Figure 8.11</link></emphasis></td><td valign="top" align="left">Dimension of managing change in digital transformation.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-12">Figure 8.12</link></emphasis></td><td valign="top" align="left">Essential characteristics of an Agile operating model.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-13">Figure 8.13</link></emphasis></td><td valign="top" align="left">Message broker Apache Kakfka, by LinkedIn.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-14">Figure 8.14</link></emphasis></td><td valign="top" align="left">Service and cost class diagram for the business case model.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-15">Figure 8.15</link></emphasis></td><td valign="top" align="left">Portal de Servi&#x000E7;os financial analysis.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-16">Figure 8.16</link></emphasis></td><td valign="top" align="left">General components of business plan.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-17">Figure 8.17</link></emphasis></td><td valign="top" align="left">Example economic evaluation projection.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-18">Figure 8.18</link></emphasis></td><td valign="top" align="left">Smart maintenance approach in Railway Sector.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F8-19">Figure 8.19</link></emphasis></td><td valign="top" align="left">Maintenance saving cost projection.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F9-1">Figure 9.1</link></emphasis></td><td valign="top" align="left">Unexpected dust accumulation on the color calibrator of a Mars rover. Without actuators for autonomous self maintenance, only the wind can help out.</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="F9-2">Figure 9.2</link></emphasis></td><td valign="top" align="left">Virtual Reality system for maintenance.</td></tr>
</tbody>
</table>
</table-wrap>
</preface>
<preface class="preface" id="preface05">
<title>List of Tables</title>
<table-wrap position="float">
<table cellspacing="5" cellpadding="5" frame="none" rules="none">
<tbody>
<tr><td valign="top" width="15%"><emphasis role="strong"><link linkend="T4-1">Table 4.1</link></emphasis></td><td valign="top" align="left">Low-cost sensors</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T4-2">Table 4.2</link></emphasis></td><td valign="top" align="left">Sample output of the Solar Sensor Check</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T5-1">Table 5.1</link></emphasis></td><td valign="top" align="left">Root Cause Analysis Techniques</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T5-2">Table 5.2</link></emphasis></td><td valign="top" align="left">RUL techniques categories</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T5-3">Table 5.3</link></emphasis></td><td valign="top" align="left">Techniques for Alerting and Prediction of Failures</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T5-4">Table 5.4</link></emphasis></td><td valign="top" align="left">ClusterID vs MachTyp contingency table</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T5-5">Table 5.5</link></emphasis></td><td valign="top" align="left">Results of multiple Monte Carlo experiments (N = 100), for different k. A low average p-significance level indicates an underlying structure for this value of k</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T5-6">Table 5.6</link></emphasis></td><td valign="top" align="left">Results of k-means algorithm repetition</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T5-7">Table 5.7</link></emphasis></td><td valign="top" align="left">Training performance metrics on the different approaches using a batch size of 10</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T5-8">Table 5.8</link></emphasis></td><td valign="top" align="left">Train and Test performance assessment on the different subsets using a batch size set to 10,000</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T5-9">Table 5.9</link></emphasis></td><td valign="top" align="left">Attributes in the two data sets</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T6-1">Table 6.1</link></emphasis></td><td valign="top" align="left">Context Awareness Approaches in a Proactive Collaborative Maintenance Platform</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T7-1">Table 7.1</link></emphasis></td><td valign="top" align="left">List of installed sensors</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T7-2">Table 7.2</link></emphasis></td><td valign="top" align="left">True negatives (OK) and false positives (ALERT, ALARM) detected</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T7-3">Table 7.3</link></emphasis></td><td valign="top" align="left">Decision variables of the model</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T7-4">Table 7.4</link></emphasis></td><td valign="top" align="left">Parameters used in the model</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T7-5">Table 7.5</link></emphasis></td><td valign="top" align="left">Possible scenarios for the maintenance cases that are created</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-1">Table 8.1</link></emphasis></td><td valign="top" align="left">Predictive maintenance solutions to client and vendor</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-2">Table 8.2</link></emphasis></td><td valign="top" align="left">Asset Manufacturer Canvas&#x02019; key dimensions</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-3">Table 8.3</link></emphasis></td><td valign="top" align="left">Asset Service Provider Canvas&#x02019; key dimensions</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-4">Table 8.4</link></emphasis></td><td valign="top" align="left">Asset Service Provider Canvas&#x02019; key dimensions</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-5">Table 8.5</link></emphasis></td><td valign="top" align="left">Strategic Business Models Archetypes</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-6">Table 8.6</link></emphasis></td><td valign="top" align="left">Impact of remote service technology on service business models in manufacturing</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-7">Table 8.7</link></emphasis></td><td valign="top" align="left">Portal de Servi&#x000E7;os cost Business Model</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-8">Table 8.8</link></emphasis></td><td valign="top" align="left">Revenue streams template for current business model and industrial use case</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-9">Table 8.9</link></emphasis></td><td valign="top" align="left">Revenue stream projections template for future PMM Business Model and industrial use case</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-10">Table 8.10</link></emphasis></td><td valign="top" align="left">Structure cost template for technological provider business model projection</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-11">Table 8.11</link></emphasis></td><td valign="top" align="left">Structure costs Template for current business model and industrial use case</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-12">Table 8.12</link></emphasis></td><td valign="top" align="left">Structure costs template for future PMM Business Model and industrial use case</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-13">Table 8.13</link></emphasis></td><td valign="top" align="left">Structure costs template for technological provider use case</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-14">Table 8.14</link></emphasis></td><td valign="top" align="left">Economic evaluation and projection template for industrial use case</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-15">Table 8.15</link></emphasis></td><td valign="top" align="left">Economic evaluation and projection template for technological provider use case</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-16">Table 8.16</link></emphasis></td><td valign="top" align="left">Economic Projection related to current and future PMM Business Model</td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="T8-17">Table 8.17</link></emphasis></td><td valign="top" align="left">Advanced maintenance service program economic projection</td></tr>
</tbody>
</table>
</table-wrap>
</preface>
<preface class="preface" id="preface06">
<title>Contents</title>
<table-wrap position="float">
<table cellspacing="5" cellpadding="5" frame="none" rules="none">
<tbody>
<tr><td valign="top"><emphasis role="strong"><link linkend="ch01">1 Introduction</link></emphasis><?lb?>Urko Zurutuza, Michele Albano, Erkki Jantunen, and Luis Lino Ferreira</td><td valign="top" align="left" width="15%"><ulink url="https://www.riverpublishers.com/pdf/ebook/chapter/RP_9788793609846C1.pdf">Download As PDF</ulink></td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="ch02">2 Business Drivers of a Collaborative, Proactive Maintenance Solution</link></emphasis><?lb?>Erkki Jantunen, Alp Ak&#x000E7;ay, Jaime Campos, Mike Holenderski, Arja Kotkansalo, Riku Salokangas, and Pankaj Sharma</td><td valign="top" align="left" width="15%"><ulink url="https://www.riverpublishers.com/pdf/ebook/chapter/RP_9788793609846C1.pdf">Download As PDF</ulink></td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="ch03">3 The MANTIS Reference Architecture</link></emphasis><?lb?>Csaba Heged&#x00171;s, Patricia Dominguez Arroyo, Giovanni Di Orio, Jos&#x000E9; Luis Flores, Karmele Intxausti, Erkki Jantunen, F&#x000E9;lix Larrinaga, Pedro Mal&#x000F3;, Istv&#x000E1;n Moldov&#x000E1;n, and S&#x000F6;ren Schneickert</td><td valign="top" align="left" width="15%"><ulink url="https://www.riverpublishers.com/pdf/ebook/chapter/RP_9788793609846C1.pdf">Download As PDF</ulink></td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="ch04">4 Monitoring of Critical Assets</link></emphasis><?lb?>Michele Albano, Jos&#x000E9; Manuel Abete, Iban Barrutia Inza, Vito &#x0010C;u&#x0010D;ek, Karel De Brabandere, Ander Etxabe, Iosu Gabilondo, &#x000C7;i&#x000E7;ek G&#x000FC;ven, Mike Holenderski, Aitzol Iturrospe, Erkki Jantunen, Luis Lino Ferreira, Istv&#x000E1;n Moldov&#x000E1;n, Jon Olaizola, Eneko S&#x000E1;enz de Argando&#x000F1;a, Babacar Sarr, S&#x000F6;ren Schneickert, Rafael Socorro, Hans Sprong, Marjan &#x00160;terk, Ra&#x000FA;l Torrego, Godfried Webers, and Achim Woyte</td><td valign="top" align="left" width="15%"><ulink url="https://www.riverpublishers.com/pdf/ebook/chapter/RP_9788793609846C1.pdf">Download As PDF</ulink></td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="ch05">5 Providing Proactiveness: Data Analysis Techniques Portfolios</link></emphasis><?lb?>Alberto Sillitti, Javier Fernandez Anakabe, Jon Basurko, Paulien Dam, Hugo Ferreira, Susana Ferreiro, Jeroen Gijsbers, Sheng He, Csaba Heged&#x00171;s, Mike Holenderski, Jan-Otto Hooghoudt, I&#x000F1;igo Lecuona, Urko Leturiondo, Quinten Marcelis, Istv&#x000E1;n Moldov&#x000E1;n, Emmanuel Okafor, Cl&#x000E1;udio Rebelo de S&#x000E1;, Ricardo Romero, Babacar Sarr, Lambert Schomaker, Arvind Kumar Shekar, Carlos Soares, Hans Sprong, S&#x000F8;ren Theodorsen, Tom Tourw&#x000E9;, Gorka Urchegui, Godfried Webers, Yi Yang, Andriy Zubaliy, Ekhi Zugasti, and Urko Zurutuza</td><td valign="top" align="left" width="15%"><ulink url="https://www.riverpublishers.com/pdf/ebook/chapter/RP_9788793609846C1.pdf">Download As PDF</ulink></td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="ch06">6 From KPI Dashboards to Advanced Visualization</link></emphasis><?lb?>Goreti Marreiros, Peter Craamer, I&#x000F1;aki Garitano, Roberto Gonz&#x000E1;lez, Manja Gorenc Novak, Ale&#x00161; Kancilija, Quinten Marcelis, Diogo Martinho, Antti Niemel&#x000E4;, Franc Novak, Gregor Papa, &#x00160;pela Poklukar, Isabel Pra&#x000E7;a, Ville Rauhala Daniel Reguera, Marjan &#x00160;terk, Gorka Urchegui, Roberto Uribeetxeberria, Juha Valtonen, and Anja Vidmar</td><td valign="top" align="left" width="15%"><ulink url="https://www.riverpublishers.com/pdf/ebook/chapter/RP_9788793609846C1.pdf">Download As PDF</ulink></td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="ch07">7 Success Stories on Real Pilots</link></emphasis><?lb?>Rafael Socorro, Mar&#x000ED;a Aguirregabiria, Alp Ak&#x000E7;ay, Michele Albano, Mikel Anasagasti, Andoitz Aranburu, Mauro Barbieri, Iban Barrutia, Ansgar Bergmann, Karel De Brabandere, Marcel Boosten, Rui Casais, David Chico, Paolo Ciancarini, Paulien Dam, Giovanni Di Orio, Karel Eerland, Xabier Eguiluz, Salvatore Esposito, Catarina F&#x000E9;lix, Javier Fernandez-Anakabe, Hugo Ferreira, Luis Lino Ferreira, Attila Frank&#x000F3;, Iosu Gabilondo, Raquel Garc&#x000ED;a, Jeroen Gijsbers, Mathias Gr&#x000E4;dler, Csaba Heged&#x00171;s, Silvia Hern&#x000E1;ndez, Petri Helo, Mike Holenderski, Erkki Jantunen, Matti Kaija, Ale&#x00161; Kancilija, F&#x000E9;lix Larrinaga Barrenechea, Pedro Mal&#x000F3;, Goreti Marreiros, Eva Mart&#x000ED;nez, Diogo Martinho, Asif Mohammed, Mikel Mondragon, Istv&#x000E1;n Moldov&#x000E1;n, Antti Niemel&#x000E4;, Jon Olaizola, Gregor Papa, &#x00160;pela Poklukar, Isabel Pra&#x000E7;a, Stefano Primi, Verus Pronk, Ville Rauhala, Mario Riccardi, Rafael Rocha, Jon Rodriguez, Ricardo Romero, Antonio Ruggieri, Oier Sarasua, Eduardo Saiz, Veli-Pekka Salo, M&#x000F3;nica S&#x000E1;nchez, Paolo Sannino, Babacar Sarr, Alberto Sillitti, Carlos Soares, Hans Sprong, Daan Terwee, Bas Tijsma, Tom Tourw&#x000E9;, Nayra Uranga, Lauri V&#x000E4;limaa, Juha Valtonen, P&#x000E1;l Varga, Alejandro Veiga, Mikel Viguera, Jaap van der Voet, Godfried Webers, Achim Woyte, Kees Wouters, Ekhi Zugasti and Urko Zurutuza</td><td valign="top" align="left" width="15%"><ulink url="https://www.riverpublishers.com/pdf/ebook/chapter/RP_9788793609846C1.pdf">Download As PDF</ulink></td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="ch08">8 Business Models: Proactive Monitoring and Maintenance</link></emphasis><?lb?>Michel I&#x000F1;igo Ulloa, Peter Craamer, Salvatore Esposito, Carolina Mej&#x000ED;a Ni&#x000F1;o, Mario Riccardi, Antonio Ruggieri, and Paolo Sannino</td><td valign="top" align="left" width="15%"><ulink url="https://www.riverpublishers.com/pdf/ebook/chapter/RP_9788793609846C1.pdf">Download As PDF</ulink></td></tr>
<tr><td valign="top"><emphasis role="strong"><link linkend="ch09">9 The Future of Maintenance</link></emphasis><?lb?>Lambert Schomaker, Michele Albano, Erkki Jantunen, and Luis Lino Ferreira</td><td valign="top" align="left" width="15%"><ulink url="https://www.riverpublishers.com/pdf/ebook/chapter/RP_9788793609846C1.pdf">Download As PDF</ulink></td></tr>
</tbody>
</table>
</table-wrap>
</preface>
<chapter class="chapter" id="ch01" label="1" xreflabel="1">
<title>Introduction</title>
<para><emphasis role="strong">Urko Zurutuza<superscript><emphasis role="strong">1</emphasis></superscript>, Michele Albano<superscript><emphasis role="strong">2</emphasis></superscript>, Erkki Jantunen<superscript><emphasis role="strong">3</emphasis></superscript>, and Luis Lino Ferreira<superscript><emphasis role="strong">2</emphasis></superscript></emphasis></para>
<para><superscript>1</superscript>Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para><superscript>2</superscript>ISEP, Polytechnic Institute of Porto, Porto, Portugal</para>
<para><superscript>3</superscript>VTT Technical Research Centre of Finland Ltd, Finland</para>
<para>Current advances in ICT are shaping the way we live and work. Current communication solutions allow people to be &#x0201C;always online, always connected&#x0201D;, and similar technologies are applied to machine communication, also known as M2M communication. In particular, research and development efforts are devoted to extending the IoT to maintenance, thus providing ubiquitous access through the Internet to industrial systems under maintenance.</para>
<para>Two cornerstones for this (r)evolution, which is part of Industry 4.0, are the utilization of CPS in maintenance contexts, and leveraging data collected in the field by means of techniques from the Artificial Intelligence family. In this context, and based on these two cornerstone technologies, the MANTIS project (Cyber Physical System based Proactive Collaborative Maintenance) was born, to realize platforms that can perform CBM and PM in real industrial contexts.</para>
<para>In this vision, physical systems (e.g., industrial machines, vehicles, renewable energy assets) and the environment they operate in are monitored continuously by a broad and diverse range of intelligent sensors, resulting in massive amounts of data that characterise the usage history, operational condition, location, movement and other physical properties of those systems. These CPS form part of a larger network of heterogeneous and collaborative systems (e.g., vehicle fleets or photovoltaic and windmill parks) connected via robust communication mechanisms able to operate in challenging environments [Jantunen et al., 2017].</para>
<para>Sophisticated distributed sensing and decision-making functions are performed at different levels in a collaborative way ranging from (i) local nodes that pre-process raw sensor data and extract relevant information before transmitting them, thereby reducing bandwidth requirements of communication, (ii) over intermediate nodes that offer asset-specific analytics to locally optimise performance and maintenance, (iii) to cloud-based platforms that integrate information from ERP, CRM and CMMS systems and execute distributed processing and analytics algorithms for global decision-making [Jantunen et al., 2018].</para>
<para>The resulting technological ecosystem can empower methodologies such as CBM, which aims at predicting the condition of machinery based on the parameters or conditions of the equipment, in which some limits are established and the behaviour of such parameters are verified through different strategies. The objective of CBM is to provide the maximum objective data of the equipment to identify and avoid possible failures that could generate non-desired downtime in advance.</para>
<para>PM moves the bar one step further, and considers how the collected data have to be processed to extract more information, for example by allowing intelligent software to learn the behaviour of equipments in terms of the monitored parameters, and thus to identify outliers that correspond to potential problems. The modelling of faults can lead to the prognosis about the condition of machinery, comprising approximate information regarding when and how faults can present themselves given the current condition of the machine, the environment it is working in, and data collected on similar machines.</para>
<para>Throughout this book, authors mention both CBM and PM terms repeatedly.</para>
<para>As we have made clear here, CBM is based on using real-time data to prioritize and optimize maintenance resources. Observing the state of the system is known as condition monitoring. Such a system determines the equipment&#x02019;s health, and acts only when maintenance is actually necessary [Wikipedia]. The work in [Jardine et al., 2006] defines CBM as a maintenance program that recommends maintenance decisions based on the information collected through condition monitoring, consisting in three main steps: data acquisition (to get condition data), data processing and maintenance decision-making (to aid in diagnosis and prognosis functions).</para>
<para>Thus, PM uses a CBM strategy, and considers the detection and correction of root cause conditions that would otherwise lead to failure.</para>
<para>This introductory chapter draws a picture of the current trends in maintenance, and then summarizes the content of the rest of the book by providing hints to what PM can be and can do for the industry.</para>
<section class="lev1" id="sec1-1">
<title>1.1 Maintenance Today</title>
<para>Maintenance in industry today focuses on aspects related to availability, profitability and safety [Holmberg et al., 2010; Paz and Leigh, 1994]. Companies wish to improve OEE, availability, performance and quality to supply better products, reducing the costs and increasing profitability and safety as much as possible [Ferri et al., 2012]. In fact, the economic importance of the machinery and equipment maintenance is huge, as hinted for example by the current average level of OEE is 50% i.e., there is a hidden factory behind every factory. The current view of the maintenance primary task is to keep the machinery and equipment constantly in working order to perform their intended functions [Swedish Standards Institute, 2010]. The definitions available in current literature are very close to each other and encompass the following basic assumptions [Mikkonen, 2009]:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Maintenance tends to ensure that machinery remain in working order or that they are restored to normal operating condition;</para></listitem>
<listitem>
<para>Maintenance includes all technical, administrative and management actions implemented during the lifetime of a machine.</para></listitem>
</itemizedlist>
<para>The importance of condition monitoring is also highlighted in traditional industries, where the shortcomings in maintenance activities often end in decreased performance and quality. This obviously results in economic losses [Mikkonen, 2009]. Traditional maintenance methods are not cost optimised, and therefore new more effective means are being developed. In fact, maintenance today is moving in the direction of intelligent maintenance, and concepts such as CBM are gaining ground where efficiency is sought by means of information technology.</para>
</section>
<section class="lev1" id="sec1-2">
<title>1.2 The Path to Proactive Maintenance</title>
<para>The main objective of MANTIS is set to &#x0201C;develop a Cyber Physical System based Proactive Maintenance Service Platform Architecture enabling Collaborative Maintenance Ecosystems&#x0201D;.</para>
<para>For an optimum maintenance of assets, different systems and stakeholders have to share information, resources and responsibilities, in other words, collaboration is required. Such a Collaborative Maintenance Ecosystem aims to:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Reduce the adverse impact of maintenance on productivity and costs;</para></listitem>
<listitem>
<para>Increase the availability of assets;</para></listitem>
<listitem>
<para>Reduce time required for maintenance tasks;</para></listitem>
<listitem>
<para>Improve the quality of the maintenance service and products;</para></listitem>
<listitem>
<para>Improve labor working conditions and maintenance performance;</para></listitem>
<listitem>
<para>Increase sustainability by preventing material losses (due to out-of-tolerance production).</para></listitem>
</itemizedlist>
<para>PM commissions corrective actions aimed at sources of failure. It is designed to extend the life of mechanical machinery as opposed to 1) making repairs when often nothing is broken, 2) accommodating failure as routine and normal, and 3) pre-empting crisis failure maintenance &#x02013; all of which are characteristics of the predictive/preventive disciplines.</para>
<para>A PM platform has to enable service-based business models and improved asset availability at lower costs through continuous process and equipment monitoring and data analysis. With this goal in mind, MANTIS also aims to identify and integrate critical information from other sources such as production, maintenance, equipment manufacturers and service providers. This service platform architecture has to take into account the needs of industries in the forefront of service-based business and operations as well as less mature ones, allowing improvements in maintenance to be achieved gradually and consistently.</para>
<para>The PM service platform consists of distributed processing chains that efficiently transform raw data into knowledge while minimising the need for transfer bandwidth. Chasing this overall objective gives raise to the need for a smart integrated domain knowledge system with advanced data monitoring, communication and analytics with self-learning capabilities, which themselves have to be overall dependable and secure. Thus, this chain includes key technologies such as (<link linkend="F1-1">Figure <xref linkend="F1-1" remap="1.1"/></link>):</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Smart sensors, actuators and cyber-physical systems capable of local pre-processing and local data storing/buffering;</para></listitem>
<listitem>
<para>Robust communication systems for harsh environments;</para></listitem>
<listitem>
<para>Distributed machine learning tools for data validation and decision-making;</para></listitem>
<listitem>
<para>Cloud-based processing, analytics and data availability;</para></listitem>
<listitem>
<para>HMI to provide the right information to the right people at the right time in the right format.</para></listitem>
</itemizedlist>
<fig id="F1-1" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F1-1">Figure <xref linkend="F1-1" remap="1.1"/></link></label>
<caption><para>Overall concept idea.</para></caption>
<graphic xlink:href="graphics/ch01_Fig1_1.jpg"/>
</fig>
</section>
<section class="lev1" id="sec1-3">
<title>1.3 Why to Read this Book</title>
<para>In this book, different current maintenance activities are discussed, and, while focusing on the particular case of CBM and PM, it also considers future trends and technologies that can support maintenance.</para>
<para>Especially for industrial readers, the book has a section focused on the economic aspects and it introduces some potential methods on how to evaluate the cost benefits of PM and justify the implementation of preventive maintenance.</para>
<para>CBM (and therefore PM), is broken up into processing phases, according to the MIMOSA standard, and later on the processing phases are related to the tools that could be used to fulfil these processing phases.</para>
<para>One further objective of the book is a focus on how the integration between different CBM systems can be done and the benefits of creating an integrated system.</para>
</section>
<section class="lev1" id="sec1-4">
<title>References</title>
<orderedlist numeration="arabic" continuation="restarts" spacing="normal">
<listitem>
<para>Ferri, G., Fumagalli, L., Jantunen, E., Salokangas, R., and Macchi M. (2012). <emphasis>Simulation tool development to support customer-supplier relationship for cbm services</emphasis>, In: 11th International Probabilistic Safety Assessment Management Conference (PSAM11), pp. 365&#x02013;374. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ferri%2C+G%2E%2C+Fumagalli%2C+L%2E%2C+Jantunen%2C+E%2E%2C+Salokangas%2C+R%2E%2C+and+Macchi+M%2E+%282012%29%2E+Simulation+tool+development+to+support+customer-supplier+relationship+for+cbm+services%2C+In%3A+11th+International+Probabilistic+Safety+Assessment+Management+Conference+%28PSAM11%29%2C+pp%2E+365-374%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Holmberg, K., Jantunen, E., Adgar, A., Arnaiz, A., Mascolo, J., and Mekid, S. (2010). <emphasis>E-maintenance.</emphasis> Springer-Verlag, London. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Holmberg%2C+K%2E%2C+Jantunen%2C+E%2E%2C+Adgar%2C+A%2E%2C+Arnaiz%2C+A%2E%2C+Mascolo%2C+J%2E%2C+and+Mekid%2C+S%2E+%282010%29%2E+E-maintenance%2E+Springer-Verlag%2C+London%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Jantunen, E., Gorostegi, U., Zurutuza, U., Larringa, F., Albano, M., di Orio, G., Mal&#x000F3;, P., and Heged&#x000FC;s, C. (2017). The way cyber physical systems will revolutionise maintenance. In <emphasis>30th Conference on Condition Monitoring and Diagnostic Engineering Management (COMADEM)</emphasis>, July 10&#x02013;13, 2017, Preston, UK. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Jantunen%2C+E%2E%2C+Gorostegi%2C+U%2E%2C+Zurutuza%2C+U%2E%2C+Larringa%2C+F%2E%2C+Albano%2C+M%2E%2C+di+Orio%2C+G%2E%2C+Mal%F3%2C+P%2E%2C+and+Heged%FCs%2C+C%2E+%282017%29%2E+The+way+cyber+physical+systems+will+revolutionise+maintenance%2E+In+30th+Conference+on+Condition+Monitoring+and+Diagnostic+Engineering+Management+%28COMADEM%29%2C+July+10-13%2C+2017%2C+Preston%2C+UK%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Jantunen, E., Gorostegui, U., Zurutuza, U., Albano, M., Ferreira, L. L., Heged&#x000FC;s, C., and Campos, J. (2018). Remote maintenance support with the aid of cyber-physical systems and cloud technology. <emphasis>Proceedings of the Institution of Mechanical Engineers, Part I: Journal of Systems and Control Engineering</emphasis>, Sage Publishing. May 2018.</para></listitem>
<listitem>
<para>Jardine, A. K., Lin, D., and Banjevic, D. (2006). A review on machinery diagnostics and prognostics implementing condition-based maintenance. <emphasis>Mechanical systems and signal processing</emphasis>, 20(7), 1483&#x02013;1510. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Jardine+%282017%29+Jardine%2C+A%2E+K%2E%2C+Lin%2C+D%2E%2C+and+Banjevic%2C+D%2E+%282006%29%2E+A+review+on+machinery+diagnostics+and+prognostics+implementing+condition-based+maintenance%2E+Mechanical+systems+and+signal+processing%2C+20%287%29%2C+1483-1510%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Mikkonen, H. (2009). <emphasis>Kuntoon perustuva kunnossapito</emphasis>, KP-Media Oy, ISBN 978-952-99458-4-9. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Mikkonen%2C+H%2E+%282009%29%2E+Kuntoon+perustuva+kunnossapito%2C+KP-Media+Oy%2C+ISBN+978-952-99458-4-9%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Paz, N. M., and Leigh, W. (1994). <emphasis>Maintenance scheduling: Issues, results and research needs.</emphasis> International Journal of Operations &#x00026; Production Management, 14(8), pp. 47&#x02013;69. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Paz%2C+N%2E+M%2E%2C+and+Leigh%2C+W%2E+%281994%29%2E+Maintenance+scheduling%3A+Issues%2C+results+and+research+needs%2E+International+Journal+of+Operations+%26+Production+Management%2C+14%288%29%2C+pp%2E+47-69%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Swedish Standards Institute. (2010). <emphasis>Maintenance &#x02013; maintenance terminology &#x02013; ss-en 13306:2010</emphasis>.</para></listitem>
<listitem>
<para>Wikipedia Wikipedia. https://en.wikipedia.org/wiki/Condition-based_maintenance</para></listitem>
</orderedlist>
</section>
</chapter>
<chapter class="chapter" id="ch02" label="2" xreflabel="2">
<title>Business Drivers of a Collaborative, Proactive Maintenance Solution</title>
<para><emphasis role="strong">Erkki Jantunen<superscript><emphasis role="strong">1</emphasis></superscript>, Alp Ak&#x000E7;ay<superscript><emphasis role="strong">2</emphasis></superscript>, Jaime Campos<superscript><emphasis role="strong">3</emphasis></superscript>, Mike Holenderski<superscript><emphasis role="strong">2</emphasis></superscript>, Arja Kotkansalo<superscript><emphasis role="strong">4</emphasis></superscript>, Riku Salokangas<superscript><emphasis role="strong">1</emphasis></superscript>, and Pankaj Sharma<superscript><emphasis role="strong">5</emphasis></superscript></emphasis></para>
<para><superscript>1</superscript>VTT Technical Research Centre of Finland Ltd, Finland</para>
<para><superscript>2</superscript>Technische Universiteit Eindhoven, The Netherlands</para>
<para><superscript>3</superscript>Department of Informatics, Linnaeus University, Sweden</para>
<para><superscript>4</superscript>Lapland University of Applied Sciences Ltd., Finland</para>
<para><superscript>5</superscript>Mechanical Engineering Department, IIT Delhi, India</para>
<para>The elevated complexity and costs of production assets combined with the requirements for high-quality manufactured products necessitate novel design and condition-based maintenance approaches that are able to provide the required levels of availability, maintainability, quality and safety while decreasing the cost of the system as a whole and throughout the production lifecycle.</para>
<section class="lev1" id="sec2-1">
<title>2.1 Introduction</title>
<para>A few years ago, maintenance was considered only a cost factor in industry, but the situation has changed radically over the last years. There are still companies whose maintenance strategy is based solely on the fact that the machine is repaired only at the stage when one of its components breaks down. For certain types of businesses, such a strategy may be working, where the degradation of a component does not generate major costs, for example.</para>
<para>However, mainly unpredictable degradation of industrial machines causes production standstills and thus high costs. The availability of spare parts is also affecting the costs, as they may not be immediately available. On the other hand, keeping all spare parts in the stock while waiting for the degradation of any component is not suitable, since the stocking itself is expensive. On top of this comes the cost of time spent on the repair of the machine.</para>
<para>The implementation of a PM strategy, and the application of a CBM platform, are not only a technological challenge but a deeper question of new business models and their acceptance. The problem is that it has been claimed that maintenance personnel tend to be conservative and not prone to change.</para>
<para>On a conceptual level, the economic impact of a CBM platform may be expressed as:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Reducing the total cost of ownership &#x02013; CBM allows preventive and corrective actions to be scheduled at the optimal time, and concurs to the implementation of a PM strategy;</para></listitem>
<listitem>
<para>Reducing maintenance costs &#x02013; Reduce equipment downtime, lower maintenance costs and improve equipment life cycle management;</para></listitem>
<listitem>
<para>Increased safety where an unpredictable catastrophic system failure could cause severe damages, e.g., an explosion.</para></listitem>
</itemizedlist>
<section class="lev2" id="sec2-1-1">
<title>2.1.1 CBM-based PM in Industry</title>
<para>One of the main reasons and the single most common cause for the damage of the machines is overloading or improper use. For the person in charge of maintenance, it is important to monitor the use of the machine and guide the user so that the machine does not wear any more than in normal use. In addition, it is important to control the use of the machine so that the machine works as efficient as possible and does not cause any unnecessary environmental burdens. Typically, the OEE is measured by performance, availability and quality. When one or more of these elements are out of operating range, profits are declining.</para>
<para>The aim of the production plant maintenance is continuous improvement of the factory assets and increased OEE. Plant maintenance enables one or more algorithms to be assigned to each asset or asset type. The algorithms can be based on an asset&#x02019;s operational status (e.g., a pump&#x02019;s total running time) or its condition (e.g., vibrations or temperature), or a combination of both. The useful set of parameters are based on analysis of the system&#x02019;s critical components and their behaviour.</para>
<para>Using real-time data from the asset, the plant maintenance system can automatically inform maintenance personnel when it is time for inspection or corrective action well in advance. This way, the maintenance can be scheduled to interfere as little as possible with the production. Job performance can be guided, e.g., with a mobile device, and add a pictorial execution instructions or exploit augmented reality. Reporting work should be carried out as detailed as possible, because only in this way it can help to increase the availability of the machine.</para>
<para>Inventory management optimization is also an important issue requiring that both location(s) and the number of spare parts must be optimized. The ordering system itself should be automatic and deliveries should go to the right place at the right time. In addition, guidance should be available, that means installation and operating instructions, as well as access to the manufacturer&#x02019;s data.</para>
<para>CBM in plant maintenance system is not intended to replace established maintenance procedures. Rather, it improves the effectiveness and increases the value of such systems. As well as helping to optimize maintenance routines, it also provides a basis for continual improvement, e.g., allowing users to compare the performance of several assets of the same type and then applying the best maintenance practice across that group.</para>
</section>
<section class="lev2" id="sec2-1-2">
<title>2.1.2 CBM-based PM in Service Business</title>
<para>Advances in information technology and the intensification of competition has forced companies to a situation where the significance of the service business in the market has increased significantly. Previously, profit was particularly due to new sales and post marketing. After-sale services were not important aspects. However, nowadays, many appliance manufacturers post-market will take up to 30&#x02013;50% of their net sales [Mikkonen et al., 2009].</para>
<para>As a result, companies are becoming interested in the services they can provide to their customers and at the same time, gather information about their own products, such as design improvements, maintenance or warranty analysis. This information makes possible to offer customers even better services during the product&#x02019;s life cycle.</para>
<para>New sophisticated data systems in production assets enables various types of new business models. One promising avenue is MaaS, which is a close relative to Power-by-the-hour concept. Drawing on the capabilities now afforded by new wireless sensors, advanced connectivity and cloud computing, OEMs can innovate in the areas of machine support and maintenance packages. By empowering remote access and real-time monitoring capabilities, new HMI (Human Machine Interface) software makes it possible to offer preventive maintenance, quick troubleshooting advice or the ability to access machines and screens remotely for real-time service and operational parameters.</para>
<para>Customers are also particularly interested in the factors affecting the overall performance of the equipment. As it has been mentioned before, the availability, efficiency and quality can be influenced and improved by CBM-based PM avoiding unforeseen problems, which may become a determinant factor in the cost of a product.</para>
<para>This business model is becoming increasingly popular among many industries. For instance, software is typically downloaded to a computer and its maintenance, or updates, are run automatically if the monthly licence fee is payed. Some advanced cars are also gathering data and communicating with the factory about their health status and need for maintenance. Manufacturers are able to use this real driving condition data as basis to their product development, an extra asset gained as side product.</para>
</section>
<section class="lev2" id="sec2-1-3">
<title>2.1.3 Life Cycle Cost and Overall Equipment Effectiveness</title>
<para>The main costs affecting to the asset within its life-cycle are the purchase price, use of the product and the costs associated with maintenance. The OEE has a major impact to the product&#x02019;s lifecycle costs throughout its life-cycle. All the decisions that are made regarding the design and manufacture of the machine affect its performance, availability, quality, safety, reliability and maintainability, and ultimately decide the purchase price and the costs associated with the ownership and disposal.</para>
<para>LCC is the tool to optimize this kind of economical challenge. It is most effective when it is carried out at an early design stage of the machine, making it possible to optimize the basic structure of the machine. It should also be updated and used in the following stages of the life cycle, when looking for significant cost uncertainties and risks (IEC 60300-3-3, 2004). In addition, LCC can be effectively applied to evaluate the costs associated with a specific activity, for example, the effects of different maintenance strategies, to cover a specific part of a product, or to cover only selected phase or phases of a product&#x02019;s life cycle. Typical LCC analysis is as follows, according to IEC-60300-3-3, 2004:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>LCC plan;</para></listitem>
<listitem>
<para>LCC model selection or development;</para></listitem>
<listitem>
<para>LCC model application;</para></listitem>
<listitem>
<para>LCC documentation;</para></listitem>
<listitem>
<para>Review of LCC results;</para></listitem>
<listitem>
<para>Analysis update.</para></listitem>
</itemizedlist>
</section>
<section class="lev2" id="sec2-1-4">
<title>2.1.4 Integrating IoT with Old Equipment</title>
<para>In the previous sections, the cost of the life cycle of a product has been discussed. Usually, when a machine becomes obsolete or needs a repair, it is necessary to make an investment to fix the problem.</para>
<para>However, with new technologies such as IoT it is possible to delay this obsolescence and to monitor the wearing and condition of the machines. The IoT means having the ability to access data about the condition of the machine or product from anywhere instantaneously. By using wireless communications, managers or maintenance technicians can measure the health of their infrastructure and take proper action.</para>
<para>IoT can help companies improve their operations by using their own data. First of all, the data needs to be collected from various sources to make a study and gather meaningful information. This is usually done by integrating different sensors in the machines and doing various tests. The collected data are then used to develop algorithms that will monitor the status of the machine. If the algorithm detects that the machine is going to have a problem or if it already has one, an error or alarm will be sent to the proper manager.</para>
<para>As an example, a manager could install a temperature sensor in an old valve, where the sensor would not have any effect on its operation, and monitor its activity to see if the valve needs to be changed or it has started to wear in the most critical parts. If that is the case, the system could send an SMS to the manager informing which valve has to be changed. Depending on the automation of the system, it could also make an order of the parts needed so that when the manager gets to the valve, the spare part is already there.</para>
<para>This way, the break-fix mentality is eliminated and substituted with a PM approach. Improving the equipment by using IoT monitoring could lead to a reduction in costs and a prolonged life cycle of the machines.</para>
</section>
<section class="lev2" id="sec2-1-5">
<title>2.1.5 CBM Strategy as a Maintenance Business Driver</title>
<para>There are many recognised business drivers and the implemented or relevant ones depend on the market where the company performs its business. However, the typical business drivers are profit and growth, and different stakeholder&#x02019;s expectations, which are normally connected with a company&#x02019;s policies and its productivity [Lozano, 2015].</para>
<para>Financial Times (lexicon.ft.com) defines business driver as &#x0201C;A descriptive rationale, ideally measurable, used to support a business vision or project to clarify why a change or completely new direction is necessary&#x0201D;.</para>
<para>The business drivers are applied to achieve effective prioritising within a consolidated programme of business initiatives [Ward and Peppard, 2002]. Therefore, the business drivers are a set of critical forces that the business must consider. They usually characterise short, medium and long-term aspects that the business must include with the aim to meet the objectives and satisfy the CSFs. Thus, the business drivers are the core constitutes of a business strategy.</para>
<para>A business strategy consists of the mission, i.e., a company&#x02019;s primary role to set a direction for the company/department to follow, namely its mission on the market. The vision is about the future aims, and it is normally visualised as a common picture that everyone agrees on about the company. The business drivers are a set of forces that need to be met to be able to compete in the specific market, and the objectives are the targets the company/departments set to achieve its vision. The strategies define how these objectives will be met. The CSFs are a small number of key areas where things must go right for the business to flourish and be able to compete under the same (or better) conditions than the competitors. In addition, the CSFs also provide information about the actions one should take as well as how they should be measured to confirm that they have been achieved. <link linkend="F2-1">Figure <xref linkend="F2-1" remap="2.1"/></link>, highlights the above mentioned. It shows as well some examples of objectives in connection with improved cost/financial indicators. In addition, it highpoints the measurements and actions (CSFs) related to the objectives.</para>
<fig id="F2-1" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F2-1">Figure <xref linkend="F2-1" remap="2.1"/></link></label>
<caption><para>The business strategy main elements.</para></caption>
<graphic xlink:href="graphics/ch02_fig001.jpg"/>
</fig>
<para>Thus, the departmental strategy must be connected with the business strategy, as is the case of the asset maintenance management strategy and overall business strategy.</para>
<para>Accordingly, in the area of industrial asset management, the maintenance department might implement a variety of maintenance strategies (see <link linkend="ch01">Chapter <xref linkend="ch01" remap="1"/></link>). The corrective is immediate or deferred, while preventive can be on condition or predetermined. The preventive maintenance (including reactive, predictive and proactive maintenance) is where the CBM falls in (is classified), and it is preferable to use whenever it is appropriate.</para>
<para>In CBM the equipment is periodically inspected by manual or automatic systems so that their condition may be assessed and the rate of degradations can be identified by data analytics. According to British Standards [BS 3811, 1993], CBM is the maintenance carried out according to the need indicated by condition monitoring. Then action is taken based on the CM information. The most common CM techniques used for CBM are visual monitoring, performance monitoring, vibration monitoring, and wear debris analysis.</para>
<fig id="F2-2" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F2-2">Figure <xref linkend="F2-2" remap="2.2"/></link></label>
<caption><para>Production &#x02013; maintenance relationship (modified from Ben-Daya and Duffuaa, 1995).</para></caption>
<graphic xlink:href="graphics/ch02_fig002.jpg"/>
</fig>
<para>The need for maintenance emerges from the production department that needs to have its equipment in good health to be able to run the production smoothly and efficiently, see <link linkend="F2-2">Figure <xref linkend="F2-2" remap="2.2"/></link>.</para>
<para>The connection between the three crucial components of any production system is the production, quality and asset management function, which are shown in <link linkend="F2-2">Figure <xref linkend="F2-2" remap="2.2"/></link>. The secondary output of production is maintenance (shown as demand for service), whose output is an increased production capacity. Both the production process and the quality of the maintenance work, which, in turn, affects equipment condition, affect the state and quality of the final product.</para>
<para>The asset management needs for service can be highlighted as business drivers, and the objective of PM is to avoid, diminish, or identify the onset of failure using diagnostic techniques [Ben-Daya and Duffuaa, 1995].</para>
<para>The expected benefits, i.e., business drivers from equipment condition monitoring, fall into two general categories, namely Financial and Soft benefits [Spare, 2001]. <link linkend="F2-3">Figure <xref linkend="F2-3" remap="2.3"/></link> shows the relationships between the business drivers, a maintenance process/es and the role of performance measurements.</para>
<fig id="F2-3" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F2-3">Figure <xref linkend="F2-3" remap="2.3"/></link></label>
<caption><para>Implementing CBM in support of business drivers (modified from Spare, 2001).</para></caption>
<graphic xlink:href="graphics/ch02_fig003.jpg"/>
</fig>
<para>It is important to first identify the business drivers that will be aligned with the business strategy, in this case the maintenance department strategy. The business drivers are usually a financial measure or other performance measures that the specific company or department might have.</para>
<para>Business drivers connected with the financial part are, for instance, to reduce maintenance costs, reduce catastrophic failures, defer replacement (extend life), and increase equipment utilisation.</para>
<para>For the case of preventive maintenance, the objective is to reduce the costs of maintenance by the monitoring of equipment health and avoid unplanned failures.</para>
<para>Preventive maintenance objective is to reduce unplanned catastrophic failures. Condition monitoring, which is part of the CBM strategy, provides the detection of failures that might be developed before or after a scheduled maintenance service task. The approach averts the result of high cost related to maintenance since it detects a failure in time, which results in a usually much easier and less costly service provided.</para>
<para>In addition, CBM provides an extended life to the equipment since the equipment is better maintained by the service given in a timely manner. Also, the CBM strategy provides an increased equipment utilisation since it results in an opportunity to increase the value of its utilisation by following a strategy that continuously monitors the health of the equipment.</para>
<para>Continuously, the so-called soft benefits are reliability/availability, as well as safety concerns. The reliability/availability are goals that need to be achieved by the maintenance department to be able to increase the profit or at least not increase the cost of maintenance and by keeping the machines running for the production department. Another aspect is safety, which is an important factor that provides acceptable conditions to work for the employees.</para>
<para>It is crucial to identify and have an understanding of different business drivers that might exist for the specific domain, and specifically the maintenance department in this case. Consequently, this should be done to be able to run the business into the right direction, i.e., by transforming the written strategy into action and implement all its parts with appropriate actions and measurements to track its deviations from the predetermined plans and act accordingly.</para>
<para>Furthermore, it is crucial to understand the link between production and maintenance departments to be able to see the interdependence between the two mentioned departments in a company and the need to adjust the business drivers for the successful implementation of the overall business strategy.</para>
</section>
</section>
<section class="lev1" id="sec2-2">
<title>2.2 Optimization of Maintenance Costs</title>
<para>Even though maintenance has the potential to provide revenue streams, one of the main pillars for advanced maintenance is still the reduction of the maintenance costs while keeping the machinery and equipment constantly in working order. The traditional maintenance methods are not cost optimised, and ICT can help in this sense, by allowing for intelligent maintenance and concepts such as CBM.</para>
<para>Nowadays, the smartphones and mobile devices play a huge role in society. It was only a matter of time that these devices would be integrated in the industrial field, and more specifically in the maintenance services. Many organizations have implemented CMMSs on their smartphones or handheld devices. As smartphones are getting more powerful, they are substituting the handheld tools that have been used until now, enhancing the CMMS experience and providing the necessary tools for the occasion.</para>
<para>The internet connection of the smartphones enables the technician to view the maintenance history of that asset and make a first assessment of the problem. If he needs a spare part, he could order it straight from his phone and have it delivered to the exact location. Once the problem has been fixed, he could make a report and upload it directly to the history database.</para>
<para>A technician that needs to scan barcodes or QR codes typically uses a barcode reader. Nevertheless, with the camera of a smartphone it is possible to substitute the reader, and it could also show the info of the asset and even take you to its location. The camera could also be used to upload images of the product to another technician to solve any issues instead of describing the problem on the phone.</para>
<para>Most of the smartphones are equipped with NFC technology. NFC is used to read RFID tags, in a similar manner to the barcodes.</para>
<para>GPS, barometer, metal detector, sound meter or vibration analyser could be some of the other uses a smartphone could have.</para>
<para>With these examples it can be seen that the smartphone can substitute a considerable amount of handheld tools. However, it must be taken into account that even if these technologies are well-developed, they might not be comparable to the quality of other tools or suitable for all the needs a company has.</para>
</section>
<section class="lev1" id="sec2-3">
<title>2.3 Business Drivers for Collaborative Proactive Maintenance</title>
<para>Corrective maintenance is carried out after failure recognition. It is aimed to restore the equipment to a state in which it can perform its required function. Corrective maintenance can be deferred or immediate. The deferred maintenance is not done immediately after a fault has occurred. The immediate maintenance refers to provide service to the equipment without any delay at a suitable time. Therefore it is also called emergency maintenance. Hedderich [1996] indicated that the maintenance philosophy must change from a purely repair function to one that focuses on the operations of the equipment. Corrective maintenance, also called as &#x0201C;Fix when failed&#x0201D; was not a good one because when things break down maintenance has failed [Blann, 2003].</para>
<para>Corrective maintenance had many drawbacks, which led to evolution of Preventive Maintenance or TBM. Preventive maintenance is &#x0201C;the maintenance carried out at predetermined interval or according to prescribed criteria and intended to reduce the probability of failure or the degradation of the functioning of an item&#x0201D; [BS 3811, 1993] [SFS-EN 13306, 2010], and can thus be based on condition or predetermined interval maintenance. The basis for the determination of intervals can be in terms of time, number of operations, mileage etc. It is most effective when historical data exist to provide statistical failure rate for the equipment, MTBF can be accurately predicted, knowledge about the failure mode/s exist, low costs are associated with the regular overhaul of the equipment and low costs spare parts are available, etc. Preventive maintenance started to emerge in 1950s when reliability engineering started gaining popularity. The well-known bath-tub curve was based on the hypothesis that all equipment go through similar kind of deterioration and therefore, similar kinds of maintenance actions are required to keep it running. This is definitely not true because the failure in the equipment is based not only on its age, but also on the operating conditions. These operating conditions vary in terms of temperature, operator expertise, environmental factors, etc. This meant that sometimes, the maintenance actions happened before the required time, and in other cases, the maintenance was scheduled a little too late. When the maintenance happens before it is required, it leads to avoidable loss of production time and maintenance resources. A delayed maintenance action can result in the machine running up to failure, thereby loss of production time and other related problems. There are some other challenges, like inadequate data on asset performance and service history, maintenance not as a top priority for management, and nonstandard maintenance processes.</para>
<para>Collaborative proactive maintenance practices provide an answer to these challenges. Proactive maintenance strategies such as predictive maintenance that uses sensors for monitoring the health of the equipment by measuring vibrations, acoustic emissions, pressure waveforms, etc. further improve maintenance performance. Optimal asset performance is not dependent on maintenance practices alone. Planned maintenance activities must be scheduled to minimally impact production requirements and schedules. This necessitates close collaboration and planning between both maintenance and operations staff. Such collaborative proactive maintenance can result in improved asset reliability, greater asset uptime and availability, lower costs of servicing assets, fewer unexpected downtimes and outages, and a higher return in invested capital.</para>
<para>Business drivers are the crucial factors which lead to success in business. In case of asset maintenance, two broad categories of business drivers can be identified. These are Financial and Operational drivers. These two broad drivers can be further sub-divided and are shown in <link linkend="F2-4">Figure <xref linkend="F2-4" remap="2.4"/></link> below.</para>
<fig id="F2-4" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F2-4">Figure <xref linkend="F2-4" remap="2.4"/></link></label>
<caption><para>Classification of business drivers.</para></caption>
<graphic xlink:href="graphics/ch02_fig007.jpg"/>
</fig>
<para>RoI is a measure used to evaluate the efficiency of an investment or to compare the efficiency of a number of different investments. RoI measures the amount of return on an investment relative to the investment&#x02019;s cost. To calculate RoI, the benefit (or return) of an investment is divided by the cost of the investment, and the result is expressed as a percentage or a ratio. In the current context, it is the overall benefit accrued by the company due to its investment in a proactive collaborative maintenance approach divided by the cost of the investments.</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg18-1.jpg"/></para>
<para>Reduction in <emphasis>Operational</emphasis> and <emphasis>Maintenance Costs</emphasis> are also important business drivers. The companies that seek to apply collaborative proactive maintenance for managing their assets look for reducing these costs. These two drivers can be analysed intelligently by studying them as a percentage of the revenues that the company is generating. This would make sense as the larger companies will spend more on operations and maintenance and yet be better off as they generate much larger revenues.</para>
<para>Collaborative proactive maintenance strategy achieves a reduction in these costs through various means. When the company is proactive in asset maintenance, it has the foresight to plan for maintenance breaks. The strategy also enables the firms to pre-position the spare parts and the technically skilled personnel at the right place and right time. This will help in reducing costs of employing costly skilled manpower. The approach will also result in reduction in the cost of spare parts by reducing the MRO inventory.</para>
<para>Corrective maintenance or <emphasis>fix-when-fail</emphasis> is an often used strategy. However, it has serious drawbacks. It results in unplanned stoppages of production when the machine fails. This unplanned stoppage can have an adverse impact on the production schedule, leading to not fulfilling the customer demands. Corrective maintenance also has other drawbacks where a damaged part can lead to secondary faults. A broken bearing or a gear can foul with other parts of the machine and damage them too. This will increase the maintenance costs. A proactive maintenance strategy can help in removing an about-to-fail spare part, thereby avoiding damage to other parts of the machine.</para>
<para>Collaborative proactive maintenance strategy results in keeping the machine in a better state. A well-maintained machine, with reliable spare parts will produce better quality product. A calibrated machine will produce the products that are within tolerances. Such machine will also have a longer usable life. This can help in reducing further investments in new capital procurements at the end of life of the current machines.</para>
<para><emphasis>Operational Drivers</emphasis> of asset maintenance are increase of <emphasis>asset reliability</emphasis> and <emphasis>availability. Asset Reliability</emphasis> is defined as the probability that a component or system will perform a required function for a given time when used under stated operating conditions. <emphasis>Asset availability</emphasis> is the asset&#x02019;s availability to be put to its intended use. A Proactive maintenance strategy will help in increasing both the availability and the reliability of the assets.</para>
<para>Another <emphasis>Operational Driver</emphasis> is increasing the <emphasis>Asset Productivity</emphasis>. It describes how effectively are the business assets deployed. This ratio typically looks at sales dollars generated per unit of resource. Resources can include inventory, fixed assets, and occasionally other tangible assets. Similar analyses may also be done not just for financial assets but also for operational assets like square footage, number of employees, etc.</para>
<para>The financial and operational drivers of boosting service and maintenance effectiveness in asset intensive industries are substantial. Preventive maintenance is effective, but proactive maintenance strategies such as predictive maintenance and reliability centred maintenance processes have a positive impact on nearly every asset performance measurement. Collaboration must be carried out between maintenance and operations staff to develop planned maintenance schedules that minimally affect production requirements. The companies must try to bring all maintenance planning and maintenance operations together under the control of one senior-level executive to strengthen maintenance processes across the entire enterprise, not just within individual departments or operations. The companies must use the latest analytics software to drive the business. Such software solutions are very useful in pinpointing the areas that need focus for improving the asset performance.</para>
<section class="lev2" id="sec2-3-1">
<title>2.3.1 Maintenance Optimisation Models</title>
<para>Maintenance optimisation models are those mathematical models whose aim is to find the optimal balance between the costs and benefits of maintenance, while taking all kinds of constraints into account [Sharma and Yadava, 2011]. In general, maintenance optimisation models cover four aspects:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>A description of the system (or component), and the function and importance of the system in keeping the business operational;</para></listitem>
<listitem>
<para>A modeling of the physical deterioration of the system in time as well as the failure behavior and consequences;</para></listitem>
<listitem>
<para>A description of the available information about the system (i.e., state space) and the decisions available to management (i.e., action space);</para></listitem>
<listitem>
<para>An objective function and a set of constraints along with an optimisation technique in order to find the best balance.</para></listitem>
</itemizedlist>
<para>The maintenance optimisation models in the literature can be classified based on different aspects such as the presence of randomness in the system, the number and interdependency of failure-prone units, discretization of time, model types, optimality criterion, methods of solution and planning time horizon. <link linkend="F2-5">Figure <xref linkend="F2-5" remap="2.5"/></link> summarizes these aspects within a proposed workflow in building maintenance optimisation models.</para>
<para>In the context of maintenance optimisation, it is common to model the uncertainty in the time-to-failure with a known probability distribution function. However, this probability distribution function is not readily available in practice, requiring data-driven maintenance optimisation policies. Alternatively, uncertainty in the failure behaviours can build on the knowledge about the failure characteristics learned from historical data. In particular, the maintenance optimisation models provide the tools that potentially rely on the output from root cause analysis, remaining useful life time analysis and alerting and prediction of assets failures in the optimal planning of maintenance and related resource allocation.</para>
<para>The optimisation methods for solving the mathematical formulations of the maintenance-planning problems often include linear and nonlinear programming, dynamic programming, Markov decision processes, decision analysis techniques, search techniques (i.e., simulation-based optimisation) and heuristic approaches.</para>
<para>Maintenance optimisation models yield various managerial outcomes [Dekker, 1996]: First of all, the structure of optimal polices, like the existence of an optimal control-limit policy, can be established. Second, models can assist in the timing aspect of maintenance activities: how often and at what extent to inspect or to repair a machine or a component; and if applicable, how often to preventively replace them. Finally, models can also be of help in determining effective and efficient schedules and plans for service engineers and spare-part stocking, taking all kinds of operational, physical and economic constraints into account.</para>
<fig id="F2-5" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F2-5">Figure <xref linkend="F2-5" remap="2.5"/></link></label>
<caption><para>Workflow in building maintenance optimisation models.</para></caption>
<graphic xlink:href="graphics/ch02_fig004.jpg"/>
</fig>
</section>
<section class="lev2" id="sec2-3-2">
<title>2.3.2 Objectives and Scope</title>
<para>In general, an <emphasis role="strong">optimal maintenance policy</emphasis> achieves one of the two objectives: minimizing system maintenance cost (e.g., the sum of failure and proactive maintenance costs) while the system reliability requirements (e.g., system availability or uptime) are satisfied, or maximizing the system reliability measures when requirements for the system maintenance cost are satisfied.</para>
<para>In line with these objectives, the maintenance decision making process (<link linkend="F2-6">Figure <xref linkend="F2-6" remap="2.6"/></link>) is composed of two main assessments followed by the selection of the appropriate decision model [Ahmad and Kamaruddin, 2012]:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Operational cost assessment:</emphasis> The aim of this assessment is to calculate the two types of operational costs, namely the failure cost, the setup cost for data-driven maintenance (e.g., the cost of collecting, storing, and processing data, the cost of installing new sensors for condition monitoring) and the preventive maintenance cost (e.g., the cost of performing inspections to detect failures or to perform data analytics to predict the failures in order to act before they are realized);</para></listitem>
<listitem>
<para><emphasis role="strong">Component assessment:</emphasis> The goal of this assessment is to classify the maintenance type of the equipment as either non repairable or repairable, and if so, to what extent, and to identify the structure and interdependency of the components in the system;</para></listitem>
<listitem>
<para><emphasis role="strong">Decision (optimisation) model selection:</emphasis> After the component assessment, the appropriate maintenance model is built to identify the optimal maintenance policy. The maintenance policy is the output of the maintenance model. Maintenance models can also be used to evaluate the performance of the practical policies that are not necessarily optimal.</para></listitem>
</itemizedlist>
<para>In the remainder of this section, we use the term <emphasis role="strong">maintenance policy</emphasis> in its most general meaning, including an inspection (information collection) policy, repair policy, or replacement policy. That is, maintenance optimisation is a unifying approach for finding the optimal subset of policy for inspection, repair, or replacement activities.</para>
<fig id="F2-6" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F2-6">Figure <xref linkend="F2-6" remap="2.6"/></link></label>
<caption><para>Maintenance decision making process.</para></caption>
<graphic xlink:href="graphics/ch02_fig005.jpg"/>
</fig>
<para>We note that the maintenance optimisation starts at the design phase of a system of components because the level of redundancy or accessibility in product design have significant impact on the maintainability of a system. Therefore, it is of practical importance to account for the total maintenance costs in the product-life cycle during new product development. This is an example of a strategic decision in maintenance optimisation. Alternatively, operational and tactical decisions in the scope of the maintenance optimisation models consider that maintenance has to be planned and scheduled once the system is in operation, in accordance with other plans (e.g., production, planned maintenance, spare parts planning) by building on the data-driven prediction tools that capture the uncertainty in the failure behaviour of a system.</para>
</section>
<section class="lev2" id="sec2-3-3">
<title>2.3.3 Maintenance Standards</title>
<para>It is essential to understand the definitions of maintenance when selecting a maintenance strategy. The starting point may be considered generally known standards where maintenance is defined as follows:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Standards PSK 6201 and PSK 7501: Maintenance is the totality of all the technical, administrative and management issues designed to maintain the target object or return it to a state where it is capable of performing the required function during its entire life cycle [PSK 6201, 2003], [PSK, 2010];</para></listitem>
<listitem>
<para>The European standard SFS-EN 13306: Maintenance consists of all technical, administrative and management issues of the lifetime of the object designed to maintain or restore the object so that the object is capable of performing the required function [SFS-EN 13306, 2010].</para></listitem>
</itemizedlist>
</section>
<section class="lev2" id="sec2-3-4">
<title>2.3.4 Maintenance-related Operational Planning</title>
<para>Maintenance optimisation models often span multiple business functions as maintenance decisions are directly linked to service-logistics operations. Based on the preliminary feedback from use-case owners, the maintenance-related operational planning can be divided into three stream:</para>
<para><emphasis role="strong">Service logistics/transportation:</emphasis> Since the capital goods are often operated at remote locations in a network, unplanned maintenance requires significant logistic effort and hence can be very costly. The data-driven planning of the routes, vehicles, and skill levels of the maintenance teams, is of interest in reducing high maintenance costs.</para>
<para><emphasis role="strong">Spare parts/service tools planning:</emphasis> Because the demand for spare parts is uncertain (as it is unknown when machines or certain components break down), and the spare parts should be delivered in a timely manner, most often the practical approach is to stock resources (spare parts, tools and service engineers) in all the places where such needs can arise. This leads to high inventory holding costs. The failure prediction algorithms bring an opportunity to enrich the spare parts/service tools planning via the joint maintenance and spare-parts planning models. Houtum and Kranenburg [2015] present accurate evaluation and optimisation algorithms for spare parts stocking policies.</para>
<para><emphasis role="strong">Operational decision-making based on remote monitoring:</emphasis> The advanced sensor and ICT technology allows acquiring the physical conditions of systems/components remotely and with less cost. Based on these condition data, significant amount of unnecessary maintenance tasks can be avoided, by taking maintenance actions only when the failure of critical components is imminent. This can be seen as a special case of predictive maintenance, through which maintenance costs can be significantly reduced in comparison to preventive maintenance (<link linkend="F2-7">Figure <xref linkend="F2-7" remap="2.7"/></link>). Remote maintenance and service-logistics planning are highly relevant here.</para>
<fig id="F2-7" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F2-7">Figure <xref linkend="F2-7" remap="2.7"/></link></label>
<caption><para>Decision framework for predictive maintenance.</para></caption>
<graphic xlink:href="graphics/ch02_fig006.jpg"/>
</fig>
</section>
</section>
<section class="lev1" id="sec2-4">
<title>2.4 Economic View of CBM-based PM</title>
<para>According to the standard SFS-EN 15341, 2007 (Maintenance. Maintenance Key Performance Indicators), total maintenance cost (often based annually related only to the maintenance activities performed on the asset/item). Includes costs referred to:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Wages, salaries and overtimes for managerial, supervision, support staff and direct staff;</para></listitem>
<listitem>
<para>Payroll added costs for the above mentioned persons (Taxes, Insurance, Legislative contributions);</para></listitem>
<listitem>
<para>Spares and material consumables charged to maintenance (including freight costs);</para></listitem>
<listitem>
<para>Tools and equipment (not capitalized or rented);</para></listitem>
<listitem>
<para>Contractors, rented facilities;</para></listitem>
<listitem>
<para>Consultancy services;</para></listitem>
<listitem>
<para>Administration costs for maintenance;</para></listitem>
<listitem>
<para>Education and training;</para></listitem>
<listitem>
<para>Costs for maintenance activities carried out by production people;</para></listitem>
<listitem>
<para>Costs for transportation, hotels, etc.;</para></listitem>
<listitem>
<para>Documentation;</para></listitem>
<listitem>
<para>CMMS and Planning Systems;</para></listitem>
<listitem>
<para>Energy and utilities;</para></listitem>
<listitem>
<para>Depreciation of maintenance capitalized equipment&#x02019;s and workshops, warehouse for spare-parts.</para></listitem>
</itemizedlist>
<para>Exclusions:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Costs for product changeover or transaction time (e.g., Exchange of dies);</para></listitem>
<listitem>
<para>Depreciation of strategic spare parts;</para></listitem>
<listitem>
<para>Downtime costs.</para></listitem>
</itemizedlist>
<para>A remote access to the plant data would reduce the travel time needed to perform the service and significantly cut costs. Remote access also allows experts from different specialities and physical locations to share data and collaborate. The analysis work could be performed remotely while traveling, from home or office or from anywhere increasing work flexibility and improving personnel efficiency.</para>
<para>A continuous data collection piloted in the use-case enables the use of automated analysis and monitoring software more effectively. The software would monitor the data collected from the plant and alert the expert to cases that would require special attention. This enhancement is also aimed at reducing personnel costs involved.</para>
<para>The new business model aims to provide:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Own fleet operation and maintenance</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Better availability by using predictive condition monitoring;</para></listitem>
<listitem>
<para>Possibility to apply platform architecture and advanced maintenance concept to operation;</para></listitem>
<listitem>
<para>More exact timing of maintenance action based on actual condition.</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>Increasing sales of O&#x00026;M services and cost effective operation support</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Possible contractual based long term maintenance agreements by using condition monitoring for energy production customers;</para></listitem>
<listitem>
<para>Optimal use of personnel for remote sites by using predictive tools and utilizing new operation concept concerning local and back office support;</para></listitem>
<listitem>
<para>Faster and optimized failure correction by using analytics and prognosis for critical components enables management of larger power plant fleet;</para></listitem>
<listitem>
<para>Systematic long term information collection for root cause analysis and investment planning of replacements.</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
<para>Estimating the revenue of a new business model is rarely accurate and due to the high variance in the potential business models, it is impossible to show a typical revenue stream projection of a remote support service.</para>
<para>Previous experiences and references show that successful implementation of a remote monitoring system can reduce the unscheduled unavailability of a component by 35%&#x02013;45%. This can have a serious impact on reducing the lost production, e.g., of a power plant.</para>
<para>Like in the traditional service model, the most significant costs of the service come from personnel costs. Utilization of machine aided monitoring these costs could potentially be reduced, but often this is not possible. In a well designed service, each member is a specialist in a certain field and their area of expertise needs to be covered.</para>
<para>Benefits of Collaborative, Proactive Maintenance solution could be:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Production price (taxes, emission taxes, tariffs, government support);</para></listitem>
<listitem>
<para>Price of energy as a function of seasonal demand (e.g., winter time demand for energy is high and so is price/revenue);</para></listitem>
<listitem>
<para>Demand as a function of time (seasonal, each plant has a maximum energy production capacity);</para></listitem>
<listitem>
<para>Investment cost of the critical component;</para></listitem>
<listitem>
<para>Maintenance work required (hours);</para></listitem>
<listitem>
<para>Work cost, based on market situation;</para></listitem>
<listitem>
<para>Reduction of unnecessary travel to remote locations -> increases sustainability by reducing emissions caused by travelling -> increases company&#x02019;s public image.</para></listitem>
</itemizedlist>
</section>
<section class="lev1" id="sec2-5">
<title>2.5 Risks in CBM Plan Implementation</title>
<para>CBM has developed rapidly as a maintenance concept and is being considered as panacea for all ailments related to machine health. But like with all things that seem miraculous, there are a few issues in successful implementation of a CBM plan. Many implementation efforts fail and condition monitoring tools too often end up at maintenance workshops cupboards, hardly ever used [Walker, 2005]. Nearly 30% of industrial equipment does not benefit from CBM [Hashemian and Bean, 2011]. This may be due to a number of reasons. The data is gathered in large amounts, often from assets that are dispersed over large distances; sometimes even across continents. The collected data needs to be integrated from multiple sources such that effective analysis could be done. In certain cases, new assets and data acquisition sources get added. Additional resources are required to integrate this new data with existing data in order to derive meaningful insights. In some CBM plans, the collected data and the derived results are required to be assessed by an expert. There is a scarcity of good experts who can recommend corrective actions in time. Therefore, even if a condition monitoring programme is in operation, failures still occur, defeating the very purpose for which the investment was made in CBM [Campos, 2009; Prakash, 2006; Rao et al., 2003]. The basic idea behind implementing a CBM plan is to detect the impending failures and initiate corrective actions that are timely. A good CBM plan will ensure that the maintenance scheduling is neither too early (leading to unnecessary maintenance) or too late (catastrophic failures resulting in disruptions of manufacturing operations). However, it must be noted from the operational viewpoint that any prospective maintenance policy based on condition information must have clear economic benefits; otherwise the initial outlay for the CM system and associated costs cannot be justified [McMillan and Ault, 2007]. There are a large number of other factors that not only impede a smooth implementation of a CBM plan but also result in a system that is ineffective and considered as an unnecessary burden in terms of costs and workload. Through an extensive literature review, this chapter identifies these issues that can hamper a successful conversion of a theoretical plan into an effective workable solution.</para>
<para>CBM Implementation is not devoid of the challenges and hurdles associated with programs that require a drastic change in the way an organization functions. In this chapter, an attempt has been made to classify these issues into four major categories. These are: Technology, People, Processes and Organizational Culture. Every organization that is planning to implement a successful CBM program must pay attention to these issues and avoid the obvious pitfalls. The risks are listed below.</para>
<section class="lev2" id="sec2-5-1">
<title>2.5.1 Technology</title>
<para>CBM is a technology rich maintenance driver that aims at taking over the functions of humans. In certain cases, only the mundane monitoring is handled by the technology while in some other advance CBM systems, even the decision making is automated. Technological barriers to a successful implementation are most important because these factors can lead to efforts getting wasted on unfruitful activities. Some of the issues are listed below.</para>
<para><emphasis role="strong">Selection of assets to Monitor:</emphasis> First and foremost, the organizations must select the assets that require condition monitoring. CBM is supposed to be applied where appropriate, not as an overall policy as some techniques are expensive and it would not be cost effective to implement everywhere [Starr, 1997]. This selection should be based on technical and economical feasibilities. The selected asset must be such that it has the parameters that can be monitored with the sensors. Also, the machine should be critical enough to warrant large investments in a CBM system.</para>
<para><emphasis role="strong">Complexity of measuring parameters:</emphasis> One of the most important step in CBM implementation is to decide the parameter that should be measured by the sensors. Most CBM implementation drives fail because the organizations start to measure &#x02018;what can be measured&#x02019; rather than &#x02018;what should be measured&#x02019;. The first challenge is to obtain effective features from many candidate features to reflect health degradation propagation in the whole life of machines [Yu, 2012]. In machine-learning-based defect detection, the accuracy of prognostics and diagnostics models subsequently dependents on the sensitivity of the features used to estimate the condition and propagation of the defects. Therefore, it is critical to devise a systematic scheme that is capable of selecting the most representative features for current machine health states [Yu, 2012; Malhi et al., 2004]. The complexity of selecting a suitable measure can be gauged from an example. Some research shows that the use of acoustic signal is better than vibration signal due to its sensitivity and accuracy [Al-Ghamd and Mba, 2006; Baydar and Ball, 2001; Tandon et al., 2007]. However, in practice, the application of acoustic signal may not appropriate due to the significant effects of noise (unwanted signals) from other equipment. In addition, alternative sources of information are important contributors to health monitoring. These sources include the OEM, ISO standards, experience of the workers, etc. The new challenge is to find ways to use these alternative sources of information in order to achieve better monitoring of assets and correct decision making [Ahmad and Kamaruddin, 2012].</para>
<para><emphasis role="strong">Complexity of sensors:</emphasis> Choice of a correct sensor with appropriate sensitivity is an important step. Selecting a costlier sensor when it is not required will make the CBM system unnecessarily expensive, which will not be able to justify the cost-benefit argument. On the other hand, selecting a cheaper non-sensitive sensor when the data being measured has minute variations that require high quality sensor can also upset the cost-benefit balance as the diagnosis of the fault may not be correct. Other ICT challenges include sensor data quality related to gathering frequency, noise, and level of details of sensor data, data availability, wireless communication problem, frequency of diagnostics and prognostics, and so on [Shin and Jun, 2015]. In addition, the technologies and technical methods for the CBM approach are still in their infancy. It means that there are some limitations in ensuring the accuracy of diagnostics and prognostics [Shin and Jun, 2015]. Numerous different techniques and technologies exist but choosing the correct one, or even remembering to make the decision in due time can be a troublesome activity which can put an entire implementation effort at risk [Bengtsson, 2004].</para>
<para><emphasis role="strong">Deciding on thresholds:</emphasis> The condition monitoring practice is based on the fact that a sensor is used to measure a parameter. When the value of the measured parameter crosses a pre-determined threshold, suitable maintenance actions are initiated. In practice however, deciding on the threshold is a complex process. The failure of each of equipment may be defined and classified in different ways. Some organizations consider failure as the physical event such as a breakage that stops production. The machine stops to function as a result of such a failure. In some other cases, a functional failure may occur which results in the final product of the machine to have quality flaws but the machine may continue to work. It is necessary for the organizations to determine threshold based on their requirements. The definition and determination of failure limits should be considered from both the entire machining process perspective (system/sub-system) and the overall output of the system (e.g., product quality characteristics) [Ahmad and Kamaruddin, 2012].</para>
<para><emphasis role="strong">Noise effects:</emphasis> Analysing waveform data is an intricate process because of noise effects, which are unwanted signals generated by other equipment. Noise must be minimised or eliminated from the data [Ahmad and Kamaruddin, 2012]. Some noise also gets generated due to the transmission medium. It is necessary to identify which data transmission type (wire or wireless) is effective in terms of cost and reliability with least noise [Shin and Jun, 2015].</para>
<para><emphasis role="strong">Data storage and monitoring period:</emphasis> In many cases, the assets that are monitored with sensors are big machines that need years of use in order to show wear in some parts, or need years to reach failures. Thus, there is a need to monitor and store huge time periods (and related data) in order to provide enough data to generate predictive models.</para>
<para><emphasis role="strong">Data cleaning:</emphasis> Large data sets are required for effective data analysis and modelling. This collected data needs to be cleaned before any analysis. This is a complex task, especially for waveform-type data.</para>
<para><emphasis role="strong">Biased thresholds:</emphasis> Determination of thresholds is a complex process. Often, the complexity increases due to biases that get introduced in determining these failure limits [Ahmad and Kamaruddin, 2012].</para>
<para><emphasis role="strong">New Assets and the complexities:</emphasis> Newly commissioned systems have no historical data. Even the OEM is not aware of the failure patters or failure rates. In such cases, it is not possible to identify the trends or failure thresholds. Such situations are no-data situations [Si et al., 2011]. The quantity and completeness of data are insufficient to fit the full statistical models. Hence, it may be a better choice to develop physics-based models with the help of subjective expert knowledge from design and manufacturing [Si et al., 2011].</para>
</section>
<section class="lev2" id="sec2-5-2">
<title>2.5.2 People</title>
<para>CBM aims at using technology to monitor and/or diagnose the faults in machines. The focus of CBM is to replace humans by machines to carry out mundane monitoring and intelligent decision making tasks. In spite of this, it is not possible to replace humans completely. The success of any CBM implementation program depends on a lot of people factors. Some of these are listed below.</para>
<para><emphasis role="strong">Training and Skills:</emphasis> There is a need to have requisite skills in the personnel that are responsible to use CBM. Familiarity with the condition monitoring system and training for analysis of the data is required to be imparted. Advanced data analysis such as frequency analysis requires advanced skills [Rastegari and Bengtsson, 2014]. Education and training is important pre-requisites for a company to increase the competence [Rastegari et al., 2013]. Other factors like technical competence and knowledge is required in order to completely implement a CBM system.</para>
<para><emphasis role="strong">Human Factors:</emphasis> Other human factors also need special emphasis. Walker [2005] lists several factors like lack of direction, unwillingness to adopt a new approach to maintenance, etc. as reasons for failure of CBM approaches. Senior technicians often treat a CBM initiative as something that is unnecessary expenditure. There are also some reservations about the frequency of monitoring. Technicians often feel that it is better to spend the money in other systems which can give more money back [Rastegari et al., 2013].</para>
</section>
<section class="lev2" id="sec2-5-3">
<title>2.5.3 Processes</title>
<para>The implementation of a CBM program is a long drawn complicated process that needs careful planning. The processes must be put in place to implement, monitor, measure and improve the system.</para>
<para><emphasis role="strong">Economic Feasibility study:</emphasis> An economic feasibility study must be carried out before the implementation process. It is imperative to define the business model for new maintenance operation and identify benefits and costs [Shin and Jun, 2015]. High data collection costs must be considered in the feasibility study [Kothamasu et al., 2006]. All these support monitoring tools directly involve high costs, and not all companies are willing to invest in them [Ahmad and Kamaruddin, 2012]. Specific computerized monitoring systems and experts are required. These requirements directly involve large company investments, especially since such companies must buy and maintain these systems, as well as provide training for their use [Ahmad and Kamaruddin, 2012].</para>
<para><emphasis role="strong">Assignments of Responsibilities:</emphasis> In order to have a more structured implementation, it is needed to define the responsibilities in early phase of the implementation [Rastegari and Bengtsson, 2014]. Implementation responsibilities must be written down as company documents and adhered to.</para>
<para><emphasis role="strong">Piecemeal implementation:</emphasis> To get the practical benefit of CBM approach, it is necessary to consider applying CBM into not only one piece of equipment but also an integrated system level [Shin and Jun, 2015]. The implementation process should consider holistic application of the technology at all levels.</para>
</section>
<section class="lev2" id="sec2-5-4">
<title>2.5.4 Organizational Culture</title>
<para>It is important to have encouraging atmosphere in the organization to successfully implement a CBM program. Some of the related factors are listed below.</para>
<para><emphasis role="strong">Top management commitment:</emphasis> One of the main challenges to implement CBM at the company is management support. The management must comprehend the importance of the CBM&#x02019;s role at the company and provide the resources needed. It needs to have a long term strategy to change the way of working from reactive maintenance to proactive maintenance [Shin and Jun, 2015]. That is why, the top management must continue to support the initiative for as long as it is necessary.</para>
<para><emphasis role="strong">Alignment of business objectives with the workforce:</emphasis> The workforce must be on the same page as the top management. The process must not look like a management fad in which the workforce has no interest.</para>
<para><emphasis role="strong">Establishment, documentation and communication of objective:</emphasis> The overall objective of CBM must be decided and documented. The objective must be communicated down to each of the stakeholder. The communication gap often results in the worker considering CBM as another top driven agenda with no fixed goals. This leads to failure or partial implementation of CBM which provides no benefits to the organization.</para>
<para><emphasis role="strong">Conflicting departmental priorities:</emphasis> There may exist conflicting priorities amongst several departments in an organization. For example, Maintenance Department may want a machine for inspection/maintenance before it fails but the production department wants high utilization of the equipment and hence does not hand over the machine to the maintenance people [Campos, 2009]. The organization must have conflict resolution rules such that the CBM implementation as well as the health of the equipment does not suffer.</para>
</section>
</section>
<section class="lev1" id="sec2-6">
<title>References</title>
<orderedlist numeration="arabic" continuation="restarts" spacing="normal">
<listitem><para>Ahmad, R. and Kamaruddin, S. (2012) &#x2018;An overview of time-based and condition-based maintenance in industrial application&#x2019;, <emphasis>Computers &#x0026; Industrial Engineering</emphasis>, 63, pp. 135&#x2013;149. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ahmad%2C+R%2E+and+Kamaruddin%2C+S%2E+%282012%29+%27An+overview+of+time-based+and+condition-based+maintenance+in+industrial+application%27%2C+Computers+%26+Industrial+Engineering%2C+63%2C+pp%2E+135-149%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Al-Ghamd, A.M. and Mba, D.A. (2006) &#x2018;Comparative experimental study on the use of acoustic emission and vibration analysis for bearing defect identification and estimation of defect size&#x2019;, <emphasis>Mechanical System and Signal Process</emphasis>, 20, pp. 1537&#x2013;1571. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Al-Ghamd%2C+A%2EM%2E+and+Mba%2C+D%2EA%2E+%282006%29+%27Comparative+experimental+study+on+the+use+of+acoustic+emission+and+vibration+analysis+for+bearing+defect+identification+and+estimation+of+defect+size%27%2C+Mechanical+System+and+Signal+Process%2C+20%2C+pp%2E+1537-1571%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Baydar, N. and Ball, A. (2001) &#x2018;A comparative study of acoustic and vibration signals in detection of gear failures using Wigner-Ville distribution&#x2019;, <emphasis>Mechanical System and Signal Process</emphasis>, 15(6), pp. 1091&#x2013;1107. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Baydar%2C+N%2E+and+Ball%2C+A%2E+%282001%29+%27A+comparative+study+of+acoustic+and+vibration+signals+in+detection+of+gear+failures+using+Wigner-Ville+distribution%27%2C+Mechanical+System+and+Signal+Process%2C+15%286%29%2C+pp%2E+1091-1107%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Ben-Daya, M. and Duffuaa, S. O. (1995) &#x2018;Maintenance and quality: the missing link&#x2019;, <emphasis>Journal of Quality in Maintenance Engineering</emphasis>, 1(1), pp. 20&#x2013;26. https://doi.org/10.1108/13552519510083110. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ben-Daya%2C+M%2E+and+Duffuaa%2C+S%2E+O%2E+%281995%29+%27Maintenance+and+quality%3A+the+missing+link%27%2C+Journal+of+Quality+in+Maintenance+Engineering%2C+1%281%29%2C+pp%2E+20-26%2E+https%3A%2F%2Fdoi%2Eorg%2F10%2E1108%2F13552519510083110%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Bengtsson, M. (2004) <emphasis>Condition Based Maintenance Systems: An Investigation of Technical Constituents and Organizational Aspects</emphasis>, Licentiate thesis M&#x00E4;lardalen University Eskilstuna. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Bengtsson%2C+M%2E+%282004%29+Condition+Based+Maintenance+Systems%3A+An+Investigation+of+Technical+Constituents+and+Organizational+Aspects%2C+Licentiate+thesis+M%E4lardalen+University+Eskilstuna%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Blann, D. R. (2003) &#x2018;Reliability as a Strategic Initiative: To Improve Manufacturing Capacity, Throughput and Profitability&#x2019;, <emphasis>Asset Management and Maintenance Journal</emphasis>, 16(2). <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Blann%2C+D%2E+R%2E+%282003%29+%27Reliability+as+a+Strategic+Initiative%3A+To+Improve+Manufacturing+Capacity%2C+Throughput+and+Profitability%27%2C+Asset+Management+and+Maintenance+Journal%2C+16%282%29%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>BS 3811. (1993) <emphasis>Glossary of Terms Used in Terotechnology,</emphasis> British Standards Institution (BSI). <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=BS+3811%2E+%281993%29+Glossary+of+Terms+Used+in+Terotechnology%2C+British+Standards+Institution+%28BSI%29%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Campos, J. (2009) &#x2018;Development in the application of ICT in condition monitoring and maintenance&#x2019;, <emphasis>Computers in Industry</emphasis>, 60, pp. 1&#x2013;20. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Campos%2C+J%2E+%282009%29+%27Development+in+the+application+of+ICT+in+condition+monitoring+and+maintenance%27%2C+Computers+in+Industry%2C+60%2C+pp%2E+1-20%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Dekker, R. (1996) &#x2018;Applications of maintenance optimization models: A review and analysis,&#x2019; <emphasis>Reliability Engineering and System Safety</emphasis>, 51, pp. 229&#x2013;2440. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Dekker%2C+R%2E+%281996%29+%27Applications+of+maintenance+optimization+models%3A+A+review+and+analysis%2C%27+Reliability+Engineering+and+System+Safety%2C+51%2C+pp%2E+229-2440%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Hashemian H. M. and Bean W. C. (2011) &#x2018;State-of-the-art predictive maintenance techniques&#x2019;, <emphasis>IEEE Transactions on Instrumentation and Measurement,</emphasis> 60(10), pp. 3480&#x2013;3492. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Hashemian+H%2E+M%2E+and+Bean+W%2E+C%2E+%282011%29+%27State-of-the-art+predictive+maintenance+techniques%27%2C+IEEE+Transactions+on+Instrumentation+and+Measurement%2C+60%2810%29%2C+pp%2E+3480-3492%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Hedderich, C. P. (1996) &#x2018;Navy Predictive Maintenance&#x2019;, <emphasis>Naval Engineers Journal</emphasis>, 10(6), pp. 41&#x2013;57. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Hedderich%2C+C%2E+P%2E+%281996%29+%27Navy+Predictive+Maintenance%27%2C+Naval+Engineers+Journal%2C+10%286%29%2C+pp%2E+41-57%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Holmberg, K., Adgar, A., Arnaiz, A., Jantunen, E., Mascolo, J., and Mekid, S. (Eds.) (2010) <emphasis>E-maintenance</emphasis>, Springer Science &#x0026; Business Media, pp. 1&#x2013;511.</para></listitem>
<listitem><para>Houtum, G. J. and Kranenburg, A. (2015) &#x2018;Spare Parts Inventory Control under System Availability&#x2019;, <emphasis>International Series in Operations Research &#x0026; Management Science</emphasis>, Springer, 2015. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Houtum%2C+G%2E+J%2E+and+Kranenburg%2C+A%2E+%282015%29+%27Spare+Parts+Inventory+Control+under+System+Availability%27%2C+International+Series+in+Operations+Research+%26+Management+Science%2C+Springer%2C+2015%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>IEC 60300-3-3. (2004) <emphasis>Dependability management &#x2013; Part 3-3: Application guide &#x2013;Life cycle costing</emphasis>, 2nd edn. Geneva, International Electrotechnical Commission. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=IEC+60300-3-3%2E+%282004%29+Dependability+management+-+Part+3-3%3A+Application+guide+-Life+cycle+costing%2C+2nd+edn%2E+Geneva%2C+International+Electrotechnical+Commission%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Kothamasu, R., Huang, S. H., and VerDuin, W. H. (2006) &#x2018;System health monitoring and prognostics: A review of current paradigms and practices&#x2019;, <emphasis>International Journal in Advances Manufacturing Technology</emphasis>, 28, pp. 1012&#x2013;1024. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Kothamasu%2C+R%2E%2C+Huang%2C+S%2E+H%2E%2C+and+VerDuin%2C+W%2E+H%2E+%282006%29+%27System+health+monitoring+and+prognostics%3A+A+review+of+current+paradigms+and+practices%27%2C+International+Journal+in+Advances+Manufacturing+Technology%2C+28%2C+pp%2E+1012-1024%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Lozano, R. (2015) &#x2018;A holistic perspective on corporate sustainability drivers&#x2019;, Corporate Social Responsibility and Environmental Management, 22, pp. 32&#x2013;44. https://doi.org/10.1002/csr.1325. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Lozano%2C+R%2E+%282015%29+%27A+holistic+perspective+on+corporate+sustainability+drivers%27%2C+Corporate+Social+Responsibility+and+Environmental+Management%2C+22%2C+pp%2E+32-44%2E+https%3A%2F%2Fdoi%2Eorg%2F10%2E1002%2Fcsr%2E1325%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Malhi, A., Yan, R. and Gao, R. X. (2004) &#x2018;PCA-based feature selection scheme for machine defect classification,&#x2019; <emphasis>IEEE Transactions on Instrumentation and Measurement</emphasis>, 53(6), pp. 1517&#x2013;1525. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Malhi%2C+A%2E%2C+Yan%2C+R%2E+and+Gao%2C+R%2E+X%2E+%282004%29+%27PCA-based+feature+selection+scheme+for+machine+defect+classification%2C%27+IEEE+Transactions+on+Instrumentation+and+Measurement%2C+53%286%29%2C+pp%2E+1517-1525%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>McMillan, D. and Ault, G. W. (2007) &#x2018;Quantification of condition monitoring benefit for offshore wind turbines&#x2019;, <emphasis>Wind Engineering</emphasis>, 31(4), pp. 267&#x2013;285. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=McMillan%2C+D%2E+and+Ault%2C+G%2E+W%2E+%282007%29+%27Quantification+of+condition+monitoring+benefit+for+offshore+wind+turbines%27%2C+Wind+Engineering%2C+31%284%29%2C+pp%2E+267-285%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Mikkonen, H., Miettinen, J., Leinonen, P., Jantunen, E., Kokko, V., Riutta, E., Sulo, P., Komonen, K., Lumme, V. E., Kautto, R., (ed.) (2009) <emphasis>Condition-based Maintenance: Handbook</emphasis>, Maintenance Publication Series, Vol. 13, Maintenance Association, Kerava.</para></listitem>
<listitem><para>Paz, N. M. and Leigh, W. (1994) &#x2018;Maintenance scheduling: Issues, results and research needs&#x2019;, <emphasis>International Journal of Operations &#x0026; Production Management</emphasis>, 14, pp. 47&#x2013;69. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Paz%2C+N%2E+M%2E+and+Leigh%2C+W%2E+%281994%29+%27Maintenance+scheduling%3A+Issues%2C+results+and+research+needs%27%2C+International+Journal+of+Operations+%26+Production+Management%2C+14%2C+pp%2E+47-69%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Prakash, O. (2006) &#x2018;Asset management through condition monitoring &#x2013; how it may go wrong: A case study&#x2019;, In <emphasis>Proceedings of the 1st World Congress on Engineering Asset Management (WCEAM).</emphasis> <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Prakash%2C+O%2E+%282006%29+%27Asset+management+through+condition+monitoring+-+how+it+may+go+wrong%3A+A+case+study%27%2C+In+Proceedings+of+the+1st+World+Congress+on+Engineering+Asset+Management+%28WCEAM%29%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>PSK 6201. (2003) <emphasis>PSK 6201 Maintenance Concepts and definitions</emphasis>, PSK Standardization Association.</para></listitem>
<listitem><para>PSK 7501. (2010) <emphasis>Key Process Indicators for Process Industry Maintenance</emphasis>, PSK Standardization Association.</para></listitem>
<listitem><para>Rao, J. S., Zubair, M. and Rao, C. (2003) &#x2018;Condition monitoring of power plants through the Internet&#x2019;, <emphasis>Integrated Manufacturing Systems</emphasis>, 14(6), pp. 508&#x2013;517. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Rao%2C+J%2E+S%2E%2C+Zubair%2C+M%2E+and+Rao%2C+C%2E+%282003%29+%27Condition+monitoring+of+power+plants+through+the+Internet%27%2C+Integrated+Manufacturing+Systems%2C+14%286%29%2C+pp%2E+508-517%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Rastegari, A. and Bengtsson, M. (2014) &#x2018;Implementation of condition based maintenance in manufacturing industry &#x2013; A pilot case study&#x2019;, <emphasis>Prognostics and Health Management (PHM)</emphasis> 22&#x2013;25 June 2014, Cheney, WA, USA. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Rastegari%2C+A%2E+and+Bengtsson%2C+M%2E+%282014%29+%27Implementation+of+condition+based+maintenance+in+manufacturing+industry+-+A+pilot+case+study%27%2C+Prognostics+and+Health+Management+%28PHM%29+22-25+June+2014%2C+Cheney%2C+WA%2C+USA%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Rastegari, A., Salonen, A., Bengtsson, M., and Wiktorsson, M. (2013) &#x2018;Condition based maintenance in manufacturing industries: Introducing current industrial practice and challenges&#x2019;, <emphasis>22nd International Conference on Production Research, ICPR 2013</emphasis>, Parana, Brazil, 28 July&#x2013;1 August. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Rastegari%2C+A%2E%2C+Salonen%2C+A%2E%2C+Bengtsson%2C+M%2E%2C+and+Wiktorsson%2C+M%2E+%282013%29+%27Condition+based+maintenance+in+manufacturing+industries%3A+Introducing+current+industrial+practice+and+challenges%27%2C+22nd+International+Conference+on+Production+Research%2C+ICPR+2013%2C+Parana%2C+Brazil%2C+28+July-1+August%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>SFS-EN 13306. (2010) <emphasis>Maintenance. Maintenance terminology</emphasis>, 2<emphasis>nd</emphasis> edn., Helsinki, Finnish Standardization Association SFS.</para></listitem>
<listitem><para>SFS-EN 15341. (2007) <emphasis>Maintenance. Maintenance Key Performance Indicators</emphasis>, Helsinki, Finnish Standardization Association SFS. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=SFS-EN+15341%2E+%282007%29+Maintenance%2E+Maintenance+Key+Performance+Indicators%2C+Helsinki%2C+Finnish+Standardization+Association+SFS%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Sharma, A. and Yadava, G. (2011) &#x2018;A literature review and future perspectives on maintenance optimization,&#x2019; <emphasis>Journal of Quality in Maintenance Engineering</emphasis>, 17(1), pp. 5&#x2013;25. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Sharma%2C+A%2E+and+Yadava%2C+G%2E+%282011%29+%27A+literature+review+and+future+perspectives+on+maintenance+optimization%2C%27+Journal+of+Quality+in+Maintenance+Engineering%2C+17%281%29%2C+pp%2E+5-25%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Shin, J. H. and Jun, H. B. (2015) &#x2018;On condition based maintenance policy&#x2019;, <emphasis>Journal of Computational Design and Engineering</emphasis>, 2, pp. 119&#x2013;127. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Shin%2C+J%2E+H%2E+and+Jun%2C+H%2E+B%2E+%282015%29+%27On+condition+based+maintenance+policy%27%2C+Journal+of+Computational+Design+and+Engineering%2C+2%2C+pp%2E+119-127%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Si, X., Wang, W., Hua, C., and Zhou, D. (2011) &#x2018;Remaining useful life estimation: A review on the statistical data driven approaches&#x2019;, <emphasis>European Journal of Operational Research,</emphasis> 213, pp. 1&#x2013;14. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Si%2C+X%2E%2C+Wang%2C+W%2E%2C+Hua%2C+C%2E%2C+and+Zhou%2C+D%2E+%282011%29+%27Remaining+useful+life+estimation%3A+A+review+on+the+statistical+data+driven+approaches%27%2C+European+Journal+of+Operational+Research%2C+213%2C+pp%2E+1-14%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Spare, J. H. (2001) &#x2018;&#x2018;Building the business case for condition-based maintenance&#x2019;, <emphasis>IEEE/PES Transmission and Distribution Conference and Exposition: Developing New Perspectives (Cat. No.01CH37294)</emphasis>, pp. 954&#x2013;956, https://doi.org/10.1109/TDC.2001.971371. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Spare%2C+J%2E+H%2E+%282001%29+%27%27Building+the+business+case+for+condition-based+maintenance%27%2C+IEEE%2FPES+Transmission+and+Distribution+Conference+and+Exposition%3A+Developing+New+Perspectives+%28Cat%2E+No%2E01CH37294%29%2C+pp%2E+954-956%2C+https%3A%2F%2Fdoi%2Eorg%2F10%2E1109%2FTDC%2E2001%2E971371%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Starr, A. (1997) &#x2018;A structured approach to the selection of condition based maintenance&#x2019;, In <emphasis>Proceedings of the 5th International Conference on Factory,</emphasis> pp. 131&#x2013;138.</para></listitem>
<listitem><para>Tandon, N., Yadava, G. S., and Ramakrishna, K. M. (2007) &#x2018;A comparison of some condition monitoring techniques for the detection of defect in induction motor ball bearings&#x2019;, <emphasis>Mechanical Systems and Signal Processing</emphasis>, 21, pp. 244&#x2013;256. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Tandon%2C+N%2E%2C+Yadava%2C+G%2E+S%2E%2C+and+Ramakrishna%2C+K%2E+M%2E+%282007%29+%27A+comparison+of+some+condition+monitoring+techniques+for+the+detection+of+defect+in+induction+motor+ball+bearings%27%2C+Mechanical+Systems+and+Signal+Processing%2C+21%2C+pp%2E+244-256%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Walker, N. (2005) &#x2018;The implementation of a condition based maintenance strategy, In <emphasis>Proceedings of the 18th International Congress of COMADEM,</emphasis> UK, Cranfield, pp. 51&#x2013;61. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Walker%2C+N%2E+%282005%29+%27The+implementation+of+a+condition+based+maintenance+strategy%2C+In+Proceedings+of+the+18th+International+Congress+of+COMADEM%2C+UK%2C+Cranfield%2C+pp%2E+51-61%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Ward, J. and Peppard, J. (2002) <emphasis>Strategic Planning for Information Systems</emphasis>, John Wiley &#x0026; Sons, Inc. New York, NY, USA &#x00A9; 2002, ISBN:0470841478. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ward%2C+J%2E+and+Peppard%2C+J%2E+%282002%29+Strategic+Planning+for+Information+Systems%2C+John+Wiley+%26+Sons%2C+Inc%2E+New+York%2C+NY%2C+USA+&#x00A9;+2002%2C+ISBN%3A0470841478%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Yu, J. (2012) &#x2018;Health condition monitoring of machines based on hidden Markov model and contribution analysis&#x2019;, <emphasis>IEEE Transactions on Instrumentation and Measurement</emphasis>, 61(8), pp. 2200&#x2013;2211. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Yu%2C+J%2E+%282012%29+%27Health+condition+monitoring+of+machines+based+on+hidden+Markov+model+and+contribution+analysis%27%2C+IEEE+Transactions+on+Instrumentation+and+Measurement%2C+61%288%29%2C+pp%2E+2200-2211%2E" target="_blank">Google Scholar</ulink></para></listitem>
</orderedlist>
</section>
</chapter>
<chapter class="chapter" id="ch03" label="3" xreflabel="3">
<title>The MANTIS Reference Architecture</title>
<para><emphasis role="strong">Csaba Heged&#x00171;s<superscript><emphasis role="strong">1</emphasis></superscript>, Patricia Dominguez Arroyo<superscript><emphasis role="strong">2</emphasis></superscript>, Giovanni Di Orio<superscript><emphasis role="strong">3</emphasis></superscript>, <!--br role="newline" /-->Jos&#x000E9; Luis Flores<superscript><emphasis role="strong">4</emphasis></superscript>, Karmele Intxausti<superscript><emphasis role="strong">4</emphasis></superscript>, Erkki Jantunen<superscript><emphasis role="strong">5</emphasis></superscript>, F&#x000E9;lix Larrinaga<superscript><emphasis role="strong">6</emphasis></superscript>, Pedro Mal&#x000F3;<superscript><emphasis role="strong">3</emphasis></superscript>, Istv&#x000E1;n Moldov&#x000E1;n<superscript><emphasis role="strong">7</emphasis></superscript>, and S&#x000F6;ren Schneickert<superscript><emphasis role="strong">8</emphasis></superscript></emphasis></para>
<para><superscript>1</superscript>AITIA International Inc., Hungary</para>
<para><superscript>2</superscript>University of Groningen, The Netherlands</para>
<para><superscript>3</superscript>FCT-UNL, UNINOVA-CTS, Caparica, Portugal</para>
<para><superscript>4</superscript>IK4-Ikerlan, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para><superscript>5</superscript>VTT Technical Research Centre of Finland Ltd, Finland</para>
<para><superscript>6</superscript>Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para><superscript>7</superscript>Budapest University of Technology and Economics, Hungary</para>
<para><superscript>8</superscript>Fraunhofer IESE, Germany</para>
<para>The purpose of this chapter is to describe the MANTIS reference architecture. The generic focus here is on an architecture that enables service-based business models and improved asset availability at lower costs through continuous process and equipment monitoring, aided by big data analysis. This architecture takes into account needs of various industries in the forefront of service-based business and operations models. It also takes into account less mature industrial domains, where improvements in maintenance can be only achieved gradually and consistently. The higher level requirements for the whole project are described by [The MANTIS Consortium, 2018] and further tuned for the architecture in [Jantunen et al., 2016].</para>
<section class="lev1" id="sec3-1">
<title>3.1 Introduction</title>
<para>The MANTIS proactive service maintenance platform and its associated architecture draws inspiration from the CPS approach. Physical systems (e.g., industrial machines, vehicles, renewable energy assets) operate in an environment, where everything is continuously monitored by a broad and diverse range of intelligent sensors.</para>
<para>This continuous, high resolution monitoring eventually results in massive amounts of data. Systems are characterized, for example, by their usage history, operational conditions, location, movements and other physical properties. These systems and machines form larger collaborative systems-of-systems over heterogeneous networks (e.g., vehicle fleets, photo-voltaic or windmill parks), and hence should be connected via robust communication mechanisms able to operate in challenging (industrial) environments. Here, sophisticated, distributed sensing and decision making functions are performed at different levels in a collaborative way ranging from:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>the local nodes (that pre-process raw sensor data and extract relevant information before transmitting it, thereby reducing bandwidth requirements of communication);</para></listitem>
<listitem>
<para>over intermediate nodes (that offer asset-specific analytics to locally optimize performance and maintenance);</para></listitem>
<listitem>
<para>into cloud-based platforms (that integrate information from ERP, CRM and CMMS systems and execute distributed processing and analytic algorithms for supporting global decision making processes).</para></listitem>
</itemizedlist>
<para>For the optimal maintenance of assets, different systems, and stakeholders will have to share information, resources, and responsibilities. In other words, collaboration is required. Such a Collaborative Maintenance Ecosystem will have to be able to reduce the adverse impacts of maintenance on productivity and costs, increase the availability of assets, reduce time required for maintenance tasks, improve the quality of the maintenance service and products, improve labor working conditions, and maintenance performance, increase sustainability by preventing material loss (due to out-of-tolerance production), and help optimizing spare part management.</para>
<para>The overall concept of MANTIS aims to provide a proactive maintenance service platform architecture that allows the precise forecasting of future performance, the prediction and prevention of imminent failures, and should also be able to schedule proactive maintenance tasks. This proactive maintenance service platform will consist of distributed processing chains that can efficiently transform raw data into knowledge while minimizing the need for transfer bandwidth, as already mentioned in <link linkend="ch01">Chapter <xref linkend="ch01" remap="1"/></link>.</para>
<para>Reference architectures provide a template, often based on the generalization of a set of solutions. This is also the case for MANTIS. These solutions may have been generalized and structured for the depiction of one or more architecture structures based on the harvesting of a set of patterns that have been observed in a number of successful implementations. Further, it shows how to compose these parts together into a solution. Reference architectures can be instantiated for a particular domain or for specific projects.</para>
<para>The role of the reference architecture in MANTIS is to provide guidance on how to instantiate an architecture for a particular MANTIS domain or specific MANTIS task, and to ensure consistency and applicability of technologies, interoperability mechanisms, data formats and models, and data analysis tools to be used in the different MANTIS use cases.</para>
<para>Based on requirements and use case descriptions, the reference service platform architecture and overall design needs to address a number of aspects. Important aspects addressed by this chapter are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Interface, protocol, and functional interoperability</emphasis> ensuring that several cooperating vendors can effectively assemble the complete MANTIS service platform. Includes the need to identify or develop standards for data semantic representation and exploitation;</para></listitem>
<listitem>
<para><emphasis role="strong">Data validation</emphasis> ensuring that data analyses are made on data that give clean, correct and useful data information about the system;</para></listitem>
<listitem>
<para><emphasis role="strong">Distributed data, and information processing, and decision-making</emphasis> ensuring consistent behavior and avoid contradicting actions, e.g., between local and distributed data analysis and decision making;</para></listitem>
<listitem>
<para><emphasis role="strong">Information validation</emphasis> ensuring that created information still is relevant for the system analyzed;</para></listitem>
<listitem>
<para><emphasis role="strong">System and service level security</emphasis> ensuring that the system incorporates means to hinder misconfiguration and can be protected from wire-tapping and various attacks;</para></listitem>
<listitem>
<para><emphasis role="strong">System engineering and re-usability</emphasis> of defined and existing services;</para></listitem>
<listitem>
<para><emphasis role="strong">System verification and validation</emphasis> of the service platform architecture and overall design, covering both functional and non-functional properties.</para></listitem>
</itemizedlist>
<section class="lev2" id="sec3-1-1">
<title>3.1.1 MANTIS Platform Architecture Overview</title>
<para>The development of a specific implementation of the reference architecture can make use of any of the generalized artifacts, described in this Chapter. All of them help in various ways to avoid having to create a whole reference architecture from scratch, and help to leverage the knowledge and experience that went into the formation and definition of the generalized models, architectures, and patterns.</para>
<para>Adopting a reference architecture within an organization accelerates delivery through the re-use of an effective solution and provides a basis for governance to ensure the consistency and applicability of technology use within an organization. In the field of software architecture, empirical studies have shown the following common benefits and drawbacks from adopting a software reference architecture within organizations [Martinez-Fernandez et al., 2015]:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>improvement of the interoperability of the software systems by establishing a standard solution and common mechanisms for information exchange;</para></listitem>
<listitem>
<para>reduction of the development costs of software projects through the reuse of common assets;</para></listitem>
<listitem>
<para>improvement of the communication inside the organization because stakeholders share the same architectural mind-set;</para></listitem>
<listitem>
<para>influencing the learning curve of developers due to the need of learning its features.</para></listitem>
</itemizedlist>
<para>The purpose of the MANTIS ecosystem is to make proactive maintenance possible in a scalable, multi-leveled way. We are targeting CBM: the processes are defined by the ISO 13374 standard [ISO, 2012].</para>
<para>In order to enable maintenance optimization and new business models within MANTIS, appropriate utility services and modules are elaborated and implemented. Within these, data mining and analytic services are created, which are mainly related to these functions [Jantunen et al., 2016]:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>RUL of components: continuous tracking of telemetry (usage) data and estimating how much time the given device or component has left before needs to be replaced;</para></listitem>
<listitem>
<para>FP: the system shall predict based on diagnostic data an inbound failure mode (different to wear-out to be detected by RUL);</para></listitem>
<listitem>
<para>RCA: when an unpredicted, complex failure occurs, the system shall deduct the actual module, the rout caused of the issue;</para></listitem>
<listitem>
<para>MSO: provide a decision making support on better maintenance planning.</para></listitem>
</itemizedlist>
<para>To facilitate the requirements and business goals, the MANTIS ARM consists of five elements from high-level design to implementation of the architecture in various use cases:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Reference Model: a reference model is an abstract framework for understanding significant relationships among the entities of some environment;</para></listitem>
<listitem>
<para>Reference Architecture: provide a template solution for the architecture (aka. architectural blueprint) for a particular domain;</para></listitem>
<listitem>
<para>Feature model: Introduces key concepts to characterize common and varying aspects in the architectures to be derived from the reference architecture;</para></listitem>
<listitem>
<para>Guidelines: discusses how the provided models, views and perspectives are to be used;</para></listitem>
<listitem>
<para>Reference applications: show the diversity of the included solution variants, and thus illustrate architecture signification features and related design decisions.</para></listitem>
</itemizedlist>
<para>The approach for architecting MANTIS use cases follows the principle of architecting for concrete stakeholder concerns (based on Architecture Drivers). These stakeholder concerns will drive the eventual architecture design, which is based in the approach follow by the SPES consortium [Pohl et al., 2012]. This approach suggests to start by delineating system and its context, then to continue with the functional decomposition of the system. The next step is the software realization. The final steps consider the hardware realization of functions and the deployment of software entities, as depicted in <link linkend="F3-1">Figure <xref linkend="F3-1" remap="3.1"/></link>.</para>
<fig id="F3-1" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-1">Figure <xref linkend="F3-1" remap="3.1"/></link></label>
<caption><para>Mantis architecture construction approach.</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_1.jpg"/>
</fig>
</section>
</section>
<section class="lev1" id="sec3-2">
<title>3.2 The MANTIS Reference Architecture</title>
<para>As discussed in <link linkend="ch02">Chapter <xref linkend="ch02" remap="2"/></link>, many of the requirement categories are related to the <emphasis>operating environment</emphasis> of the MANTIS architecture: communication restrictions and expectations, design principles, the need for web clients, integration of legacy human-machine interfaces, and so on. These have not been addressed by MANTIS on an implementation level (since being a reference architecture model), although, when designing an installation, we have to keep in mind that the final, integrated systems has to cover these as well. These categories included (i) data handling, (ii) event handling, (iii) guarantee-related, and (iv) security issues.</para>
<para>The most important implementation-related requirements are aimed towards scalability and fault tolerance in the data collection and processing. The inputs of the platform are coming from so-called edge devices, and the output is utilized by various enterprise systems and maintenance operations personnel. It is worth noting that the scope of MANTIS platform architecture does not include or target the actual life-cycle management of the devices. To do that, MANTIS relies on the already existing corporate systems, and resources. However, these interactions with external systems is planned and designed into the framework via standardized secure communications between platform modules using the interoperability guidelines set [Di Orio et al., 2018].</para>
<section class="lev2" id="sec3-2-1">
<title>3.2.1 Related Work and Technologies</title>
<para>Every novel result is based on previous work, which acts as background for the novel advances. In fact, background information lays the foundations for the MANTIS architecture, which is built over novel and existing technologies that are composed to allow for the MANTIS maintenance strategies. Moreover, other related work acts as reference and comparison for the MANTIS architecture. This section describes this plethora of information to support the description and discussion on the architecture.</para>
<section class="lev3" id="sec3-2-1-1">
<title>3.2.1.1 Reference architecture for the industrial internet of things</title>
<para>CPSs are nowadays built together within some form of IoT architectures. A &#x0201C;usual&#x0201D; IoT application employs various <emphasis>things</emphasis> collecting enormous amounts of data from a number of places and sending them to an <emphasis>IT cloud</emphasis> for a specific purpose. A survey of 39 IoT platforms [Mineraud et al., 2016] concluded in a generic architecture and common characteristics of IoT platforms. <link linkend="F3-2">Figure <xref linkend="F3-2" remap="3.2"/></link> depicts a generalized commercial IoT platform in its fullest form, and two possibilities are shown. The various IoT modules and services can be deployed on local premises &#x02013; or within a global IT cloud, depending on the restrictions made by the use case.</para>
<fig id="F3-2" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-2">Figure <xref linkend="F3-2" remap="3.2"/></link></label>
<caption><para>A generalized IoT framework [Mineraud et al., 2016].</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_2.jpg"/>
</fig>
<para>The generic modules in such a platform are the following [Mineraud et al., 2016]:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Sensor or actuator nodes, i.e., &#x0201C;motes&#x0201D; or &#x0201C;things&#x0201D; that create and then send in the data &#x02013; or act based on the received data;</para></listitem>
<listitem>
<para>Gateways that &#x0201C;hide&#x0201D; constrained devices that might be communicating via non-IP based networking (&#x0201C;legacy&#x0201D;) and/or incapable of implementing the platform interface on their own;</para></listitem>
<listitem>
<para>A platform interface that receives the data from the devices and passes it to other modules (gateway and data distributor);</para></listitem>
<listitem>
<para>Data storage, often distributed, which is accessible by other modules of the platform;</para></listitem>
<listitem>
<para>Various service modules that can access the historical or even the current inbound data streams and generate insights and various processing tasks (i.e., &#x0201C;big data applications&#x0201D;);</para></listitem>
<listitem>
<para>Graphical interfaces for operators to manage the system and validate the output (i.e., &#x0201C;Business Insights&#x0201D;).</para></listitem>
</itemizedlist>
<para>The data gathering and processing viewpoint shown by <link linkend="F3-3">Figure <xref linkend="F3-3" remap="3.3"/></link> corresponds to that of the IIoT reference architecture proposed by the IIC [Industrial Internet Consortium, 2017]. However, three additional architecture patterns are proposed to better suite the targeted environments, extending the general standalone IoT solutions. These include (i) a three-tier architecture pattern; aided by (ii) gateway-mediated edge processing; and a (iii) layered databus pattern.</para>
<fig id="F3-3" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-3">Figure <xref linkend="F3-3" remap="3.3"/></link></label>
<caption><para>The industrial internet of things reference architecture.</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_3.jpg"/>
</fig>
<para>This latter term is related to one of fundamental value added of the IIoT approach: enhancing <emphasis>&#x0201C;legacy&#x0201D; production systems</emphasis> by &#x0201C;making them smart&#x0201D; with additional, usually non-invasive components. This additional device (i.e., gateway) shall utilize e.g., the management interfaces of the machines, and represent the functionalities provided there using IP-based interfaces. Nevertheless, this results in the physical machines being connected to the network, their operations offered as &#x0201C;services&#x0201D; &#x02013; hence creating CPSs [Cengarle et al., 2013] out of them.</para>
<para>Therefore, we also have to create a logical space that implements a common schema, while it also has to provide a &#x0201C;language&#x0201D; used in the communications between endpoints (i.e., translation between various data description ontologies into one understanding). Such a logical data bus design pattern hence supports communication between applications and devices: semantics and translation are the basis for interoperability within MANTIS, and in IIoT, in general.</para>
<para>Moreover, this architecture is therefore dissected into three Tiers. The first one is the Edge Tier, where the sensors and actuators are located (e.g., production floors). In here, we are tapping out information from the communications within the (real-time) control loops between the given CPSs (cf. ISA95 systems [International Electrotechnical Commission, 2003&#x02013;2007]). This way, we are collecting mostly process telemetry, then aggregating and preprocessing it locally. This is usually supported by an <emphasis>application gateway</emphasis> that provides the connectivity: it bridges to a WAN towards the platform level(s). It also acts as an endpoint for the WAN, while isolating the local network of edge nodes (i.e., the involved local CPSs). This architecture pattern allows for localizing operations and controls (i.e., edge analytics and computing). Its main benefit, however, is that this way, we are breaking down the complexity of IIoT systems, so that they may scale up both in the numbers of managed assets as well as in networking.</para>
<para>The access network enables connectivity for data and control flows between the edge and the platform tiers. It may be a corporate network, or an overlay private network over the public Internet or a 4G/5G network.</para>
<para>The Platform Tier receives the streams of telemetry data from the Edge tier. It is also executing the control commands coming from the Enterprise Tier, and may forward some of these commands to the Edge Tier in a cloud-to-device manner. It consolidates and analyzes the data flows from the Edge Tier and other systems. It provides management functions for devices and assets (e.g., Over The Air firmware updates for the application gateways). It also offers non-domain specific services such as data query and analytics. The functional blocks of the cloud platform are the same as in any generic IoT platform.</para>
<para>Meanwhile, the Enterprise Tier receives the processed data flows (i.e., business insights) coming from the Edge devices towards Platform Tiers. It might also issue control commands to the Platform and Edge Tiers. This tier is the main beneficiary of the IIoT system. However, the utilization of a well-built MANTIS platform is also not an easy task on the corporate side either. The issues are presented and tackled in <link linkend="ch08">Chapter <xref linkend="ch08" remap="8"/></link>.</para>
</section>
<section class="lev3" id="sec3-2-1-2">
<title>3.2.1.2 Data processing in Lambda</title>
<para>The primary purpose of any (I)IoT systems is to create value added by processing the collected data in a cloud platform. To do so, there are many paradigms, software stacks (both open source and commercial), consultant firms. However, in general, the data processing usually follows the same logic. Within MANTIS, the Lambda architecture [Hausenblas and Bijnens, 2017] is considered, however, there are many, similar &#x0201C;competitors&#x0201D; of it as well [Kappa, 2018].</para>
<fig id="F3-4" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-4">Figure <xref linkend="F3-4" remap="3.4"/></link></label>
<caption><para>The Lambda data processing architecture.</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_4.jpg"/>
</fig>
<para>According to the generalized Lambda architecture pattern [Hausenblas and Bijnens, 2017] defined by industry experts, data can be processed either as soon as it reaches the platform (stream processing), or later on, on demand fetched from storage (batch processing). <link linkend="F3-4">Figure <xref linkend="F3-4" remap="3.4"/></link> depicts the overview of a generic analytic platform.</para>
<para>In here, the &#x0201C;speed layer&#x0201D; comprises of stream processing technologies, that are processing inbound data real time. This is an event-driven programming paradigm, where the processing functionality receives touples periodically, and executes the same function over them (e.g., creating a counter for a specific message type or classifying them using a well-taught machine learning algorithm). The other type of processing is asynchronous to the inbound data, and can be called on batched, already stored datasets. These tasks are run once at a time and might take long to complete &#x02013; such as the training phase of a machine learning algorithm. An other major responsibility of the batch layer is to maintain the (distributed) data storage, aided by the serving (database) layer. These modules are naturally part of any IIoT application. Many commercial products and platforms support these operations.</para>
<para>The batch layer has two major tasks: (a) managing historical data; and (b) recomputing results such as machine learning models. Specifically, the batch layer receives arriving data, combines it with historical data and recomputes results by iterating over the entire combined data set. The batch layer operates on the full data and thus allows the system to produce the most accurate results. However, the results come at the cost of high latency due to high computation time. The speed layer is used in order to provide results in a low-latency, near real-time fashion. The speed layer receives the arriving data and performs incremental updates to the batch layer results. Thanks to the incremental algorithms implemented at the speed layer, computation cost is significantly reduced. This is an event-driven programming paradigm, where the processing functionality receives touples periodically, and executes the same function over them (e.g., creating a counter for a specific message type or classifying them using a well-taught machine learning algorithm).</para>
</section>
<section class="lev3" id="sec3-2-1-3">
<title>3.2.1.3 Maintenance based on MIMOSA</title>
<para>OSA-CBM has developed an system architecture for condition-based maintenance, i.e., a way to enhance the modularization of different vendor systems, while not locking customers into a single-source solution. The MIMOSA has since the middle of the 1990&#x02019;s hosted open conventions for information exchange between plant and machinery information systems, namely a way to enhance, amongst other things, the compatibility issue between different vendor products [MIMOSA consortium, 2016].</para>
<para>The OSA-CBM and MIMOSA are two major standard organizations and they claim that an accepted non-proprietary open system architectural standard is important, since it would bring an improved ease of upgrading system components, a broader supplier community, more rapid technology development, and reduced prices [Lebold and Thurston, 2001].</para>
<para>One of MIMOSA&#x02019;s most valuable contributions are the development of a CRIS. It is a relational database model for different data types that need to be processed in a CBM application. The system interfaces are defined according to the database schema based on CRIS. The interfaces&#x02019; definitions developed by MIMOSA are an open data exchange convention to use for data sharing in today&#x02019;s CBM systems. In addition, defined by MIMOSA Cris is also MIMOSA&#x02019;s OSA-EAI, which provides an open exchange standard, for technology types, in key asset management areas, such as asset register management, work management, diagnostic and prognostic assessment, vibration and sound data, oil, fluid and gas data, thermographic data and reliability information.</para>
<para>Besides supporting all the data needed for a CBM application it also considers for instance the CMMS handling, to be precise work management of a maintenance department. The structure of data in a relational database is predefined by the layout of the tables and the fixed names and types of the columns, which is the case of the MIMOSA CRIS database model. In addition, during 2012 the OSA-EAI V3.2.3 released a complete UML model and XML schema implementation called CCOM-ML for the CCOM, in addition to continued support and updates for CRIS.</para>
<fig id="F3-5" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-5">Figure <xref linkend="F3-5" remap="3.5"/></link></label>
<caption><para>MIMOSA data model diagram [MIMOSA consortium, 2016].</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_5.jpg"/>
</fig>
<para>Within MANTIS, MIMOSA [MIMOSA consortium, 2016] is providing the common understanding and data ontology between partners and applications (<link linkend="F3-5">Figure <xref linkend="F3-5" remap="3.5"/></link>). It is presented as a defining standard format for the data exchange, while it also provides the data meta-model structure together with the definition of the ontologies of the data. In fact, one of the greatest benefits in using MIMOSA is this definition of semantics and ontologies so that parties developing their MANTIS solutions do not need to worry about how different types of information need to get linked together. In here, therefore, MIMOSA serves the role of the common data bus [Industrial Internet Consortium, 2017], while remaining loosely coupled. MANTIS has developed a full stack of message models extending the MIMOSA ontology that are designed to facilitate communications between edge and cloud, and also between the various cloud modules.</para>
<fig id="F3-6" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-6">Figure <xref linkend="F3-6" remap="3.6"/></link></label>
<caption><para>A three-part CBM architecture in the light of the OSA-CBM [Lebold and Thurston, 2001].</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_6.jpg"/>
</fig>
<para>Moreover, MANTIS follows the ISO-17359 standard in terms of the scope of functionality as a specific, maintenance-related IIoT implementation. According to this standard, a CBM system should be composed of various functional blocks, which then corresponds well to a general IoT system architecture with edge computing, as the implementations of the MANTIS architecture. In <link linkend="F3-6">Figure <xref linkend="F3-6" remap="3.6"/></link> can the three parts be visualized against the OSA-CBM architecture.</para>
</section>
</section>
<section class="lev2" id="sec3-2-2">
<title>3.2.2 Architecture Model and Components</title>
<para>As <link linkend="F3-7">Figure <xref linkend="F3-7" remap="3.7"/></link> suggests, the architecture follows the IEC IIoT [Industrial Internet Consortium, 2017] reference architecture model in general, in the sense of using the edge computing paradigm in connection with the gateway mediated pattern; and the MANTIS architecture is also planned for multiple tier levels. However, certain features are added to support additional, maintenance-related tasks, as well.</para>
<fig id="F3-7" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-7">Figure <xref linkend="F3-7" remap="3.7"/></link></label>
<caption><para>Overview of the MANTIS reference architecture.</para></caption>
<graphic xlink:href="graphics/ch03_figure3_7.jpg"/>
</fig>
<section class="lev3" id="sec3-2-2-1">
<title>3.2.2.1 Edge tier</title>
<para>Within the MANTIS use cases, there are primarily two main types of edge level devices: closed, fully fledged (i.e., production) sites and standalone devices (e.g., vehicles or outdoor measurement points). These two cases require completely different approaches, and completely different design in the edge-cloud interface.</para>
<para><emphasis>&#x02219; Standalone Devices</emphasis></para>
<para>In this case, there can be standalone machines, CPSs, vehicular system or other outdoors systems that are equipped with various sensors. Their communication capability is mostly wireless (i.e., via mobile networks) and therefore limited due to the radio interface capabilities. To these cases, MANTIS proposes an alleviated implementation of the edge-cloud interface, still relying on the MIMOSA ontology model. In these use cases, the standalone edge devices implement some pre-processing and aggregation algorithms, and only transmit an extract of the high frequency information available locally in order to save traffic.</para>
<para>Meanwhile, it is also possible that besides the periodical uploads from these devices, the platform level can still request additional, on demand, temporary data streams from the devices. This might help with the verification of a prediction provided by an analytic module. This measure is also implemented to save bandwidth. However, this might be limited in some use cases, since the nature of the edge device being a low power embedded system operating on battery (e.g., a sensor mote), attached to a machine. It is worth noting here from a development point of view, that these use cases require extensive training and familiarization phases, where all available data from all possible sources are collected. This phase is required in order to develop the necessary local analytic models that allows to decide what pre-processing is viable on-board (what information is necessary to transmit and what is redundant).</para>
<para>Furthermore, in these cases, the standalone devices (motes) are also expected to have intelligent functions on their own. These include intelligent sensor management, self diagnostics, and resilience to the unreliable nature of the communications. MANTIS does not propose limitations on these matters, but provides guidelines based on the lessons learned in the sensor developments of the project.</para>
<para><emphasis>&#x02219; Cyber-Physical Production Systems</emphasis></para>
<para>In many of the cases, the Edge Tier includes complex CPSoS, based on legacy production machines (&#x0201C;made smarter&#x0201D;) that are are connected through closed, self-contained networks. The primary data source consists in these CPSs, by means of their continuous (telemetry) output. Within MANTIS, building the automation based on the CPS context is out of scope.</para>
<para>Here, the choice fell on the gateway-mediated edge connectivity design, where a gateway is responsible for communications towards the platform level. These systems are usually not constrained by processing or power limitations. Rather the communications need to be efficient for different purposes: to increase the scalability of the overall platform. There was also the need to design various pre-processing and analytical modules that have to be put locally for at least three reasons. Firstly, raw process telemetry information is too much to send to any outside platform. Secondly, companies are reluctant to share critical real-time information about their core business, while also legal restrictions might apply [Donnelly, 2015]. Finally, maintenance personnel is usually located on-site. In a sense, therefore, these edge setups are complete on their own.</para>
</section>
<section class="lev3" id="sec3-2-2-2">
<title>3.2.2.2 Platform tier</title>
<para>When realizing the Platform Tier, MANTIS initially employs five modules. These are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>the edge broker;</para></listitem>
<listitem>
<para>multi-purpose (distributed) data storage and management;</para></listitem>
<listitem>
<para>stream processors;</para></listitem>
<listitem>
<para>batch processors;</para></listitem>
<listitem>
<para>HMI.</para></listitem>
</itemizedlist>
<para>The speed and batch layers are dedicated towards the three main maintenance-related objectives. The development of these modules are iterated over two phases: first an off-line, manual establishment of the algorithms with expert and data analyst knowledge is carried out, and then the developed solutions are deployed into the on-line system, taking the specific use case dependent constraints into account. The modules in the Platform Tier are generally intended to scale well: there can be multiple modules of each type resulting in a large distributed system. In such cases, a big data processing platform is needed for the implementation. Possible implementations build on commercial solutions, but open source implementations can also be applied. Within MANTIS, the following platforms have been used, without being exhaustive:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Microsoft Azure [Microsoft, 2017];</para></listitem>
<listitem>
<para>Amazon AWS [Amazon, 2017];</para></listitem>
<listitem>
<para>Apache ecosystem (Kafka [Apache Community, 2017], Storm [Apache Community, 2017] and Spark [Apache, 2017]);</para></listitem>
<listitem>
<para>Wapice IoT-Ticket [Wapice, 2018].</para></listitem>
</itemizedlist>
<para>The <emphasis>Edge Broker</emphasis> is responsible for keeping the direct communication with the edge level devices and gateways. It provides translation between the data format used in the edge-cloud interface, and within the platform level. Its primary purpose is to forward the inbound data towards the various cloud modules. It also includes the addition of all the required asset information to the upstream data to be MIMOSA compliant. Based on the added information, all the processing and database nodes in the platform can identify and process the inbound data. Moreover, since the edge-cloud interface is not restricted to one implementation, the edge broker usually implements multiple transport protocols. Therefore, edge devices can communicate via publish-subscribe protocols [Curry, 2004] (such as MQTT) or request-response type of protocols (such as RESTful HTTP). In a sense, the edge broker is fulfilling the role of an enterprise service bus [IBM, 2011], and here consists of three major components:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Protocol facades (e.g., an MQTT broker or an HTTP server, to receive the messages from the edge);</para></listitem>
<listitem>
<para>Message parser and translator (from the edge-cloud interface to the data distributor feed);</para></listitem>
<listitem>
<para>Client to the data distributor module (to push the translated message into the various platform modules).</para></listitem>
</itemizedlist>
<para>Within MANTIS, the Edge Broker has been implemented in a multitude of ways. One cornerstone of this module is that it might be use case and edge device dependent. The MANTIS platform does not wish to rule out legacy or COTS implementations for brown field use cases: the cost of deployment for MANTIS is intentionally kept to the minimum. Therefore, the edge broker is modular, and its main purpose is to translate between the various ways of communication formats within the framework, using the interoperability messaging schema of the framework. Moreover, since bidirectional communication is expected for some use cases in the edge-cloud interface, the edge broker is addressable by the cloud modules and can send messages (i.e., commands) towards the edge devices, as well.</para>
<para>The <emphasis>Data Management</emphasis> system (or data distributor) is needed when a large system is being implemented, to ensure scalability and robustness. In smaller deployments, the edge broker can forward the inbound data to other modules in the cloud, directly as well. Basically, this module is a message oriented middleware on its own. Its purpose is to collect the inbound data stream from the edge brokers, and build a data pipeline towards the other modules in the platform. One popular solution for data distribution is the Apache Kafka [Apache Community, 2017]. Section 3.3 further discusses the issues and design decisions to be made regarding data management and collection tasks necessary to build such IIoT big data systems as MANTIS.</para>
<para>The main data storage and everything connected to data descriptions within MANTIS is based on MIMOSA. The reference implementation of this domain ontology is provided in a Microsoft SQL database, deploying the MIMOSA structure. MANTIS has developed a RESTful HTTP interface for the database as well. This database is used for handling various types of information: from raw sensory data to the scheduled maintenance events, as discussed in section 3.2.1.3. This central database provides storage of the historical data (per asset and measurement point) for the analytic modules. The MANTIS reference HMI solutions also utilize it to fetch all information required for the overview of the system.</para>
<para>It is worth noting, that fully fledged production edge systems (i.e., CPPSs) can also have their own local storage, local MIMOSA instance. This enables easy operation and implementation: every level/tier utilizes its own MIMOSA instance, and when further interaction is needed between levels, it can happen via simple database synchronization, using the same semantics. This is one of the great advantages brought by MIMOSA, besides the implemented standard-compliant domain expertise (operational management) and affiliated information ontologies.</para>
<para>The <emphasis>Stream Processing</emphasis> functionality is required for several maintenance functions and features, that can be executed in real-time. Such functions include the detection and triggering of different types of events, based on simple rules such as thresholds. A typical example is when a measurement exceeds a threshold indicating a failure condition, and further investigation is needed to confirm the failure (hence fault prediction and root cause analysis). Moreover, various KPIs for predictive maintenance are also computed on-the-fly by the stream processor. Such KPIs include for example the RUL of the main components.</para>
<para>The <emphasis>Batch processors</emphasis> are designated to run asynchronous tasks (such as machine learning jobs) on big bulks of data. Such functions here include training root-cause analysis, prognosis estimation on historical data and possible recalibration of the applied machine learning models when sufficient new information has been collected. These typically require further information or historical data, fetched from the MIMOSA storage. These processes might be triggered by the stream processors during runtime or run periodically, and they perform complex tasks that are not needed to be real-time. An example scenario here is the detection of a possible failure: a value in the streaming data passing a threshold initiates a longer (i.e., batch) analysis on the system logs, for example looking for an known failure pattern beforehand.</para>
</section>
<section class="lev3" id="sec3-2-2-3">
<title>3.2.2.3 Enterprise tier</title>
<para>The Enterprise Tier consists of the following elements:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis>Analysis Applications</emphasis> provide result dashboards, as well as analysis request HMI for the operators and other experts at the enterprise level;</para></listitem>
<listitem>
<para><emphasis>Service Management Applications</emphasis> enable configuration and tracking of the services provided by the overall architecture in the given domain with all of its applications;</para></listitem>
<listitem>
<para><emphasis>Service Execution Applications</emphasis> support service deployment and execution;</para></listitem>
<listitem>
<para><emphasis>Management Applications</emphasis> enable configuration and tracking of the status for the <emphasis>platform</emphasis>, interfacing the Cloud- and Edge management functions at the Platform Tier;</para></listitem>
<listitem>
<para><emphasis>Edge IDE</emphasis> supports the configuration of the Edge Tier equipment and network setup through an Integrated Development Environment.</para></listitem>
</itemizedlist>
<para>The Enterprise tier provides the usual features of <emphasis>HMI applications</emphasis> such as presentation and processing of information, and moreover it adds mechanisms for explanation and adaptability based on user and application models. Therefore, these are knowledge-based systems for decision support as well. While MANTIS strongly emphasizes autonomy, self-testing, and self-adaptation, the human role remains one of the important factors in the system operation. It is however twofold: controlling, which comprises continuous and discrete tasks of open- and closed-loop activities and problem solving which includes the higher cognitive tasks of fault management and planning. These issues are further described in the next Chapters, as considerations are made for the generic parts (e.g., HMI) and installation-specific issues as well.</para>
<para>The HMI also benefits from the MIMOSA based implementation of the MANTIS architecture. As all information are stored in a database in a well-defined format, the relevant information can be easily extracted and presented in a unified manner. Furthermore, participating in the distribution process by subscribing to the relevant channels, an efficient HMI can be implemented. It supports decision making by proactively pushing relevant information to the right people at the right time, by intelligently filtering and summarizing information to prevent information overload through context awareness, by automatically and dynamically scheduling and adapting maintenance plans, thereby keeping the human in the loop at all times.</para>
</section>
<section class="lev3" id="sec3-2-2-4">
<title>3.2.2.4 Multi stakeholder interactions</title>
<para>One major issue tackled by MANTIS is the realization of multi-stakeholder integration and support for collaborative (maintenance) decision making: external and other corporate internal parties should be able access information tailored for them. One exemplary use case is the establishment of the necessary collaboration between the supplier of replacement parts and the service departments, since these have high stakes in the maintenance operations, as described in [Jantunen et al., 2018]. This requires a service-oriented approach [Bell, 2008].</para>
<para>Multiple Platform or Enterprise Tiers are enabled to access information coming from one single production site or edge device, in a controlled way. The same goes vice versa, one edge deployment shall be able to locate and connect to additional (external) services, once local decision making algorithms are deployed. An example case for this might be that a production plant shall be able to inquire replacement part orders if it detects the need for it (based on RUL estimation). All this can be aided by the architecture, so that multiple stakeholders can run-time receive and request information they need, in an asynchronous way. For this matter, the integration of the Arrowhead framework [Delsing, 2017] is proposed since it also supports other capabilities that are useful for advanced maintenance operations, such as Quality of Service (Albano, 2017).</para>
</section>
</section>
</section>
<section class="lev1" id="sec3-3">
<title>3.3 Data Management</title>
<para>Since MANTIS is proposing big data based analytic solutions to solve Maintenance 4.0 problems, the data storage solutions are essential to discuss and design into the framework&#x02019;s core architecture. Cisco has forecasted that by 2020, 92 percent of the workloads will be processed by cloud data centers of which 68 percent will be in public data centers [Cisco, 2016]. Within project MANTIS, Big Data are collections of data entries having the following characteristics [Munshi and Yasser, 2017; Laney, 2001; Assun&#x000E7;o et al., 2015] (while adding the newest and oriented interpretation [Fan and Bifet, 2013; Grover and Kar, 2017] dimensions as well):</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis>Volume</emphasis>: big data implies enormous volumes independently of the data source: machine or environmental sensors, event logs, external data sources;</para></listitem>
<listitem>
<para><emphasis>Variety</emphasis>: the diversity in source and format of the data collections. Although data repositories or processing may allow a restricted amount of heterogeneity, it is safe to include variety as one of the implicit features of big data as a whole;</para></listitem>
<listitem>
<para><emphasis>Velocity</emphasis> (i.e., data generation rate): Notwithstanding that collections and repositories may be static at a certain point, data is generated over time and that is what is referred to as velocity. Data might also be geographically distributed to make geographic static classifications, however such static representations of geographic distribution in an isolated or discrete point in time are more on the statistical side than in the big data spectrum;</para></listitem>
<listitem>
<para><emphasis>Variability</emphasis> in the units, data structures and formats that hold an equivalent representation of a measurement, state or data entries from different sources. Data transformations have to guarantee that input as output are equally faithful to the information they account for;</para></listitem>
<listitem>
<para><emphasis>Value</emphasis> is the engineering process behind that converts any piece of information into a mercantile asset that has enterprise value;</para></listitem>
<listitem>
<para><emphasis>Veracity</emphasis> makes reference to the quality of data. Raw data is often preprocessed to fit a particular data process. The veracity of the data refers to the appropriate handling of data formats and preprocessing transformations. Data transformations should always keep the relative real representation of the truth status they represent.</para></listitem>
</itemizedlist>
<para>For data to be turned into an economic asset, so called Big Data, engineers, architects and scientists are involved in the data gathering and engineering processes that collect and convert the terabytes of information into meaningful added value. One of the well recognized obstacles that prevents potential useful data from being exploited is that 60% to 80% of data mining efforts is strictly dedicated to the preparatory work. This is yet another motivation to carefully study the needs of a data exploitation infrastructure and put attention to data engineering process before turning data into insights and building a MANTIS (or alike) architecture deployment.</para>
<section class="lev2" id="sec3-3-1">
<title>3.3.1 Data Quality Considerations</title>
<para>Thorough the data life cycle, data transformations and data source integration processes, original data entries might suffer from inadequate handling, leading to misleading data analysis results. Some highlighted threats to data integrity are found in the literature [Singh et al., 2010].</para>
<para>Completeness here makes reference to whether data values are missing or that some data points are noisy. Besides that, completeness is a data quality issue that goes beyond the strict integrity of the data entries themselves. It also makes reference to the context information necessary to make adequate interpretations to the data. Meanwhile, validity refers to the correctness and reasonableness of data as the capability to faithfully represent the piece of reality it describes. As a title of example, it implies that data is describing reality in a scientifically accepted manner. Accuracy is the exactness with which data represents the real world. The concern here is whether a specific piece or collection of data or reading &#x0201C;makes sense&#x0201D; in comparison with the actual state of what originates that data. Meanwhile integrity: this entails in the broader sense the lack of corruption of data through the entire life cycle.</para>
</section>
<section class="lev2" id="sec3-3-2">
<title>3.3.2 Utilization of Cloud Technologies</title>
<para>Cloud services do not necessarily have to be outsourced to third parties entirely and in industrial cases they are not. Within an organization, cloud services can be provided internally, as private clouds; as shared, as hybrid clouds or fully hosted in a third party infrastructure, as commercial solutions. This classification entails the following [Lenk et al., 2009; Hashem et al., 2015]:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Private/Local clouds: all services, infrastructure, platform and software on which a solution runs, are hosted entirely by the party that exploits the solution. This does not imply that all the services in the system are proprietary solutions, there can be commercial, licensed products (services) hosted in a private environment;</para></listitem>
<listitem>
<para>Hybrid clouds: the solution is partially hosted by a third party service provider and partially hosted by the same entity that exploits the service. The distribution of who hosts what does not affect the classification;</para></listitem>
<listitem>
<para>Public or Commercial clouds: these are fully hosted by a service provider that is a different entity than the entity exploiting or consuming the service. In this case the differentiation between service provider and user/consumer is complete.</para></listitem>
</itemizedlist>
<para>It is possible, however, that a particular implementation has a service provider to fully host one of the layers of a system architecture (i.e., infrastructure, platform, software), in which case, the specific terminology to refer to the cloud hosting scheme can be used differently for each of those layers.</para>
<para>MANTIS promotes the utilization of distributed data storage solutions, regardless on how the solution shares its resources between the various types of CSP and on what level it is utilizing a COTS technology (i.e., provided infrastructure, software or platform &#x0201C;as a Service&#x0201D; [Bermbach et al., 2017]). The interoperability issues and how they can be tackled the MANTIS way when using different solutions and products in the data pipelines is presented in Section 3.4.</para>
</section>
<section class="lev2" id="sec3-3-3">
<title>3.3.3 Data Storages in MANTIS</title>
<para>The <emphasis>Central Data Storage</emphasis> is the main storage system where all maintenance related data is collected. It should be a scalable, possibly cloud based. Its implementation should be possible using both open source tools (e.g., Apache big data framework) or vendor specific (like Microsoft Azure). The adoption of MIMOSA (see Section 3.2.1.3) is recommended in the MANTIS architecture.</para>
<para>High granularity measurement data is usually not needed to be uploaded to the central database. However, <emphasis>local databases</emphasis> are maintained (in a CPPS for example), and in case when needed, the data can be accessed from the local database. Such high granularity measurement data can be for example raw vibration measurement data.</para>
<para>Although, in some cases local storage is not practical or it is not possible. However, diagnostics and prediction would benefit of the high granularity data. In these cases, it can be considered the possibility of ad-hoc data request in between the tiers of the MANTIS architecture (e.g., between edge-platform, where the CPS performs a high granularity measurement and provides the data as a response).</para>
<para>The MANTIS reference architecture proposes to encapsulate the actual storage technology behind a respective data management facade in order to mitigate the impact of technology changes of the data storage on the data analysis functions. There are several other databases that may be needed for the main functions of MANTIS (i.e., RUL, RCA, FP and MSO). Such databases include for example:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Environmental database (e.g., weather): to complete the diagnostics data environmental conditions may be taken into consideration;</para></listitem>
<listitem>
<para>Component stock database at subcontractors: needed for efficient planning;</para></listitem>
<listitem>
<para>Other examples: Safety/regulatory databases, etc.</para></listitem>
</itemizedlist>
<para>Scalability is the service property of the system to adapt to larger workloads (storage or cloud service) such as to &#x0201C;serve X% more requests when deployed on X% more resources&#x0201D; [Bermbach et al., 2017]. This is at least the general concept of scalability, which is also known as vertical scalability.</para>
<para>However, another aspect of scalability concerns the way a system is able to accommodate new functional requirements over time. This characteristic has been given the name of horizontal scalability, it played an important role in the evolution from RDBMS to NoSQL paradigms, with a lower threshold for system adaptation in the latter.</para>
<para>It is also worth noting that many enterprises are not only migrating their database systems to the cloud but also shifting their original relational databases to non structured NoSQL databases. One of the main reasons to adopt this new paradigm is the ability of the services in the digital system to handle real-time data, with high agility and flexibility. The problem is well posed in [Hecht and Jablonski, 2011]: &#x0201C;In the past SQL databases were used for nearly every storage problem, even if a data model did not match the relational model well. The object-relational impedance mismatch is one example, the transformation of graphs into tables another one for using a data model in a wrong way. This leads to increasing complexity by using expensive mapping frameworks and complex algorithms&#x0201D;.</para>
</section>
<section class="lev2" id="sec3-3-4">
<title>3.3.4 Storage Types</title>
<para>MANTIS relies on already existing data sources, or promotes the development of new data collection and storage if legacy sources are insufficient for the targeted analytic services.This section introduces some of the most popular storage solutions classified in a taxonomy of storage infrastructures. Although there are more technologies available both commercial and open source, the following have been researched to take in consideration for possible use. In this section, the following database types are considered for implementation of the MANTIS platform:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Traditional SQL databases: Microsoft SQL, MySQL, etc;</para></listitem>
<listitem>
<para>Big Data File Systems: Google File System, Hadoop File System, Disco File System;</para></listitem>
<listitem>
<para>NoSQL: Key-Value, Column-Oriented, Documents, Graph.</para></listitem>
</itemizedlist>
<section class="lev3" id="sec3-3-4-1">
<title>3.3.4.1 Big data file systems</title>
<para>Big data file systems in this section are easily scalable and can support a variety of data formats, and databases in both structured and unstructured data models. These are the most widely spread and flexible data storage systems, also the most rudimentary in terms of pre-packaged functionality.</para>
<para>The Google File System is a scalable distributed file system for large distributed data-intensive applications. It provides fault tolerance while running on inexpensive commodity hardware, and it delivers high aggregate performance to a large number of clients [Ghemawat et al., 2015]. A Google File System cluster consists of a single master and multiple chunkservers and is accessed by multiple clients. Chunkservers store chunks on local disks as Linux files and read or write chunk data specified by a chunk handle and byte range.</para>
<para>The HDFS [Apache, 2017] is also a distributed file system designed to run on commodity hardware. It has many similarities with existing distributed file systems. However, the differences from other distributed file systems are significant. HDFS is highly fault-tolerant and is designed to be deployed on low-cost hardware. HDFS provides high throughput access to application data and is suitable for applications that have large data sets. HDFS relaxes a few POSIX requirements to enable streaming access to file system data.</para>
<para>Disco Distributed Filesystem (DDFS) [Nokia, 2017] provides a distributed storage layer for Disco. DDFS is designed specifically to support use cases that are typical for Disco and MapReduce in general: Storage and processing of massive amounts of immutable data. This makes it very suitable for storing, for instance: log data, large binary objects (photos, videos, indices), or incrementally collected raw data such as web crawls. Although DDFS stands for Disco Distributed filesystem, it is not a general-purpose POSIX-compatible filesystem.</para>
</section>
<section class="lev3" id="sec3-3-4-2">
<title>3.3.4.2 NoSQL databases</title>
<para>In line with what the acronym indicates (Not Only SQL) NoSQL data storage systems integrate structured and not-structured or semistructured data structures. NoSQL is becoming increasingly popular for its compatibility with the object oriented programming paradigm that many application use nowadays [Assun&#x000E7;o et al., 2015]. NoSQL technologies avoid the architectural friction called Impedance Mismatch [Sahafizadeh and Nematbakhsh, 2015] between object oriented data generation at edge and hard-shaped relational database information storages. In here, the data types under consideration can be enumerated as (i) key-value, (ii) column-oriented, (iii) document-oriented, (iv) graph-oriented and (v) time-series.</para>
<para>Cassandra [Apache, 2017] is an open source distributed storage for managing key-value typed data. The properties mentioned in [Han et al., 2011] for Cassandra are the flexibility of the schema, supporting range query and high scalability. Cassandra, among others, sadly has the potential for denial of service attacks because it performs one thread per one client and it does not support inline auditing [Noiumkar and Chomsiri, 2014].</para>
<para>Other key-value typed database implementations are the Voldemolt<footnote id="fn3_1" label="1"><para>Project Voldemort. A distributed database. Online: http://www.project-voldemort.com</para></footnote>, Redis<footnote id="fn3_2" label="2"><para>Redislabs. Redis. Online: http://redis.io/</para></footnote> and DynamoDB<footnote id="fn3_3" label="3"><para>Amazon Web Services, Inc. Amazon DynamoDB. Online: http://aws.amazon.com/ dynamodb</para></footnote>.</para>
<para>Meanwhile, HBase<footnote id="fn3_4" label="4"><para>Apache HBase, Online: https://hbase.apache.org/</para></footnote> is an open source column oriented database modeled after Google big table and implemented in Java. Hbase can manage structured and semi-structured data and it uses distributed configuration and write ahead logging.</para>
<para>HyperTable<footnote id="fn3_5" label="5"><para>HyperTable. Online: http://www.hypertable.org</para></footnote> is also an open source high performance column oriented database that can be deployed on HDFS. Hypertable does not support data encryption and authentication [Noiumkar and Chomsiri, 2014]. Eventhough Hypertbale uses HQL which is similar to SQL, it has no vulnerabilities for the injection [Noiumkar and Chomsiri, 2014]. Additionally, no denial of service vulnerability is reported to work against Hypertable [Noiumkar and Chomsiri, 2014].</para>
<para>MongoDB<footnote id="fn3_6" label="6"><para>MongoDB. Online: https://www.mongodb.com/</para></footnote> belongs to the third category here, namely it is a document-based database. It supports complex datatypes and has high speed access to huge data [Han et al., 2011]. All data in MongoDB is stored as plain text and there is no encryption mechanism to encrypt data files.</para>
<para>There are also databases that are tailored to store time series. In here, arrays of numbers are indexed by time (a range of datetime). In some fields these time series are called profiles, curves, or traces. A time series of stock prices might be called a price curve. A time series of energy consumption might be called a load profile. A log of temperature values over time might be called a temperature trace. Popular implementations are eXtremeDB<footnote id="fn3_7" label="7"><para>eXtremeDB. Online: http://www.mcobject.com/extremedbfamily.shtml</para></footnote> and Graphite<footnote id="fn3_8" label="8"><para>Graphite. Online: https://graphiteapp.org/</para></footnote>.</para>
<para>Other document-typed NoSQL database implementations are the CouchDb and DynamoDB<footnote id="fn3_9" label="9"><para>DynamoDB. Online: http://aws.amazon.com/dynamodb</para></footnote>. Meanwhile, Neo4J<footnote id="fn3_10" label="10"><para>Neo Technology, Inc. Neo4j. Online: http://neo4j.com</para></footnote> is an open source graph database, for example.</para>
<fig id="F3-8" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-8">Figure <xref linkend="F3-8" remap="3.8"/></link></label>
<caption><para>MANTIS interoperability proposed approach.</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_8.jpg"/>
</fig>
</section>
</section>
</section>
<section class="lev1" id="sec3-4">
<title>3.4 Interoperability and Runtime System Properties</title>
<para>The MANTIS approach for interoperability specifications and guidance definition builds up on the main assumption that the identification of a reference model for the interoperability of CPS systems cannot be established without any link to the concrete, instantiated architectures. Therefore, in order to extract the main requirements and models for interoperability and its level of application, the steps presented in <link linkend="F3-8">Figure <xref linkend="F3-8" remap="3.8"/></link> have been followed. These steps provide the interoperability requirements that are used to create reference models for interoperability, i.e., to define specification and guidance to respond to the main interoperability requirements. The main steps of the proposed approach are the following:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Use Case Analysis: characterization of the use case concrete architecture in which the MANTIS platform will be integrated;</para></listitem>
<listitem>
<para>Concrete System Unified Description: unique description of the overall system, i.e., concrete use case architecture plus MANTIS platform;</para></listitem>
<listitem>
<para>Platform Interoperability needs: identification of the interoperability issues at platform;</para></listitem>
<listitem>
<para>Edge Interoperability needs: identification of the interoperability issues at edge level;</para></listitem>
<listitem>
<para>Component Interoperability needs: identification of the interoperability issues at component level;</para></listitem>
<listitem>
<para>Base technology: identification of the base technologies.</para></listitem>
</itemizedlist>
<para>In here, various reference models for interoperability are taken into account and utilized. It is necessary to identify elements on the tools and technology level that could potentially help the integration of the MANTIS platform within the use case ecosystems. Interoperability needs to be handled on four different levels that are component, edge, platform, and on the global level as well.</para>
<para>The component level focuses on how physical entities should be virtualized, i.e., how physical entities can be &#x0201C;cyberized&#x0201D; in terms of the functionalities and/or services that they are able to provide or in other words how to create MANTIS-enabled CPS. The edge level issues revolve around a set of physical entities belonging to the same local system logically represented by a LAN. At this level, the data extracted from physical entities is used to model and analyze the behavior of the system. The edge level also includes a sub-level that is the component level where physical entities are analyzed singularly. At the machine and component sub-level the data extracted from a physical entity is used to model and analyze the behavior of a physical entity singularly.</para>
<para>The platform level needs to describe the information exchange and data integration in the cyberspace. At this level, the data coming from the edge level is organized in order to be processed by MANTIS digital artifacts, i.e., integration between digital artifacts that are responsible to analyze the data provided by CPS located at edge level. Moreover, the global interoperability reference model needs to provide a unique model that is the confluence of the platform, edge and component interoperability reference models.</para>
<section class="lev2" id="sec3-4-1">
<title>3.4.1 Interoperability Reference Model</title>
<para>Inspired by IoT-A [IoT-A Reference Architecture Model, 2018], a MANTIS-ARM has been created to provide the cornerstone for designing, developing and deploying MANTIS-enabled solutions.</para>
<para>The reference architecture provides views and perspectives on distinct architectural aspects that provide the foundation for building compliant architectures. A perspective here defines a collection of activities, tactics, and guidelines that are used to ensure that a system exhibits a particular set of related quality properties that require consideration across a number of the system&#x02019;s architectural views [International Electrotechnical Commission, 1993]. Therefore, the interoperability perspective is something orthogonal to the several other views defined within the MANTIS reference architecture building blocks: interoperability is considered in all tiers of the architecture.</para>
<para>Considering a complex system (e.g., industrial production system) it is composed of a huge number of machines and their related components. Thus, the machine or component level comprises of physical entities that are part of the same functional unit. At machine and component level the main interoperability issues are related to</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>the definition of the granularity level for CPSs;</para></listitem>
<listitem>
<para>the design and development of communication library for extracting raw data from physical entities and appending it into the cyber entity.</para></listitem>
</itemizedlist>
<para>The edge level comprises physical entities that belong to the same local network and functional area. The topology and the intrinsic characteristics of the edge level strictly depend on the particular architectural pattern that is used for designing the MANTIS platform. As a matter of fact, the edge level can be as simple as an elementary gateway that delivers data to the platform level (where the intelligence is installed and deployed) or as complex as a network of digital artifacts that provide advanced edge data analytics and knowledge generation functionalities as well as transmission of the data to the platform level for more accurate and resource consuming tasks. In both of the cases, the main interoperability issues are related to:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>the integration of the CPS entities into the edge: It implies the virtualization of the physical entities into CPS (component level interoperability) to be integrated within the edge;</para></listitem>
<listitem>
<para>data extraction, translation and pre-processing from the available edge nodes;</para></listitem>
<listitem>
<para>provisioning of the data extracted from the CPSs (i.e., edge nodes) to the platform level.</para></listitem>
</itemizedlist>
<para>The platform level receives, processes and forwards commands from/to the edge level. It provides more complex and resource consuming data analytics and knowledge generation functionalities wrapped into digital artefacts than the edge level. At platform level data received from the edge level are organized according to a common ontology and/or data model that supports the data flow and exchange between the digital artefacts. Therefore, at platform level there are two main interoperability issues. The first one is concerned with the semantic data representation and exchange, to allow the digital artefacts of data analyitics to process the data. The second interoperability issue is on how to represent the knowledge models generated by the digital artefacts of data analytics, in order to enable the usage of such models back to the edge level for local control.</para>
<para>Finally, there are several interoperability issues that are orthogonal to the considered interoperability levels, i.e., models, guidelines and specifications that can be applied to all the interoperability levels without any restriction. These topics reveal interoperability issues that are related on how CPS and, more in general, digital artefacts are connected together. This includes the definition of the communication protocol and message exchange pattern to use in both edge and platform levels and the definition of an ontology of events to support systems interactions at both edge and platform levels.</para>
</section>
<section class="lev2" id="sec3-4-2">
<title>3.4.2 MANTIS Interoperability Guidelines</title>
<para>The MANTIS Interoperability Guidelines have been structured into three main parts:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Conceptual integration (modeling and design stage): it is focused on concepts and their relationships, models and meta-models. It provides the modeling foundation for systemizing the relevant interoperability aspects for the specific application domain;</para></listitem>
<listitem>
<para>Application integration (guidance to instantiate the models): it is focused on methodologies, guidance and patterns to support the design and development of their own MANTIS concrete instantiations;</para></listitem>
<listitem>
<para>Technical integration (specific implementation and integration): it is focused on technical aspects related to the networking (protocols, connectivity, etc.), hardware (CPU/memory power preferably low-cost and low-power consumption) and more in general to integration of heterogeneous data sources.</para></listitem>
</itemizedlist>
<section class="lev3" id="sec3-4-2-1">
<title>3.4.2.1 Conceptual and application integration</title>
<para>The model for conceptual integration (see <link linkend="F3-9">Figure <xref linkend="F3-9" remap="3.9"/></link>) has been created following a model-driven development approach to enable the design of interoperable and interconnected CPS-populated systems. It starts with the definition of a domain model (see <link linkend="F3-10">Figure <xref linkend="F3-10" remap="3.10"/></link>) that is aimed to capture the essence of the CPS while enabling the specification of the services and interfaces that the CPS must provide. The domain model is then complemented with the semantic data representation and exchange model and the system interaction model. The former is aimed to define and specify the structure of all the data and/or the information handled by CPS at the network level. The latter is aimed to define and specify the relevant events produced/consumed within the MANTIS platform, as well as, the distinct patterns for system interactions.</para>
<fig id="F3-9" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-9">Figure <xref linkend="F3-9" remap="3.9"/></link></label>
<caption><para>Model for conceptual integration.</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_9.jpg"/>
</fig>
<fig id="F3-10" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-10">Figure <xref linkend="F3-10" remap="3.10"/></link></label>
<caption><para>Domain model for MANTIS CPS.</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_10.jpg"/>
</fig>
<para>Architectural aspects are related to the use case specific, concrete architectural pattern used to design the MANTIS platform. Interoperability issues are different in number, type and location if a cloud-based or an edge-based pattern is used and/or a facade, broker or mediator [Martin, 2002] pattern is applied. While the CPS level issues are related to the design and development of a given CPS, i.e., to provide guidance and guidelines on how to virtualize physical entities (i.e., machines) in terms of services/functionalities, especially for those that are low-tech. This is connected to the information level that is related to the description of the data and the messages/structures exchanged, processed and stored. The messages/structures exchanged here are connected to a MIMOSA-compliant database that models the specific application context by using the MIMOSA OSA-EAI standard. Finally,interaction modeling is related to the definition and the identification of the necessary MEPs, i.e., how messages/structures are exchanged within the MANTIS platform.</para>
<para>The semantic data representation and exchange (see <link linkend="F3-11">Figure <xref linkend="F3-11" remap="3.11"/></link>) is aimed to describe the structure of all the data and/or information handled by cyber entities at a network level. Thus, the provided model details the way information should be modeled inside the cyber entity of a CPS and represents a necessary condition to guarantee that all the data circulating within the MANTIS platform cyberspace satisfies a well-defined structure to assure interoperability between different digital artifacts.</para>
<fig id="F3-11" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-11">Figure <xref linkend="F3-11" remap="3.11"/></link></label>
<caption><para>Semantic Data representation and information exchange model.</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_11.jpg"/>
</fig>
<para>In here, the user can be a human person or some kind of a Digital Artefact (e.g., a Service, an application, or a software agent) that needs to interact with a Physical Entity, where a digital artifact is a software component.</para>
<para>A Cyber Entity is represented in the digital world of Physical Entities. The Cyber Entities have two fundamental properties: (i) they are digital artifacts and (ii) they are the synchronized representation of a given set of aspects of their Physical Entities. Any change in the Physical Entity affects the Cyber Entity and vice-versa.</para>
<para>A resource is a software component that provides data from or is used in the actuation of Physical Entities. Since it is the Functionality that makes a Resource accessible, the relations between Resources and Cyber Entities are modeled as associations between Cyber Entities and Functionality. Resources can run on Devices or somewhere in the network (Network Resources). Devices host the technological interface (Native Communication Library) for interacting with, or gaining information about the Physical Entity.</para>
<para>The obvious similarities between IoT and CPS-based systems are always pushing the merging of the two research streams. The main aspects are represented by the elements <emphasis>Cyber Entity, Functionality Description and Association</emphasis>. A <emphasis>Cyber Entity</emphasis> is the cyber counterpart of a <emphasis>Physical Entity</emphasis> and the <emphasis>Functionality Description</emphasis> describes the set of functionalities the cyber entities are sharing within the virtual space. Actually, a functionality can be mapped to a service if the SOA technology [Thomas, 2008] is used or a skill/capability in a MAS [Jacques, 1999]. The Association is used to establish the connection between the <emphasis>Attribute</emphasis> of a cyber entity and the <emphasis>Functionality Description</emphasis>. As an example, for a temperature sensor a functionality could be the getTemperature function that provides information about the temperature <emphasis>Attribute</emphasis> value. A cyber entity can have zero to many different attributes.</para>
<para>Each <emphasis>Attribute</emphasis> has a name (attributeName), a type (attributeType), and one to many values (<emphasis>Value Containers</emphasis>). The attributeType specifies the semantic type of an attribute, for example, that the value represents temperature. Each value container groups one <emphasis>Value</emphasis> and zero to many <emphasis>Metadata</emphasis> fields that belongs to the given value. The metadata can, for instance, be used to save the timestamp of the value, or other quality parameters, such as accuracy or the unit of measurement. The cyber entity is also connected to the functionality description via the <emphasis>Functionality Description &#x02013; Cyber Entity</emphasis> association. Additionally, it may contain one (or more) <emphasis>Resource Description(s)</emphasis>. Finally, the resource description might contain information about the physical entity. The concept of Value within the MANTIS information model is specified according to the OSA-CBM open standard (see <link linkend="F3-12">Figure <xref linkend="F3-12" remap="3.12"/></link>).</para>
<fig id="F3-12" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-12">Figure <xref linkend="F3-12" remap="3.12"/></link></label>
<caption><para>Value concept specification.</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_12.jpg"/>
</fig>
</section>
<section class="lev3" id="sec3-4-2-2">
<title>3.4.2.2 System interaction model</title>
<para>Significant actions, incidents or episodes need to be registered and stored in the MANTIS platform that promotes monitoring or data analysis for performance improvement. Those events store data relevant to the entities related in the process (e.g., temperature of a surface) but additionally must collect both spatial and temporal information and associate them to entities/measures (e.g., Cyber Entity controlling rolling sheets or temperature). Events might be triggered and/or should be created after a given situation (such as after sensor reading, an operational action, a breakdown or other maintenance actions).</para>
<section class="lev4" id="sec3-4-2-2-1">
<title>3.4.2.2.1 MANTIS event model</title>
<para>Events are a fundamental part within MANTIS for supporting system interactions (at cyber level) at both edge and platform levels while interlinking data and automatic machine data processing. Consequently, a generic event model (see <link linkend="F3-13">Figure <xref linkend="F3-13" remap="3.13"/></link> has been designed to provide the skeleton for the definition of all the events produced/consumed within the MANTIS platform.</para>
<fig id="F3-13" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-13">Figure <xref linkend="F3-13" remap="3.13"/></link></label>
<caption><para>MANTIS event model (based on the IoT-A event information model).</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_13.jpg"/>
</fig>
<para>The base type of all events is the <emphasis>Abstract CEP Event</emphasis> type. It provides the skeleton and basic information for modelling all the events within the MANTIS platform. This basic information is the <emphasis>DateTime</emphasis>, the <emphasis>EventDescription</emphasis> and the <emphasis>MeasurementLocation</emphasis>. The concepts/classes DateTime and <emphasis>MeasurementLocation</emphasis> are defined to adhere to the OSA-CBM standard (the same is for the unique identifier type UUID).</para>
<para>There are two very generic concrete events type: i) the <emphasis>Simple CEP Event</emphasis>, that contains atomic event information and ii) the <emphasis>Complex CEP Event</emphasis> that contains information derived by a complex event processing application. The <emphasis>Value Container</emphasis>, <emphasis>Value</emphasis> and <emphasis>MetaData</emphasis> concepts/classes are used to model the content of each event that in turn can be as simple as a temperature value to complex strings that are serialized objects.</para>
</section>
<section class="lev4" id="sec3-4-2-2-2">
<title>3.4.2.2.2 Patterns for interactions</title>
<para>The definition of the most suited MEPs in the context of distributed computed is a typical problem [Martin, 2002]. MEPs refers to the way messages are exchanged between distributed components. It is worth noting that the selection and usage of a message exchange pattern can affect the way digital artifact should be implemented in order to be interoperable. In the context of MANTIS, three type of MEPs are considered, namely:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Push/Fire-and-forget: messages are sent in between digital artefacts and with CPSs. The sender sends the message and the receiver receives the message and ends the message exchange activity;</para></listitem>
<listitem>
<para>Request/Response-Reply: request response messages are sent in between digital artefacts and with CPSs. The sender sends a request message and the receiver receives the message and informs the sender with a response;</para></listitem>
<listitem>
<para>Publish/Subscribe: messages are sent in between digital artefacts and with CPSs in the form of events. The event source publishes a topic and all the digital artefacts and/or CPS that are interested to the topic subscribe to it and will receive events.</para></listitem>
</itemizedlist>
<para>The types of CPS and/or more in general digital artefacts that can be found within the MANTIS platform are: MANTIS-enabled that are supposed to be natively MANTIS platform compatible. These CPSs and/or digital artefacts already support the MANTIS interoperability specifications and can be immediately integrated within the MANTIS platform. These digital artefacts and/or resources need mechanisms and/or additional work in order to be integrated within the MANTIS platform. In this case, external adapters are needed to harmonize and bring together them within the MANTIS platform, i.e., translations from native to common protocols used in MANTIS.</para>
<fig id="F3-14" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-14">Figure <xref linkend="F3-14" remap="3.14"/></link></label>
<caption><para>Model for technical integration.</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_14.jpg"/>
</fig>
</section>
</section>
<section class="lev3" id="sec3-4-2-3">
<title>3.4.2.3 Implementation integration</title>
<para>The model for implementation integration (see <link linkend="F3-14">Figure <xref linkend="F3-14" remap="3.14"/></link>) has been developed to capture and show how the provided interoperability models can be related to a concrete MANTIS platform instantiation with specifying and/or establishing technologies used.</para>
<para>The business model describes the specific domain for the software solution that needs to be implemented. The software model represents the instantiation of the semantic data representation and exchange model and system interaction model in the specific domain derived from the business model. Therefore, in the software model all the necessary aspects of the conceptual integration are included. Finally, the technical architecture represents and describes the concrete environment, infrastructure and related technologies for supporting the platform/application.</para>
<para>The technical framework of the MANTIS platform can be distilled into a model (see <link linkend="F3-14">Figure <xref linkend="F3-14" remap="3.14"/></link>) focused on the interoperability perspective. The model is represented by a 4-tier model that covers all the necessary issues and aspects that developers need to consider whenever they want to implement MANTIS-compliant systems, i.e., data representation and exchange, system interactions (event models and patterns for interaction), data transformation and translation and services and functionalities definition, as well as, physical entities virtualization (domain model). A central part of the framework is the MIMOSA data storage that acts as a facilitator for the design and implementation of maintenance applications within the business tier.</para>
</section>
</section>
</section>
<section class="lev1" id="sec3-5">
<title>3.5 Information Security Model</title>
<para>Traditionally, security has been just a commodity but along the years this perception has being changing to become an integral and an inseparable part of any system. Indeed, security nowadays is a functional requirement to become interoperable with many existing systems. In this sense, MANTIS addresses these requirements and has been not only focusing in providing functionalities but being secure by design. This aim can only be accomplished by means of a modern secure information model and the most suitable access control information system. In order to achieve this goal is necessary to understand the basic pillars of the information security [Rahalkar, 2016]:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis>Integrity</emphasis>: to maintain the completeness and accuracy of data over its entire lifecycle;</para></listitem>
<listitem>
<para><emphasis>Confidentiality</emphasis>: to guarantee the privacy of data over its entire lifecycle to unauthorized individuals, entities or processes;</para></listitem>
<listitem>
<para><emphasis>Availability</emphasis>: to guarantee that the information is available when is needed.</para></listitem>
</itemizedlist>
<para>Nevertheless, these principles impose various requirements towards the architecture. The system must be defined with having</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>its every relevant element supplied with a digital identity;</para></listitem>
<listitem>
<para>a specific information model for managing different levels of confidentiality;</para></listitem>
<listitem>
<para>that is enforced by a security policy model.</para></listitem>
</itemizedlist>
<para>From these requirements the right process for obtaining a realistic and effective security management system involves the following processes:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Process 1. Digitization or the process of obtaining a unique and distinguishable digital identity;</para></listitem>
<listitem>
<para>Process 2. Definition of an information model;</para></listitem>
<listitem>
<para>Process 3. Definition of a control access policy specification;</para></listitem>
<listitem>
<para>Process 4. Definition of additional requirements associated with MANTIS.</para></listitem>
</itemizedlist>
<para>Finally, practical consideration is detailed in order to fulfill with existing security technologies such as advanced threat detection techniques.</para>
<section class="lev2" id="sec3-5-1">
<title>3.5.1 Digital Identity</title>
<para>In MANTIS every object, subject and action must have a digital reflect recorded. Therefore, it is necessary to establish a specification and classification of the elements of the system, that can be involved in the processing of the information. In this context, digital identities are the key to be able to establish effective and realistic security policies, without them it is not possible to control the behavior of the system. Having in mind, MANTIS associates every element of the platform with one of the following categories:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis>Subject</emphasis> is the element in charge of requesting operations (actions) with objects. These are the actors of the system. In many situations the subjects are processes intermediated by users but there are other situations where the processes are not associated with users;</para></listitem>
<listitem>
<para><emphasis>Action</emphasis> is the definition of an operation; every operation must be defined in order to control the behavior of the system;</para></listitem>
<listitem>
<para><emphasis>Object</emphasis> are the elements, which receive the actions. In this category, certain elements can be subjects and/or objects such as processes.</para></listitem>
</itemizedlist>
<para>Nowadays, the process of giving an identity to an object/subject is performed by generating a digital certificate, and an unique identifier [Vacca, 2004], as depicted in <link linkend="F3-15">Figure <xref linkend="F3-15" remap="3.15"/></link>.</para>
<fig id="F3-15" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-15">Figure <xref linkend="F3-15" remap="3.15"/></link></label>
<caption><para>Elements having their digital identity (certificate).</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_15.jpg"/>
</fig>
</section>
<section class="lev2" id="sec3-5-2">
<title>3.5.2 Information Model</title>
<para>In previous steps, the digital identity for every element of MANTIS was defined. Now, it is necessary to define the security classifications of information while taking into account the potential of the platform under two challenging situations:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Industrial environments. The companies try to maintain the availability of the system as well as the confidentiality of the information of the processes (key performance indication, KPI), the model of machines, the technology used, etc. In addition to this, there are two factors that become very important:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>The integrity of the industrial processes is very important for taking right decisions on processes;</para></listitem>
<listitem>
<para>The system must respect operational safety systems. MANTIS cannot interfere with real-time systems, since that may produce personal injuries or catastrophic losses.</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>Medical environments. The personal health information is one of the most critical assets, very restrictive legislation exist a in many countries related to that (c.f. GDPR [Donnelly, 2015]). MANTIS should implement a model, which guarantees this confidentiality of the information.</para></listitem>
</itemizedlist>
<fig id="F3-16" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-16">Figure <xref linkend="F3-16" remap="3.16"/></link></label>
<caption><para>Elements of the BLP model [Hansche et al., 2003].</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_16.jpg"/>
</fig>
<para>The assessment of these two challenging environments leads the use of restrictive security information models. A priori, the first candidate is the most restrictive information model known as Bell LaPadula (BLP) [Hansche et al., 2003]. This model is used in government and military applications and it is focused in enforcing the access control to confidential data. This model has the following properties, see <link linkend="F3-16">Figure <xref linkend="F3-16" remap="3.16"/></link>:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>The simple security property. This establishes that a subject of a specific level cannot read information at a higher security level;</para></listitem>
<listitem>
<para>The &#x0002A;(star) property. This establishes that a subject of a specific level cannot write to any object at a lower security level;</para></listitem>
<listitem>
<para>The discretionary security property. In this the specification of the discretionary access control is made by means an access matrix;</para></listitem>
<listitem>
<para>Security levels introduced: Top Secret, Secret, Confidential and Unclassified.</para></listitem>
</itemizedlist>
</section>
<section class="lev2" id="sec3-5-3">
<title>3.5.3 Control Access Policy Specification</title>
<para>The information model requires an access control policy to ensure security of the system. With this aim in mind, the original model specified (BLP) establishes the policies as a matrix where the access to every element is specified. The main drawback of the original specification here is related to the inherent complexity of MANTIS, which requires a more complex specification of security, and at the same time, a way to facilitate the management of the security. The direct application of this matrix will conclude in a huge matrix and poses significant problems to manage the policy in real life. With the aim of overcoming this drawback, an initial approach is trying to facilitate with the use of the concept of Roles. It is a mechanism for grouping sets of subjects with the same level of security but with some other interesting properties in terms of manageability:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Encapsulating the organizational functions/duties of a user;</para></listitem>
<listitem>
<para>Different roles can be defined, each for different types of competences, which are then assigned to users;</para></listitem>
<listitem>
<para>Realizing the security principle of &#x0201C;least privilege&#x0201D; [Rahalkar, 2016];</para></listitem>
<listitem>
<para>It is consistent with BLP model.</para></listitem>
</itemizedlist>
<para>Therefore, in MANTIS, the first approach to manage the security will be the use of roles and BLP for establishing the security policy. The NIST establishes some subdivisions of the original model such as: Core RBAC, Hierarchical RBAC, Constraint RBAC and Consolidated Model [Ferraiolo and Kuhn, 1992]. An RBAC model can formally be described by the tuple RBAC=&#x0003C; <emphasis>U,R,P,O</emphasis>,>, the most important elements of this tuple are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>U: User;</para></listitem>
<listitem>
<para>R: Role;</para></listitem>
<listitem>
<para>P: Permission;</para></listitem>
<listitem>
<para>O: Object.</para></listitem>
</itemizedlist>
<para>The model that best suits to MANTIS is the <emphasis>Hierarchical and Constrained RBAC</emphasis> model [Ferraiolo and Kuhn, 1992], which supports challenging environments and situations. The joint use with the BLP model facilitates the management of the security of the system. MANTIS uses the standard for specifying the security policy called eXtensible Access Control Markup Language (XACML) [OASIS, 2018]. This standard provides</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>a Policy Language;</para></listitem>
<listitem>
<para>a Request and Response Language;</para></listitem>
<listitem>
<para>Standard data-types, functions, combining algorithms.</para></listitem>
</itemizedlist>
<para>It is extensible, where there can be privacy profiles, with architecture defining the major components in an implementation. The structure of a security policy is specified in <link linkend="F3-17">Figure <xref linkend="F3-17" remap="3.17"/></link> and the relevant, general terms within XACML are the following:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Resource: Data, system component or service;</para></listitem>
<listitem>
<para>Subject: An actor who requests to access certain Resources;</para></listitem>
<listitem>
<para>Action: An action on resource;</para></listitem>
<listitem>
<para>Environment: The set of attributes that are relevant to an authorization decision and are independent of a particular subject, resource or action;</para></listitem>
<listitem>
<para>Attributes: Characteristics of a subject, resource, action or environment;</para></listitem>
<listitem>
<para>Target: Defines conditions that determine whether policy applies to request.</para></listitem>
</itemizedlist>
<fig id="F3-17" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-17">Figure <xref linkend="F3-17" remap="3.17"/></link></label>
<caption><para>Policy hierarchy.</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_17.jpg"/>
</fig>
</section>
<section class="lev2" id="sec3-5-4">
<title>3.5.4 Additional Requirements</title>
<para>MANTIS is an agnostic reference architecture, which can be applied in many environments, but like many technologies it is necessary to have high adaptibility. In this sense, many situations might exist where the classic security information model and their corresponding access control policies are instantiated; however there might be other situations involve relevant additional changes (e.g., the integration of new sensors in every machine that the company sells). This innocent change might cause a nightmare in terms of efficiency when only classic security access controls are in place. Taking these aspects in MANTIS into account, it will be necessary to consider the nature of changes to be introduced for improving the efficiency of the security management.</para>
<para>MANTIS considers not only the joining of the RBAC [Ferraiolo and Kuhn, 1992] and BLP [Hansche et al., 2003], but the use of modern approaches, that is, an ABAC model. The solution relies on the use of additional PKI [Vacca, 2004] in a different way to support the sharing of responsibilities. With these, emphasis is put on having a controlled but shared access model. The access to information is conditioned by the security policy defined in the cloud, managed by the company. This allows controlling situations such as validity periods for accessing the information but also it would be beneficial to the original company to be able to control external parties interesting in their data. This leads to work with a federated PKI for sharing Certification Authorities and to introduce the concept of secret sharing schemes. An increase in the number of sensors also requires additional considerations. In order to alleviate this, MANTIS introduces the concept of attributes in final elements for controlling the access by using existing RBAC based policies.</para>
</section>
</section>
<section class="lev1" id="sec3-6">
<title>3.6 Architecture Evaluation</title>
<para>Software and systems design, in its core essence, is the creative activity of software engineers making principal design decisions about a software system to be built or to be evolved. It translates the concerns and drivers in the problem space into design decisions and solution concepts in the solution space. Architecture evaluation is a valuable, useful, and a worthwhile instrument to manage risks in software engineering. It provides confidence for decision-making at any time in the lifecycle of a software system [Knodel and Naab, 2016].</para>
<section class="lev2" id="sec3-6-1">
<title>3.6.1 Architecture Evaluation Goals, Benefits and Activities</title>
<para>Architecture evaluation is the key quality-engineering instrument in software and systems design. Its goal is to make sure that the resulting systems really exhibit the desired qualities. To this end, it pursues two major objectives:</para>
<para>On the one hand, architecture evaluation aims at improving the overall quality of software and systems design. Architecture evaluations challenge the decisions made. They help clarifying quality requirements and enable to analyse the adequacy of the architecture solution. Further, they allow predictions of the impact of the architecture solutions and consequently the decisions made on the quality of the resulting system. As the architecture sets the course for the resulting system, quality problems and drawbacks of decisions can be detected early. Thus, architecture evaluation serves to mitigate risks. Eventually, it enables the improvement of the architecture by correcting and adapting the decisions made. Furthermore, architecture evaluation comes along with the side effect of increased architecture awareness in the development organization. While reasoning about and communicating the decisions made, their understanding in the organization is improved. Architecture evaluation reveals the rationales for the decisions, and their justification allows achieving a common understanding in the development organization. In short, architecture evaluation determines how well suited the architecture of the system for its purpose is.</para>
<para>On the other hand &#x02013; once having a well-designed architecture &#x02013; architecture evaluation aims as well at preserving architectural decisions and quality in the evolution of software systems. The follow-up activities in the lifecycle of a software system first translate architectural decisions into component models, detailed design models, and eventually source code including data structures and algorithms. At later points in time, evolving requirements and change requests yield modified system artefacts. However, to reap the architectural investment benefits, the managed software system lifecycle needs to enforce and preserve the architectural decisions made. This sustainment of architectural decisions assures that the architecture is in fact the conceptual tool to cope with challenges in its evolution. To be able to evaluate the decision enforcement, architecture documentation must trace all decisions to the system artefacts. Traceability of architectural decisions breaks down into the accuracy of their description and the distance of the system to its architecture.</para>
<para>With this said, the benefits of architecture evaluation turn out to be an improved software architecture, improved architecture documentation, and improved implementations of architectural solutions. The evaluation activities help to mitigate risks by raising the likelihood to detect problems early and to clarify the required system qualities. Evaluation also improves the understanding of design decisions and leads to a higher awareness of the architecture in organizations. As needed for input, the traces from architectural decisions to the system artefacts help their preservation and allow for higher compliance in implementations.</para>
<para>To achieve the above-mentioned objectives, MANTIS follows the RATE approach [Knodel and Naab, 2016], developed by Fraunhofer. RATE is a compilation and collection of best practices of existing evaluation approaches tailored towards a pragmatic (or rapid) application in industry. It comprises five checks, whereby each check serves for a distinct purpose. All checks follow the same working principle: to reveal findings to confirm and improve the system quality and/or the artifact quality.</para>
</section>
<section class="lev2" id="sec3-6-2">
<title>3.6.2 Concepts and Definitions</title>
<para>RATE uses the following concepts and definitions in its analysis:</para>
<para><emphasis>Stakeholder</emphasis>: A stakeholder in a software architecture is a person, group, or entity, with an interest in or concerns about the realization of the architecture [ISO/IEC/IEEE, 2011].</para>
<para><emphasis>Concern</emphasis>: A concern about an architecture is a requirement, an objective, an intention, or an aspiration that the stakeholder has for the architecture [ISO/IEC/IEEE, 2011].</para>
<para>Usually the stakeholder concerns are not consolidated or validated. This makes dealing with the stakeholder concerns challenging, since concerns might be ambiguous, conflicting with other stakeholders&#x02019; concerns, and are likely to be incomplete. In case a concern is specific, unambiguous, and measurable, it is possible to call it a requirement concern; otherwise, architecture scenarios can capture concerns. In either case, stakeholder concerns form the product and drive the architecture, which explains the need for compensation of missing and too complex concerns, for their aggregation and consolidation as well as their negotiation in case of ambiguity and inconsistency. This the architect does by deriving architecture drivers from the concerns.</para>
<para><emphasis>Architecture Driver</emphasis>: In general, it can be drawn a distinction between four main classes of architecture drivers: business goals, functional requirements, constraints, and quality requirements. Each of these classes might have its individual stakeholders that articulate concerns belonging to that particular class. In other words, all drivers originate from stakeholders in one way or another. The identification and analysis of stakeholders for further requirements elicitation and their stake within the architecture development is therefore key to any architecture definition or evaluation.</para>
<para><emphasis>Business Goals</emphasis>: are the first class (and most abstract) of architectural drivers. Business goals are goals that are important for the overall enterprise that is developing the respective architecture or has placed an order to build the system. Usually the business goals are quite abstract and are only partially depending on the architecture under consideration. However, the business goals are the most essential ones, since without the business goals there would be no need to think about creating an architecture of (a set of) products that end up in supporting a business goal. Examples for business goals are time to market (denoting the strategy in terms of time), the market scope, or costs.</para>
<para><emphasis>Functional Requirements</emphasis>: are drivers for the architecture as well. However, there are differences in functional requirements: some drive the architecture &#x02013; some do not. It depends on characteristics like &#x0201C;Does this particular function separate us from competitors&#x02019; products?&#x0201D; In some sense, the functional requirements that make the product unique and worth building are the ones that influence the architecture development the most. These kinds of functional requirements the architecture needs to explicitly support; otherwise, the endeavour of building an architecture would be meaningless.</para>
<para><emphasis>Quality Requirements</emphasis>: Quality is not only about correctness of functionality. Successful software systems have to assure additional properties such as performance, security, extensibility, maintainability, and so forth. In general, it is possible to distinguish between run-time and development-time quality attributes. Run-time quality attributes can be measured by watching the respective system in operation. Examples for run-time quality attributes are performance, security, safety, availability, and reliability. Development-time quality attributes can be measured by watching a team in operation. Examples for development-time quality attributes are extensibility, modifiability, and portability. One problem, however, is that there is no standard measurable meaning of quality attributes. Besides naming issues, people also tend to create new notions on their own. The solution towards this problem is to utilize the so-called architectural scenarios (see <emphasis>Architecture Scenario</emphasis>) that make the meaning of the quality attribute in the system context clearer and that lower the chance for misinterpretations.</para>
<para><emphasis>Constraints</emphasis>: One important but easily overlooked input for software and systems design are constraints that influence the design decisions of subsequent steps. Constraints can be organizational, technical, regulatory, or political. Organizational constraints might arise from the resources available for a particular system development effort. Technical constraints might arise from legacy systems that are already deployed in the field. Regulatory constraints usually stem from obligations to comply with particular standards. Depending on the domain, there might be different standards to consider. Political constraints are most of the time disguised as technical constraints. There might be different (more or less reasonable) roots for the existence of the constraint, however, since the source of the constraint is most likely higher management there is only low negotiability of the constraint from the perspective of the architects. Making them explicit, however, provides a solid basis for subsequent decision making in design.</para>
<para><emphasis>Architecture Scenario</emphasis>: An architectural scenario is a crisp, concise description of a situation that the system is likely to face, along with a definition of the response required of the system [Rozanski and Woods, 2011]. Architecture scenarios could be used to document both, functional and quality drivers in a measurable way, but are especially used for capturing software and system qualities. Functional requirements are usually clearer than quality requirements. However, it is the quality requirements that drive the architecture most. Therefore, it is crucial to elicit required qualities using scenarios in a measurable way, so that architects or evaluators can find a baseline to work with. The scenarios are the input for creating, designing and evolving architectural solutions, which have to be preserved in follow-up activities. Thereby scenarios evolve over distinct states: Unknown, Elicited, Designed, Documented, Implemented, and Sustained. Depending on the state of the scenario, different types of architecture evaluations are possible. Architecture scenarios should be documented in a structured way (cf. [Clements et al., 2010]), rendering data on the following aspects:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Scenario: Representative name (and ID) of the scenario;</para></listitem>
<listitem>
<para>Quality: Related quality attribute;</para></listitem>
<listitem>
<para>Environment: Context applying to this scenario (if possible provide quantifications);</para></listitem>
<listitem>
<para>Stimulus: The event or condition arising from this scenario (if possible provide quantifications);</para></listitem>
<listitem>
<para>Response: The expected reaction of the system to the scenario event;</para></listitem>
<listitem>
<para>Response Measure: The measurable effects showing if the scenario is fulfilled by the architecture.</para></listitem>
</itemizedlist>
</section>
<section class="lev2" id="sec3-6-3">
<title>3.6.3 Architecture Evaluation Types</title>
<para>RATE (<link linkend="F3-18">Figure <xref linkend="F3-18" remap="3.18"/></link>) comprises five checks on three evaluation levels, whereby each check focuses on different aspects. The main levels are stakeholder level, architecture level and implementation level. In the stakeholder level, architecture drivers are validated, in the architecture level, the architecture and its documentation are checked and in the implementation level, the compliance and code quality is checked.</para>
<fig id="F3-18" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F3-18">Figure <xref linkend="F3-18" remap="3.18"/></link></label>
<caption><para>Architecture evaluation with Fraunhofer RATE [Knodel and Naab, 2016].</para></caption>
<graphic xlink:href="graphics/ch03_Fig3_18.jpg"/>
</fig>
<para>All checks performed come with a related rating of the confidence level of the findings.</para>
<para>The rating comprises</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>the severity of findings that expresses the criticality of the findings aggregated over all findings per goal;</para></listitem>
<listitem>
<para>the balance of findings that expresses the ratio of positive vs. negative findings aggregated per goal.</para></listitem>
</itemizedlist>
<para>The combination of the ratings results in a rating of N/A (Not Applicable), or an assignment to one of the target achievement levels NO, PARTIAL, LARGE, and FULL where each check defines its own target achievements.</para>
<para><emphasis>Driver Integrity Check (DIC)</emphasis>: The goal of the Driver Integrity Check is to get confidence that an architecture is built based on a set of architecture drivers that is agreed among stakeholders and to clarify unclear or not agreed architecture drivers. DIC also aims to compensate not elicited requirements and to aggregate a large set of requirements into a manageable set for an architecture evaluation. Inputs of the DIC are stakeholder information (if available), existing documentation (if available), and a template for documenting architecture drivers (mandatory). It is crucial to elicit required qualities using scenarios in a measurable way, so that architects or evaluators can find a baseline to work with. One can use architecture scenarios to document both, functional and quality drivers in a measurable way.</para>
<para><emphasis>Solution Adequacy Check (SAC)</emphasis>: The Solution Adequacy Check can be done as soon as there is a first idea of the architecture, and tries to answer the question how well suited the architecture for the intended purpose is. The SAC</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>can determine whether an architecture permits or precludes the achievement of targeted functional and quality requirements;</para></listitem>
<listitem>
<para>can determine whether an architecture or part of it is concrete enough or redundant for the purpose it will be used for;</para></listitem>
<listitem>
<para>enables the identification of problematic design decisions;</para></listitem>
<listitem>
<para>enables the early prediction;</para></listitem>
<listitem>
<para>enables a timely reaction and correction.</para></listitem>
</itemizedlist>
<para>That is, the main purpose of the SAC is to gain confidence, to predict the future behaviour of the system or to get some evidences. Inputs for the SAC are architecture drivers and architecture documentation. Besides findings on the adequacy of architecture decisions to fulfil architecture drivers (explicit rationales, risks, trade-offs, assumptions) SAC puts out revised architecture decisions, driver solutions, and diagrams.</para>
<para><emphasis>Documentation Quality Check (DQC)</emphasis>: The Documentation Quality Check can be done as soon as there is a first draft of the architecture description, and tries to answer the question how well documented the architecture solution for its audience and purposes is. The DQC</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>can determine whether or not an architecture documentation allows understanding the solution;</para></listitem>
<listitem>
<para>enables information sharing on architecture;</para></listitem>
<listitem>
<para>enables consistency checks;</para></listitem>
<listitem>
<para>enables the detection of gaps in architecture and its documentation.</para></listitem>
</itemizedlist>
<para>The main purpose of the DQC is to ease information sharing within architectural stakeholders. Inputs for the DQC are the documentation purposes, architecture documents, models, wikis, sketches, API documentation, and the targeted audience. It feeds back findings on adequacy of the documentation and its adherence to best practices.</para>
<para><emphasis>Architecture Compliance Check (ACC)</emphasis>: The Architecture Compliance Check can be done as soon as there is a first skeleton of the implementation, however, it should be iterated, as there may be modified implementations during the lifecycle of the system. The ACC tries to answer the question how well realized the architectural solution by the implementation is. The ACC</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>can determine whether or not the implementation violates architectural solutions;</para></listitem>
<listitem>
<para>enables traceability;</para></listitem>
<listitem>
<para>enables compliance checking (as-is vs. planned intention);</para></listitem>
<listitem>
<para>enables delta tracking (as-is vs. envisioned target).</para></listitem>
</itemizedlist>
<para>The main purposes for ACC are to identify the structural and behavioural compliances between the architecture and the implemented system. Inputs for the ACC are architecture documents, models, wikis, sketches, API documentation, source code, and the running system (if applicable). It results in findings on the compliance of the implementation with respect to the intended architecture, convergences, divergences (violations) and absences (violations).</para>
<para><emphasis>Code Quality Check (CQC)</emphasis>: The Code Quality Check&#x02019;s main goal is to gather data about the source code base. As such, the CQC is not a direct part of architecture evaluation. However, reasoning about quality attributes (in particular maintainability) requires the CQC results in order to make valid statements about the software system under evaluation. CQC helps to</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>improve the implementation of a software system;</para></listitem>
<listitem>
<para>monitor (and fix) anomalies over time;</para></listitem>
<listitem>
<para>derive common metrics and coding best practices;</para></listitem>
<listitem>
<para>define team-specific coding guidelines;</para></listitem>
<listitem>
<para>improve the overall understanding of the code base;</para></listitem>
<listitem>
<para>make the development organization more robust.</para></listitem>
</itemizedlist>
<para>Inputs for the CQC are the source code, and build scripts (if applicable). It puts out findings on quality of the source code, best practice violations, code clones, quality warnings (maintainability, security), code metrics, and more.</para>
</section>
</section>
<section class="lev1" id="sec3-7">
<title>3.7 Conclusions</title>
<para>Proactive maintenance for the CPS domain requires solutions that cover data gathering, storage, processing, feedback and presentation to the human operator. While examples of custom-tailored systems are appearing, this chapter presents a generic platform together with specific toolset to cover the problem space. Typical target areas of proactive maintenance include FP, calculation of the RUL, RCA, among others.</para>
<para>Beside providing an architectural view on data gathering and handling for the given area, the MANTIS architecture for CPS-based proactive maintenance provides solutions for the Egde Tier, for the Platform Tier, and for the Enterprise Tier, as well. The main building blocks of the Edge Tier are physical sensors and actuators, as well as local, edge-level data processing entities. These also communicate with the elements of a system-wide view at the Platform Tier. Depending on the targeted area (i.e., FP, RUL, RCA), stream processing or batch processing entities handle the data and provide meaningful output that either gets feed back to the physical entities as control information, or gets presented towards the Enterprise Tier for further processing or action (e.g., ordering spare equipment, scheduling jobs, visualizing trends, etc.). There are various issues to tackle on each of the tiers, and even within the communications between the various actors and tiers. Therefore, the interoperability and security aspects must be focal points when such an installation is made, and hence their place in this Chapter.</para>
<para>Moreover, such an architecture has to be evaluated and has to facilitate the requirements of every actor and stakeholder. To this end, evaluation techniques are instruments to increase the confidence level in the architecture solutions, where the confidence level expresses trust in the architecture designed and the system derived from it. In addition, architecture evaluation allows the prediction of the impact of decisions made and helps secure the consistency across decisions in descriptions. It helps to check the structural and behavioral compliance of the implemented as-is architecture with the intended one, and makes possible the assertion of qualities in the system at execution. Therefore, architecture evaluation is the key quality-engineering instrument in software and systems design, and it provides confidence for decision-making at any time in the life cycle of a software system.</para>
<para>This architecture already has been successfully utilized in various use-cases from industrial utility vehicles (forklifts) through railway control system to CPPSs [Ferreira et al., 2017; Heged&#x000FC;s et al., 2018], as part of the ECSEL MANTIS project [The MANTIS Consortium, 2018]. These use case instantiations are presented in <link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link>.</para>
</section>
<section class="lev1" id="sec3-8">
<title>References</title>
<orderedlist numeration="arabic" continuation="restarts" spacing="normal">
<listitem>
<para>Albano, M., Barbosa, P. M., Silva, J., Duarte, R., Ferreira, L. L., and Delsing, J. (2017). Quality of service on the Arrowhead Framework. In <emphasis>2017 IEEE 13th International Workshop on Factory Communication Systems (WFCS),</emphasis> (pp. 1&#x02013;8). IEEE. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Albano%2C+M%2E%2C+Barbosa%2C+P%2E+M%2E%2C+Silva%2C+J%2E%2C+Duarte%2C+R%2E%2C+Ferreira%2C+L%2E+L%2E%2C+and+Delsing%2C+J%2E+%282017%29%2E+Quality+of+service+on+the+Arrowhead+Framework%2E+In+2017+IEEE+13th+International+Workshop+on+Factory+Communication+Systems+%28WFCS%29%2C+%28pp%2E+1-8%29%2E+IEEE%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Amazon. (2017) Amazon web services. https://aws.amazon.com/products/ analytics/.</para></listitem>
<listitem>
<para>Apache. (2017) Cassandra database. http://cassandra.apache.org/.</para></listitem>
<listitem>
<para>Apache. (2017) Hadoop distributed file system. http://hadoop.apache. org/.</para></listitem>
<listitem>
<para>Apache. (2017) Spark. https://spark.apache.org/.</para></listitem>
<listitem>
<para>Apache Community&#x02019;. (2017) The kafka distributed streaming platform. http://kafka.apache.org/.</para></listitem>
<listitem>
<para>Apache Community. (2017) Storm stream processor. http://storm.apache. org/.</para></listitem>
<listitem>
<para>Assun&#x000E7;&#x00E3;o, M. D., Calheiros, R. N., Bianchi, S., Netto, M. A. S., and Buyya, R. (2015) &#x02018;Big data computing and clouds: Trends and future directions&#x02019;, Journal of Parallel and Distributed Computing, 79, pp. 3&#x02013;15. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Assun%E7o%2C+M%2E+D%2E%2C+Calheiros%2C+R%2E+N%2E%2C+Bianchi%2C+S%2E%2C+Netto%2C+M%2E+A%2E+S%2E%2C+and+Buyya%2C+R%2E+%282015%29+%27Big+data+computing+and+clouds%3A+Trends+and+future+directions%27%2C+Journal+of+Parallel+and+Distributed+Computing%2C+79%2C+pp%2E+3-15%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Bell, M. (2008) Introduction to Service-Oriented Modeling. Wiley and Sons.</para></listitem>
<listitem>
<para>Bermbach, D., Wittern, E., and Tai, S. (2017) Cloud Service Benchmarking: Measuring Quality of Cloud Services from a Client Perspective. Springer.</para></listitem>
<listitem>
<para>Cengarle, M., Bensalen, S., McDermid, J., Passerone, R., Sangiovanni-Vincetelli, A., and Torngren, M. (2013) <emphasis>Characteristics, Capabilities, Potential Applications of Cyber-physical Systems: A Preliminary Analysis</emphasis>. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Cengarle%2C+M%2E%2C+Bensalen%2C+S%2E%2C+McDermid%2C+J%2E%2C+Passerone%2C+R%2E%2C+Sangiovanni-Vincetelli%2C+A%2E%2C+and+Torngren%2C+M%2E+%282013%29+Characteristics%2C+Capabilities%2C+Potential+Applications+of+Cyber-physical+Systems%3A+A+Preliminary+Analysis%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Cisco. (2016) Global cloud index: Forecast and methodology. https://www. cisco.com/c/en/us/solutions/collateral/service-provider/ global-cloud-index-gci/white-paper-c11-738085.html.</para></listitem>
<listitem>
<para>Clements, P.,<etal/> (2010) <emphasis>Documenting Software Architectures: Views and Beyond</emphasis>. p. 608. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Clements%2C+P%2E%2C+%282010%29+Documenting+Software+Architectures%3A+Views+and+Beyond%2E+p%2E+608%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Curry, E. (2004) &#x02018;Message-oriented middleware,&#x02019; Middleware for Communications, pp. 1&#x02013;26. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Curry%2C+E%2E+%282004%29+%27Message-oriented+middleware%2C%27+Middleware+for+Communications%2C+pp%2E+1-26%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Delsing, J. (2017) IoT Automation: Arrowhead Framework, CRC Press.</para></listitem>
<listitem>
<para>Di Orio, G., Mal&#x000F3;, P., Barata, J., Albano, M., and Ferreira, L. L. (2018, July). Towards a Framework for Interoperable and Interconnected CPS-populated Systems for Proactive Maintenance. <emphasis>In 2018 IEEE 16th International Conference on Industrial Informatics (INDIN)</emphasis>, (pp. 146&#x02013;151). IEEE. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Di+Orio%2C+G%2E%2C+Mal%F3%2C+P%2E%2C+Barata%2C+J%2E%2C+Albano%2C+M%2E%2C+and+Ferreira%2C+L%2E+L%2E+%282018%2C+July%29%2E+Towards+a+Framework+for+Interoperable+and+Interconnected+CPS-populated+Systems+for+Proactive+Maintenance%2E+In+2018+IEEE+16th+International+Conference+on+Industrial+Informatics+%28INDIN%29%2C+%28pp%2E+146-151%29%2E+IEEE%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Donnelly, C. (2015) Eu data protection regulation: What the ec legislation means for cloud providers. http://www.computerweekly.com/.</para></listitem>
<listitem>
<para>Fan, W. and Bifet, A. (2013) &#x02018;Mining big data: Current status, and forecast to the future,&#x02019; ACM sIGKDD Explorations Newsletter, 14(2), pp. 1&#x02013;5. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Fan%2C+W%2E+and+Bifet%2C+A%2E+%282013%29+%27Mining+big+data%3A+Current+status%2C+and+forecast+to+the+future%2C%27+ACM+sIGKDD+Explorations+Newsletter%2C+14%282%29%2C+pp%2E+1-5%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Ferraiolo, D. F. and Kuhn, D. R. (1992) &#x02018;Role-based access control,&#x02019; In <emphasis>15th National Computer Security Conference</emphasis>, Baltimore, October 13&#x02013;16, pp. 554&#x02013;563. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ferraiolo%2C+D%2E+F%2E+and+Kuhn%2C+D%2E+R%2E+%281992%29+%27Role-based+access+control%2C%27+In+15th+National+Computer+Security+Conference%2C+Baltimore%2C+October+13-16%2C+pp%2E+554-563%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Ferreira, L. L., Albano, M., Silva, J., Martinho, D., Marreiros, G., di Orio, G., Mal&#x000F3;, P., and Ferreira, H. (2017) A pilot for proactive maintenance in industry 4.0. In 13th IEEE International Workshop on Factory Communication Systems (WFCS 2017), pp. 1&#x02013;9. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ferreira%2C+L%2E+L%2E%2C+Albano%2C+M%2E%2C+Silva%2C+J%2E%2C+Martinho%2C+D%2E%2C+Marreiros%2C+G%2E%2C+di+Orio%2C+G%2E%2C+Mal%F3%2C+P%2E%2C+and+Ferreira%2C+H%2E+%282017%29+A+pilot+for+proactive+maintenance+in+industry+4%2E0%2E+In+13th+IEEE+International+Workshop+on+Factory+Communication+Systems+%28WFCS+2017%29%2C+pp%2E+1-9%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Ghemawat, S., Gobioff, H., and Leung, S.-T. (2015) <emphasis>Systems and Methods for Replicating Data</emphasis>. US Patent 9,047,307. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ghemawat%2C+S%2E%2C+Gobioff%2C+H%2E%2C+and+Leung%2C+S%2E-T%2E+%282015%29+Systems+and+Methods+for+Replicating+Data%2E+US+Patent+9%2C047%2C307%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Grover, P. and Kar, A. K. (2017) &#x02018;Big data analytics: A review on theoretical contributions and tools used in literature&#x02019;, Global Journal of Flexible Systems Management, pp. 1&#x02013;27. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Grover%2C+P%2E+and+Kar%2C+A%2E+K%2E+%282017%29+%27Big+data+analytics%3A+A+review+on+theoretical+contributions+and+tools+used+in+literature%27%2C+Global+Journal+of+Flexible+Systems+Management%2C+pp%2E+1-27%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Han, J., Haihong, E., Le, G., and Du, J. (2011) &#x02018;Survey on nosql database,&#x02019; In 2011 6th International Conference on Pervasive Computing and Applications (ICPCA), pp. 363&#x02013;366. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Han%2C+J%2E%2C+Haihong%2C+E%2E%2C+Le%2C+G%2E%2C+and+Du%2C+J%2E+%282011%29+%27Survey+on+nosql+database%2C%27+In+2011+6th+International+Conference+on+Pervasive+Computing+and+Applications+%28ICPCA%29%2C+pp%2E+363-366%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Hashem, I. A. T., Yaqoob, I., Anuar, N. B., Mokhtar, S., Gani, A., and Khan, S. U. (2015) &#x02018;The rise of big data on cloud computing: Review and open research issues,&#x02019; Information Systems, 47, pp. 98&#x02013;115. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Hashem%2C+I%2E+A%2E+T%2E%2C+Yaqoob%2C+I%2E%2C+Anuar%2C+N%2E+B%2E%2C+Mokhtar%2C+S%2E%2C+Gani%2C+A%2E%2C+and+Khan%2C+S%2E+U%2E+%282015%29+%27The+rise+of+big+data+on+cloud+computing%3A+Review+and+open+research+issues%2C%27+Information+Systems%2C+47%2C+pp%2E+98-115%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Hansche, S., Berti, J., and Hare, C. (2003) <emphasis>Official (ISC)2 Guide to the CISSP Exam</emphasis>. CRC Press. p. 104. ISBN 978-0-8493-1707-1. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Hansche%2C+S%2E%2C+Berti%2C+J%2E%2C+and+Hare%2C+C%2E+%282003%29+Official+%28ISC%292+Guide+to+the+CISSP+Exam%2E+CRC+Press%2E+p%2E+104%2E+ISBN+978-0-8493-1707-1%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Hausenblas, M. and Bijnens, N. (2017) &#x02018;The lambda architecture website,&#x02019; http://lambda-architecture.net/. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Hausenblas%2C+M%2E+and+Bijnens%2C+N%2E+%282017%29+%27The+lambda+architecture+website%2C%27+http%3A%2F%2Flambda-architecture%2Enet%2F%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Hecht, R. and Jablonski, S. (2011) &#x02018;Nosql evaluation: A use case oriented survey,&#x02019; In 2011 International Conference on Cloud and Service Computing (CSC), pp. 336&#x02013;341. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Hecht%2C+R%2E+and+Jablonski%2C+S%2E+%282011%29+%27Nosql+evaluation%3A+A+use+case+oriented+survey%2C%27+In+2011+International+Conference+on+Cloud+and+Service+Computing+%28CSC%29%2C+pp%2E+336-341%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Heged&#x000FC;s, C., Ciancarini, P., Franko, A., Kancilija, A., Moldovan, I., Papa, G., Poklukar, S., Riccardi, M., Sillitti, A., and Varga, P. (2018) &#x02018;Proactive maintenance of railway switches,&#x02019; In <emphasis>Proceedings of the 5th International Conference on Control, Decision and Information Technology (CoDIT)</emphasis>, Thessaloniki, Greece.</para></listitem>
<listitem>
<para>IBM. &#x02018;The enterprise service bus, re-examined: Updating concepts and terminology for an evolved technology,&#x02019; https://www.ibm.com/developerworks/websphere/ techjournal/1105_flurry/1105_flurry.html.</para></listitem>
<listitem>
<para>Industrial Internet Consortium. (2017) The industrial internet of things reference architecture.</para></listitem>
<listitem>
<para>International Electrotechnical Commission. (1993) <emphasis>Information Technology &#x02013; Vocabulary &#x02013; Part 1: Fundamental terms</emphasis>.</para></listitem>
<listitem>
<para>International Electrotechnical Commission. (2003&#x02013;2007). <emphasis>Enterprise-control System Integration</emphasis>. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=International+Electrotechnical+Commission%2E+%282003-2007%29%2E+Enterprise-control+System+Integration%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>IoT-A Reference Architecture Model. (2018) http://open-platforms.eu/ standard_protocol/iot-a-architectural-reference-model/.</para></listitem>
<listitem>
<para>ISO. (2012) Condition monitoring and diagnostics of machines &#x02013; data processing, communication and presentation.</para></listitem>
<listitem>
<para>ISO/IEC/IEEE. (2011) &#x02018;Systems and software engineering &#x02013; architecture description,&#x02019; ISO/IEC/IEEE 42010:2011(E) (Revision of ISO/IEC 42010:2007 and IEEE Std 1471-2000), pp. 1&#x02013;46. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=ISO%2FIEC%2FIEEE%2E+%282011%29+%27Systems+and+software+engineering+-+architecture+description%2C%27+ISO%2FIEC%2FIEEE+42010%3A2011%28E%29+%28Revision+of+ISO%2FIEC+42010%3A2007+and+IEEE+Std+1471-2000%29%2C+pp%2E+1-46%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Jacques, F. (1999) <emphasis>Multi-agent Systems: An Introduction to Distributed Artificial Intelligence</emphasis>. Addison-Wesley. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Jacques%2C+F%2E+%281999%29+Multi-agent+Systems%3A+An+Introduction+to+Distributed+Artificial+Intelligence%2E+Addison-Wesley%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Jantunen, E., Di Orio, G., Hegedus, C., Varga, P., Moldovan, I., Larrinaga, F., Becker, M., Albano, M., and Malo, P. (2018) Maintenance 4.0 world of integrated information. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Jantunen%2C+E%2E%2C+Di+Orio%2C+G%2E%2C+Hegedus%2C+C%2E%2C+Varga%2C+P%2E%2C+Moldovan%2C+I%2E%2C+Larrinaga%2C+F%2E%2C+Becker%2C+M%2E%2C+Albano%2C+M%2E%2C+and+Malo%2C+P%2E+%282018%29+Maintenance+4%2E0+world+of+integrated+information%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Jantunen, E., Zurutuza, U., Ferreira, L. L. and Varga, P. (2016) &#x02018;Optimising maintenance: What are the expectations for cyber physical systems,&#x02019; In 2016 3rd International Workshop on Emerging Ideas and Trends in Engineering of Cyber-Physical Systems (EITEC), pp. 53&#x02013;58. IEEE. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Jantunen%2C+E%2E%2C+Zurutuza%2C+U%2E%2C+Ferreira%2C+L%2E+L%2E+and+Varga%2C+P%2E+%282016%29+%27Optimising+maintenance%3A+What+are+the+expectations+for+cyber+physical+systems%2C%27+In+2016+3rd+International+Workshop+on+Emerging+Ideas+and+Trends+in+Engineering+of+Cyber-Physical+Systems+%28EITEC%29%2C+pp%2E+53-58%2E+IEEE%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Kappa. (2018) &#x02018;The kappa architecture site,&#x02019; https://www.talend.com/ blog/2017/08/28/lambda-kappa-real-time-big-data-architectures/. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Kappa%2E+%282018%29+%27The+kappa+architecture+site%2C%27+https%3A%2F%2Fwww%2Etalend%2Ecom%2F+blog%2F2017%2F08%2F28%2Flambda-kappa-real-time-big-data-architectures%2F%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Knodel, J. and Naab, M. (2016) <emphasis>Pragmatic Evaluation of Software Architectures</emphasis>, p. 132. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Knodel%2C+J%2E+and+Naab%2C+M%2E+%282016%29+Pragmatic+Evaluation+of+Software+Architectures%2C+p%2E+132%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Laney, D. (2001) &#x02018;3d data management: Controlling data volume, velocity and variety,&#x02019; META Group Research Note, 6, p. 70. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Laney%2C+D%2E+%282001%29+%273d+data+management%3A+Controlling+data+volume%2C+velocity+and+variety%2C%27+META+Group+Research+Note%2C+6%2C+p%2E+70%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Lebold, M. and Thurston, M. (2001) &#x02018;Open standards for condition-based maintenance and prognostic system,&#x02019; In Proceedings of MARCON 2001 &#x02013; Fifth annual maintenance and reliability conference, Gatlinburg, USA. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Lebold%2C+M%2E+and+Thurston%2C+M%2E+%282001%29+%27Open+standards+for+condition-based+maintenance+and+prognostic+system%2C%27+In+Proceedings+of+MARCON+2001+-+Fifth+annual+maintenance+and+reliability+conference%2C+Gatlinburg%2C+USA%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Lenk, A., Klems, M., Nimis, J., Tai, S., and Sandholm, T. (2009) &#x02018;What&#x02019;s inside the cloud? An architectural map of the cloud landscape,&#x02019; In Proceedings of the 2009 ICSE Workshop on Software Engineering Challenges of Cloud Computing, pp. 23&#x02013;31. IEEE Computer Society. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Lenk%2C+A%2E%2C+Klems%2C+M%2E%2C+Nimis%2C+J%2E%2C+Tai%2C+S%2E%2C+and+Sandholm%2C+T%2E+%282009%29+%27What%27s+inside+the+cloud%B4+An+architectural+map+of+the+cloud+landscape%2C%27+In+Proceedings+of+the+2009+ICSE+Workshop+on+Software+Engineering+Challenges+of+Cloud+Computing%2C+pp%2E+23-31%2E+IEEE+Computer+Society%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Martin, R. C. (2002) <emphasis>Agile Software Development: Principles, Patterns, and Practices</emphasis>. Prentice Hall. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Martin%2C+R%2E+C%2E+%282002%29+Agile+Software+Development%3A+Principles%2C+Patterns%2C+and+Practices%2E+Prentice+Hall%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Martinez-Fernandez, S., Dos Santos, P. M., Ayala, C., Franch, X., and Travassos, G. (2015) &#x02018;Aggregating empirical evidence about the benefits and drawbacks of software reference architectures,&#x02019; In ACM/IEEE International Symposium on Empirical Software Engineering and Measurement (ESEM), Beijing. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Martinez-Fernandez%2C+S%2E%2C+Dos+Santos%2C+P%2E+M%2E%2C+Ayala%2C+C%2E%2C+Franch%2C+X%2E%2C+and+Travassos%2C+G%2E+%282015%29+%27Aggregating+empirical+evidence+about+the+benefits+and+drawbacks+of+software+reference+architectures%2C%27+In+ACM%2FIEEE+International+Symposium+on+Empirical+Software+Engineering+and+Measurement+%28ESEM%29%2C+Beijing%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Microsoft. (2017) &#x02018;Azure web services,&#x02019; https://azure.microsoft.com.</para></listitem>
<listitem>
<para>MIMOSA consortium. (2016) &#x02018;The mimosa project site,&#x02019; http://www.mimosa. org/.</para></listitem>
<listitem>
<para>Mineraud, J., Mazhelis, O., Su, X., and Tarkoma, S. (2016) &#x02018;A gap analysis of internet-of-things platforms,&#x02019; Computer Communications, 89, pp. 5&#x02013;16. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Mineraud%2C+J%2E%2C+Mazhelis%2C+O%2E%2C+Su%2C+X%2E%2C+and+Tarkoma%2C+S%2E+%282016%29+%27A+gap+analysis+of+internet-of-things+platforms%2C%27+Computer+Communications%2C+89%2C+pp%2E+5-16%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Munshi, A. A. and Yasser, A.-R. I. M. (2017) &#x02018;Big data framework for analytics in smart grids,&#x02019; Electric Power Systems Research, 151, pp. 369&#x02013;380. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Munshi%2C+A%2E+A%2E+and+Yasser%2C+A%2E-R%2E+I%2E+M%2E+%282017%29+%27Big+data+framework+for+analytics+in+smart+grids%2C%27+Electric+Power+Systems+Research%2C+151%2C+pp%2E+369-380%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Noiumkar, P. and Chomsiri, T. (2014) &#x02018;A comparison the level of security on top 5 open source nosql databases,&#x02019; In The 9th International Conference on Information Technology and Applications (ICITA2014). <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Noiumkar%2C+P%2E+and+Chomsiri%2C+T%2E+%282014%29+%27A+comparison+the+level+of+security+on+top+5+open+source+nosql+databases%2C%27+In+The+9th+International+Conference+on+Information+Technology+and+Applications+%28ICITA2014%29%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Nokia. (2017) &#x02018;Disco distributed file system,&#x02019; https://disco.readthedocs. io.</para></listitem>
<listitem>
<para>OASIS. (2018) &#x02018;eXtensible Access Control Markup Language (XACML),&#x02019; https://www.oasis-open.org/committees/tc_home. php?wg_abbrev=xacml#CURRENT. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=OASIS%2E+%282018%29+%27eXtensible+Access+Control+Markup+Language+%28XACML%29%2C%27+https%3A%2F%2Fwww%2Eoasis-open%2Eorg%2Fcommittees%2Ftc%5Fhome%2E+php%B4wg%5Fabbrev%3Dxacml%23CURRENT%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Pohl, K., H&#x000F6;nninger, H., Achatz, R., and Broy, M. (2012) Model-Based Engineering of Embedded Systems: The SPES 2020 Methodology. Springer, 2012. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Pohl%2C+K%2E%2C+H%F6nninger%2C+H%2E%2C+Achatz%2C+R%2E%2C+and+Broy%2C+M%2E+%282012%29+Model-Based+Engineering+of+Embedded+Systems%3A+The+SPES+2020+Methodology%2E+Springer%2C+2012%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Rahalkar, S. A. (2016) <emphasis>Information Security Basics. In Certified Ethical Hacker (CEH) Foundation Guide</emphasis>, Apress, Berkeley, CA, pp. 85&#x02013;95. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Rahalkar%2C+S%2E+A%2E+%282016%29+Information+Security+Basics%2E+In+Certified+Ethical+Hacker+%28CEH%29+Foundation+Guide%2C+Apress%2C+Berkeley%2C+CA%2C+pp%2E+85-95%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Rozanski, N. and Woods, E. (2011) <emphasis>Software Systems Architecture: Working with Stakeholders Using Viewpoints and Perspectives</emphasis>, p. 678. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Rozanski%2C+N%2E+and+Woods%2C+E%2E+%282011%29+Software+Systems+Architecture%3A+Working+with+Stakeholders+Using+Viewpoints+and+Perspectives%2C+p%2E+678%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Sahafizadeh, E. and Nematbakhsh, M. A. (2015) &#x02018;A survey on security issues in big data and nosql,&#x02019; Advances in Computer Science: an International Journal, 4(4), pp. 68&#x02013;72. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Sahafizadeh%2C+E%2E+and+Nematbakhsh%2C+M%2E+A%2E+%282015%29+%27A+survey+on+security+issues+in+big+data+and+nosql%2C%27+Advances+in+Computer+Science%3A+an+International+Journal%2C+4%284%29%2C+pp%2E+68-72%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Singh, R., Singh, K.,<etal/> (2010) &#x02018;A descriptive classification of causes of data quality problems in data warehousing,&#x02019; International Journal of Computer Science Issues, 7(3), pp. 41&#x02013;50. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Singh%2C+R%2E%2C+Singh%2C+K%2E%2C+%282010%29+%27A+descriptive+classification+of+causes+of+data+quality+problems+in+data+warehousing%2C%27+International+Journal+of+Computer+Science+Issues%2C+7%283%29%2C+pp%2E+41-50%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>The MANTIS Consortium. (2018) &#x02018;The mantis project website,&#x02019; http://www. mantis-project.eu/.</para></listitem>
<listitem>
<para>Thomas, E. R. L. (2008) <emphasis>SOA: Principles of Service Design.</emphasis> Upper Saddle River: Prentice Hall. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Thomas%2C+E%2E+R%2E+L%2E+%282008%29+SOA%3A+Principles+of+Service+Design%2E+Upper+Saddle+River%3A+Prentice+Hall%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Vacca, J. R. (2004) <emphasis>Public Key Infrastructure: Building Trusted Applications and Web Services</emphasis>, CRC Press. ISBN 978-0-8493-0822-2. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Vacca%2C+J%2E+R%2E+%282004%29+Public+Key+Infrastructure%3A+Building+Trusted+Applications+and+Web+Services%2C+CRC+Press%2E+ISBN+978-0-8493-0822-2%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Wapice. (2018) IoT-Ticket. https://iot-ticket.com/.</para></listitem>
</orderedlist>
</section>
</chapter>
<chapter class="chapter" id="ch04" label="4" xreflabel="4">
<title>Monitoring of Critical Assets</title>
<para><emphasis role="strong">Michele Albano<superscript><emphasis role="strong">1</emphasis></superscript>, Jos&#x000E9; Manuel Abete<superscript><emphasis role="strong">2</emphasis></superscript>, Iban Barrutia Inza<superscript><emphasis role="strong">2</emphasis></superscript>, Vito &#x0010C;u&#x0010D;ek<superscript><emphasis role="strong">3</emphasis></superscript>, Karel De Brabandere<superscript><emphasis role="strong">4</emphasis></superscript>, Ander Etxabe<superscript><emphasis role="strong">5</emphasis></superscript>, Iosu Gabilondo<superscript><emphasis role="strong">5</emphasis></superscript>, &#x000C7;i&#x000E7;ek G&#x000FC;ven<superscript><emphasis role="strong">6</emphasis></superscript>, Mike Holenderski<superscript><emphasis role="strong">6</emphasis></superscript>, Aitzol Iturrospe<superscript><emphasis role="strong">2</emphasis></superscript>, Erkki Jantunen<superscript><emphasis role="strong">7</emphasis></superscript>, Luis Lino Ferreira<superscript><emphasis role="strong">1</emphasis></superscript>, Istv&#x000E1;n Moldov&#x000E1;n<superscript><emphasis role="strong">8</emphasis></superscript>, Jon Olaizola<superscript><emphasis role="strong">2</emphasis></superscript>, Eneko S&#x000E1;enz de Argando&#x000F1;a<superscript><emphasis role="strong">2</emphasis></superscript>, Babacar Sarr<superscript><emphasis role="strong">4</emphasis></superscript>, S&#x000F6;ren Schneickert<superscript><emphasis role="strong">9</emphasis></superscript>, Rafael Socorro<superscript><emphasis role="strong">10</emphasis></superscript>, Hans Sprong<superscript><emphasis role="strong">11</emphasis></superscript>, Marjan &#x00160;terk<superscript><emphasis role="strong">3</emphasis></superscript>, Ra&#x000FA;l Torrego<superscript><emphasis role="strong">5</emphasis></superscript>, Godfried Webers<superscript><emphasis role="strong">11</emphasis></superscript>, and Achim Woyte<superscript><emphasis role="strong">4</emphasis></superscript></emphasis></para>
<para><superscript>1</superscript>ISEP, Polytechnic Institute of Porto, Porto, Portugal</para>
<para><superscript>2</superscript>Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para><superscript>3</superscript>XLAB, Ljubljana, Slovenia</para>
<para><superscript>4</superscript>3e, Belgium</para>
<para><superscript>5</superscript>IK4-Ikerlan, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para><superscript>6</superscript>Technische Universiteit Eindhoven, The Netherlands</para>
<para><superscript>7</superscript>VTT Technical Research Centre of Finland Ltd, Finland</para>
<para><superscript>8</superscript>Budapest University of Technology and Economics, Hungary</para>
<para><superscript>9</superscript>Fraunhofer IESE, Germany</para>
<para><superscript>10</superscript>Acciona Construcci&#x000F3;n S.A., Spain</para>
<para><superscript>11</superscript>Philips Medical Systems Nederland B.V., The Netherlands</para>
<para>Sensors used in the MANTIS project must be able to measure the physical phenomenon relevant for the assets&#x02019; condition. Examples of this include temperature, light intensity, pressure, fluid flow, velocity, and force among others. Anyhow, it is not a trivial problem for a given installation, where an adequate measurement solution must be chosen or developed in order to accurately and robustly acquire data about the physical process related to each MANTIS use cases [Jantunen et al., 2017].</para>
<para>Another relevant matter is the cost of the monitoring solution. Industry is always aiming for cost savings and a better market positioning. Therefore, new technological solutions such as WSNs have become a strategic asset in this context, increasing the interest of the industrial companies. This type of sensor networks is used to share information with the purpose of increasing productivity, gathering data for developing future technological improvements and/or detecting/predicting maintenance issues. Moreover, even when a single sensor is considered instead of a WSN, the use of wireless communications provides flexibility, installation ease, weight reduction, which makes them suitable for many applications, conditions and situations.</para>
<para>Industrial environments usually have hostile site conditions, both for the sensors themselves and for the wireless communication systems, and a section is devoted to the analysis of the issues and the solutions raising in these environments.</para>
<para>Finally, a section is focused on the intelligent functions that can be offered by CPS, both to preprocess collected data and to support the CPS itself.</para>
<section class="lev1" id="sec4-1">
<title>4.1 The Industrial Environment</title>
<para>There are situations in which the integration of sensing devices and their associated processing systems within the framework of a hostile environment pose challenges that have to be overcome. Viable solutions must take into account different factors related to the factory and to the environment where the monitoring process takes place. Examples of such situations are described below.</para>
<section class="lev2" id="sec4-1-1">
<title>4.1.1 Extreme High/Low Temperatures (Ovens, Turbines, Refrigeration Chambers etc.)</title>
<para>Silicon based conventional electronic devices fail to perform within specifications when subject to external temperatures over 150<superscript>&#x02218;</superscript>C, as conventional CMOS substrates are not designed to withstand such high temperatures, and the same applies to batteries. Regarding plastic materials like cable covers made of polyethylene, they can suffer severe damage, exposing conductive parts, which increases the risk of short circuits or malfunctioning in general.</para>
<para>Opposite to this, temperatures below zero can drain batteries faster, and short circuits can also appear due to ice and frost formations between conductive parts of the circuits. Any amplifying device used in the sensor can suffer from increased gain due to cold temperatures. In addition to this, any cabling deployed between sensing devices and data processing units can break the thermal insulation of the premises. Finally, a wide temperature span can induce thermic stress in the sensing components, and substrates withstanding extremely high and low temperatures (generally of a ceramic type) must be considered.</para>
</section>
<section class="lev2" id="sec4-1-2">
<title>4.1.2 High Pressure Environments (Pneumatic/Hydraulic Systems, Oil Conductions, Tires etc.)</title>
<para>Sensors or devices inside structures subject to high hydraulic and pneumatic pressure can suffer structural damage, as they are subject to mechanical stress. Furthermore, not only the sensing device is subject to mechanical stress, but the containment vessel itself is subject to pressure, and thus can have its structural integrity compromised, for example if cables are allowed to get outside the housing. Fluids or air used in the industrial process could also escape from sealing O-rings. In the case of tire pressure sensors, for example, small volume, lightweight devices are necessary, as they are introduced on a rubber frame that could be torn if a sharp metallic piece impacted into the tire fabric.</para>
</section>
<section class="lev2" id="sec4-1-3">
<title>4.1.3 Nuclear Radiation (Reactors or Close and Near-By Areas)</title>
<para>Radiation is a direct threat not only for human individuals, but also for electronic systems (e.g., memory contents can be changed when subject to radiation). In these case, only rad-hard devices that are manufactured to withstand the impact of these high energy particles can be used for the development of intelligent sensors, or sensors with electronic circuitry. Additional elements like cabling coming out of the radiation zone can imply a direct biological hazard to humans, as containment sealing is not guaranteed and the components involved become radioactive wastes. All of these increase the development, deployment and operative costs. Additionally, the replacement of complex and bulky equipment and cabling can become a difficult, costly and time consuming task. If the adopted approach looks for a wireless solution, it is necessary to bear in mind that the containment walls of such premises are usually made of thick reinforced concrete, which means high signal loss, requiring additional amplification for thru-wall communications.</para>
</section>
<section class="lev2" id="sec4-1-4">
<title>4.1.4 Abrasive or Poisonous Environments</title>
<para>This situation can be similar to the extreme temperature case. Corrosive liquids and gases can degrade plastic covers and generate short circuits in the sensor circuit. The metallic parts and electronic components themselves are also subject to corrosion, compromising sensor performance. Like in the previous situations, an appropriate cabling scheme will usually increase costs.</para>
</section>
<section class="lev2" id="sec4-1-5">
<title>4.1.5 Presence of Explosive Substances or Gases</title>
<para>It is obvious that in the case of potentially explosive environments, it is of utmost importance the control upon the electric signals that are present on the premise where the sensing process is to be carried out. Electric fields due to excessively high signaling voltages in cables in a flammable gas environment, or current and voltage surges and spikes (due to thunder strikes) travelling down the cables can ignite the explosive substances, if these locations are physically hardwired to the exterior.</para>
</section>
<section class="lev2" id="sec4-1-6">
<title>4.1.6 Rotating or Moving Parts</title>
<para>The inherent dynamical nature of certain systems (rotors, assembly chains, clutches) persuades the system out of using cables, for obvious reasons, unless additional elements like electric brushes are added. In addition to this, in situations where high rpm regimes are attained (e.g., turbines or motors), the inertial mass, shape and even the orientation of the inserted sensor structure have to be considered carefully beforehand and in direct relation to the particular application for which they are to be used. The reason for this is not only to obtain a reliable and accurate measurement from the sensor, but also not to compromise the structural integrity and performance of the component or part that is being measured (e.g., sensors or parts of sensors impacting on the frame structure, damaging it).</para>
</section>
</section>
<section class="lev1" id="sec4-2">
<title>4.2 Industrial Sensor Characteristics</title>
<para>Sensors used in advanced maintenance operations can be classified into three main categories, based on how they are acquired. The first category is the <emphasis>off-the-shelf sensors</emphasis>, which are the most commonly employed in the industry, are mass-produced, are cheaper when bought, and target the most common environmental data, which are temperature, acceleration, light, force, audio, humidity and proximity according to [Beigl et al., 2004].</para>
<para>Specialized kinds of sensors make up the category of the <emphasis>custom sensors</emphasis>, can be found in specific applications, and are usually not mass produced, their structure presents a high degree of customization, and they retrieve very specific environmental data. Among the plethora of the custom sensors, there are sensors capable of performing crack detection, torque measurement, analyse wear of material and retrieve oil status. All of these environmental data are more complex to detect and can be collected using different approaches. For example, crack detection through non-destructive methods can be performed using different techniques like radiography, ultrasonic, penetrating liquid, and magnetic particle inspection.</para>
<para>The last category is the <emphasis>soft sensors</emphasis> (also called virtual sensors), which is a technology used to distil more effective and accurate information out of collected data. Soft sensors make use of readings collected either by a sensor network, or from a single sensor. The soft sensors operate based on data analyses and advanced modelling techniques in order to provide new data related to the physical processes. Data are combined from multiple sources (e.g., temperature, humidity, CO<subscript>2</subscript>) and process models are applied to compute new outputs, based on not only current sensor values, but also on its time series.</para>
<para>Many work have already surveyed off-the-shelf sensors, such as [Beigl et al., 2004], and they will not be further described in here. The rest of this section considers selected topics regarding custom sensors and soft sensors. In particular, PWS is considered, followed by MEMS sensors, and soft sensor computational trust.</para>
<section class="lev2" id="sec4-2-1">
<title>4.2.1 Passive Wireless Sensors</title>
<para>Often, a hostile environment comprises several of the aforementioned characteristics. Plus, it is usually difficult to have regular access to the physical location of the sensor, for example for battery replacement or system reset after latching. For each and every sensing case in a hostile environment, sensor performance must be guaranteed under all circumstances. It is advisable thus to make use of a physical working principle, implemented in a simplified architecture, built from durable and robust components to reduce the risk of errors or malfunctions. In addition to this, a reduced consumption or, equivalently, the capability to operate in a passive way is desirable also to minimize maintenance service. This not only reduces costs, but also the probability of errors committed by human operators. On this regard, PWS devices come out as an appropriate and affordable solution that can overcome many of the difficulties that the hostile environments pose, due to their simple working principle and robustness, which is highlighted by their definition itself:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Passive:</emphasis> No need for batteries or power supplies. Energization of the sensing structure relies on the impinging EM field power. Several solutions for this energy harvesting principle are available, like inductively or capacitively coupled antennas or rectennas for DC voltage generation;</para></listitem>
<listitem>
<para><emphasis role="strong">Wireless:</emphasis> No cables are needed to convey information from one communication device to the other. All exchanges are carried out via electromagnetic waves. Only antennas are needed for this purpose, decreasing in size as the involved frequencies are located at higher bands.</para></listitem>
</itemizedlist>
<para>Surface acoustic wave (SAW) devices or bulk acoustic wave (BAW) devices are feasible options for the implementation of PWS in case there is a need for measuring a temperature in a given location, due to their inherent robust properties and simple operation. The working principles of a SAW device is first reviewed, as BAW&#x02019;s operation is similar to that of SAW&#x02019;s, and then 3 examples of PWS temperature measurement systems are described.</para>
<para><emphasis>SAW devices for temperature measurement</emphasis></para>
<para>A SAW device, represented in <link linkend="F4-1">Figure <xref linkend="F4-1" remap="4.1"/></link>, is a small element that consists of a piezoelectric crystal substrate with a metallic pattern printed on its surface in a hairpin-like form, with an input port and an output port. An electric signal applied to this substrate generates a mechanical wave propagating in the surface of the structure (BAW devices operate in a similar fashion, but generating a wave propagating in the bulk of the substrate, rather than in the surface). In a reciprocal way, a mechanical wave moving on the surface can generate an electrical signal that can be captured across the terminals of a metallic port that has been deposited on the piezoelectric substrate. The equivalent circuit of such a component can be approximated by a combination of resistance, inductance and capacitance, thus showing a resonant behavior. As the resonance frequency value depends on the physical properties of the piezoelectric crystal (with these properties varying with temperature) this very resonance frequency of a SAW device can be used as an indicator of the temperature of the environment on which the SAW device is embedded.</para>
<fig id="F4-1" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-1">Figure <xref linkend="F4-1" remap="4.1"/></link></label>
<caption><para>SAW device&#x02019;s RLC equivalent circuit.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_1.jpg"/>
</fig>
<para>It is possible to add antennas to the described device, make an EM wave impinge upon it and then analyze the frequency of the reradiated wave. If the frequency of the &#x0201C;interrogating&#x0201D; wave lies close to the resonance of the SAW element, energy will be stored in the device in the form of an acoustic wave, and for a given frequency bandwidth centered at the resonant frequency. The majority of this stored energy will be contained in this resonant frequency, and when the excitation signal ceases, the device will progressively radiate back the stored energy (as this will convert from a mechanical wave to an electric signal at the antenna terminals, for reasons explained above). An analysis of this received &#x0201C;response&#x0201D; will provide an estimation of the temperature of the SAW device, as the maximum amplitude of the received signal&#x02019;s spectrum will give the value of the resonance frequency. On this regard, the implementation of a group of intelligent functions (mainly signal processing algorithms like filtering and transform calculations, to be explained in this chapter) is necessary in the edge server of the sensing system.</para>
<fig id="F4-2" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-2">Figure <xref linkend="F4-2" remap="4.2"/></link></label>
<caption><para>Signal processing scheme.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_2.jpg"/>
</fig>
<para>As can be deduced from the explanation above, also represented in <link linkend="F4-2">Figure <xref linkend="F4-2" remap="4.2"/></link>, such a scheme allows for the measurement of temperature in a wireless fashion, allowing for a simple remote temperature sensing method. Nonetheless, in the case of measurements to be carried out in high temperature environments (such as turbines and ovens) the sensing SAW devices that provide the readings need to be designed to withstand such hostile conditions, so that these extreme temperatures do not compromise the device&#x02019;s performance and, hence, the temperature estimation. Different manufacturers providing high-temp SAW devices have been identified [Wisen-tech; Syntonics; SAW].</para>
<para><emphasis>PWS temperature measurement examples</emphasis></para>
<para>Environetix has built a dedicated PWS sensor network for hostile environments, and has applied it to helicopter motor monitoring. They have also developed high temperature PWS devices for measurements in rotating parts. These sensors have been tested in JetCat turbine engine. A second example can be found at Wireless Sensor Technologies LLC. This company has developed a temperature measuring system for turbine blades, based on SAW sensors. The sensors were designed to operate in hot sections of gas turbines and the research project was sponsored by the U.S Air Force, Navy and the Department of Energy of the United States [Environetix].</para>
<para>On the framework of the MANTIS collaborative project, a research group at the MGEP (Mondragon Goi Eskola Politeknikoa, Arrasate-Mondrag&#x000F3;n, Spain) has developed a system for temperature measurement, based on the concepts explained above. The objective consisted on the measurement and monitoring of the temperatures at the surfaces of heating resistances inside a concrete-block curation oven. The mentioned resistance surface temperatures can range from some 20<superscript>&#x02218;</superscript>C up to well in excess of 200<superscript>&#x02218;</superscript>C at the hottest curation process steps. It is known that the properties of the cured blocks correlate with the curation temperature, which, at the same time, is related to the surface temperatures on the surfaces of the heating resistances. Thus, it is of paramount importance to collect a measurement as close to the real surface temperature as possible in order to carry out a real time manufacturing process monitoring. Several issues have to be considered in this case for a correct temperature sensing:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis>Wireless operation.</emphasis> The high temperatures present inside the oven preclude the use of cables with plastic covers, which can rapidly degrade after few thermal cycles;</para></listitem>
<listitem>
<para><emphasis>Passive sensor implementation</emphasis>. The necessary capability to withstand high temperatures renders commercial electronics, batteries and plastic battery holders, if not unusable, at least not recommendable;</para></listitem>
<listitem>
<para><emphasis>Accurate surface temperature measurement</emphasis>. The sensing element has to be in close contact with the surface under measurement via a low thermal resistivity path. For this reason, if the sensing device is located at a certain distance from the heat source, the sensing structure has to be small in size to avoid heat dispersion in the path between the hot surface and the sensing element (by thermal radiation to the air or by thermal diffusion to other components).</para></listitem>
</itemizedlist>
<para>The size constraint is fulfilled by the use of a COTS SAW resonator component (SAW Components GmbH) with a QFN encapsulation. Designed to work from -50<superscript>&#x02218;</superscript>C up to 275<superscript>&#x02218;</superscript>C in the ISM 343 MHz band, it allows for a temperature sensitivity of 15 KHz/<superscript>&#x02218;</superscript>C. Two sensor prototypes have been constructed for this purpose, and they can be seen in <link linkend="F4-3">Figures <xref linkend="F4-3" remap="4.3"/></link> and <link linkend="F4-4"><xref linkend="F4-4" remap="4.4"/></link>.</para>
<fig id="F4-3" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-3">Figure <xref linkend="F4-3" remap="4.3"/></link></label>
<caption><para>First prototype of a SAW sensor.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_3.jpg"/>
</fig>
<fig id="F4-4" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-4">Figure <xref linkend="F4-4" remap="4.4"/></link></label>
<caption><para>Second prototype of a SAW sensor.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_4.jpg"/>
</fig>
<para>The sensor on <link linkend="F4-3">Figure <xref linkend="F4-3" remap="4.3"/></link> consists of a SAW device with two metallic arms (conforming a dipole) soldered together in a low cost FR-4 substrate. This low cost material does not, of course, withstand high temperature. As a dipole, it cannot be laid in metallic surfaces, as it would preclude the EM energy radiation and absorption from and towards the sensor, but has been used nonetheless as a proof of concept and for validation purposes.</para>
<para>The device in <link linkend="F4-4">Figure <xref linkend="F4-4" remap="4.4"/></link> is a prototype of the planned temperature sensor for the heating resistances. It consists of a small metallic surface (FR-4 substrate) on which the SAW device is soldered in the upright position. A single metallic arm (monopole) is soldered to one of the terminals of the resonator, the other being soldered to the flat metallic surface (ground) that will mimic the remaining arm of a dipole to sustain the resonance in the SAW device. In the final implementation, the SAW device was soldered vertically to a thin sheet of copper, and the vertical monopole arm included a supporting structure consisting of an additional metallic arm to ground to convert the short circuit at the ground level to an open circuit in the monopole end, thus leaving the monopole&#x02019;s electrical behavior unchanged. Before placing the sensor upon the heating resistor, the bottom part of the copper sheet got covered with thermal grease to keep the thermal resistance from the hot surface to the SAW device to a minimum value. Tests proved that the SAW device is in perfect thermal equilibrium with the surface under measurement, and resonant frequency value readings provide an accurate estimation for the temperature of the heating resistance surface.</para>
<fig id="F4-5" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-5">Figure <xref linkend="F4-5" remap="4.5"/></link></label>
<caption><para>Standalone SDR platform, antennas and sensor.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_5.jpg"/>
</fig>
<fig id="F4-6" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-6">Figure <xref linkend="F4-6" remap="4.6"/></link></label>
<caption><para>Developed GUI showing real time temperature measurements.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_6.jpg"/>
</fig>
<para>An external system has been devised to first interrogate and then frequency-analyze the response from the temperature sensor. This system, based on a Software Defined Radio (SDR) architecture, implements the interrogation and response processing systems in a Nutaq&#x02019;s ZeptoSDR&#x00A9; unit [NutaQ]. The hardware solution is visible in <link linkend="F4-5">Figure <xref linkend="F4-5" remap="4.5"/></link>, and it provides the designer with the digital signal processing capabilities of a reconfigurable logic device like an FPGA and the high frequency operation possibilities offered by an incorporated RF front-end mezzanine card.</para>
<para>Two quasi-Yagi antennas were designed and simulated in-house, and they provide enough gain, directivity and required linear polarization to adequately interrogate the sensor and listen to its response. <link linkend="F4-6">Figure <xref linkend="F4-6" remap="4.6"/></link> shows a graphic user interface (GUI) that has been developed in Matlab&#x00A9;. This GUI communicates via Ethernet to the SDR unit, and extracts the temperature measurement to show it in an appropriate manner. A graph shows the temperature variation in real time. Data can also be stored in binary files for further data analysis. It is worth noting the flexibility that such a system architecture offers, as the parameters of the interrogating signal can be accommodated to those of the sensor in use. Thus, as long as the SAW devices used for sensing fall inside the operative frequency band of the ZeptoSDR unit (300MHZ &#x02013; 3GHz), all that is needed to carry out the temperature estimation is an appropriate set of transmitting and receiving antennas.</para>
</section>
<section class="lev2" id="sec4-2-2">
<title>4.2.2 Low-Cost Sensor Solution Research</title>
<para>New sensor technologies like Micro Electro Mechanical System (MEMS) sensors and printed sensors are becoming more popular nowadays. The MEMS sensors take advantage of manufacturing processes that are used commonly with the family of integrated circuits called semiconductor manufacturing processes. The used semiconductor manufacturing methods include for example wet etching, dry etching, molding and plating [Angell et al., 1983]. With the semiconductor manufacturing processes it is possible to build many kinds of sensors cost-effectively, and get the price down for the single sensor. Wide variety of MEMS sensors can be acquired commercially, including, but not limited to, accelerometers, gyroscopes, temperature sensors and pressure sensors. The MEMS commonly integrate also capabilities other than just sensing to the same package such as amplification, pre-processing and analogue-to-digital conversion by combining micromechanics and microelectronics. One advantage of MEMS is their small size and lightness in weight. MEMS structures can be really tiny.</para>
<fig id="F4-7" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-7">Figure <xref linkend="F4-7" remap="4.7"/></link></label>
<caption><para>Ten year market forecast for printed sensors made in 2014 (in $ million) [Chansin, 2014].</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_7.jpg"/>
</fig>
<para>Selection of printed sensors is not as wide as with MEMS sensors but the variety is increasing by the day. For example, currently at least temperature sensors, pressure sensors, gas sensors, strain gauges, parts of accelerometer sensors and moisture sensors can be found as printed sensors. Chansin [2014] from IDTechEx divides the printed sensors into 9 categories, and <link linkend="F4-7">Figure <xref linkend="F4-7" remap="4.7"/></link> shows the ten year market forecast for the printed sensors divided into these categories. Printed sensors are made using common printing methods and equipment. Conductor, semiconductor, dielectric and/or insulator inks are printed on surfaces, leading to flexible and really low-cost sensors. Altogether manufacturing processes have improved in sensor production, decreasing the price of sensors. <link linkend="T4-1">Table <xref linkend="T4-1" remap="4.1"/></link> shows multiple examples of different low-cost sensors with their prices.</para>
<table-wrap position="float" id="T4-1">
<label><link linkend="T4-1">Table <xref linkend="T4-1" remap="4.1"/></link></label>
<caption><para>Low-cost sensors [Junnola, 2017]</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left">Manufacturer</td>
<td valign="top" align="left">Model</td>
<td valign="top" align="left">Type</td>
<td valign="top" align="left">Sensor Type</td>
<td valign="top" align="left">Price</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Rohm Semiconductor</td>
<td valign="top" align="left">KXTJ3-1057</td>
<td valign="top" align="left">MEMS</td>
<td valign="top" align="left">Accelerometer</td>
<td valign="top" align="left">0,49 &#x20AC;</td>
</tr>
<tr>
<td valign="top" align="left">STMicroelectronics</td>
<td valign="top" align="left">LIS2HH12TR</td>
<td valign="top" align="left">MEMS</td>
<td valign="top" align="left">Accelerometer</td>
<td valign="top" align="left">0,49 &#x20AC;</td>
</tr>
<tr>
<td valign="top" align="left">TE Connectivity</td>
<td valign="top" align="left">1007158-1</td>
<td valign="top" align="left">Printed Sensor; Piezofilm</td>
<td valign="top" align="left">Accelerometer</td>
<td valign="top" align="left">3,40 &#x20AC;</td>
</tr>
<tr>
<td valign="top" align="left">Allegro MicroSystems, LLC</td>
<td valign="top" align="left">LLC ACS711EEXLT-31AB-T</td>
<td valign="top" align="left">Integrated Circuit</td>
<td valign="top" align="left">Current Transducer</td>
<td valign="top" align="left">0,56 &#x20AC;</td>
</tr>
<tr>
<td valign="top" align="left">Melexis Technologies NV</td>
<td valign="top" align="left">MLX91209LVA-CAA-000-CR</td>
<td valign="top" align="left">Integrated Circuit</td>
<td valign="top" align="left">Current Transducer</td>
<td valign="top" align="left">1,63 &#x20AC;</td>
</tr>
<tr>
<td valign="top" align="left">Bosch Sensortec</td>
<td valign="top" align="left">BMP280</td>
<td valign="top" align="left">MEMS</td>
<td valign="top" align="left">Pressure Sensor</td>
<td valign="top" align="left">1,17 &#x20AC;</td>
</tr>
<tr>
<td valign="top" align="left">EPCOS (TDK)</td>
<td valign="top" align="left">B58601H8000A35</td>
<td valign="top" align="left">MEMS</td>
<td valign="top" align="left">Pressure Sensor</td>
<td valign="top" align="left">1,17 &#x20AC;</td>
</tr>
<tr>
<td valign="top" align="left">Microchip Technology</td>
<td valign="top" align="left">MTCH101T-I/OT</td>
<td valign="top" align="left">Integrated Circuit</td>
<td valign="top" align="left">Proximity Sensor</td>
<td valign="top" align="left">0,36 &#x20AC;</td>
</tr>
<tr>
<td valign="top" align="left">Semtech Corporation</td>
<td valign="top" align="left">SX9300IULTRT</td>
<td valign="top" align="left">Integrated Circuit</td>
<td valign="top" align="left">Proximity Sensor</td>
<td valign="top" align="left">0,73 &#x20AC;</td>
</tr>
<tr>
<td valign="top" align="left">NXP</td>
<td valign="top" align="left">LM75BDP</td>
<td valign="top" align="left">Integrated Circuit</td>
<td valign="top" align="left">Temperature Sensor</td>
<td valign="top" align="left">0,22 &#x20AC;</td>
</tr>
<tr>
<td valign="top" align="left">Texas Instruments</td>
<td valign="top" align="left">LMT88DCKR</td>
<td valign="top" align="left">Integrated Circuit</td>
<td valign="top" align="left">Temperature Sensor</td>
<td valign="top" align="left">0,17 &#x20AC;</td>
</tr>
</tbody>
</table>
</table-wrap>
</section>
<section class="lev2" id="sec4-2-3">
<title>4.2.3 Soft Sensor Computational Trust</title>
<para>This intelligent soft sensor focuses on the interpretation of the errors that occur during the production process in terms of frequency and severity of the different errors in the shaver production plant described in Section 7.1.</para>
<para>In the shaver production plant, tooling quality is one of the factors affecting product (shaver cap) quality. Tooling quality drops as the number of products generated increases and due to problems like a (series of) short circuit(s). When the production monitoring system observes such a problem, it raises an error. There is a list of error codes in the system resulting from different measurements. There is an order of importance among the errors that are considered relevant for tool damage. Increasing number and frequency of errors in a time interval is also considered critical for tooling quality. Operators can stop the process and change the tool due to raised errors and due to finding visible distortions in products upon examination. The changed tool is inspected: if it is faulty it is discarded, otherwise it is added back to the stock.</para>
<para>There is room for improvement in this process. Firstly, there can be inconsistencies among operators in interpreting the errors. Moreover, as the shift changes, operators tend to look into the errors raised in their shift time only. A model is presented in the following. It uses error data and the expert knowledge about the relative order of importance among error codes as input and raises a flag when the errors imply a critical tool condition. This soft sensor combines insight coming from different errors and provides some sort of data fusion.</para>
<fig id="F4-8" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-8">Figure <xref linkend="F4-8" remap="4.8"/></link></label>
<caption><para>Error codes over time, colored based on the tool IDs.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_8.jpg"/>
</fig>
<para>The dataset under consideration consists of <emphasis>process data</emphasis> that were collected for several months (162 days) for one of the machines in the shaver production plant and <emphasis>maintenance logs</emphasis> serving as the ground truth for why a tool was changed.</para>
<para>A record in the process data contains several process parameters that were measured in the making of a single product, including the time of the production, error code (see <link linkend="F4-8">Figure <xref linkend="F4-8" remap="4.8"/></link>), the ID of the tool used, the numbers of products produced with this tool since it was last replaced and since the first time this tool was ever used. During normal operation the error code parameter contains a 0, otherwise it contains a specific error code corresponding to a certain error type.</para>
<para>A record in the maintenance log contains the time of the tool change, the number of products produced with that tool at the time of the change, and a note written by the operator indicating the reason why the tool was sent for maintenance.</para>
<para>For tooling quality, not all error signals are considered equally relevant. The codes that operators take into account in practice are 50, 150, 52, 186, 182 and 188. According to the production experts, in a scale of 1 to 5 (5 meaning the most and 1 meaning the least critical), their importance levels are 5, 4, 2, 2, 1 and 1, respectively. This grading can be described by the function:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq4-1.jpg"/></para>
<para><emphasis>Theoretical background:</emphasis></para>
<para>Trust is &#x0201C;<emphasis>the degree of justifiable belief a trustor has that, in a given context, a trustee will live up to a given set of statements about its behavior</emphasis>&#x0201D; [Bui et al., 2014]. Computational trust aims to quantify the subjective probability (a trust value in [0,1]) that a <emphasis>trustor</emphasis> A attributes to the truth of a <emphasis>trust statement</emphasis> P that has the following form: &#x0201C;The trustee T satisfies a predicate &#x02118; within context c in the time period &#x003B4;&#x0201D;. In a trust statement, time and context are usually not given explicitly.</para>
<para>The inputs to compute the trust value <emphasis>tv<subscript>P,i</subscript></emphasis> are: positive evidence <inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg108-1.jpg"/>, negative evidence <inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg108-2.jpg"/>, prior trust <emphasis>a<superscript>A</superscript></emphasis> and the weight of prior trust <emphasis>W<superscript>A</superscript></emphasis> (see Equation 4.2). Positive evidence and negative evidence are cumulative, and updated after each observation <emphasis>i</emphasis> via update rules. These rules capture the anticipated behavior of a trust curve. The weight <emphasis>W<superscript>A</superscript></emphasis> determines how fast the effect of prior trust decays.</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq4-2.jpg"/></para>
<para>This is a Bayesian trust model. For further information about Bayesian trust models, see [J&#x00F8;sang, 2016; Ries, 2009].</para>
<para><emphasis>Application:</emphasis></para>
<para>The trust framework was applied to the afore-mentioned production plant dataset. The goal is to make the operators&#x02019; job easier in a way that reduces inconsistencies across different operators and shifts. Computing trust based alerts can support stability across different operators. Moreover, computed trust will not be impacted by shift changes of people. After each observation, the updated trust value can be compared to a threshold. When the computed trust is below the threshold, the operator can be notified. The trust statement of interest is: &#x0201C;The tooling is in good state for production.&#x0201D; The trustee (T) refers to &#x0201C;the tool&#x0201D;, and &#x02118; refers to &#x0201C;&#x02026;is in good state for production&#x0201D;. Domain knowledge is inserted into the trust computation through the error importance levels and the design of the update rules given next.</para>
<para>An update rule is implemented by means of the constants &#x003B7;,&#x003B3;,&#x003B6;,&#x003C8; &#x02208; (0, 1), which are chosen to decay the evidence over time. The constants are used together with e<subscript><emphasis>i</emphasis></subscript>, which is the error code at observation <emphasis>i</emphasis>, and are used to implement the set of formulas in Equation 4.3.</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq4-3.jpg"/></para>
<para>After each time the tool is replaced, the positive and negative evidences are reset back to zero, implying that the trust value is reset back to the prior trust value (see Equation 4.2). In case the error code is in X, the positive evidence is expected to decay quickly. Also, each time error code is not in X, the negative evidence should decay, but slowly. The following coefficients were chosen to make this possible: &#x003B3; = .99, &#x003B7; = .999, &#x003B6; = .4, &#x003C8; = .999. The prior trust <emphasis>a<superscript>A</superscript></emphasis>is set to .95 and <emphasis>W<superscript>A</superscript></emphasis> is set to 10.</para>
<fig id="F4-9" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-9">Figure <xref linkend="F4-9" remap="4.9"/></link></label>
<caption><para>Error codes over time, and corresponding trust values.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_9.jpg"/>
</fig>
<para><link linkend="F4-9">Figure <xref linkend="F4-9" remap="4.9"/></link> shows the behavior of trust over a shorter interval. The tool was changed three times (on November 3 at 03:24, November 6 at 03:36, and November 9 at 20:53) in that interval and, according to maintenance logs, the reason was error code 186 two of the times, and distortions found upon visual inspection once. There is a small interval with relatively low trust values, on November 7 around 21 pm. At that time, there is an accumulation of error codes 186 and 52, but the operator chose to take no action.</para>
<para>A set of 13 instances were inspected, where the tool was changed to see if the computed trust was low in the 50 minutes leading to that change. According to the maintenance log, 8 changes were due to errors. Among these for 6 of them the trust was also relatively low. In the remaining 2, despite the operators note that the change was due to errors, these errors were not concentrated in a small enough time interval around a tool change, hence they did not result in low trust. Among the 5 changes that were due to other reasons (due to the preventive maintenance schedule, due to visible distortion in products or other) in two cases, there was also some accumulation of error codes prior to change, hence trust was also low.</para>
<para>Some errors imply deteriorating tool condition, which should result in a tool change. The proposed approach was evaluated by comparing the times of low trust in the tooling being in a good state with the operator notes in the maintenance logs. Mostly, tool changes due to error codes coincide with time intervals with low trust value. There are cases where accumulation of errors did not result in a tool change, or tool changes were due to errors when in fact there were not so many errors. The trust computation shows that there are inconsistencies of operators about how and when the error signals result in a tool change. Supporting operators with the proposed trust framework can anticipate the need for tool changes, improving the timely maintenance of tools and reducing the number of faulty products.</para>
</section>
</section>
<section class="lev1" id="sec4-3">
<title>4.3 Bandwidth Optimization for Maintenance</title>
<para>Condition monitoring approaches require sensors in order to gather information about the different components and elements that take part in the performance of the machine. This information may be used locally for estimating the state of the components, but has to ultimately be conveyed from sensors to edge servers or processing centres for data analysis, but bandwidth is, in most cases, a limited resource. Roughly speaking, the bandwidth occupied by a signal can be related to the highest relevant frequency component that conforms the signal waveform. Thus, if one is to reach the maximum data throughput that the channel can offer, a natural choice would consist on trying to reduce the time duration of the waveform for the individual piece of information. This way, more and more information is packed in a given time duration and throughput increases. But as waveform change rate increases, so does the required bandwidth. Fast changing signals occupy a wider bandwidth, and less spectrum-demanding signals have to vary in a slower fashion. With this scheme in mind, it is needed to think of approaches that address the bandwidth optimization issue in an appropriate way. Three such approaches are described below.</para>
<section class="lev2" id="sec4-3-1">
<title>4.3.1 Reduced Data Amount and Key Process Indicators (KPI)</title>
<para>One way for decreasing the necessary bandwidth is the reduction in the amount of exchanged data. This way, less spectrum is utilized for the required data throughput and both systems, sensor and edge server, can be of a simpler nature, at least with regards to data processing capabilities. This may lead to a reduced system cost and complexity, but with the drawback of compromised accuracy and reliability of the obtained measurements, as important pieces of information may have to be discarded for transmission. A much better approach could be the reduction of the amount of data by discarding the undesired data and sending only key indicators of the condition of the machines by means of KPIs, which are fundamental for the design and maintenance phase of a machine, as they allow identifying faults, their underlying causes, and any effect that can propagate to other systems. Usually, the monitored machines or parts are comprised of components that are related with each other, so that the deviation of a component from its normal performance causes additional, measurable effects in other components. KPIs are extracted from qualitative models of devices under observation, taking advantage of measurements and estimated magnitudes and parameters. For this purpose, the most meaningful magnitudes of the machines are analysed, highlighting the performance of the system as shown below and described for example in [Mehdi et al., 2015]. See <link linkend="F4-10">Figure <xref linkend="F4-10" remap="4.10"/></link> for a graphical representation of the process.</para>
<fig id="F4-10" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-10">Figure <xref linkend="F4-10" remap="4.10"/></link></label>
<caption><para>BW optimisation by KPI extraction.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_10.jpg"/>
</fig>
<para>On a first stage, all the raw data collected from installed sensors is pre-processed in order to obtain appropriate signals of the measured magnitudes for their later analysis. On the second stage, the preprocessed data is processed through data analytic tools made available by software sensors. At this stage, the condition monitoring of the machines is performed, extracting magnitudes and parameters of interest. At this point, these magnitudes and parameters themselves could be used as KPIs, as they monitor directly the condition of the analysed machine. On the final stage of the data analysis, a high level abstraction data reduction is still possible, by combining the most relevant magnitudes and parameters extracted from the previous stage. These KPIs may not have a straightforward physical interpretation, but they show the working condition at which the machine is. In this way, KPIs can be used to effectively reduce the amount of data that will be exchanged with the main processing and decision centre, optimising the use of bandwidth.</para>
</section>
<section class="lev2" id="sec4-3-2">
<title>4.3.2 Advanced Modulation Schemes</title>
<para>In the previous approach, a decrease in the quantity of transmitted data was used for the reduction of the required transmission spectrum bandwidth. Second scheme focuses on the time percentage on which the physical medium is occupied for data exchange, so that alternative sensors can communicate in different time slots.</para>
<para>Communication is performed by the transmission over the wireless of wired medium of a symbol, which is the minimal entity that can represent some digital data to be sent. The bandwidth optimization technique is based on the concept of modulation, which is the way that data bits are associated to the symbols.</para>
<para>The bandwidth optimization can be addressed from a frequency, phase or amplitude point of view, or by a combination of them, by means of applying different modulations to the signal used to carry over the data. In fact, while the most obvious signal modulation is able to send a single bit with each signal, for example by encoding a 1 with the presence of a wireless wave and a 0 with its absence, it is possible to use a richer alphabet of symbols to encode the data, and this way send more than one bit at a time. On this regard, an N-fold increase in the number of the allowed discrete symbols used for the modulation allows to send a larger (log <emphasis>N</emphasis>) number of bits at the same time while occupying a single symbol duration, i.e., by means of one symbol. More information on modulation schemes and related subjects can be found in [Proakis, 1995; Oppenheim and Willsky, 1997; Haykin, 1994; Proakis, 1995].</para>
<para>It is worth mentioning that the increase in the data transmission rate achieved by means of such modulation processes comes at the expense of more complex sensor and data processing systems, involving electronics related to modulating and demodulating processes, adding protocol complexity to the signal processing methods and thus increasing the total development and deployment cost. On this regard, despite commercial transceivers are readily available in the market and render the development of such systems from scratch unnecessary, their protocols are generally proprietary and restricted to work at certain limited frequency bands (such as ISM). In addition to this, the power supply requirement of such electronically advanced sensors has to be considered before opting for a complex modulation scheme, as in several measurement systems, low power consumption is an important constraint.</para>
</section>
<section class="lev2" id="sec4-3-3">
<title>4.3.3 EM Wave Polarization Diversity</title>
<para>The third approach for the bandwidth optimization arises from the fact that antennas respond differently to electromagnetic fields whose spatial field distributions show diverging field orientations. This phenomenon is known as <emphasis role="strong">polarization</emphasis> [Mehdi et al., 2015] and implies that, for antennas that lie in a given plane, such as the ones represented in <link linkend="F4-11">Figure <xref linkend="F4-11" remap="4.11"/></link>, maximum EM wave power reception occurs for the plane parallel to that of the antenna. On the contrary, a minimum fraction (ideally zero) of the incoming EM field power is captured by the antenna if the relative angle between incoming field and antenna approaches 90 degrees. With this behaviour in mind, it is possible to think of an arrangement of a couple of antennas, orthogonal to each other and both making use of the same transmission frequency value. As long as the data receiving part of the system deploys a pair of also relatively perpendicular antennas (with spatial orientations respectively parallel to those of the transmitting antennas), an effective doubling of the data transmission rate is at hand, at the expense of doubled antenna and cabling deployment cost. Nonetheless, present antennas fabricated in printed circuit form (integrating the associated electronic circuitry) can achieve the desired goal at a reduced cost. Finally, it is of importance to consider situations in which the spatial orientation of the antennas changes with time (e.g., Rotating or translating frames for the sensors and/or data receiving equipment). In those cases, it is necessary to ensure that the relative orientation of the transmitting and receiving antennas does not change, as the aforementioned dynamic displacement could lead to a situation in which transmitting sensor antennas are spatially paired with a wrong receiving antenna, and thus different sensor information would not be received and processed by the appropriate system. Several classical texts on antennas [Balanis, 1982; Krauss and Marhefka, 2002; Stutzman and Thiele, 2013] can be consulted for a more in depth knowledge on the subject.</para>
<fig id="F4-11" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-11">Figure <xref linkend="F4-11" remap="4.11"/></link></label>
<caption><para>Linearly polarized quasi &#x02013; Yagi antennas: Maximum alignment (left) and maximum misalignment (right).</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_13.jpg"/>
</fig>
</section>
</section>
<section class="lev1" id="sec4-4">
<title>4.4 Wireless Communication in Challenging Environments</title>
<para>New technological solutions such as WSNs have become a strategic asset in industrial context. This type of sensor networks are used to share information with the purpose of increasing productivity, gathering data for developing future technological improvements or detecting/predicting maintenance issues. Besides, the use of wireless communications provides flexibility, installation ease, weight reduction, etc. which makes them suitable for many applications, conditions and situations.</para>
<para>Industrial environments usually have hostile site conditions, both for the sensors (dust, oil, heat, corrosive products, vibrations, etc.) and for the wireless communication systems (interferences, metallic environments, etc.). These conditions must be considered before and during the design and development process of a WSN as they could make an impact on the resultant performance. Besides, there are also requirements such as synchronization, low power consumption or data security that can constrain the design process of the wireless sensors to be deployed within an application.</para>
<para>In this section different possible networking solutions are analyzed in order to describe a design methodology to be followed when a wireless sensor solution has to be deployed within a specific challenging environment. The design methodology aims to gather as many technological alternatives and communication solutions as possible.</para>
<section class="lev2" id="sec4-4-1">
<title>4.4.1 Design Methodology Basis</title>
<para>The proposed design methodology is targeted at wireless networks in industrial scenarios, and it is organized as follows:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis>Requirements and challenges identification.</emphasis> This task consists in the network requirements definition according to the specifications and needs of the industrial application, to investigate the constraints and limitations of the resulting network;</para></listitem>
<listitem>
<para><emphasis>Characterization of the medium.</emphasis> This is an essential step during the development of the communication system, as well as the available bandwidth measurements. This provides information about the main parameters that can negatively affect the wireless signal propagation inside the factory. Moreover, mathematical representation of the environment or wireless channel can be obtained;</para></listitem>
<listitem>
<para><emphasis>Interference detection.</emphasis> It is mandatory to know about the presence of other possible users of the spectrum in the area where the network is located. Interference level information would provide a view on the minimum requirements for the device under design, especially to develop a system robust enough to handle and maintain the communication in a specific interference level;</para></listitem>
<listitem>
<para><emphasis>Power and energy saving aspects.</emphasis> Wireless devices usually are battery powered devices, and consequently they have a limited autonomy level. Therefore, low power and energy cost aspects have to be taken into account before and during the network design and development in order to extend the network lifetime;</para></listitem>
<listitem>
<para><emphasis>Development of the device.</emphasis> Considering all the information gathered from the previous tasks, it is time to select, design and develop the components, the topology and the PHY and MAC layers required by the wireless communication system. There are many examples proposed in the literature, so the adequate alternative should be chosen and modified depending on needs and the requirements determined in the previous steps;</para></listitem>
<listitem>
<para><emphasis>Validation.</emphasis> Once the network is designed and deployed, it is necessary to carry out its validation. As a first step, performance tests using a channel emulator are highly valuable, since this type of device enables the designer to test different propagation parameters and effects in a lab environment, testing if the designed system is reliable enough for the industrial site it is destined to;</para></listitem>
<listitem>
<para><emphasis>Performance tests.</emphasis> As a last step, it is necessary to make performance tests in the target location to know whether the network fulfils the specified initial design requirements or not.</para></listitem>
</itemizedlist>
</section>
<section class="lev2" id="sec4-4-2">
<title>4.4.2 Requirement and Challenge Identification</title>
<para>This step collects the requirements and challenges that can be found in a project related to wireless communication systems.</para>
<para>The industrial environment has always been a very demanding scenario for technology, especially for wireless communications. There are usually elements such as metallic surfaces, electro-magnetic interferences, rotating or moving elements, vibrations, etc. that influence directly or indirectly the wireless devices or the electromagnetic waves used for communication.</para>
<para>The large metal surfaces that are usually located in industrial environments distort the wireless signal, which leads to reflections, scattering and diffraction. The metal surfaces can make the signal travel via different paths, which then suffers from multipath interference.</para>
<para>Moreover, in industrial spaces there is also massive electromagnetic noise. Noise is usually caused by hardware thermal effects, other wireless networks or industrial machinery which can alter the correct behavior of the network at the industrial site. <link linkend="F4-12">Figure <xref linkend="F4-12" remap="4.12"/></link> shows some processes and devices which cause interference in industrial environments.</para>
<fig id="F4-12" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-12">Figure <xref linkend="F4-12" remap="4.12"/></link></label>
<caption><para>Interferences from different processes and devices in industry.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_14.jpg"/>
</fig>
<para>Sometimes it is possible to used wired technologies to protect communication from these problems, but many scenarios do not allow for cables and must use wireless nodes. For example, machines can have rotating or moving parts such as rotors, assembly chains or clutches.</para>
</section>
<section class="lev2" id="sec4-4-3">
<title>4.4.3 Channel Measurement</title>
<para>The main objective of the channel measurement task is to reduce the effort involved in the selection of the most suitable PHY and MAC layers during the following steps of the methodology as well as to obtain detailed information about the specific industrial environment. This reduces the time and effort required for the development, while maximizing the performance of the resulting network.</para>
<para>Most of channel characterization techniques are based on stimulus-response measurements. The channel is excited with a predefined and known signal and it is received on the other side of the communication. Then, the differences between the acquired and emitted signals are observed to determine the characteristics of the channel. As a result, a mathematical model for the channel is obtained. This model is to be used both in the PHY and MAC selection and as well in the validation process, for example while using a channel emulator.</para>
<para>The most important statistical parameters within a channel model are the following:</para>
<para><emphasis role="strong">Multipath characterization</emphasis> &#x02013; The multipath effect causes signal amplitude variations because of the arrival of the waves from different paths. Signals usually suffer similar attenuations but different phase modification, which leads to constructive or destructive interference in the receiver. The characterization of this effect may be useful to place the antennas and avoid destructive interferences. To compensate the effect of multipath, the equalization technique is commonly employed. Furthermore, other solutions such as directional antennas provoke less multipath signal components and consequently less fading and delay spread.</para>
<para><emphasis role="strong">Coherence bandwidth measurement</emphasis> &#x02013; This is a statistical measurement of the range of frequencies over which the channel can be considered &#x0201C;flat&#x0201D;, which is the frequency interval over which two frequencies of a signal are likely to experience comparable or correlated amplitude fading.</para>
<para>The obtained results help in the further selection of the PHY and MAC layers of the communication system to be deployed: wide-band/narrow-band, techniques to mitigate fading (diversity, equalization, OFDM, MIMO), techniques to mitigate interference, etc.</para>
<para><emphasis role="strong">3D modelling</emphasis> &#x02013; Simulation algorithms based on the ray tracing method-of-images enhanced by double refraction modelling can be used to identify the correct placement of the receiver and transmitter antennas. The algorithm takes as input a 3D model provided as a list of triangulated surfaces and a list of convex edges, outputs a signal loss value in dB for each surface in the receptor plane. An example is represented in <link linkend="F4-13">Figure <xref linkend="F4-13" remap="4.13"/></link>, where the colour scale is from 0 dB loss (black) to 100 dB (white).</para>
<fig id="F4-13" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-13">Figure <xref linkend="F4-13" remap="4.13"/></link></label>
<caption><para>Example of a 3D model.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_15.jpg"/>
</fig>
</section>
<section class="lev2" id="sec4-4-4">
<title>4.4.4 Interference Detection and Characterization</title>
<para>The current use of the wireless spectrum by the nearby elements or devices has an influence in the design of a wireless network, so interference measurement and interference modeling is useful to characterize wireless behaviour in a specific environment. Therefore, it is essential to study the primary users of the desired frequencies in order to avoid possible interferences and poor network performance.</para>
<para>The main causes of external interference are the electric interference and other RF wireless networks. Furthermore, these interfering signals might have been transmitted maliciously by a jammer. RF interference can be a very serious threat especially for low power wireless networks.</para>
<para>Interference may be also caused by industrial machinery, such as motorized devices or computers. Electrical motors and relays are one of the most interfering elements, especially during the switching instants. In addition, other types of high-voltage devices can also be an interference source due to defective insulation.</para>
<para>To solve RF interference problems, the most common alternatives involve eliminating the source of the interference, or modifying the transmission frequency. However, this is not always possible, so researchers are developing new mechanisms to behave dynamically in case of RF interference presence.</para>
</section>
<section class="lev2" id="sec4-4-5">
<title>4.4.5 PHY Design/Selection and Implementation</title>
<para>The PHY layer is the first and lowest layer of the OSI model, and consequently it has a considerable impact on the final operation and performance of the network. Therefore, the physical layer should be carefully designed/selected to satisfy the requirements defined during the first step of the methodology.</para>
<para>Most standards address PHY and MAC layers together, hence a decision regarding the PHY layer has an impact regarding which MAC layers can be used in the system.</para>
<section class="lev3" id="sec4-4-5-1">
<title>4.4.5.1 Single/multi carrier</title>
<para>The first aspect a designer should decide is whether the physical layer is a single carrier or multicarrier based physical layer. In single carrier based physical layers, just one carrier is used to carry all the information. They are simpler than multicarrier based alternatives and are a better fit for low consumption but low performance solutions.</para>
<para>On the other hand, multicarrier based physical layers are designed to operate with multiple carrier signals at different frequencies to send the data. This approach has got performance advantages and is more robust against narrow-band and multipath interference. On the other hand, the main limitation of multicarrier physical layers is the difficulty to synchronize the carriers correctly. Besides they require more bandwidth and the PHY layer&#x02019;s complexity has a direct impact on power consumption.</para>
</section>
<section class="lev3" id="sec4-4-5-2">
<title>4.4.5.2 High performance/low power</title>
<para>This is a dual view with respect to the previous section. An engineer should also take into account the difficulties that may appear to achieve the desired performance with low power consumption. For example, multi carrier modulations are more complex and they are more demanding on the energy. However, they provide a higher data rate and a high robustness against interference comparing to single carrier modulations. Therefore, the data rate and the results from the interference detection step must be taken into account in this decision.</para>
</section>
</section>
<section class="lev2" id="sec4-4-6">
<title>4.4.6 MAC Design/Selection and Implementation</title>
<para>The MAC sublayer is part of the Data Link Layer of the OSI model and there are many alternatives to design or select a MAC layer for the wireless communication system to be developed.</para>
<para>Many aspects can be discussed within the selection/development of a MAC layer: low power consumption, synchronization and real-time features, cognitive features, security, reliability, etc. Moreover, the selected MAC layer must be compatible with the PHY layer chosen in the previous step.</para>
<section class="lev3" id="sec4-4-6-1">
<title>4.4.6.1 Real-time/deterministic MACs</title>
<para>There are many mechanisms to access the medium for wireless technologies, though the techniques can be classified into two major groups. On the one hand, there are channel partitioning based Media Access Control mechanisms where the channel is divided into several parts so that several nodes can access the channel in a multiple manner. In this way, nodes can only communicate on their assigned division. The main advantage of channel partitioning is the ease to organize the communications between nodes, but this alternative is not suited for sporadic and bursty communications. On the other hand, random access based MACs can be also divided depending on contention capacities, such as in the case of ALOHA or CSMA.</para>
<para>The MAC layer has to be chosen depending on the requirements of the application as well as the characteristics of the environment. To ensure determinism and real-time communications, partitioning based media access should be used. The random access based MACs do not ensure transmissions bounded in time because they can allow the presence of collisions among the nodes in the network. However, the partitioning based MACs are more complex due to the fact that they require coordination among the different nodes in the network.</para>
</section>
<section class="lev3" id="sec4-4-6-2">
<title>4.4.6.2 Low-power MACs</title>
<para>Apart from the most common options and standards, there are many Low Power MACs for WSNs in the literature. The studied MACs can be divided into the following three categories: asynchronous, synchronous and multichannel.</para>
<para>The asynchronous MACs are aimed for networks whose nodes have different active/sleep schemes. The nodes of this type of networks are asleep most of the time and they wake up to just communicate occasionally. The fact that each node has its own scheme requires communication establishment process with long communication preambles in order to make the receiver detect the transmitter node. Therefore asynchronous MACs can result in an adequate alternative for low traffic networks. The main drawbacks of asynchronous MACs are the low spectral efficiency and channel inhibition due to long preamble as well as poor or none time synchronization, which is usually required by many industrial applications.</para>
<para>The synchronous protocols are thought to obtain instantaneous wake up between the nodes and, on the contrary of asynchronous MACs, the devices of the network have a predefined behaviour. Only predefined transmitter and receiver access the channel at a determined moment, while the other devices remain asleep or just listen to the channel until their communication time. The energy consumption is usually a main issue for synchronous MACs, thus their techniques try to avoid the principal energy wastes: collisions, overhead and overhearing. Synchronization can result in delay reduction and throughput improvement, although it requires an additional overhead to the communication to send information related to the clocks and synchronization. This requires the implementation of algorithms to manage the temporal aspects of the network.</para>
<para>Multichannel MACs are used to enable parallel transmissions by using different channels for the communication between groups of nodes. The tendency is to combine TDMA and FDMA in order to achieve a higher communication rate and consequently a better overall network performance. Therefore, the nodes must be able to handle frequency changes over time, which usually require more expensive technologies.</para>
</section>
<section class="lev3" id="sec4-4-6-3">
<title>4.4.6.3 High level protocols for error mitigation</title>
<para>Communication in a challenging environment has to cope with problems in the communication process, which can take the form of errors in the transmitted data, fluctuating bandwidth, and intermittent connections. Current approaches to cope with the hurdle can be divided into two families:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis>Message caching</emphasis>: this set of approaches is based on caching the messages to be exchanged. The message is kept in the cache and the communication process is repeated until the recipient is reached. Most of these techniques are based on brokers that act as intermediaries in the communication process. These techniques help only with errors on the channel from the broker to the receiver, and do not support communication from sender to the broker. The Advanced Message Queuing Protocol (AMQP), the Message Queuing Telemetry Transport (MQTT), the Java Message Service (JMS), the Data Distribution System (DDS) and the Open Platform Communications Unified Architecture (OPC-UA) are some of the most common implementations;</para></listitem>
<listitem>
<para><emphasis>Cognitive communication</emphasis>: it is based on the use of context information to drive the mechanisms of communication protocols. Context information is exchanged over the application layers, and it is used to drive the lower layers in order to mitigate the noisy and congested spectrum bands, for example by controlling which wireless frequencies are used over time, yielding reliable and high capacity links for wireless communication.</para></listitem>
</itemizedlist>
</section>
</section>
<section class="lev2" id="sec4-4-7">
<title>4.4.7 System Validation</title>
<para>The main objective of system validation is to measure the performance of the designed wireless communication system and verify if it fulfils the requirements identified in the first steps of the design process. This validation can be done using instrumentation such as channel emulators or/and via tests in the target location.</para>
<section class="lev3" id="sec4-4-7-1">
<title>4.4.7.1 Channel emulation</title>
<para>Channel emulators enable to reproduce controlled and programmable channel conditions so that product performance can be evaluated in realistic conditions before product release or delivery. Information obtained from the channel measurement step is used in this emulation process, such as parameters on Pathloss, Multipath, Doppler or Noise.</para>
<para>This is shown for example in <link linkend="F4-14">Figure <xref linkend="F4-14" remap="4.14"/></link>, which depicts the insertion of these effects during the emulation of the channel.</para>
<fig id="F4-14" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-14">Figure <xref linkend="F4-14" remap="4.14"/></link></label>
<caption><para>Channel emulator scheme.</para></caption>
<graphic xlink:href="graphics/ch04_ch04_fig14.jpg"/>
</fig>
<para>The validation by channel emulation can provide useful information about the networks behaviour or performance. The main drawback of wireless channel emulators is the limited support for a high number of devices.</para>
</section>
<section class="lev3" id="sec4-4-7-2">
<title>4.4.7.2 Performance tests</title>
<para>Once the communication system has been installed in the target location, the initial design requirements satisfaction needs to be verified. These are the main topics to be taken into account:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis>Channel selection</emphasis>: Depending on the wireless spectrum used, and especially in 2.4 GHz frequency bands, it is important to select the best channel to avoid interferences;</para></listitem>
<listitem>
<para><emphasis>Coverage verification</emphasis>: For each wireless device installed, it is needed to know how much signal power is arriving to the installation point. Using this information, usually some range or ranges can be defined:</para>
<table-wrap position="float">
<table cellspacing="5" cellpadding="5" frame="box" rules="all">
<tbody>
<tr>
<td valign="top" align="left">signal > -x</td>
<td valign="top" align="left">&#x02192; Good coverage</td>
</tr>
<tr>
<td valign="top" align="left">-x > signal > -y</td>
<td valign="top" align="left">&#x02192; Normal coverage</td>
</tr>
<tr>
<td valign="top" align="left">-y > signal</td>
<td valign="top" align="left">&#x02192; Poor coverage</td></tr>
</tbody>
</table>
</table-wrap>
<para>After these verifications, some corrective measures could be needed. Some examples could be to move the sensor to a new spot, to use an antenna extender or to add a repeater to improve the coverage area;</para></listitem>
<listitem>
<para>Regarding the topology, it is recommended that every sensor should have at least two good neighbours (with good coverage) to ensure that if any problem occurs with one link, the communication is guaranteed by routing over other links;</para></listitem>
<listitem>
<para><emphasis>Packet reception verification</emphasis>. Once the coverage checking process is completed and eventual problems described in previous paragraphs are solved, it is necessary to ensure that data is transmitted continuously and without losses that could show any kind of coverage problems. Ideally, the test period should have enough time to cover at least: all possible climate conditions (rain, sun, etc.), day and night and several days operating in normal environmental condition with all equipment and staff working as usual.</para></listitem>
</itemizedlist>
</section>
</section>
</section>
<section class="lev1" id="sec4-5">
<title>4.5 Intelligent Functions in the Sensors and Edge Servers</title>
<para>In the present section, the main categories and their subcategories regarding intelligent functions are presented, the main distinction being whether the function falls in the &#x0201C;intelligent&#x0201D; or &#x0201C;smart&#x0201D; category. Nonetheless, other possible categorizations arise, and are briefly discussed. Later on, a list of intelligent and smart functions is given, depending on the way the data is handled, and details on the implementations are also given, based on real-life sensors developed for various use cases.</para>
<para><emphasis role="strong">Smart vs Intelligent</emphasis></para>
<para>A sensor is a combination of one or many sensing elements, an analog interface, an analog to digital converter (ADC) and a bus interface all in one housing. A <emphasis>smart</emphasis> sensor can be defined as a sensor that, in addition to the pure sensing, provides communication and preprocessing, which are needed in most Maintenance 4.0 applications, and possibly more complex capabilities pertaining to the following categories:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>outlier detection (from historical data);</para></listitem>
<listitem>
<para>false value detection (from data);</para></listitem>
<listitem>
<para>combine and choose the best value.</para></listitem>
</itemizedlist>
<para>An <emphasis>intelligent</emphasis> sensor has got some level of self-awareness, and for example can provide functions such as self-testing, self-identification, self-validation, self-adaptation, etc. (see <link linkend="F4-15">Figure <xref linkend="F4-15" remap="4.15"/></link>). These functions can also support the maintenance of the sensor itself, or they could give additional information about the accuracy, the confidence of the measurement result or about the sensor&#x02019;s own &#x0201C;health status&#x0201D;, and can support in more sophisticated ways the distributed collaborative decision-making concept for monitoring and proactive/predictive maintenance.</para>
<fig id="F4-15" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-15">Figure <xref linkend="F4-15" remap="4.15"/></link></label>
<caption><para>Schematic outline of a process measured by a sensor with sensing elements A and B.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_17.jpg"/>
</fig>
<para>The most common intelligent function found in the industry is self-validation. In many use cases the intelligent functions are implemented at a higher level than the sensor level, therefore they are not directly associated to the sensor. In the edge tier, this set of functions is commonly implemented in the Raspberry Pi based gateways, where considerable processing power is present.</para>
<para><emphasis role="strong">Domains of the measurements</emphasis></para>
<para>Most sensing aspects relate to six well-defined physical domains, listed in <link linkend="F4-16">Figure <xref linkend="F4-16" remap="4.16"/></link>.</para>
<fig id="F4-16" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-16">Figure <xref linkend="F4-16" remap="4.16"/></link></label>
<caption><para>Sensing Physical Domains.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_18.jpg"/>
</fig>
<para>Most of the use cases of the MANTIS project monitor environmental data such as environmental temperature, humidity. More often than not, use cases have special sensors, not shared by other use cases. This was expected, as different use cases have case specific sensors.</para>
<para>However, there are some sensors that are common for many industrial use cases. Such examples include:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Oil quality;</para></listitem>
<listitem>
<para>Vibration;</para></listitem>
<listitem>
<para>Temperature (not environmental);</para></listitem>
<listitem>
<para>Air pressure sensor;</para></listitem>
<listitem>
<para>Power sensor (electrical);</para></listitem>
<listitem>
<para>Current sensor.</para></listitem>
</itemizedlist>
<para>More information on this topic can be found in [Albano, 2018].</para>
<para><emphasis role="strong">Communication</emphasis></para>
<para>The sensors can communicate through a wired or wireless technology.</para>
<para>From the communication perspective, the sensors can use unidirectional communication (just upstream, uploading data) or bidirectional, where sensors can also receive commands. The bidirectional communication enables additional intelligent functions and interaction between different levels of processing, such as collaborative decision making.</para>
<para>The sensors may be able to communicate directly to the cloud, or they may require a gateway for communication. In many cases the sensor is only capable to send the actual sample, using simple wired/wireless communication. In these cases a gateway is needed which can run a more complex communication stack capable of uploading data to the cloud with the required context and security. The analysis of the collected data showed that many of the use cases have complex devices, such as gateways, that are used as aggregation point for sensors and perform smart/intelligent functions and communication.</para>
<fig id="F4-17" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-17">Figure <xref linkend="F4-17" remap="4.17"/></link></label>
<caption><para>Categories table header.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_19.jpg"/>
</fig>
<para><emphasis role="strong">Function categories</emphasis></para>
<para>It is possible to compile a table for the categories of smart and intelligent functions, to find commonalities along the categories. This approach can provide a further level of refinement to three subcategories cited above (self-calibration, self-diagnostics and self-testing), and the same is true for the smart functions.</para>
<para>The categories table is reported in <link linkend="F4-17">Figure <xref linkend="F4-17" remap="4.17"/></link>. No straightforward hierarchical characterization and listing of sensors emerge from the table, and the proposed taxonomy considers &#x0201C;intelligent&#x0201D; and &#x0201C;smart&#x0201D; as the two main categories for intelligent function distinctions, considering additional characteristics (those regarding aspects like communication or measurement domain) as implementation details for a given function. In the following sections, the defined intelligent functions are explained and practical examples are given.</para>
<section class="lev2" id="sec4-5-1">
<title>4.5.1 Intelligent Function: Self-Calibration</title>
<para>Sensors are used to measure different physical magnitudes, and their outputs are correlated to the absolute values of the latter. This correlation can consist in an absolute value correspondence between the output value and the measured property, or can also consist on a relative change representing the variation of the magnitude under measurement. Whatever the case, be it absolute or relative measurement, one desirable property of a sensor is that of being able to compare its present output value against a predefined input, the latter emulating the effect that the magnitude under measurement exerts upon the sensor. This way it is possible to check whether the outcome of the sensing follows the expected performance. Sensors that incorporate the ability of self-calibrating function substitute thus the actual magnitude under measurement with known input values and check the corresponding output, in order to apply the needed corrections.</para>
<section class="lev3" id="sec4-5-1-1">
<title>4.5.1.1 Practical application: Press machine torque sensor</title>
<para>The strain gauges used in the press machine for the torque sensing, represented in <link linkend="F4-18">Figure <xref linkend="F4-18" remap="4.18"/></link> and further described in <link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link> Section 3, require a good initial settlement to enhance dynamic range and avoid signal saturation. Two techniques have been implemented to achieve this initial settlement: auto-zeroing and offset cancellation. The self-calibration intelligent function is handled by the microcontroller through programmable interface electronics in an iteration loop until both the pursued zero value and the elimination of the offset are reached.</para>
<fig id="F4-18" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-18">Figure <xref linkend="F4-18" remap="4.18"/></link></label>
<caption><para>Torque sensor.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_20.jpg"/>
</fig>
</section>
<section class="lev3" id="sec4-5-1-2">
<title>4.5.1.2 Practical application: X-ray tube cathode filament monitoring</title>
<para>In the monitoring of health equipment, described in <link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link> Section 9, an X-ray tube is connected to a so-called High Voltage generator to enable and control its operation. There are three basic functions that the generator has to perform to enable the X-ray generation:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>It needs to supply the high voltage (in the range of 40 kV to 125 kV) between the cathode and the anode;</para></listitem>
<listitem>
<para>It needs to supply a heating current to the cathode to control the amount of electrons emitted;</para></listitem>
<listitem>
<para>It needs to power the electromotor to spin the anode disk, so the electrons hit an ever moving &#x0201C;focal track&#x0201D; to avoid too much damage in one position.</para></listitem>
</itemizedlist>
<para>The X-ray tube can electronically be seen as a passive component without any intelligence. The control intelligence and sensing are mainly built into the generator. In the use case on X-ray tube monitoring some extra intelligence was added to the sensing already available.</para>
<para>The X-ray tubes are used for making medical images. Depending on the type of image desired, the properties of the patient, the technique used and considerations of physical and regulatory limitations, an optimal regulation strategy was designed. This makes it possible to obtain very similar imaging results for a wide range of patients with different weights and bone structures. It is important for the Image Quality that images are taken at correct set points for tube voltage and emission current. The first time a X-ray tube is used a number of properties of this tube are measured and stored to enable the generator to actually reach the desired settings. The usage of the tube will cause the actual behavior to deviate increasingly from the stored properties. During a semi-annual recalibration cycle, the properties need to be measured again. The self-calibration intelligent function implemented in the context of the MANTIS project, was able to remove the need for this recalibration cycle.</para>
<para>There are three reasons why it is desirable to eliminate the need for recalibration:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>For the recalibration it is necessary to produce X-ray for a non-medical purpose. For reasons of radiation-safety it is highly desirable to limit this as much as possible;</para></listitem>
<listitem>
<para>The further the tube starts to deviate from the stored properties the higher the probability that regulatory constraints or image quality are compromised;</para></listitem>
<listitem>
<para>The recalibration takes time of field service engineers and the faster they can perform their jobs the lower the cost is.</para></listitem>
</itemizedlist>
<para>The proposed intelligent function limits itself to the direct relationship between a particular high voltage generator and the X-ray tube connected to it. The algorithm provides means to automatically adjust the initially established properties to the changes caused by the wear of X-ray tube parts. The algorithm to perform this adjustment runs inside the generator itself. It also provides a means to monitor X-ray tube wear.</para>
<fig id="F4-19" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-19">Figure <xref linkend="F4-19" remap="4.19"/></link></label>
<caption><para>Filament heat/emission correction.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_21.jpg"/>
</fig>
<para>The algorithm is represented in <link linkend="F4-19">Figure <xref linkend="F4-19" remap="4.19"/></link>, and it proceeds as follows. When a filament heating current (I<subscript>fa</subscript>) is applied, given a tube voltage, this produces an actual emission current (I<subscript>ea</subscript>), this allows to look-up the corresponding filament current at this tube voltage in the adaptation table: (I<subscript>ft</subscript>). For the next run it is possible to find the set-point for the filament current: (I<subscript>fn</subscript>) by looking-up the filament current for the desired emission current and tube voltage: (I<subscript>fd</subscript>) with the formula:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq4-4.jpg"/></para>
<para>The factor: (I<subscript>fa</subscript>) / (I<subscript>ft</subscript>) that actually adjusts the adaptation table is a correction factor called the C-factor (F<subscript>c</subscript>).</para>
</section>
<section class="lev3" id="sec4-5-1-3">
<title>4.5.1.3 Practical application: Compressed air system</title>
<para>The Air pressure sensor called Sphere Avisense, represented in <link linkend="F4-20">Figure <xref linkend="F4-20" remap="4.20"/></link> and used in the use case on pultrusion described in <link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link> Section 2, supports the user by displaying the different steps to follow for analog outputs calibration. The output calibration procedure consists in measuring the current for 5mA and 19mA current set by the sensor. After entering those real values, a correction is done on board to calibrate the output. This calibration needs to be done for the 2 analog outputs and then must be stored locally.</para>
<fig id="F4-20" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-20">Figure <xref linkend="F4-20" remap="4.20"/></link></label>
<caption><para>Sphere Avisense.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_22.jpg"/>
</fig>
</section>
</section>
<section class="lev2" id="sec4-5-2">
<title>4.5.2 Intelligent Function: Self-Testing (Self-Validating)</title>
<para>The performance of wireless sensor networks could be greatly improved by self-testing or self-validation techniques. Fault tolerance is a highly desirable property for any control system. Fault-tolerant controllers typically rely on some sort of fault detection algorithm, and self-testing or validating devices extend this concept, supplying the user with an estimate of measurement reliability as well as measurement value and its associated uncertainty. Generally, this type of function relies on data analysis techniques applied to aggregated information, gathered from several distinct sensors. This aggregated data can be analysed and mathematically processed to obtain a basis for the identification and quantification of inconsistent data, and evaluation of reference values and associated uncertainties. Self-validation thus becomes more important when various data from multi-sensors or multi-measurements are sent to the system.</para>
<section class="lev3" id="sec4-5-2-1">
<title>4.5.2.1 Practical application: Oil tank system</title>
<fig id="F4-21" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-21">Figure <xref linkend="F4-21" remap="4.21"/></link></label>
<caption><para>Oil condition sensor.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_23.jpg"/>
</fig>
<para>The algorithm of Software Based Testing with Compressed is used for the Oil Condition Sensor in <link linkend="F4-21">Figure <xref linkend="F4-21" remap="4.21"/></link> and employed in the monitoring of pultrusion process (<link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link> Section 2). The algorithm combines SBST testing with CS and can improve some limits such as limited time testing, limited energy and limited processing capability in WSNs. The framework contains six major steps:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Step1. The SBST produces test vectors for each WN and test vectors development to all units of WN such as power supply, communication and sensing units;</para></listitem>
<listitem>
<para>Step2. The test-result is compared with a result known in each WN to uncover any fault;</para></listitem>
<listitem>
<para>Step3. The BS collects all compressive test-results of all WNs with recovers the original test-result for each WN;</para></listitem>
<listitem>
<para>Step4. The WN is implemented UT that is Test Driven Development (TDD) that produces different test vector than SBST program and new test-result is written on WN;</para></listitem>
<listitem>
<para>Step5. The WN is sent the new test-result to BS.</para></listitem>
</itemizedlist>
</section>
<section class="lev3" id="sec4-5-2-2">
<title>4.5.2.2 Practical application: Air and water flow and temperature sensor</title>
<fig id="F4-22" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-22">Figure <xref linkend="F4-22" remap="4.22"/></link></label>
<caption><para>Air/water Flow and temp: SA5000, IFM electronic.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_24.jpg"/>
</fig>
<para>The sensor in <link linkend="F4-22">Figure <xref linkend="F4-22" remap="4.22"/></link> is used to monitor the pultrusion process, described in <link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link> Section 2. The sensor incorporates a red/green display for clear identification of the acceptable range. If the measured value is outside the measuring range or in the event of an internal error, the current or frequency signals indicated in <link linkend="F4-23">Figure <xref linkend="F4-23" remap="4.23"/></link> are provided.</para>
<para>For measured values outside the display range or in case of a fault, messages are displayed</para>
<para>The analogue signal in case of a fault is adjustable:</para>
<itemizedlist mark="none" spacing="normal">
<listitem>
<para>[FOU] = On determines that the analogue signal goes to the upper final value (22 mA) in case of an error.</para></listitem>
<listitem>
<para>[FOU] = OFF determines that the analogue signal goes to the lower final value (3.5 mA) in case of an error.</para></listitem>
</itemizedlist>
<fig id="F4-23" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-23">Figure <xref linkend="F4-23" remap="4.23"/></link></label>
<caption><para>Air/water/temperature sensor: Error signals for self-test result indications.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_25.jpg"/>
</fig>
</section>
<section class="lev3" id="sec4-5-2-3">
<title>4.5.2.3 Practical application: Sensors for the photovoltaic plants</title>
<para>When reviewing irradiance sensors in the field, it was found that they often tend to underestimate the solar resource. This phenomenon has been identified by analysing data from a random selection of 88 sensors recorded by 3E, who is the use case owner and built the monitoring platform SynaptiQ, and comparing it with irradiation data derived from satellite images, which has proven to have a very low error for annual irradiation. The results, collected on the use case described in <link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link> Section 7, show that 50% of all sensors tended to underestimate the yearly irradiation by 7% or more. It is therefore possible to conclude that many resource assessments or PR calculations based on these sensors and data will be biased in the same way.</para>
<para>A Solar Sensor Check has been developed and applied to check the integrity of irradiance sensors based on measurements and indicate if their measurements are wrong or imprecise. The PV Sensor Check checks for a selection of most important faults and imprecisions in an automated process based on measurements. It returns a conclusion on whether a fault could be detected or not. If a fault is detected, the Sensor Check specifies and quantifies the error and indicates the most probable root causes (see <link linkend="T4-2">Table <xref linkend="T4-2" remap="4.2"/></link>), so that the user can decide on alleviation actions. The effects considered in the developed work are plausibility (data completeness, minimum and maximum values, overall data bias), soiling (reduction on the measured irradiance relative to the real irradiance) and shading (reduction on the measured irradiance in dependence of the sun position in a highly nonlinear way).</para>
<table-wrap position="float" id="T4-2">
<label><link linkend="T4-2">Table <xref linkend="T4-2" remap="4.2"/></link></label>
<caption><para>Sample output of the Solar Sensor Check</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left">Check</td>
<td valign="top" align="left">Fault Illustrator</td>
<td valign="top" align="left">Fault Illustrator Value</td>
<td valign="top" align="left">Conclusion</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Recording: maximum irradiance</td>
<td valign="top" align="left">Maximum Irradiance</td>
<td valign="top" align="left">1126 W/m<superscript>2</superscript></td>
<td valign="top" align="left">OK</td>
</tr>
<tr>
<td valign="top" align="left">Recording: minimum irradiance</td>
<td valign="top" align="left">Minimum Irradiance</td>
<td valign="top" align="left">0.75 W/m<superscript>2</superscript></td>
<td valign="top" align="left">OK</td>
</tr>
<tr>
<td valign="top" align="left">Recording: sensor data complete</td>
<td valign="top" align="left">Daytime recording fraction</td>
<td valign="top" align="left">68.5 %</td>
<td valign="top" align="left">Not OK</td>
</tr>
<tr>
<td valign="top" align="left">Total irradiation</td>
<td valign="top" align="left">Mean bias error</td>
<td valign="top" align="left">-6.48 %</td>
<td valign="top" align="left">Not OK</td>
</tr>
<tr>
<td valign="top" align="left">Clock synchronization</td>
<td valign="top" align="left">Time shift</td>
<td valign="top" align="left">0 min</td>
<td valign="top" align="left">OK</td>
</tr>
<tr>
<td valign="top" align="left">Sensor orientation</td>
<td valign="top" align="left">Estimated azimuth &#x00026; tilt</td>
<td valign="top" align="left">58<superscript>&#x02218;</superscript>, 10<superscript>&#x02218;</superscript>->-20<superscript>&#x02218;</superscript>, 29<superscript>&#x02218;</superscript></td>
<td valign="top" align="left">Not OK</td>
</tr>
<tr>
<td valign="top" align="left">Sensor calibration: offset</td>
<td valign="top" align="left">Sensor offset</td>
<td valign="top" align="left">1.8 W/m<superscript>2</superscript></td>
<td valign="top" align="left">OK</td>
</tr>
<tr>
<td valign="top" align="left">Sensor calibration: slope</td>
<td valign="top" align="left">Sensor slope</td>
<td valign="top" align="left">0.968</td>
<td valign="top" align="left">OK</td>
</tr>
</tbody>
</table>
</table-wrap>
</section>
</section>
<section class="lev2" id="sec4-5-3">
<title>4.5.3 Intelligent Function: Self-Diagnostics</title>
<para>Broadly speaking, self-diagnosis can be thought of as the process by which a sensor self-applies a method to detect or evaluate a failure in its functioning. These methods often rely on &#x0201C;pulling&#x0201D; different state indicators from the sensors (e.g., connectivity metrics, current or voltage consumption, historical measurements&#x02026;) and conducting then the pertinent analysis upon the gathered data. The sensor can be represented for example as a state-machine that, in accordance with the given inputs, transits from state to state, reporting (based on constructed fault detectors) back to the central analysis unit and taking decisions based on the estimated performance and possible failures. The state of the sensor can be checked for detection of lost calibration. For example, if after the self-diagnosis a sensor presents offset behavior, an alarm should be triggered. Other sensors correlated to this sensor need to be checked to distinguish a sensor anomaly from a component and/or compressor anomaly. If redundant or highly correlated sensors are available, relationship between the sensor can be used to perform further tests.</para>
<section class="lev3" id="sec4-5-3-1">
<title>4.5.3.1 Practical application: Environmental parameters</title>
<para>A number of alarms can be configured for the sensor in <link linkend="F4-24">Figure <xref linkend="F4-24" remap="4.24"/></link>. In the use case of pultrusion described in <link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link> Section 2, two different alarms can be configured for each of the three different parameters measured by the device: minimum (low) and maximum (high) levels can be specified in the device.</para>
<fig id="F4-24" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-24">Figure <xref linkend="F4-24" remap="4.24"/></link></label>
<caption><para>ZED-THL-M ZigBee sensor for temperature, humidity and light.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_26.jpg"/>
</fig>
<para>The device is continuously verifying the voltage of the battery. If the level of mV falls under a certain threshold (where the device still runs but the situation starts to be dangerous), an alarm is raised.</para>
</section>
<section class="lev3" id="sec4-5-3-2">
<title>4.5.3.2 Practical application: Intelligent process performance indicator</title>
<para>The production process of shavers, described in <link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link> Section 1, consists of several physical manufacturing processes. Electrical, chemical and mechanical elements are working together in order to produce the products, making it a highly complex process where interactions between different signals can be easily overlooked when just monitoring every signal individually. A soft sensor that combines all different signals and processes them together will give better insight in these interaction effects via computational intelligence. This sensor fusion mechanism will deal with signals of disparate sources that do not have to originate from identical sensors.</para>
<para>The PCA algorithm in combination with the Hotelling&#x02019;s T<superscript>2</superscript> score is used to get insight in the interaction effect of all different process parameters. The process parameters are first preprocessed by the PLC-PMAC system, after which they are stored in a database to serve as input to train algorithm. To train the model, the data is extracted from this database and analyzed to make sure that the historical dataset consists of data that indicates only normal process behavior, without any deviations or outliers. This is an important step, since this data will serve as a reference for future predictions.</para>
<para>For real-time calculations the trained PCA model is deployed on a server and data from the PLC-PMAC system is fed into this model in order to obtain the new weighted scores that indicate how close new observations are related to the historical dataset. Because the PCA algorithm is a &#x02018;white box&#x02019; algorithm it can have self-diagnostics abilities. For example, the model can be instrumental to determine the root causes for fluctuations, trends and outliers.</para>
<para>This results in a single value to monitor if the machine is still operating within its stable operating window and triggers trends and outliers. Furthermore, the interaction effects between parameters can be taken into account in a multivariate manner.</para>
</section>
</section>
<section class="lev2" id="sec4-5-4">
<title>4.5.4 Smart Function: Formatting</title>
<para>Data formatting consists of using predefined data &#x0201C;shapes&#x0201D; or types, for exchange of information between the sensor and the central unit. The sensor data, usually in analogue format, has to be translated to the mentioned digital representation, and this conversion usually involves an appropriately chosen representation format that avoids transmitted data increase due to unnecessarily complicated protocols or data length in excess of what actual accuracy requires, like the number of bits/bytes used in the codification of the data sample.</para>
<section class="lev3" id="sec4-5-4-1">
<title>4.5.4.1 Practical applications: Compressed air system</title>
<para>The sensor in <link linkend="F4-18">Figure <xref linkend="F4-18" remap="4.18"/></link> on page 128 is used in the pultrusion use case (<link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link> Section 2) and it sends the collected data via a Modbus field bus protocol. In the present case, the data formatting can be either Float or int32.</para>
</section>
</section>
<section class="lev2" id="sec4-5-5">
<title>4.5.5 Smart Function: Enhancement</title>
<para>In addition to simply collecting data, it would be desirable that a sensor developed further data processing function upon the data. This functionality not only increases added value on the performance of the sensor but also relaxes the burden of overall data processing requirements on the central node(s). These are, in general, the units upon which relies the duty of calculation and decision, and an appropriate mathematical manipulation of the locally sensed quantities would shorten the reaction times upon data collection events and improve the overall sensor network performance. Various functions like averages, moving windows, transforms and statistical indicators fall in this category.</para>
<section class="lev3" id="sec4-5-5-1">
<title>4.5.5.1 Practical application: Air and water flow and temperature sensor</title>
<para><emphasis>Hysteresis or window function</emphasis>: When the hysteresis function of the sensor in <link linkend="F4-21">Figure <xref linkend="F4-21" remap="4.21"/></link> on page 131, used in the pultrusion use case of <link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link> Section 2, is set (<link linkend="F4-25">Figure <xref linkend="F4-25" remap="4.25"/></link>), the set point SP and the reset point rP are defined. The rP value must be lower than the SP value. The distance between SP and rP is at least 4% of the final value of the measuring range (= hysteresis). If only the set point is changed, the reset point is changed automatically, since in this way the difference remains constant.</para>
<fig id="F4-25" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-25">Figure <xref linkend="F4-25" remap="4.25"/></link></label>
<caption><para>Window function (Hysteresis) indications.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_27.jpg"/>
</fig>
<para>By setting the window function, it is possible to use the curve to compute proper values for the upper limit value FH and the lower limit value FL. The distance between FH and FL is at least 4% of the final value of the measuring range. FH and FL have a fixed hysteresis of 0.25% of the final value of the measuring range. This keeps the switching status of the output stable if the flow rate varies slightly.</para>
<para><emphasis>Damping of measured value</emphasis>: The damping time is the number of seconds that are waited before considering new values after there has been a change in the output values. In fact, whenever a signal gets out of the limits, there is a change in the outputs, the displayed values and the process value transfer via the IO Link interface. The damping time is able to stabilize these elements against flow values that change suddenly.</para>
</section>
<section class="lev3" id="sec4-5-5-2">
<title>4.5.5.2 Practical application: Railway strain sensor</title>
<para>The focus in the use case on railway maintenance (<link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link> Section 6) is on optical strain sensors used in train check points. Train &#x0201C;check points&#x0201D; are installations at certain locations in a railway network that monitor key parameters of passing trains to establish a number of safety relevant features of a train and to produce an automated warning in case safety critical limits are exceeded. These train parameters are derived from the strain introduced into the rail by the train&#x02019;s axle loads and by dynamic parameters, such as wheel out of roundness. This strain can be measured by fiber-Bragg-grating (FBG) sensors attached to the rail.</para>
<para>The core functionality of the FBG sensor is to measure strain applied to the sensor. The measured strain is derived from an optical resonant peak in a reflected light spectrum. The core measurement parameter is represented by the position of the resonance peak in the spectrum, measured in wavelength units.</para>
<para>Laser pulses are sent to the FBG by a device (Interrogator) via an optical fiber cable. The reflected light is then analyzed for the presence and location of the resonant peak in the spectrum. A number of FBGs, characterized by their individual resonant frequencies, can be &#x0201C;daisy chained&#x0201D; along one optical fiber cable for a finer measurement of the strain, by investigating for the presence of a resonant peak at different frequencies.</para>
</section>
<section class="lev3" id="sec4-5-5-3">
<title>4.5.5.3 Practical application: Conventional energy production</title>
<para>The monitoring of rolling element bearings, such as the ones described in <link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link> Section 8, is based on using vibration measurements made with accelerometers. Since measurements of this kind are made at high frequency i.e., the data is collected at e.g., 10 kHz frequency, a lot of data is collected, assuming the measurements are made continuously. Consequently, it is natural to try to do the signal analysis, diagnosis and prognosis close to the monitored machine in order to avoid the need to send enormous amounts of data to a central processing unit. In the Mantis project these analysis functions are processed in a Raspberry processor which is located close to the accelerometer.</para>
<para>The commonly known envelope analysis [Randall, 2011] is used for signal analysis for the detection of possible bearing faults. With envelope analysis the indication of upcoming failures can be detected at an early stage, as well as the fault type i.e., outer or inner race etc. can be detected. The used band pass frequencies are tuned based on the geometry of the bearing.</para>
<para>The diagnosis of a bearing fault is based on comparison of the amplitudes of the envelope spectrum at the bearing fault frequencies to the amplitude levels at neighboring frequencies. The used mathematical formulas are simple enough that the functions can easily be handled at local level (Raspberry) and naturally, if needed, as Web services and at a service center.</para>
</section>
</section>
<section class="lev2" id="sec4-5-6">
<title>4.5.6 Smart Function: Transformation</title>
<para>In some cases, the estimation of a magnitude is calculated in terms of manipulation of acquired data that refers to a different type of measurement. Data has thus to be transformed from one domain to the other via models, estimators and other mathematical tools. The transformation could range from a simple proportionality factor up to a complex signal processing algorithm involving Fourier transforms, statistical models, Kalman filters etc. From a different point of view, transformation can also be considered as the process by which commands or data in a given protocol have to be translated to a different one. An example of this can consist on a group of neighboring sensors, interconnected via some type of field bus (e.g., Modbus). These sensors can then be connected to a second type of network, such as 802.3 based Industrial Ethernet. There is a clear need for a &#x0201C;gateway&#x0201D; that transforms commands and information from one realm to the other and vice-versa.</para>
<section class="lev3" id="sec4-5-6-1">
<title>4.5.6.1 Practical application: Pressure drop estimation</title>
<fig id="F4-26" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-26">Figure <xref linkend="F4-26" remap="4.26"/></link></label>
<caption><para>Parametric model used to estimate the relationship between the air velocity and the pressure drop.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_28.jpg"/>
</fig>
<para>The aim of the present function is to determine the relationship between the pressure drop and the flow velocity with a data driven approach and to track the change of this relationship overtime in order to follow the degradation of the filter. An ARX parametric model (see <link linkend="F4-26">Figure <xref linkend="F4-26" remap="4.26"/></link>) was used to model the relationship between the input (air velocity) and output (pressure drop). A recursive least square approach with forgetting factor was used to estimate the model parameters of the system.</para>
<para>The Kalman filter was used to recursively estimate the model parameters whenever new data was acquired. The principle of the Kalman filter is that the estimation of new parameters <inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg139-2.jpg"/> at time <emphasis role="strong"><emphasis>k+1</emphasis></emphasis> depends on the predicted error and on the previous parameter estimation <inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg139-2.jpg"/> at time <emphasis role="strong"><emphasis>k</emphasis></emphasis>.</para>
<para>Considering equation <emphasis>y</emphasis>= Z&#x1D717;+e, when new data arrived in form of input and output, the new problem consisted in estimating the new parameters <emphasis role="strong"><emphasis>w</emphasis></emphasis> based on the old data [<emphasis role="strong">y Z</emphasis>] and the newly gathered data [<emphasis>&#x01EF9; <inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg139-4.jpg"/></emphasis>]. The problem was solved by minimising the 2-norm of the following equation:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq4-5.jpg"/></para>
<para>The formula can be computed by using the Woodbury-Sherman-Morrison equation (Sherman and Morrison 1950) and after some simplifications, the formula can be described as follows:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq4-6.jpg"/></para>
<para>where <inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg139-1.jpg"/> is the error of the prediction and <emphasis role="bolditalic">G</emphasis> is the gain:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq4-7.jpg"/></para>
<para>By estimating the model parameters, the relationship between the sensor and the data is learned and the deviation between the estimated model and the actual measurement can be calculated by using the model residuals:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq4-8.jpg"/></para>
<para>The model obtained can be therefore used to recalculate the pressure drop at a reference air flow (50% of the flow range), allowing tracking the degradation of the component, or for anomaly detection.</para>
</section>
</section>
<section class="lev2" id="sec4-5-7">
<title>4.5.7 Smart Function: Fusion</title>
<para>Health monitoring strategies usually require that more than a single magnitude be measured by a single or multiple sensors. In case a sensor is devoted to measuring different quantities, a data fusion process could be pertinent. In this case, several data from various different measured magnitudes are sent together to the central node(s) in a single transmission. Plus, if the aforementioned measurements are of different nature (analog and digital), an appropriate data formatting could also be necessary.</para>
<section class="lev3" id="sec4-5-7-1">
<title>4.5.7.1 Practical application: Off-road and special purpose vehicle</title>
<para>The purpose of the practical case described in <link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link> Section 5 is to perform a fusion between data obtained from an inertial measurement unit (IMU) and an ultra-wideband localization device (UWBL), for indoor location services. The proposed algorithm is based on the IMU and UWBL measurements. This sensor fusion efficiently combines the advantages of both methods. Accelerometer and gyroscope measurements are useful to track the inertial move of the vehicle with a really high refresh rate, but suffer from the measurement error and perturbations, such as drift. UWB measurements provide global positions in the navigation environment, and are accurate enough to determine the IMU measurement drift. The sensor fusion can be solved using extended Kalman filter principles. An initial estimation on the position and direction is calculated. Then, when new data are available, comparison of estimated and actual position and direction values gives correction factors that are further utilized by the filter to improve the estimation quality for the following iterations. Refer to <link linkend="F4-27">Figure <xref linkend="F4-27" remap="4.27"/></link> for a graphical representation of the process.</para>
<fig id="F4-27" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-27">Figure <xref linkend="F4-27" remap="4.27"/></link></label>
<caption><para>Kalman filter based process for sensor data fusion.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_29.jpg"/>
</fig>
<fig id="F4-28" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-28">Figure <xref linkend="F4-28" remap="4.28"/></link></label>
<caption><para>e-Alert controller to monitor environmental conditions in a medical device.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_30.jpg"/>
</fig>
</section>
<section class="lev3" id="sec4-5-7-2">
<title>4.5.7.2 Practical application: MR magnet monitoring (e-Alert sensor)</title>
<para>The e-Alert sensor is described in <link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link> Section 9 and in [Albano, 2018], and it is a stand-alone sensor, which can autonomously, 24/7, monitor environmental conditions such as temperatures, humidity, magnet field, mains power, in the vicinity of a Philips MRI system. The e-Alert controller is wall-mounted in the technical room of the MRI system and its sensors are physically connected to those parts of the MRI system where environmental conditions are measured. The e-Alert sensor (<link linkend="F4-28">Figure <xref linkend="F4-28" remap="4.28"/></link>) is based on a Raspberry Pi mini-computer. The embedded software is developed specifically for this purpose, and is built on GPLed libraries and APIs. The sample values and logs are stored on an internal SD-card.</para>
<para>The temperature sensors are off-the-shelf one-wire sensors. These sensors are connected to an interface box (max 8 sensors per interface box). The interface box is connected to one of the inputs of the e-Alert sensor. Multiple interface boxes can be daisy-chained. This provides a scalable sensor platform that can be tailored for the specific device under monitoring.</para>
<para>As represented in <link linkend="F4-29">Figure <xref linkend="F4-29" remap="4.29"/></link>, the e-Alert sensor acquires sensor values once per minute and checks these values against configured control limits. To avoid false positives, a sensor value must exceed the control limits for a number of consecutive samples before an alert is sent. In that case, the e-Alert sensor sends an E-Mail or text message to the configured alert receivers.</para>
<fig id="F4-29" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F4-29">Figure <xref linkend="F4-29" remap="4.29"/></link></label>
<caption><para>e-Alert control sensor context diagram.</para></caption>
<graphic xlink:href="graphics/ch04_Fig4_31.jpg"/>
</fig>
</section>
<section class="lev3" id="sec4-5-7-3">
<title>4.5.7.3 Practical application: MR critical components</title>
<para>The performance of critical components in a medical device is monitored, as described in <link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link> Section 9. Intelligent components were developed, able to record their own state in real-time and provide access to this data via a software interface. The next generation high-power amplifiers will offer capabilities that enable data-driven diagnostics. Recorded data can be summarized as follows:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Identification &#x02013; critical parts of the amplifier contain identification data such as serial numbers, firmware versions. This data can be used to track component and firmware changes;</para></listitem>
<listitem>
<para>Firmware upgrade &#x02013; amplifier firmware can be upgraded via a remote connection. This capability can be used to upgrade amplifiers in the field, without physical presence of a field service engineer;</para></listitem>
<listitem>
<para>Monitoring &#x02013; physical characteristics (e.g., temperatures, voltages, and currents) are measured periodically and stored in memory as time-series data. This capability can be used to define the exact conditions of the amplifier;</para></listitem>
<listitem>
<para>State logging &#x02013; the state of the internal state machine of the amplifiers&#x02019; firmware is stored in memory as time-series data. These data can be used to reproduce the exact conditions under which the amplifier goes into an error state;</para></listitem>
<listitem>
<para>Clock synchronization &#x02013; the internal clock of the amplifier is sync&#x02019;d periodically with the clock of the medical device.</para></listitem>
</itemizedlist>
<para>The medical device periodically retrieves the data that is stored in the amplifier&#x02019;s memory and combines these data with data from other components in the medical device, to reconstruct the operational conditions of the amplifier in the medical device.</para>
</section>
</section>
</section>
<section class="lev1" id="sec4-6">
<title>References</title>
<orderedlist numeration="arabic" continuation="restarts" spacing="normal">
<listitem><para>Albano, M., Ferreira, L.L., Di Orio, G., Mal&#x00F3;, P., Webers, G., Jantunen, E., Gabilondo, I., Viguera, M., Papa, G. and Novak, F. (2018) Sensors: the Enablers for Proactive Maintenance in the Real World. In <emphasis>2018 5th International Conference on Control, Decision and Information Technologies (CoDIT)</emphasis> (pp. 569&#x2013;574). IEEE. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Albano%2C+M%2E%2C+Ferreira%2C+L%2EL%2E%2C+Di+Orio%2C+G%2E%2C+Mal%F3%2C+P%2E%2C+Webers%2C+G%2E%2C+Jantunen%2C+E%2E%2C+Gabilondo%2C+I%2E%2C+Viguera%2C+M%2E%2C+Papa%2C+G%2E+and+Novak%2C+F%2E+%282018%29+Sensors%3A+the+Enablers+for+Proactive+Maintenance+in+the+Real+World%2E+In+2018+5th+International+Conference+on+Control%2C+Decision+and+Information+Technologies+%28CoDIT%29+%28pp%2E+569-574%29%2E+IEEE%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Angell J. B., Terry S. C. and Barth P. W. (1983) &#x2018;Silicon micromechanical devices,&#x2019; <emphasis>Scientific American</emphasis>, 248, pp. 44&#x2013;55. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Angell+J%2E+B%2E%2C+Terry+S%2E+C%2E+and+Barth+P%2E+W%2E+%281983%29+%27Silicon+micromechanical+devices%2C%27+Scientific+American%2C+248%2C+pp%2E+44-55%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Balanis C. (1982) <emphasis>Antenna Theory: Analysis and Design</emphasis>, Wiley. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Balanis+C%2E+%281982%29+Antenna+Theory%3A+Analysis+and+Design%2C+Wiley%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Beigl M., Krohn A., Zimmer T. and Decker C. (2004) &#x2018;Typical sensors needed in ubiquitous and pervasive computing,&#x2019; In <emphasis>Proceedings of the First International Workshop on Networked Sensing Systems (INSS &#x2019;04)</emphasis>. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Beigl+M%2E%2C+Krohn+A%2E%2C+Zimmer+T%2E+and+Decker+C%2E+%282004%29+%27Typical+sensors+needed+in+ubiquitous+and+pervasive+computing%2C%27+In+Proceedings+of+the+First+International+Workshop+on+Networked+Sensing+Systems+%28INSS+%2704%29%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Bui V., Verhoeven R. and Lukkien J. (2014) &#x2018;Evaluating trustworthiness through monitoring: The foot, the horse and the elephant,&#x2019; In <emphasis>International Conference on Trust and Trustworthy Computing</emphasis>, Springer International Publishing, pp. 188&#x2013;205. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Bui+V%2E%2C+Verhoeven+R%2E+and+Lukkien+J%2E+%282014%29+%27Evaluating+trustworthiness+through+monitoring%3A+The+foot%2C+the+horse+and+the+elephant%2C%27+In+International+Conference+on+Trust+and+Trustworthy+Computing%2C+Springer+International+Publishing%2C+pp%2E+188-205%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Chansin G. (2014) <emphasis>Printed and Flexible Sensors 2014&#x2013;2024: Technologies, Players, Forecasts</emphasis>, Cambridge: IDTechEx Ltd.</para></listitem>
<listitem><para>El-Thalji I. and Jantunen E. (2015) &#x2018;Fault analysis of the wear fault-development in rolling bearings&#x2019;, <emphasis>Engineering Failure Analysis</emphasis> 57, pp. 470&#x2013;482. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=El-Thalji+I%2E+and+Jantunen+E%2E+%282015%29+%27Fault+analysis+of+the+wear+fault-development+in+rolling+bearings%27%2C+Engineering+Failure+Analysis+57%2C+pp%2E+470-482%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Environetix. www.environetix.com</para></listitem>
<listitem><para>G&#x00FC;ven&#x2013;Ozcelebi C., Holenderski M. J. and Lukkien J. J. (2017) &#x2018;An application of trust in predictive maintenance,&#x2019; <emphasis>13th Conference of Telecommunication, Media and Internet Techno-Economics (CTTE 2017)</emphasis>.</para></listitem>
<listitem><para>Haykin S. (1994) <emphasis>Digital Communication Systems</emphasis>, Wiley.</para></listitem>
<listitem><para>Jantunen, E., Gorostegi, U., Zurutuza, U., Larringa, F., Albano, M., di Orio, G., Mal&#x00F3;, P., and Heged&#x00FC;s, C. (2017). The way cyber physical systems will revolutionise maintenance. In <emphasis>30th Conference on Condition Monitoring and Diagnostic Engineering Management (COMADEM)</emphasis>, July 10&#x2013;13, 2017, Preston, UK. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Jantunen%2C+E%2E%2C+Gorostegi%2C+U%2E%2C+Zurutuza%2C+U%2E%2C+Larringa%2C+F%2E%2C+Albano%2C+M%2E%2C+di+Orio%2C+G%2E%2C+Mal%F3%2C+P%2E%2C+and+Heged%FCs%2C+C%2E+%282017%29%2E+The+way+cyber+physical+systems+will+revolutionise+maintenance%2E+In+30th+Conference+on+Condition+Monitoring+and+Diagnostic+Engineering+Management+%28COMADEM%29%2C+July+10-13%2C+2017%2C+Preston%2C+UK%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>J&#x00F8;sang A. (2016) <emphasis>Subjective Logic: A Formalism for Reasoning Under Uncertainty</emphasis>, Springer. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=J%5Bx00F8%3B%5Dsang+A%2E+%282016%29+Subjective+Logic%3A+A+Formalism+for+Reasoning+Under+Uncertainty%2C+Springer%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Junnola, J. &#x201C;The suitability of low-cost measurement systems for rolling element bearing vibration monitoring&#x2019;&#x201D;, Master thesis, University of Oulu, available online at: http://jultika.oulu.fi/Record/nbnfioulu-201705041660</para></listitem>
<listitem><para>Krauss J. and Marhefka R. (2002) <emphasis>Antennas for All Applications</emphasis>, McGraw Hill. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Krauss+J%2E+and+Marhefka+R%2E+%282002%29+Antennas+for+All+Applications%2C+McGraw+Hill%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Mehdi G., Naderi D., Ceschini G., and Mikhail Roshchin S. (2015) &#x2018;Model-based reasoning approach for automated failure analysis: An industrial gas turbine application [Siemens],&#x2019; <emphasis>Annual Conference of the Prognostics and Health Management Society</emphasis>. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Mehdi+G%2E%2C+Naderi+D%2E%2C+Ceschini+G%2E%2C+and+Mikhail+Roshchin+S%2E+%282015%29+%27Model-based+reasoning+approach+for+automated+failure+analysis%3A+An+industrial+gas+turbine+application+%5BSiemens%5D%2C%27+Annual+Conference+of+the+Prognostics+and+Health+Management+Society%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>NutaQ ZeptoSDR</para></listitem>
<listitem><para>Oppenheim A. and Willsky A. (1997) <emphasis>Signals and Systems</emphasis>, Pearson Prentice Hall.</para></listitem>
<listitem><para>Proakis J. (1995) <emphasis>Digital Communications</emphasis>, McGraw Hill. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Proakis+J%2E+%281995%29+Digital+Communications%2C+McGraw+Hill%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Proakis J. G., and Manolakis D. G. (2007) <emphasis>Digital Signal Processing: Principles, Algorithms and Applications</emphasis>, Pearson Prentice Hall.</para></listitem>
<listitem><para>Randall R. (2011) <emphasis>Vibration-based Condition Monitoring</emphasis>, Wiley. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Randall+R%2E+%282011%29+Vibration-based+Condition+Monitoring%2C+Wiley%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Ries S. (2009) Trust in ubiquitous computing (Doctoral dissertation, Technische Universit&#x00E4;t). <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ries+S%2E+%282009%29+Trust+in+ubiquitous+computing+%28Doctoral+dissertation%2C+Technische+Universit%E4t%29%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>SAW Components</para></listitem>
<listitem><para>Stutzman W. and Thiele G. (2013) <emphasis>Antenna Theory and Design</emphasis>, Wiley. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Stutzman+W%2E+and+Thiele+G%2E+%282013%29+Antenna+Theory+and+Design%2C+Wiley%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem><para>Syntonics</para></listitem>
<listitem><para>Wisen-tech. www.wisen-tech.com</para></listitem>
</orderedlist>
</section>
</chapter>
<chapter class="chapter" id="ch05" label="5" xreflabel="5">
<title>Providing Proactiveness: Data Analysis Techniques Portfolios</title>
<para><emphasis role="strong">Alberto Sillitti<superscript>1</superscript>, Javier Fernandez Anakabe<superscript>2</superscript>, Jon Basurko<superscript>3</superscript>, Paulien Dam<superscript>4</superscript>, Hugo Ferreira<superscript>5</superscript>, Susana Ferreiro<superscript>6</superscript>, Jeroen Gijsbers<superscript>7</superscript>, Sheng He<superscript>8</superscript>, Csaba Heged&#x00171;s<superscript>9</superscript>, Mike Holenderski<superscript>10</superscript>, Jan-Otto Hooghoudt<superscript>11</superscript>, I&#x000F1;igo Lecuona<superscript>2</superscript>, Urko Leturiondo<superscript>3</superscript>, Quinten Marcelis<superscript>12</superscript>, Istv&#x000E1;n Moldov&#x000E1;n<superscript>13</superscript>, Emmanuel Okafor<superscript>8</superscript>, Cl&#x000E1;udio Rebelo de S&#x000E1;<superscript>5</superscript>, Ricardo Romero<superscript>6</superscript>, Babacar Sarr<superscript>14</superscript>, Lambert Schomaker<superscript>8</superscript>, Arvind Kumar Shekar<superscript>15</superscript>, Carlos Soares<superscript>5</superscript>, Hans Sprong<superscript>7</superscript>, S&#x000F8;ren Theodorsen<superscript>12</superscript>, Tom Tourw&#x000E9;<superscript>16</superscript>, Gorka Urchegui<superscript>17</superscript>, Godfried Webers<superscript>7</superscript>, Yi Yang<superscript>11</superscript>, Andriy Zubaliy<superscript>16</superscript>, Ekhi Zugasti<superscript>2</superscript>, and Urko Zurutuza<superscript>2</superscript></emphasis></para>
<para><superscript>1</superscript>Innopolis University, Russian Federation<?lb?><superscript>2</superscript>Mondragon University, Arrasate-Mondrag&#x000F3;n, Spain<?lb?><superscript>3</superscript>IK4-Ikerlan, Arrasate-Mondrag&#x000F3;n, Spain<?lb?><superscript>4</superscript>Philips Consumer Lifestyle B.V., The Netherlands<?lb?><superscript>5</superscript>Instituto de Engenharia de Sistemas e Computadores do Porto, Portugal<?lb?><superscript>6</superscript>Tekniker, Spain<?lb?><superscript>7</superscript>Philips Medical Systems Nederland B.V., The Netherlands<?lb?><superscript>8</superscript>University of Groningen, The Netherlands<?lb?><superscript>9</superscript>AITIA International Inc., Hungary<?lb?><superscript>10</superscript>Technische Universiteit Eindhoven, The Netherlands<?lb?><superscript>11</superscript>Aalborg University, Denmark<?lb?><superscript>12</superscript>Ilias Solutions, Belgium<?lb?><superscript>13</superscript>Budapest University of Technology and Economics, Hungary<?lb?><superscript>14</superscript>3e, Belgium<?lb?><superscript>15</superscript>Bosch, Germany<?lb?><superscript>16</superscript>Sirris, Belgium<?lb?><superscript>17</superscript>Mondragon Sistemas De Informacion, Spain</para>
<section class="lev1" id="sec5-1">
<title>5.1 Introduction</title>
<para>Data analysis is of paramount importance in the management of proactive maintenance. This chapter provides a deep analysis of the different techniques that can be adopted when dealing with the automation of maintenance processes. In particular, it focuses on three aspects of the maintenance that we have considered as the cornerstones of PM:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Root cause analysis:</emphasis> aims to identify the causes of failures that have occurred in the past and avoid their appearance in the future. Basically, it provides a set of approaches intended at building a knowledgebase that relies on past experience to identify the core causes behind a problem in the system under investigation. This kind of analysis is useful when performing post-mortem analysis of the failures that have been reported and learn from them;</para></listitem>
<listitem>
<para><emphasis role="strong">Identification of the remaining useful life:</emphasis> aims to estimate the operational life of a component to support different activities including;</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>The proper design of a system based on the operational constraints and the expectations of the users to guarantee its usability;</para></listitem>
<listitem>
<para>The definition of a proper plan of maintenance activities to avoid unexpected down times.</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para><emphasis role="strong">Alerting and predicting of failures:</emphasis> aims to identify possible failures before they actually happen and/or provide an alert about a set of still acceptable conditions that are unusual and that may result in a failure of the system if they are not managed properly in a timely fashion.</para></listitem>
</itemizedlist>
<para>Besides these three main aspects, there are a number of additional ones that could be considered to create a comprehensive proactive maintenance environment. Since modern systems are very complex and taking decisions about maintenance requires taking into account many aspects, the decision-making approach needs the participation of multiple stakeholders and the implementation of collaborative decision-making strategies.</para>
<para>In a collaborative process, entities share information, resources and responsibilities, risks and rewards to jointly plan, implement, and evaluate a program of activities to achieve a common goal. Collaboration usually involves mutual engagement of participants to solve a problem together, which implies mutual trust. Coordination, that is, the act of working together harmoniously, is one of the main components of collaboration (<link linkend="F5-1">Figure <xref linkend="F5-1" remap="5.1"/></link>). In MANTIS, the common goal is the maintenance optimization of assets and the different systems and stakeholders that take part in maintenance tasks, will have to share information, resources and responsibilities. From the process automation point of view, a key aspect of collaborative automation includes a single, unified environment for the presentation of information to the operator as well as the ability to present information in context to the right people, at the right time, from any access point within the system. This implies that the Proactive Maintenance Service Platform Architecture of MANTIS requires to collect data from all equipment regardless of type or age, store and organize the collected data over time, perform advanced analytics and provide decision support. Thus, MANTIS pursues a collaborative maintenance ecosystem.</para>
<fig id="F5-1" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-1">Figure <xref linkend="F5-1" remap="5.1"/></link></label>
<caption><para>The collaboration concept.</para></caption>
<graphic xlink:href="graphics/ch05_fig001.jpg"/>
</fig>
<para>Maintenance management can nowadays take advantage of many information sources for faster and more accurate prognosis. Data can be extracted and stored from machines and components. Machine based data analysis may be used locally to optimize the machine&#x02019;s maintenance. Better results may be expected if these particular data are combined with other data sources such as that from the MES, ERP, cloud-based open-data sources and the machine operators.</para>
<para>Proactive and collaborative decision support is an integral part of a proactive maintenance strategy. Condition monitoring is only one part of the equation, since the actual assets&#x02019; condition needs to be combined with other relevant information, such as customer needs, ambient conditions, business models and service contracts in order for a service team to make the right decision and take the appropriate actions. The goal of MANTIS is to provide the service team with an intuitive, proactive and context-aware system for industrial maintenance that supports maintenance by proactively pushing relevant information to the right people at the right time, by intelligently filtering and summarizing information to prevent information overload through context awareness, by automatically and dynamically scheduling and adapting maintenance plans, thereby keeping the human in the loop at all times. The chapter is organized as follows:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Section 5.2 focuses on root cause analysis providing a short introduction to the theoretical background and a catalogue of techniques that have been demonstrated to be useful in this kind of analysis. Such techniques are mainly based on statistical and machine learning approaches. Moreover, a set of real-world applications are shortly introduced;</para></listitem>
<listitem>
<para>Section 5.3 deals with the identification of the remaining useful life of components. The section provides a short theoretical background, a catalogue of the useful approaches, and an analysis of some use cases to demonstrate the applicability of the described techniques analysing different modelling approaches;</para></listitem>
<listitem>
<para>Section 5.4 investigates how to alert and predict failures. The section provides an extensive catalogue of useful techniques. The presented techniques are also mainly based on statistical and machine learning approaches;</para></listitem>
<listitem>
<para>Section 5.5 provides some real examples that come from different application domains where most of the techniques previously presented have been applied with valuable results.</para></listitem>
</itemizedlist>
</section>
<section class="lev1" id="sec5-2">
<title>5.2 Root Cause Failure Analysis</title>
<para>RCA is a methodology to identify the primary cause of a failure in a system. RCA is the function that makes PM possible, detecting and correcting root conditions that would otherwise lead to failure. Once the root cause is identified, corrective action can be taken to make sure that the issue does not re-occur.</para>
<section class="lev2" id="sec5-2-1">
<title>5.2.1 Theoretical Background</title>
<para>Data-driven RCA uses data and data analysis techniques to identify root causes, based on the observable states of the system. These observable states may be directly or indirectly related to identifiable components in the system. Based on the design of the system, it is necessary to define the relevant features in the available data and to identify the components that are most likely involved in the failure. The latter may require, for example, expert domain knowledge, historical maintenance data and historical performance data. If data analysis reveals that a single component caused the (majority of) issues, then the RCA is complete; otherwise the RCA is re-iterated with an updated set of features and/or components.</para>
<para>The CRISP-DM defines a methodology for structured data mining. For RCA, the following phases are used: business understanding, data understanding, data preparation, modelling.</para>
<section class="lev3" id="sec5-2-1-1">
<title>Business understanding</title>
<para>Due to the size/complexity of systems and the amount of available data, it is necessary to narrow down the scope of RCA data analysis a-priori. Expert domain knowledge is required to scope the RCA in terms of machines, failure modes, parts and data sources. This is an iterative approach; insights in the data helps the expert to refine the scope. When the system is complex, expert knowledge may not be adequate to identify a-priori the relevant system components and data features. Probabilistic graphical models could, for example, be used to identify the most probable components and features.</para>
</section>
<section class="lev3" id="sec5-2-1-2">
<title>Data understanding</title>
<para>There is a wide variety of visualization and statistical techniques to provide an overview of the data. Graphs and statistical characteristics are instrumental in understanding the data (-distributions) and to identify the relevant features in the data in the next analysis phases. Examples of data visualization are PCA and t-Distributed Stochastic Neighbour Embedding. These techniques can provide insights of the distribution and classification of data in the multi-dimensional feature space. Examples of statistical techniques are forward/backward selection, trend extraction, and classification. The purpose of these techniques is to reduce the multi-dimensional feature space to a manageable number of features, which represent the system adequately for RCA purposes.</para>
<para>For specific RCA cases, the influence of features evolves over time while the system evolves. This adds time as an additional dimension. There are statistical techniques that include implicitly time as dimension, but this raises the risk that the statistics get biased by the more recent data.</para>
</section>
<section class="lev3" id="sec5-2-1-3">
<title>Data preparation</title>
<para>The collected data usually originate from a variety of data sources. These can comprise machine data, e.g., sensor data obtained by sampling in real-time, or machine logs that represent the internal condition of the system. Service maintenance records are another source of data. These data contain, for example, specifics of parts replaced (quantity, quality). Each data source may require specific pre-processing and transformation techniques before it can be used by data analysis algorithms. Examples include aligning timestamp formats, correcting for missing data and erroneous data and the aggregation of noisy data.</para>
</section>
</section>
<section class="lev2" id="sec5-2-2a">
<title>Modelling</title>
<para>Once the data features are identified, it is possible to develop algorithms (models) that identify the specific components -which are the most probable root cause of a specific issue - and the failure modes. The input of the algorithms are the data, which are prepared specifically for each algorithm. The output of the algorithm is a statistical measure that can be used to identify the specific component as the root cause.</para>
</section>
<section class="lev2" id="sec5-2-2">
<title>5.2.2 Techniques Catalogue</title>
<para>There is a wide variety of modelling techniques available. <link linkend="T5-1">Table <xref linkend="T5-1" remap="5.1"/></link> shows an excerpt of techniques.</para>
<table-wrap position="float" id="T5-1">
<label><link linkend="T5-1">Table <xref linkend="T5-1" remap="5.1"/></link></label>
<caption><para>Root Cause Analysis Techniques</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left">Modelling technique</td>
 <td>Algorithm</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Classification</td>
 <td>Random forest classification, Naive Bayes, Support Vector Machine (SVM), Limit and trend checking</td>
</tr>
<tr>
<td valign="top" align="left">Regression</td>
 <td>Partial least square regression (PLS), Random forest regression, Least absolute deviations regression, Logistic regression, Bayesian network (BN)</td>
</tr>
<tr>
<td valign="top" align="left">Neural network</td>
 <td>Artificial Neural network, recurrent neural network, convolutional neural network</td>
</tr>
<tr>
<td valign="top" align="left">Unsupervised learning</td>
 <td>Hierarchical clustering, K-means clustering, attribute-oriented induction</td>
</tr>
<tr>
<td valign="top" align="left">Pattern analysis</td>
 <td>Hidden Markov model (HMM), expectation maximization,</td></tr>
</tbody>
</table>
</table-wrap>
<fig id="F5-2" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-2">Figure <xref linkend="F5-2" remap="5.2"/></link></label>
<caption><para>Overview for RCA.</para></caption>
<graphic xlink:href="graphics/ch05_fig002.jpg"/>
</fig>
<para>Based on research of practical systems, a selected set of algorithms is described in the next sections. For each algorithm, the main aspects are outlined, and some practical examples are given. Some of the described algorithms are used for RCA only; other algorithms are used for root cause analysis and for failure prediction (<link linkend="F5-2">Figure <xref linkend="F5-2" remap="5.2"/></link>). Some of the examples illustrate how to embed the algorithm in the system, while other examples illustrate how to the use the algorithm outside the system.</para>
<section class="lev3" id="sec5-2-2-1">
<title>5.2.2.1 Support vector machine</title>
<para>A SVM is a supervisory learning algorithm that can be used for binary classification. It uses feature data as input and it calculates the most probable class. This algorithm is trained by providing feature data that is already classified, i.e., labelling. The algorithm then constructs a hyper plane or a set of hyper planes based on linear combination of features in the multi-dimensional feature space and optimizes the distance between the feature values and the hyper plane(s). For some problems, it may be required to first linearize the feature values by a (non-linear) transformation. Once the hyper plane(s) have been constructed, the algorithm is applied to a test dataset. The percentage of correctly classified cases is a measure for the performance of the algorithm.</para>
<para>SVM is capable to handle high-dimensional feature spaces adequately and is insensitive to outliers in feature values. Domain expertise can be embedded in the transformation of feature values. Furthermore, SVM has proven its use in numerous application domains, such as text mining, face recognition, and image processing.</para>
</section>
<section class="lev3" id="sec5-2-2-2">
<title>5.2.2.2 Limit and trend checking</title>
<para>Limit and trend checking are a classical algorithm where (derived) feature values are compared against limits (one-sided or two-sided). Feature values that are within the limits represent normal behaviour of the system, feature values that are outside the limits can be caused by defective component(s). Limit and trend checking are commonly integrated in process automation systems. These algorithms are in many cases sufficient to prevent larger failure or damage. However, faults are detected rather late and a detailed component diagnosis is mostly not possible. The limits can be based on design and/or historical data and this represents, in general, a trade-off between early detection (narrow limits) and inherent statistical variation of the system (wide limits). Limits may be fixed values, may vary in time or may vary with other feature values, depending on the characteristics of (components of) the system.</para>
<section class="lev4">
<title>Example: press machine</title>
<para>In a press machine, the forces in the ram (which consist of 2 rods) are measured during the stamp. A misalignment (eccentricity) in the rod forces creates undesired loads in the press and may lead to premature wear and future malfunction of the press. Piezoelectric sensors measure the force in the press rods. During each stroke, the eccentricity is calculated and compared against limits. When the eccentricity is outside limits, a maintenance action is triggered by the press.</para>
</section>
<section class="lev4">
<title>Example: photovoltaic plant</title>
<para>For a photovoltaic plant, the performance loss ratios are determined by each energy conversion component. These ratios are aggregated for a given time span, e.g., one day up to several months. These ratios are compared against configured limits. The limits are derived from the technical specifications of the components in the system and from field measurements of &#x0201C;known good&#x0201D; photovoltaic plants. Furthermore, the limits are corrected for weather conditions in the given time span. When the performance is outside the limits, a maintenance action is triggered.</para>
</section>
<section class="lev4">
<title>Example: health equipment</title>
<para>In a health equipment, the ambient conditions are monitored to make sure that the system operates within its operational limits. Sensors continuously measure ambient temperature, humidity and cooling water temperatures. The health equipment compares these values against configured limits. When the ambient conditions are outside the limits, a maintenance action is triggered by the health equipment. The (wide) limits are based on the design of the health equipment. The limits can be localized (narrowed) using the configuration capabilities on the health equipment, if local environment operational conditions justify this.</para>
</section>
</section>
<section class="lev3" id="sec5-2-2-3">
<title>5.2.2.3 Partial least squares regression</title>
<para>PLS is an algorithm that constructs a hyper plane or a set of hyper planes based on linear combination of features in the multi-dimensional feature space and minimizes the distance between the feature values and the hyper plane(s), based on mathematical projections to a lower-dimensional feature space. The weight of a feature in the linear fit is a measure of the relative impact of the component, associated to that feature. The higher the weight, the higher the detection capability of a defective component.</para>
<para>The upside of this algorithm is that it can handle a large (more than 10) number of features, it can handle correlated features effectively, and it can handle missing data points. The downside is that the output of the algorithm may be difficult to interpret, for example confusing causality with co-linearity.</para>
<para>PLS was initially developed for econometrics and chemo-metrics. Since then, it has been applied in numerous domains, such as education, manufacturing, marketing and social sciences.</para>
<section class="lev4">
<title>Example: shaver production plant</title>
<para>In a shaver production plan, the product quality is measured and is expressed as 4 geometry features (Y). The production line quality is measured and expressed as 12 sensors features (X). Product features are measured on a sample-basis; production line features are measured continuously. A unique product identifier is used to join the product features and the production line features, furthermore, short terms sensor failures are removed from the production line features. The dataset was fed into the PLS algorithm where production line features are the independent variables (X) and the product quality features are the dependent variables (Y). This provided the weighting factors and the root mean square error. It was concluded that the root mean square error is acceptable for each of the product features, given the acceptable bandwidth of product quality. This enabled the shaver production plant to implement the algorithm in the production line to monitor product quality in real-time.</para>
</section>
</section>
<section class="lev3" id="sec5-2-2-4">
<title>5.2.2.4 Bayesian network</title>
<para>A BN is a probabilistic directed acyclic graph of nodes where the nodes represent observable stochastic variables of a system. Such graphs can be used to calculate the probability distribution of specific conditions or states, where multiple nodes interact. More specifically in the case of RCA, nodes can represent machine components and the node state can represent the failure modes and non-failure states of the corresponding components [3]. One can therefore calculate the probability of a state (for example failure or no failure) of a component based on the states of the other dependent nodes (marginal probability). The process of building a BN graph uses expert domain knowledge to identify the primary components and their interactions. In general, many components work independently and their failure does not propagate to all other components. This simplifies the BN and allows for computation that is more efficient.</para>
<para><emphasis>Common Applications:</emphasis></para>
<para>Because the BNs are a general probabilistic inference method, it can be applied to a wide range of problems where objects can be assigned a probability of being in a state. Examples of application include:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Medicine (diagnosing disease or illness based on symptoms);</para></listitem>
<listitem>
<para>Document classification (classifying a document belonging to a given subject based on the word content);</para></listitem>
<listitem>
<para>Image processing (for example assigning a pixel to a region based on its colour components and the region membership of its surrounding pixels);</para></listitem>
<listitem>
<para>Spam filters (determining if a message is spam or not);</para></listitem>
<listitem>
<para>Biology (inferring the network), etc.</para></listitem>
</itemizedlist>
<para><emphasis>Strengths and challenges</emphasis>:</para>
<para>The BN model is a very good fit for the RCA problem. It is only necessary to identify the components and their failure modes. The components are then mapped to the nodes and the states of each node are used to hold the conditional probabilities. The problem with constructing this model however is twofold. The first is determining the state dependency between each pair of components (establishing the network) and the second is acquiring enough failure data for each and every component (conditional probabilities). This requires good domain knowledge and much effort. Since it was used a very small model (modelled only fail or non-fail states), calculating the marginal probabilities could be done efficiently.</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Advantages:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Intuitive model construction;</para></listitem>
<listitem>
<para>Allows for on-line model updates;</para></listitem>
<listitem>
<para>General model applicable to many problems.</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>Disadvantages:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Difficult to automatically infer the network;</para></listitem>
<listitem>
<para>Requires detailed domain specific knowledge to construct the network;</para></listitem>
<listitem>
<para>Has high computational costs.</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
<section class="lev4">
<title>Example: sheet metal machinery</title>
<para>One of the subsystems in a sheet metal machine is the hydraulics subsystem. Domain expertise is used to define a set of hydraulics failure modes, which are easy to understand and easy to describe. To further reduce the feature dimensional space, it is opted to use a binary representation of the state of the hydraulics subsystem: healthy or faulty. As there was no live data available from a working sheet metal machine, a BN simulation was built using hypothetical probability distributions and using a limited set of components, such as motors, encoders, actuators, sensors, valves. Each node in the graph represents one of these components. An interactive User Interface was developed to select the state of each node; the BN then calculates the failure probability distribution of each of the nodes in the network. The node with the highest probability is the most likely defective component. This interactive tool can be used to obtain a list of most probable defective components, based on real observations.</para>
</section>
</section>
<section class="lev3" id="sec5-2-2-5">
<title>5.2.2.5 Artificial neural network</title>
<para>An ANN is a collection of computational units called artificial neurons that mimic the human brain cells. Each neuron performs a weighted summation of its input values and applies an activation function, such as ReLU, sigmoid, Softmax. The output of a neuron is input to other connected neurons. The neurons are organized into layers: an input layer where the input values are fed into, the hidden layers that perform the learning process and the output layer that performs the final decision making process. The variability in number of layers, number of neurons and the activation functions results in a wide variety of possible ANN architectures. Similarly to SVM, the ANN algorithm is trained using a training data set and is validated using a test data set.</para>
<para>In recent years, the advances in computation power and effective implementations have enabled ANN to learn through large datasets and in numerous application domains.</para>
<section class="lev4">
<title>Example: health equipment</title>
<para>To research predictive maintenance capabilities for healthcare equipment, ANN was used to predict part replacement, with ANN training based on historical machine data. Historical part replacements were extracted from service business data. The matching machine errors in an observation window of 14 days before (failure sequence) and after the part replacement (non-failure sequence) was retrieved from machine logs. The machine data were grouped into features and represented as 1-dimensional and two-dimensional images and provided as input to the ANN. To test the influence of the chosen ANN architecture, five deep learning architectures were tested: artificial ANN, 1D convolution ANN, 2D convolution ANN, LeNet ANN, LSTM. For each architecture, the ANN is trained using 70% of the feature values while the remaining 30% is used to test the ANN. To judge the performance of each architecture, the percentage of correctly predicted part replacements is determined. All the above-mentioned ANN architectures yielded a percentage between 50% and 70%. This percentage does not significantly increase when doubling the number of features or doubling the historical period of time.</para>
</section>
<section class="lev4">
<title>Example: metal cutting machine</title>
<para>To research predictive maintenance capabilities for a metal cutting machine, ANN was used to predict the texture classification of machined products. The hypothesis is that machine defects may lead to surface defects in the machined products. The surface could be measured using cameras in the production line and the resulting images could be inspected in real-time. A ANN is a well known algorithm to classify image data. For that purpose, a convolution neural network has been built and trained. The Northeastern University, or NEU, surface defect database, publically available, was used to train the ANN. This database contains labelled images that cover multiple texture classes and contains hundreds of images per texture class. Each texture type is related to a certain, known, root cause. The ANN was tested using another dataset, containing images from machined products. The ANN yielded a performance of 95% of correctly classified textures. This is a good starting point to further explore how to improve maintenance of the metal cutting machine, based on visual inspection of metal surfaces.</para>
</section>
</section>
<section class="lev3" id="sec5-2-2-6">
<title>5.2.2.6 K-means clustering</title>
<para>The purpose of clustering is to group features in such a way that features of elements inside one cluster are more similar to each other than to the elements in other clusters. Clustering is useful to gain insights on the internal structure in data. In the clusters, it is easier to detect recurrent patterns and underlying rules. K-means clustering is a widely used algorithm that groups data points around centroids or means. The algorithm iteratively determines the centroids by minimizing the distance between data points and the centroids. The number of iterations depends on the dimensions of feature-space and the number of means. There is a variety of strategies how to select the centroids best, e.g., farthest point selection, K-means++.</para>
<section class="lev4">
<title>Example: press machine</title>
<para>In a press machine, the acceleration of the ram is measured during stamping. The acceleration is a measure for the vibration state of the press. The frequency spectrum is calculated from the accelerometer data. The peaks in the spectrum are identified using a peak detection algorithm. Data of multiple ram cycles are fed into a K-means cluster algorithm. The K-means algorithm is designed to detect the highest peak clusters in the frequency spectrum, covering the different press operating conditions. The higher the peak, the more it can contribute to a vibration induced failure. From the data analysis, it was concluded that 4 main peaks in the frequency spectrum can be clearly identified and can be related to excessive vibrations in the press. This knowledge is an enabler to implement this algorithm in the press machine in real-time and to trigger a maintenance action when if required.</para>
</section>
<section class="lev4">
<title>Example: special purpose vehicles</title>
<para>To research predictive maintenance capabilities in special purpose vehicles, log messages from hundreds of these vehicles have been analysed. The data set contains log messages from multiple vehicle types, multiple customers and multiple service providers. Each vehicle produces a time-series of log messages for what is considered as a stochastic process. The K-means algorithm is designed to detect log message patterns as function of vehicle type, customer and service provider. It was concluded that there is a strong relation between log message patterns and vehicle type but that there is no relation between log message patterns and service provider. Depending on the clustering size, there is a weak relation between log message patterns and customer; this indicates a weak impact of vehicle utilization.</para>
</section>
</section>
<section class="lev3" id="sec5-2-2-7">
<title>5.2.2.7 Attribute oriented induction</title>
<para>AOI is considered a hierarchical clustering algorithm for knowledge discovery in databases. Specifically, it is considered a rule-based concept hierarchy algorithm, due to the fact that the representation of the knowledge is structured in different generalization-levels of the concept hierarchy. The execution of the AOI algorithm follows an iterative process where each variable or attribute will have its own hierarchy tree. Later, data must change from one generalization-level to another, generalizing all the data in the dataset. That step is denoted concept-tree ascension. AOI is an algorithm whose power resides on defining thresholds and generalization hierarchies for the attributes of the data. This means that it is an algorithm where domain expert feedback is very useful. Basically, AOI must be trained first with healthy data, and build the base tree using these healthy data. Later, in the test part it is checked whether the evaluated data creates the same tree or it creates a different one. If the data creates the same tree, the asset can be considered to be healthy; if new instances are created, the asset had probably some issues in the meantime.</para>
<para>This would be the first level of maintenance: the anomaly detection. In order to provide RCA, there should be a database of each type of damage, to allow to correlate the cluster apparition with the damage itself.</para>
<section class="lev4">
<title>Example: clutch brake machine</title>
<para>AOI was successfully applied to a clutch brake machine monitoring use case. Brake pads are consumable parts of the clutch brake, and it is important to know when they wear out. Historical data from a clutch break machine was extracted. These data comprise information of pressure, rotating speed, air pressure and trigger, of a clutch brake both with complete brake pads and with worn out brake pads. First, using information provided by the domain expert, the generalizations of the AOI algorithm were established. Using data with the healthy clutch brake, the AOI algorithm was trained. Later, using the test data, the anomaly detection part of the AOI was used in order to model the wear of brake pads. Finally, thanks to the wear model, the solution was able to provide a RUL estimation for the asset.</para>
</section>
</section>
<section class="lev3" id="sec5-2-2-8">
<title>5.2.2.8 Hidden Markov model</title>
<para>A HMM is a statistical Markov model in which the system being modelled is assumed to be a Markov process with unobserved (hidden) states. The HMM is represented by a collection of states, state transitions probabilities and observations. It is the task of the HMM model to calculate, in the best way, the probability for a particular sequence of observations, without prior knowledge of the states. Similarly to other algorithms, such as SVM and ANN, the HMM is trained using a training data set and is validated using a test data set.</para>
<section class="lev4">
<title>Example: off-road and special purpose vehicles</title>
<para>To research predictive maintenance capabilities for forklift trucks, HMM was used to predict the probability of part replacement, based on historical machine data. Historical part replacements were extracted from service technician logs, based on text keywords. The matching machine data in an observation window of 30 days before the part replacement data (failure sequence) were retrieved from machine error logs. Machine errors after a part replacement (non-failure sequence) were also retrieved from machine error logs. These data sets were joined and fed to the HMM algorithm to train it. The performance of the HMM algorithm was determined using a test data set, randomly selected failure sequences and non-failure sequences.</para>
</section>
</section>
</section>
</section>
<section class="lev1" id="sec5-3">
<title>5.3 Remaining Useful Life Identification of Wearing Components</title>
<para>The RUL is the useful operational life left on an asset at a particular instance in time. Depending on the scientific or engineering field, RUL definitions can also include the usable or productive keywords, when defining the end of the asset useful life [ISO 13381-1, 2004].</para>
<section class="lev2" id="sec5-3-1">
<title>5.3.1 Theoretical Background</title>
<para>An essential question for design and operation of technical components is their expected useful lives. The estimation of useful life is important for designers and manufacturers for design improvements and marketing purposes, as well as product users for making decisions on which product to buy. Furthermore, the estimation of remaining useful life after the product has been put into operation is important for making decisions on maintenance, repair, reinvestment and new investments. RUL estimation is also important in relation to possibly existing service agreements between the users and manufacturers and possible guarantee periods. In this sense, the estimation of RUL of any asset is the first step towards CBM. A RUL prognostic model will typically provide at least two components: the RUL estimation and an associated confidence limit. Models that are more informative provide for instance an estimate of the failure distribution function.</para>
</section>
<section class="lev2" id="sec5-3-2">
<title>5.3.2 Techniques Catalogue</title>
<para>Various RUL models classification systems exist within the scientific literature. For instance [Si et al., 2011] classify RUL prediction models into two main types depending on whether they use: 1) direct condition monitoring data, or 2) indirect condition monitoring data. The first type is then divided into models which model the state evolution as a continuous process or alternatively as a discrete state space; while the second type is divided into the subcategories: a) filtering type of models, 2) covariate-based hazard models, and 3) hidden Markov model based methods. [Welte and Wang, 2014] classify RUL prediction models into physical, stochastic, data-driven and artificial intelligence model, while [Sikorska et al., 2011] categorize models into four main groups and a varying number of subgroups where the latter consist of the various RUL techniques. The classification system of Sikorska et al. is applied in the setup of this section and <link linkend="T5-2">Table <xref linkend="T5-2" remap="5.2"/></link> shows the four main classification categories together with the techniques that fall within each group.</para>
<table-wrap position="float" id="T5-2">
<label><link linkend="T5-2">Table <xref linkend="T5-2" remap="5.2"/></link></label>
<caption><para>RUL techniques categories</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left">Classification</td>
<td valign="top" align="left">Techniques</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Physical modeling</td>
<td valign="top" align="left">Application specific</td>
</tr>
<tr>
<td valign="top" align="left">Artificial Neural networks</td>
<td valign="top" align="left">RUL forecasting, Parameter estimation</td>
</tr>
<tr>
<td valign="top" align="left">Life expectancy models</td>
<td valign="top" align="left">Trend extrapolation, Auto regressive mean average methods, Proportional hazard models, reliability function, (Hidden) Markov models, Kalman filter, Particle filters</td>
</tr>
<tr>
<td valign="top" align="left">Knowledge based models</td>
<td valign="top" align="left">Expert systems, Fuzzy rules</td></tr>
</tbody>
</table>
</table-wrap>
</section>
<section class="lev2" id="sec5-3-3">
<title>5.3.3 Physical Modelling</title>
<para>The application of RUL techniques depend on expert knowledge that is distilled into a model of the physical asset being evaluated. The model families used in this context can differ between applications, and a few selected families are described in this section.</para>
<section class="lev3" id="sec5-3-3-1">
<title>5.3.3.1 Industrial automation</title>
<para>Manufacturing defects combined with severe working conditions and lack of maintenance, accelerate structural damage in the press head that can lead to failures in the form of fractures in the structural components of the press. The structural components, such as the press head, are typically welded steel parts. The welds are one of the sensitive parts where cracks can initiate because of notch stress concentration, residual stress and base material properties degradation.</para>
<para>Crack growth can be divided into three stages: initiation, stable propagation and fracture after an unstable or fast propagation. The first stage is difficult to predict and difficult to detect during regular maintenance service by traditional methods. Cracks are usually only detected when large enough to be visually localized during inspection. In such cases, a corrective maintenance action is taken to repair the failure. For this application, two degradation models have been used to predict the RUL: classical high cycle fatigue damage and crack propagation according to Paris&#x02019; law.</para>
<section class="lev4">
<title>High cycle damage: RUL to crack initiation</title>
<para>Classical high cycle fatigue is used to estimate the RUL to a certain threshold damage value, assuming that it is associated with a predefined initial crack length. Three methods oriented to welded structures recommended by the [International Institute of Welding, 2008] are used to evaluate the damage. The steps consist of 1) the stress evolution is calculated by Finite Element models applying the real forces measured at the press rods, see also <link linkend="F5-3">Figure <xref linkend="F5-3" remap="5.3"/></link>, 2) once the stress time history is calculated, the stress cycles during operation are obtained with the standard rain-flow cycle counting algorithm, then 3) the fatigue damage due to <emphasis>n<subscript>i</subscript></emphasis> constant stress cycles <emphasis>&#x003B4;&#x003C3;<subscript>i</subscript></emphasis> is obtained by the corresponding structural detail SN curve: <emphasis>D<subscript>i</subscript></emphasis> = <emphasis>n<subscript>i</subscript>/N<subscript>i</subscript></emphasis>, where N<emphasis><subscript>i</subscript></emphasis> indicates the maximum allowable stress cycles of range <emphasis>&#x003B4;&#x003C3;<subscript>i</subscript></emphasis> that a structural detail can withstand before failure for a given S-N curve characterized by <emphasis>C</emphasis> and <emphasis>m</emphasis> (slope). The SN-curve can have a different slope depending on the stress range, that is <emphasis>N<subscript>i</subscript></emphasis> = <emphasis>C/(&#x003B4;&#x003C3;<subscript>i</subscript>)<superscript>m</superscript></emphasis>. The total damage D due to different stress cycles is calculated according to Miner&#x02019;s rule, <emphasis>D</emphasis> = &#x003A3;<emphasis>D<subscript>i</subscript></emphasis> = &#x003A3;<emphasis>n<subscript>i</subscript>/N<subscript>i</subscript></emphasis>.</para>
<para>Finally, the RUL, considered as the estimated cycles (time) to the end of the first stage (crack initiation), is estimated at each critical zone setting up a damage threshold and calculating the remaining cycles. A fatigue damage indicators map is created in the studied component in order to identify the most probable crack initiation locations due to the real forces history applied in the press.</para>
<fig id="F5-3" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-3">Figure <xref linkend="F5-3" remap="5.3"/></link></label>
<caption><para>Press machine head critical zones and measured forces at the rods.</para></caption>
<graphic xlink:href="graphics/ch05_fig003.jpg"/>
</fig>
</section>
<section class="lev4">
<title>Crack growth, Paris&#x02019; law and particle filters</title>
<para>During the stable propagation stage, in the case of one-dimensional fracture, the crack growth rate is governed by a power law such as Paris&#x02019; law. This model gives a measure of the crack growth rate proportional to the stress intensity factor</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-1.jpg"/></para>
<para>with <emphasis>a</emphasis> the crack length, <emphasis>N</emphasis> the number of stress cycles; <emphasis>C</emphasis> and <emphasis>m</emphasis> material specific parameters &#x00394;K=K<emphasis><subscript>max</subscript></emphasis>-K<emphasis><subscript>min</subscript></emphasis>, the stress intensity factor, which is a function of the applied stress (load) range &#x00394;<emphasis>&#x003C3;</emphasis>, &#x00394;<emphasis>K(a)=Y(a)</emphasis> &#x00394;<emphasis>&#x003C3; (&#x003C0;a)</emphasis><superscript>1/2</superscript>, and <emphasis>Y(a)</emphasis> the geometry function taking into account the geometry of the surrounding of the crack.</para>
<para>Under the hypothesis of a crack of certain length started at a critical zone of the structural component, it is possible to estimate when the crack will reach a threshold length for a given loading. If experimental data on crack length propagation are available, the estimation of the RUL can be improved combining the measurements and the physics based crack propagation model. In this case, a particle filter technique has been applied to obtain the RUL. The technique handles the error associated to the measurements and to the model and updates the RUL, and it is supplemented with an interval of confidence every time a new observation becomes available.</para>
<para>In <link linkend="F5-4">Figure <xref linkend="F5-4" remap="5.4"/></link>, RUL estimation based on real data from a component under constant stress loading is shown. The initial crack size is 14 mm and the maximum crack size is set to 32 mm. The observed crack propagation is given by the blue line. An initial RUL estimation is done from the initial crack size based on the physical model (blue dotted line).</para>
<fig id="F5-4" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-4">Figure <xref linkend="F5-4" remap="5.4"/></link></label>
<caption><para>RUL calculation at 4.5 104 cycles.</para></caption>
<graphic xlink:href="graphics/ch05_fig004.jpg"/>
</fig>
</section>
</section>
<section class="lev3" id="sec5-3-3-2">
<title>5.3.3.2 Fleet&#x02019;s maintenance</title>
<para>Within MANTIS one of the goals is to find ways towards predictive fleet maintenance. With respect to RUL prediction, the main goal is to develop a method to optimize the maintenance protocols for defence vehicles and other types of complex machines, such that maintenance paradigms can be modified from a &#x0201C;time-of-usage&#x0201D; and a &#x0201C;number-of-driven-kilometres&#x0201D; approach to an approach taking into the account the severity of usage during operation. Thereby, the predictive power for when maintenance is truly necessary to be performed on the vehicle in order to prevent failure is increased.</para>
<para>An algorithm for determining the remaining useful life for a vehicle as one whole unit is introduced here. The goal of the algorithm is to estimate the wear induced on a vehicle by surface induced vibrations. The wear is quantified by calculating a wear index that is weighted with both the magnitude and the duration of an excitation. Damages due to the long-term use of a structure are typically associated with fatigue failure. This failure mode occurs when a structural member is exposed to a repeated loading a critical number of times. The fatigue strength of a certain material is often represented by an SN-curve. An SN-curve shows the mean number of constant stress-cycles it takes to break the material versus the applied, constant, stress magnitude. The critical number of cycles as a function of the applied constant stress level often follows, approximately, an exponential function of the form: <emphasis>N<subscript>i</subscript> = (S<subscript>0</subscript>/S<subscript>i</subscript>)<superscript>m</superscript></emphasis>, where <emphasis>N<subscript>i</subscript></emphasis> is the critical number of cycles, <emphasis>S</emphasis><subscript>0</subscript> a constant, <emphasis>S<subscript>i</subscript></emphasis> the constant stress magnitude applied, and <emphasis>m</emphasis> an exponent. In real-life, structures are exposed to excitations of varying amplitudes. Cycle counting then yields a histogram of cycles binned according to amplitude range. In this case, to compute the cumulative damage due to the cycles with various stress levels, Miners rule is applied, which states that:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-2.jpg"/></para>
<para>with <emphasis>n<subscript>i</subscript></emphasis> the number of cycles with stress level <emphasis>S<subscript>i</subscript>; N<subscript>i</subscript></emphasis> denotes the number of cycles after which (approximately) failure will occur when a constant stress level of <emphasis>S<subscript>i</subscript></emphasis> is applied and k denotes the number of discrete stress levels possibly applied to the structure. When <emphasis>D</emphasis> > 1 it is typically assumed that the structural member is expected to fail.</para>
<para>The concept behind the developed RUL algorithm is to estimate the wear of a vehicle to surface excitation in a manner similar to the Miner&#x02019;s rule damage estimation described above. However, as it is not feasible to measure the applied stress directly, as a measure of the stress the force applied to the structure of the rigid vehicle frame is used. To that purpose, the acceleration of the vehicle is measured with a 3D accelerometer sensor, and the amplitude of the force follows then directly from this measurement. The overall concept has some challenges. First of all the wear has to be calculated for the vehicle as one unit, although it consists of multiple structural members. Further, the loads cannot be measured directly, and instead the measured parameters are the accelerations. Finally, no parameters for the SN-model are available, moreover, these parameters are also expected to depend on other factors such as the location of sensor on the vehicle, the suspension system and the vehicle type.</para>
<para>To overcome many of these challenges a <emphasis>position translation algorithm</emphasis> has been developed. The <emphasis>position translation algorithm</emphasis> transforms the observations made at one location of the vehicle, the physical position of the sensor, to find the impact at another location of the vehicle, the physical position of the desired place of impact for the RUL computation. The so-called <emphasis>Inertial Measurement Unit</emphasis> measures translational accelerations and rotational velocities, and is located in the moving reference frame mounted on the vehicle frame close to the centre of gravity of the vehicle. Denoting the reference frame by x&#x02019; y&#x02019; z&#x02019;, the task is to transform the data measured in the reference frame to a similar aligned reference frame x&#x0201D; y&#x0201D; z&#x0201D; defined by displacing the reference frame by a displacement vector <emphasis>s&#x02032;<subscript>p</subscript></emphasis>. The prime in <emphasis>s&#x02032;<subscript>p</subscript></emphasis> indicates that the displacement vector for the new double primed reference frame is written in terms of the primed vehicle fixed system. Denoting the angular rotation vector as <emphasis>&#x003C9;</emphasis>&#x02019; =[<emphasis>&#x003C9;<subscript>x</subscript></emphasis>&#x02019;, <emphasis>&#x003C9;<subscript>y</subscript></emphasis>&#x02019;, <emphasis>&#x003C9;<subscript>z</subscript></emphasis>&#x02019;]<emphasis><superscript>T</superscript></emphasis>, which is the vector filled with data from the <emphasis>Inertial Measurement Unit</emphasis>, then the transformation of the translational accelerations from the primed to the double primed system are governed by the following relation <inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg164-1.jpg"/>, in which <emphasis>&#x003C9;<subscript>M</subscript></emphasis> is the so called skewed symmetric matrix defined through <emphasis>&#x003C9;</emphasis> as</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-3.jpg"/></para>
<para>The numerical differentiation of the angular velocities is necessary to estimate the angular acceleration of the primed system.</para>
</section>
<section class="lev3" id="sec5-3-3-3">
<title>5.3.3.3 Eolic systems</title>
<para>A degradation model is applied to predict the RUL of REBs in wind turbine gearboxes. The model can be viewed as a hybrid of a statistical model and a qualitative physics-based model. The objective of qualitative physics-based models is typically to select a proper damage indicator to construct a statistical model. The statistical model can then be fitted to the damage indicator data. Prior knowledge concerning probable values for the model parameters can be incorporated into the model by applying a Bayesian approach. By gathering an increasing amount of on-line monitoring data over time, the uncertainty related to the value of the model parameters decreases by applying a Bayesian updating method. In parallel the probability density function describing the remaining useful life of the REBs, which uses the parameter estimates, is updated accordingly.</para>
<section class="lev4">
<title>Implementation</title>
<para>Among four vibration features extracted from the full frequency domain vibration spectrum, the so-called High Frequency Peak Value was selected as the damage indicator to apply for the diagnosis and the prognosis of REBs. An exponential statistical model with multiplicative error terms is used, originally proposed by [Gebraeel, 2005] and [Gebraeel, 2003]. The mathematical form of the model is <emphasis>S(t<subscript>i</subscript>) = &#x1D703; exp[&#x003B2;t<subscript>i</subscript> + &#x1D700;(t<subscript>i</subscript>)]</emphasis>, in which the parameters &#x1D703; and <emphasis>&#x003B2;</emphasis> are assumed to follow, respectively, a log-normal and a normal distribution. Generally, the logarithm of the raw vibration signal <emphasis>S</emphasis>(<emphasis>t<subscript>i</subscript></emphasis>) is taken to fit the model to the data.</para>
<fig id="F5-5" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-5">Figure <xref linkend="F5-5" remap="5.5"/></link></label>
<caption><para>Illustration of Updated pdf of RUL for a REB.</para></caption>
<graphic xlink:href="graphics/ch05_fig005.jpg"/>
</fig>
<para>Given a specific time <emphasis>t<subscript>k</subscript></emphasis>, RUL denoted by <emphasis>T</emphasis> is defined as the duration from <emphasis>t<subscript>k</subscript></emphasis> to the critical time reaching the critical threshold denoted by <emphasis>D</emphasis>. The probabilistic distribution of <emphasis>T</emphasis> is equivalent to the logarithm of High Frequency Peak Value to reach <emphasis>D</emphasis>, and can be expressed by: <emphasis>P</emphasis> (<emphasis>T &#x0003C; t</emphasis>&#x0007C;<emphasis>L</emphasis><subscript>1</subscript><emphasis>,L</emphasis><subscript>2</subscript><emphasis>,&#x02026;,L<subscript>k</subscript></emphasis>) = <emphasis>P</emphasis>[<emphasis>L</emphasis>(<emphasis>t<subscript>k</subscript></emphasis>+ <emphasis>T</emphasis>) = <emphasis>D</emphasis>&#x0007C;<emphasis>L</emphasis><subscript>1</subscript><emphasis>,L</emphasis><subscript>2</subscript><emphasis>,&#x02026;,L<subscript>k</subscript></emphasis>], with <emphasis>L</emphasis><subscript>1</subscript>, <emphasis>L</emphasis><subscript>2</subscript>,&#x02026;, <emphasis>L<subscript>k</subscript></emphasis> the logarithm of the observations. The data related to a number of REBs was used to infer prior information for both the model parameters and the pdf of RUL. For each of these REBs, part of the signals was approximated by an exponential trend, over a period from the start-up of increase in vibration signal amplitude (day <emphasis>t<subscript>s</subscript></emphasis>) to the occurrence of cut-off (day <emphasis>t<subscript>e</subscript></emphasis>), and used to fit the statistical model. The statistics of <emphasis>&#x1D703;</emphasis> and <emphasis>&#x003B2;</emphasis> of this specific REB is estimated based upon the other <emphasis>N</emphasis>-1 REBs, also known as leave-one-out cross validation, and is used as the prior distribution of <emphasis>&#x1D703;</emphasis> and <emphasis>&#x003B2;</emphasis> for this specific REB. Bayesian updating is performed four times with the interval of 10 days (e.g., first Bayesian updating performed at Day (<emphasis>t<subscript>s</subscript></emphasis>+10), and so on), with the updated pdfs of one specific REB illustrated in <link linkend="F5-5">Figure <xref linkend="F5-5" remap="5.5"/></link>.</para>
</section>
</section>
<section class="lev3" id="sec5-3-3-4">
<title>5.3.3.4 Medical systems</title>
<para>The activity concerns the wear of cathode filaments in X-ray tubes that are used in interventional X-ray systems. X-ray tubes are the most expensive replacement parts of an interventional X-ray systems and therefore of major concern for the service organization. Because of the major impact, in terms of downtime and costs related to an X-ray tube replacement, it is important for Philips to be able to predict an upcoming failure of X-ray tubes accurately in order to provide in time replacement. A failure analysis was performed and the dominant failure mode turned out to be a blown cathode filament.</para>
<para>X-ray tube cathode filaments are heated to a high temperature during operation, in order to emit sufficient electrons to produce the desired X-ray dose. The high temperature, however, also makes the Tungsten&#x02014;of which the filaments are made of&#x02014;to evaporate. Thereby, a hot-spot forms locally at a certain location, where the generated heat and the evaporation increase exponentially over time, until the material melts at the hot-spot resulting in the opening of the filament, see also [Webers et al., 2016] and [Horster et al., 1971]. The physics of failure of the X-ray tube is very similar to the physics of failure for incandescent lamps [Rice, 1997]. Understanding the stress applied and having a damage indicator at hand are two prerequisites for building a physical model for RUL prediction of the filament. From the previous literature and experience, the filament stressor (which is the filament temperature) and a proper damage indicator (which is the filament resistance) are known. However, monitoring the damage increase over time by simply plotting the filament resistance as a function over time is problematic for two reasons: 1) individual interventional X-ray systems are daily used with various settings, meaning that is different filament currents are applied from run to run, implying that time is not a proper aging factor to apply as independent variable. 2) The filament resistance cannot be measured directly.</para>
<para>The first issue is dealt with by using a method similar to Miner&#x02019;s rule [10]. Instead of using the accumulated run time, an accumulation of linear damage (<emphasis>W</emphasis>) is used</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-4.jpg"/></para>
<para>with the sum applied over the total number of runs, and</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-5.jpg"/></para>
<para>where <emphasis>t<subscript>i,j</subscript></emphasis> is the time duration of run <emphasis>i</emphasis>, carried out with an applied current of A<subscript><emphasis>j</emphasis></subscript>, j = 1,&#x02026;,k, and k the total number of possible Ampere values to be applied; and <emphasis>T<subscript>j</subscript></emphasis> the mean lifetime for a filament when a constant current A<emphasis><subscript>j</subscript></emphasis> is applied to it.</para>
<fig id="F5-6" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-6">Figure <xref linkend="F5-6" remap="5.6"/></link></label>
<caption><para>ln(c-factor) vs. linear wear from lab experiments.</para></caption>
<graphic xlink:href="graphics/ch05_fig006.jpg"/>
</fig>
<para>The second issue is overcome by using the <emphasis>c-factor</emphasis> as a damage indicator rather than the filament resistance. The relationship between the resistance and the c-factor is given below. If the resistance of the filament at a particular temperature is <emphasis>R</emphasis><subscript>0</subscript> at the start of use, while after being used for <emphasis>i</emphasis> runs the resistance at the same temperature has changed to <emphasis>R<subscript>i</subscript></emphasis>, then the c-factor for that run is:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-6.jpg"/></para>
<para>The value of <emphasis>c<subscript>i</subscript></emphasis>can be derived for each run, without involving measurement of <emphasis>R<subscript>i</subscript></emphasis>. Two small modifications are made to the two variables in order to use them for monitoring the degradation over time. For the &#x00394;<emphasis>W</emphasis> calculation, the mean lifetime <emphasis>T<subscript>j</subscript></emphasis>&#x02014; mean lifetime for a filament when a constant current A<emphasis><subscript>j</subscript></emphasis> is applied to it&#x02014; is not used. In fact, the sum is divided by the mean life time for a current <emphasis>A<subscript>j</subscript></emphasis>/<emphasis>c<subscript>i</subscript></emphasis>. Without going into detail, by dividing the applied filament current by the c-factor, the accumulation of wear per unit time remains constant if the user keeps on using the system in the same way, and the new variable is referred as linear wear (<emphasis>LW</emphasis>), that is</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-7.jpg"/></para>
<para>with the summation over all runs <emphasis>i</emphasis>, and</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-8.jpg"/></para>
<para>with <emphasis>T&#x02019;<subscript>j</subscript></emphasis> the mean lifetime for a filament for which a constant continuous current <emphasis>A<subscript>j</subscript></emphasis>/<emphasis>c<subscript>i</subscript></emphasis> is applied. The need for this modification is a consequence of using a constant heating current during the lab experiments rather than constant heating power, which would have resulted in a constant filament temperature during a single experiment. In <link linkend="F5-6">Figure <xref linkend="F5-6" remap="5.6"/></link>, the logarithm of the c-factor is plotted as a function of linear wear for lab data.</para>
<para>It was observed that the slope of the linear part of the curve and the slope at time of failure are correlated, meaning that the relative large initial negative slope implies relative large negative slope at time of failure. This observation was quantified and incorporated in the final RUL estimation method.</para>
<para>The RUL estimation method entails now roughly the following steps</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>For the linear part of the curve a simple linear regression is used to determine the slope of this segment. For the curved segment a third order polynomial fit is applied, leading to <emphasis>c-factor= a LW</emphasis><superscript>3</superscript><emphasis>+ b LW</emphasis><superscript>2</superscript><emphasis>+d LW + e</emphasis>, with <emphasis>LW</emphasis> representing linear wear, and <emphasis>a</emphasis>, <emphasis>b</emphasis>, <emphasis>d</emphasis> and <emphasis>e</emphasis> coefficients that need to be found experimentally;</para></listitem>
<listitem>
<para>The first derivative with respect to the linear wear is taken, resulting in the quadratic function: <emphasis>c-factor&#x02019;= 3a LW</emphasis><superscript>2</superscript><emphasis>+ 2b LW +d.</emphasis> This quadratic function needs to fulfil two requirements. First, it needs to be a negative parabola, i.e., downward opening and second the top of the parabola should be located before the split of the two segments;</para></listitem>
<listitem>
<para>If the quadratic function is found, and writing the empirical found relation between the slope (<emphasis>a<subscript>ini</subscript></emphasis>) of the initial part of the curve and at failure (<emphasis>a<subscript>end</subscript></emphasis>) as <emphasis>a<subscript>end</subscript></emphasis> = <emphasis>p a<subscript>ini</subscript>+ q</emphasis>, then by solving <emphasis>c-factor&#x02019;=a<subscript>end</subscript></emphasis> for <emphasis>LW</emphasis>, gives the prediction for when failure will occur;</para></listitem>
<listitem>
<para>Finally, by extrapolating the observed increase of linear damage with time, an estimate can be found in terms of calendar date when the <emphasis>LW</emphasis> obtained under step 3) presumably is reached. Similarly, the latter can be done to obtain two confidence limits for which the filament will fail with 95% confidence.</para></listitem>
</itemizedlist>
</section>
</section>
<section class="lev2" id="sec5-3-4">
<title>5.3.4 Artificial Neural Networks</title>
<para>ANNs are Machine Learning algorithms inspired by biological nervous systems such as the human brain. ANNs process information using a set of highly interconnected nodes, also referred as neurons, organized into layers. The structure of ANN&#x02019;s are typically split into three types of layers: one <emphasis>input layer</emphasis>; one or more <emphasis>hidden layers;</emphasis>and one <emphasis>output layer</emphasis> (see <link linkend="F5-7">Figure <xref linkend="F5-7" remap="5.7"/></link>). The input layer receives the data and is connected to the first hidden layer, which in turn is connected either to the next hidden layer (and so on) or to the output layer. The output layer returns the ANN&#x02019;s predictions.</para>
<fig id="F5-7" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-7">Figure <xref linkend="F5-7" remap="5.7"/></link></label>
<caption><para>Schema of an artificial neural network, from [Haeusser, 2017].</para></caption>
<graphic xlink:href="graphics/ch05_fig007.jpg"/>
</fig>
<para>A node links to other nodes by weighted connections. At each node, an activation function combines these weights into a single value, which may limit signal propagation to the next nodes. These weights, therefore, enforce or inhibit the activation of the network&#x02019;s nodes. Neural networks can detect complex nonlinear relationships between dependent and independent variables and require very little feature engineering.</para>
<section class="lev3" id="sec5-3-4-1">
<title>5.3.4.1 Deep neural networks</title>
<para>DNNs (see <link linkend="F5-8">Figure <xref linkend="F5-8" remap="5.8"/></link>) brought major advances in solving some problems that were previously difficult to overcome in the artificial intelligence field [LeCun et al., 2015]. They have proved to be good at finding intricate structures in high dimensional data, which makes them relevant for many fields of study. Despite the fact that there is no clear border between what distinguishes a DNN from the others, a simple definition is that, a DNN contains many hidden layers in the network [Schmidhuber, 2015].</para>
<section class="lev4">
<title>Common applications</title>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>General classification [Baxt, 1990; Widrow et al., 1994];</para></listitem>
<listitem>
<para>General regression problems [Refenes et al., 1994];</para>
<fig id="F5-8" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-8">Figure <xref linkend="F5-8" remap="5.8"/></link></label>
<caption><para>Deep Neural Network with 3 hidden layers, from [Nielsen, 2017].</para></caption>
<graphic xlink:href="graphics/ch05_fig008.jpg"/>
</fig>
</listitem>
<listitem>
<para>Prediction of medical outcomes [Tu, 1996];</para></listitem>
<listitem>
<para>Environmental problems [Maier and Dandy, 2000];</para></listitem>
<listitem>
<para>Stock market index predictions [Moghaddam et al., 2016];</para></listitem>
<listitem>
<para>Remaining useful life (RUL) [Ali et al., 2015];</para></listitem>
<listitem>
<para>DNNs for image processing [Egmont-Petersen et al., 2002];</para></listitem>
<listitem>
<para>DNNs for audio [Lee et al., 2009];</para></listitem>
<listitem>
<para>DNNs for text [Collobert and Weston, 2008];</para></listitem>
<listitem>
<para>DNNs for regression tasks and remaining useful life [Tamilselvan and Wang, 2013].</para></listitem>
</itemizedlist>
</section>
<section class="lev4">
<title>Strengths and challenges</title>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Some of the strengths of ANNs are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>ANNs are effective at modelling complex non-linear systems;</para></listitem>
<listitem>
<para>Once an ANN has been trained, it processes the data efficiently;</para></listitem>
<listitem>
<para>ANNs (and DNNs in particular) require little to no feature engineering [Tran et al., 2011].</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>Some of the challenges are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Training a model usually incurs high computational and storage costs;</para></listitem>
<listitem>
<para>Lack of interpretability (&#x0201C;black boxes&#x0201D;) despite efforts to extract knowledge from them [Andrews et al., 1995; van der Maaten and Hinton, 2008; Merrienboer et al., 2014];</para></listitem>
<listitem>
<para>Many hyper parameters need to be tuned;</para></listitem>
<listitem>
<para>Many configurations are possible;</para></listitem>
<listitem>
<para>DNNs typically require a lot of data to be trained [Chilimbi et al., 2014];</para></listitem>
<listitem>
<para>Collecting large labeled datasets required for training the DNNs is a challenge.</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
<para>Research has been carried out in order to reduce the training time by using simple learning functions [Guo and Morris, 2017] or by transfer learning [Schmidhuber, 2015]. In regards to hyper-parameter tuning, several approaches have been proposed to automate this process; for example [Domhan et al., 2015].</para>
</section>
<section class="lev4">
<title>Implementation</title>
<para>The use-case considered has a dataset that represents a time-series of sensor readings from a fuel system. Each sample provides information from several sensors and the current health state of the system (range from 0% to 100%). Exposing these sensors to extreme driving or ambient conditions for several years can lead to their failure or deviations in measurement. The goal is to predict the current health state of the fuel system as accurately as possible while being robust to sensor failure and deviations. Experiments based on small feed-forward neural network with only 3 hidden layers (Input layer: 161 neurons; Hidden layer 1: 128 neurons; Hidden layer 2: 256 neurons; Hidden layer 3: 128 neurons; Output layer: 6 neurons) and using drop-out and noise injection were able to maintain accurate predictions with as much as 40% of incorrect sensor data (using simulated errors).</para>
</section>
</section>
</section>
<section class="lev2" id="sec5-3-5">
<title>5.3.5 Life Expectancy Models</title>
<para>In some cases, RUL computation leverages on models of life expectancy of assets. This is the case, in particular, when time series are used to express collected data over data, and to drive the prediction regarding particular parameters that can be useful to obtain a reliable RUL.</para>
<section class="lev3" id="sec5-3-5-1">
<title>5.3.5.1 Time series analysis with attribute oriented induction</title>
<para>A time series is a collection of data measured over time and represents the evolution of a certain quantity over a certain period. The main feature of a time series is that the collected data values are successive in time and typically strongly correlated with neighbouring observations. In many scenarios the goal is to predict the future outcomes of a variable of interest. The overall research field dealing with time series is referred to as time series analysis. Commonly, in order to analyse a time series the signal is: 1) decomposed in various parts, that is in a <emphasis>trend</emphasis> component, a <emphasis>cyclic</emphasis> (seasonal) component (both nonstationary) as well as in the <emphasis>random</emphasis> (stationary) component that remains after removing the cyclic and trend effect from the observed series; 2) analysis is performed on the various parts, which are then 3) recombined in order to make predictions. One of the most utilized time series forecasting models are autoregressive and moving average models (ARIMA). The AR part of ARIMA indicates that the evolving variable of interest is regressed on its own lagged, i.e., previous, values. The MA part indicates that the regression error is actually a linear combination of error terms whose values occurred contemporaneously and at various times in the past. The &#x0201C;I&#x0201D; (for &#x0201C;integrated&#x0201D;) in the name ARIMA indicates that in many applications, instead of analysing the original data, the data is preliminary differentiated, one or multiple times. The purpose of each of these features is to construct an ARIMA model to fit the data as well as possible.</para>
<para>AOI is a data mining hierarchical clustering algorithm. The main characteristic of AOI is the capacity to describe the information in more general concepts/terms, reducing the dimension of the data. Its power resides on combining the monitored data with domain expert knowledge, in order to provide a more trusty knowledge generation.</para>
<para>The execution of the AOI algorithm follows an iterative process where each variable or attribute will have its own hierarchy-tree. Later this data must change from one generalization-level to another, generalizing all the data in the dataset. That step is denoted concept-tree ascension. AOI is an algorithm whose power resides on defining thresholds and generalization hierarchies for the attributes of the data. This means that it is an algorithm where domain expert feedback is necessary. Based on the defined hierarchy trees, generalizations are performed iteratively in order to conform similarity clusters. The higher the generalization level on the hierarchy tree is, the more general&#x02014;but thereby also the more ambiguous&#x02014;the cluster description gets. The more general the cluster, the lower the value of significance of the cluster. Such clusters refer to different states of the behaviour of the monitored asset.</para>
<section class="lev4">
<title>Implementation</title>
<para>Time Series analysis with AOI was applied in relation to a clutch brake machine monitoring use case. Two main phases are employed on RUL calculation: (i) quantification, and (ii) time series analysis. The aim of quantification is to describe the signal(s) to be forecasted in a measurable (numeric) way. When quantification is applied over a set of signals, its information representation becomes simpler and easy to deal with, but also some information-loss occurs. With the help of AOI, different work-cycles (clutch or brake cycles) were labelled by a training process with a value corresponding to their level of normality, called Normality Factor, according to the ambiguity values of clusters. Thereby, a minimum value for a work cycle to be considered as normal is calculated based on training phase data. Finally, the evolution of the Normality Factor value can be modelled with a time series based ARIMA model, in order to check when a work-cycle, which does not meet the normality conditions, occurs.</para>
</section>
</section>
<section class="lev3" id="sec5-3-5-2">
<title>5.3.5.2 Application to a pump</title>
<para>A methodology for quantifying and monitoring the evolution of the performance of a pump has been formulated, in order to take maintenance decisions based on this information. The performance of a pump is influenced by two main causes: the wear/tear of its components and the clogging situations. In both cases, maintenance tasks may be needed to recover the best pump performance. A maintenance task is recommended when the performance decreases more than 20%. Therefore, in this context, RUL is understood as the time (in days) until a pump&#x02019;s performance decreases more than 20%. The methodology is based on data typically available for pumps: the operation frequency (Hz) and the pumped flow rate (m<superscript>3</superscript>/h).</para>
<para>First of all, the performance of the pump is estimated each day using the available frequency and flow rate historical data. Later, the evolution of the pump performance is used for estimating the RUL. Thus, the trend of the pump performance is calculated applying a linear regression algorithm. Finally, this linear regression is used for predicting the evolution of the pump and calculating the RUL, i.e., the time until the performance reaches the value of &#x02013;20%.</para>
<para>The RUL estimation error has been analysed for three clogging situations which vary in duration and magnitude. For each clogging period, <link linkend="F5-9">Figure <xref linkend="F5-9" remap="5.9"/></link> shows the RUL estimation error on the Vertical Axis, as a function of the remaining days to clogging situation (Horizontal Axis). In particular, day 0 corresponds to the day of a -20% performance.</para>
<para>The RUL estimation error ended up being between &#x000B1;5 days during the analysed situations, which is considered good enough taking into account the duration of the clogging situations and the randomness of its nature.</para>
<fig id="F5-9" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-9">Figure <xref linkend="F5-9" remap="5.9"/></link></label>
<caption><para>RUL error vs remaining days to clogging situation.</para></caption>
<graphic xlink:href="graphics/ch05_fig009.jpg"/>
</fig>
</section>
<section class="lev3" id="sec5-3-5-3">
<title>5.3.5.3 Application to industrial forklifts</title>
<para>This use-case used data collected from industrial forklifts. The overall goal is to provide RUL estimates for forklift tires. Two different datasets were available to achieve this goal. The Fleet Manager system provides data on the usage by the on-board computers of different forklifts and generates 10-minute aggregated data reports. It contains fields like driven time, distance travelled, number of direction changes and consumed energy amount. The second data source are the service reports from service technicians, in which they reported in free text both the problems identified in a forklift and the repairs that were performed.</para>
<para>Text analysis was used to identify and extract the dates from the service reports when a tire replacement occurred due to the tire being worn-out (and not due to any other cause). Thereafter, the Fleet Manager data set was further aggregated into time intervals where forklifts used a given set of tires. The regression model was built using this pre-processed dataset. The Fleet Manager data set was also clustered into three driving profiles by using expert knowledge, which resulted in useful categorical variables in the regression model.</para>
<para>The clustering was done in two steps. In the first step, a centroid-based k-means algorithm was used, and then the final 3 driving profile classifications were created with a hierarchical clustering method. This was needed to leverage the advantages of both algorithms, while minimalizing their drawbacks. For example, the k-means clustering algorithm tends to create equally sized convex clusters, which is not physically meaningful in relation to this data set. A more detailed analysis of these clustering solutions, highlighting the strength and weaknesses of both can be found in [Kaushik and Mathur, 2014].</para>
<para>The second step involved building an <emphasis>ensemble tree-based</emphasis> regression model, called <emphasis>Gradient Boosted Decision Trees</emphasis> [Friedman, 1999]. Like other ensemble regression models (e.g., Random Forest), this algorithm provides good defence against overfitting at the cost of interpretation, in the sense that the predictions of the model are less transparent. The model handles different variable types well, and it is insensitive to correlated variables. With adequate parameter optimization, it is shown in [Caruana and Niculescul-Mizil, 2005] that the <emphasis>Gradient Boosted Decision Trees</emphasis> provides great accuracy compared to many other classical machine learning models. The model training included cross-validation to reduce the level of overfitting due to the limited amount of data points, and hyper parameter optimization to find the optimal algorithm parameters (e.g., number of trees or tree depth). The results were evaluated based on the Root Mean Square Error measure.</para>
</section>
<section class="lev3" id="sec5-3-5-4">
<title>5.3.5.4 Application to a gearbox</title>
<para>One of the main objectives in this use-case in relation to RUL prediction is in improving the failure prediction of a gearbox by incorporating an additional covariate besides produced GWh to a failure model, and to test whether or not incorporating this additional covariate adds predictive power. The additional covariate is allowed to change with time and so it is a so-called time dependent variable.</para>
<para>The additional covariate is incorporated by applying proportional hazard modelling with a time-dependent covariate. The Incorporation of a time-dependent covariate within a proportional hazard model, requires the knowledge at each failure time <emphasis>t</emphasis>the value of the time dependent covariates of all assets that did not fail until time <emphasis>t</emphasis>. The latter demand is in practice in many situations hard to fulfil. Through the last two decades, applying time dependent covariates has become more and more common and standard software is available in for instance [R Core Team, 2014] to incorporate them [Therneau, 2015; Therneau and Atkinson, 2017; Fox and Weisberg, 2010].</para>
<para>Denoting <emphasis>w</emphasis> the produced energy [GWh], then for an object with one single GWh-dependent covariate the hazard function takes the form [Thomas and Reyes, 2014]</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-9.jpg"/></para>
<para>with <emphasis>z(w)</emphasis> the value of the covariate at n of <emphasis>w</emphasis>, and w &#x02265; 0. By the hazard rate the survival function conditional under <emphasis>z(w)</emphasis> is</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-10.jpg"/></para>
<para>with the cumulative hazard conditional under <emphasis>z(w)</emphasis> defined by</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-11.jpg"/></para>
<para>The latter two equations taken together imply that to define the survival function for an asset at <emphasis>w</emphasis>=0&#x02014;or at the current value of <emphasis>w</emphasis>&#x02014;it is necessary to know the future values of the covariate as a function of <emphasis>w</emphasis>, which is often an impossible requirement to fulfil [Fisher and Lin, 1999]. For more details concerning including time-dependent covariates in a proportional hazard model, please refer to [Fisher and Lin, 1999] and [Collett, 2003].</para>
</section>
</section>
<section class="lev2" id="sec5-3-6">
<title>5.3.6 Expert Systems</title>
<para>An extended Expert system can be used as support for both RUL and RCA algorithms. The extension includes a Petri net [Aghasaryan, 1997] based execution scheduling, which allows for further enhancements of the system. The method entails introducing a supporting function in order to use an expert system along with a RUL estimation algorithm. In this use case the method is applied together with a Proportional hazard model based RUL estimation of tire wear (see Section 5.3.5.3).</para>
<para>The objective of the Expert System is to verify and make decisions based on the result of the RUL estimation algorithm. The RUL estimation is performed online, and updated every time a new sensor measurement set arrives. However, the estimated RUL needs to be validated, and then a decision is taken. The Expert System is performing the validation, and makes a decision accordingly. A rule-based system is selected as it scales well, provides results faster than case based systems [Simpson and Sheppard, 1998], and does not require large amounts of labelled training data like machine learning based methods. The proposed method mimics the behaviour of a human expert as follows:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>The Expert System is triggered by a rule;</para></listitem>
<listitem>
<para>It extracts key parameters from the trigger description and based on this, a course of actions is selected;</para></listitem>
<listitem>
<para>It initiates elementary investigation checks, search routines and possible correlated processes in a simultaneous manner;</para></listitem>
<listitem>
<para>Once a result of a check is available, new routines are started using the new pieces of information;</para></listitem>
<listitem>
<para>It continue to perform checks and tests until a result is found.</para></listitem>
</itemizedlist>
<para>The course of actions to be executed for a specific trigger is based mainly on expert knowledge. For each monitored component, a different rule set is used, specific to the component.</para>
<para>Petri nets present an efficient way to implement data flow-driven programming. They can be used to mimic the simultaneous data processing capabilities of a human expert. System specialists make ad-hoc plans or follow predefined procedures to find the root cause of an alarm. Consciously or not, they fetch further input data e.g., environment data, to start measurement or analysis processes. Simultaneous processes finish asynchronously, which forces the expert to make decisions on what to do next: what kind of further checks can be initiated using the gathered input data? In short, the expert&#x02019;s behaviour that consists in scheduling elementary checks to solve a RCA problem can be naturally modelled with Petri nets. In the use-case, the Petri net describes rules, extended with elementary checks. The elementary checks are typically basic checks like database lookups, more detailed data requests as well as they may include active measurements like on-demand vibration analysis. The various triggers associated with the definition of such Petri nets are based mainly on expert knowledge.</para>
<section class="lev4">
<title>Implementation</title>
<para>In a complex system the input of RCA and RUL calculation is big data aggregated from multiple distinct sources, and the analysis of such big data is performed in the cloud. For implementation, Microsoft Azure has been chosen which provides extensive functionality for collecting and processing data from different sources. Azure IoT Hub is used to handle incoming chunks of data such as real-time measurements. The failure prediction is performed in Azure Stream Analytics based on the trigger conditions, which need to be uploaded to Azure Storage in advance. The IoT Hub receives the error logs and other measurements from the trucks and forwards this data to Stream Analytics, which checks for trigger conditions. A simple rule set can be implemented in Stream Analytics directly. The Stream Analytics process is defined in stream analytic query language, which has a similar syntax to SQL. When a failure is predicted, a Web based remote application will be informed, which does the actual verification. The remote application implements the Petri-net based scheduler and the elementary checks. The output of the application can be visualized in Azure Power BI, and it also integrates in the Mantis HMI (see <link linkend="ch06">Chapter <xref linkend="ch06" remap="6"/></link>).</para>
</section>
</section>
</section>
<section class="lev1" id="sec5-4">
<title>5.4 Alerting and Prediction of Failures</title>
<para>Alerting is a method to trigger a corrective action. The purpose of the corrective action is to make sure that the failure does not re-occur. Data driven alerting is based on algorithms that detect and predict failures. These algorithms process historical data and, based on domain knowledge, produces a failure probability (see <link linkend="F5-10">Figure <xref linkend="F5-10" remap="5.10"/></link>). Based on business understanding, these probabilities are used to define the precise corrective action.</para>
<section class="lev2" id="sec5-4-1">
<title>5.4.1 Theoretical Background</title>
<para>There is a wide variety of algorithms available to detect and predict failures. The algorithm selection is based on business understanding, data understanding, and data preparation.</para>
<para>The CRISP-DM defines a methodology for structured data mining. For prediction and alerting of failures, the following phases are used: modelling, evaluation, deployment.</para>
<section class="lev3" id="sec5-4-1-1">
<title>Modelling</title>
<para>Once the relevant data features have been identified, the algorithm to predict failures can be developed. The output of the algorithm is a statistical measure that indicates the failure probability. Based on business rules, an alert can be created if the probability warrants this.</para>
</section>
<section class="lev3" id="sec5-4-1-2">
<title>Evaluation</title>
<para>$$$The model is applied to a selected set of real-life cases. These cases are evaluated to judge whether there is a sufficient match between prediction and reality.</para>
<fig id="F5-10" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-10">Figure <xref linkend="F5-10" remap="5.10"/></link></label>
<caption><para>Overview for Alerting and prediction of failures.</para></caption>
<graphic xlink:href="graphics/ch05_fig010.jpg"/>
</fig>
</section>
<section class="lev3" id="sec5-4-1-3">
<title>Deployment</title>
<para>The model is brought into production after successful evaluation. The model can be implemented in the system or outside the system (depending on the required data sources and computation capabilities) to create alerts in real-time.</para>
</section>
</section>
<section class="lev2" id="sec5-4-2">
<title>5.4.2 Techniques Catalogue</title>
<para>There is a wide variety of modelling techniques available. <link linkend="T5-3">Table <xref linkend="T5-3" remap="5.3"/></link> shows an excerpt of techniques.</para>
<table-wrap position="float" id="T5-3">
<label><link linkend="T5-3">Table <xref linkend="T5-3" remap="5.3"/></link></label>
<caption><para>Techniques for Alerting and Prediction of Failures</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left">Modelling technique</td>
<td valign="top" align="left">Algorithm</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Pre-processing</td>
<td valign="top" align="left">Nearest Neighbor Cold-deck Imputation</td>
</tr>
<tr>
<td valign="top" align="left">Classification / Pattern Recognition</td>
<td valign="top" align="left">SVM, LDA, Pattern Mining</td>
</tr>
<tr>
<td valign="top" align="left">Dimensionality reduction</td>
<td valign="top" align="left">Temporal Pattern Mining, PCA</td>
</tr>
<tr>
<td valign="top" align="left">Probabilistic graphical models</td>
<td valign="top" align="left">Hidden Semi-Markov model with Bayes Classification</td>
</tr>
<tr>
<td valign="top" align="left">Neural networks</td>
<td valign="top" align="left">DNN, Autoencoders, Convolutional neural network with Gramian Angular Fields, RNN with LSTM</td>
</tr>
<tr>
<td valign="top" align="left">Time Series Analysis</td>
<td valign="top" align="left">Change detection</td>
</tr>
<tr>
<td valign="top" align="left">Statistical tests</td>
<td valign="top" align="left">Fisher&#x02019;s exact test, Bonferroni correction, Hypothesis testing using univariate parametric statistics, Hypothesis testing using univariate non-parametric statistics, Mean, thresholds, normality tests</td>
</tr>
</tbody>
</table>
</table-wrap>
<para>Based on research on practical systems, a selected set of algorithms is described in the next sections. For each algorithm, the common application, strengths and challenges are outlined, and some practical examples are given.</para>
<section class="lev3" id="sec5-4-2-1">
<title>5.4.2.1 Nearest neighbour cold-deck imputation</title>
<section class="lev4">
<title>General description</title>
<para>Many machine learning algorithms assume that all records in the data set are complete, i.e., there is no missing data. Industrial data, however, often contains gaps e.g., due to misconfigured sensors, connectivity problems or manual input. One approach for addressing the missing data problem is to impute the missing values. Let us assume that the data is provided as a table, with columns representing the features and rows representing the records. The common hot-deck methods include mean imputation (the missing values in a column are filled in with the mean of that column), fixed value imputation, random imputation (the missing values are filled by randomly sampling the given values in the column), considering donor samples from the <emphasis>same</emphasis> data set. Nearest neighbor cold-deck imputation fills in the missing values from donor records that are selected from a <emphasis>different</emphasis> data set. It selects the donors by searching for the nearest neighbours using meta-features describing the original data set or its column. The meta-features can be derived from the available data samples (e.g., mean, range, ...) or external to the data (e.g., specification of the data source, conditions under which the data was collected &#x02026;).</para>
</section>
<section class="lev4">
<title>Common applications</title>
<para>Cold deck imputation in general is commonly used in for imputing missing data in structured collections of data such as tables and questionnaires.</para>
</section>
<section class="lev4">
<title>Strengths and challenges</title>
<para>Nearest Neighbour Cold-deck Imputation performs better than the traditional hot deck imputation methods for data sets for which a significant proportion of the data is missing. The performance of Nearest Neighbour Cold-deck Imputation also depends on the meta-features used for selecting the nearest neighbour donors, and how they partition the donor space. If donors are selected based on meta-features derived from the data set, then (similarly to hot deck imputation) the more data is missing the less accurate the imputation. If external meta features are used, then the imputation accuracy does not depend on the amount of missing data, but on the relationship between the external features and the underlying data generating process.</para>
</section>
</section>
<section class="lev3" id="sec5-4-2-2">
<title>5.4.2.2 Support vector machine</title>
<para><emphasis role="strong"><emphasis>General description</emphasis></emphasis></para>
<para>A SVM is a supervised learning model that can be used for binary classification and regression analysis. Supervised learning requires labelled training data consisting of input objects (e.g., a feature vector) and a desired output in order to infer the model parameters from the data and to be able to predict new incoming data. A SVM constructs a hyperplane or a set of hyperplanes to separate classes in the feature space. SVM defines optimal hyperplanes for linearly separable patterns and can be extended to patterns that are not linearly separable, by transforming the original data to a new space by using a kernel function.</para>
<para>A hyperplane is defined as <emphasis>f</emphasis> (<emphasis>x</emphasis>) = <emphasis>b</emphasis> + <emphasis>w<superscript>T</superscript>x</emphasis>where <emphasis>w</emphasis> is the weight vector, <emphasis>b</emphasis> is the bias and <emphasis>x</emphasis> the training dataset. Infinite hyperplanes can be used to separate the classes by scaling <emphasis>w</emphasis> and <emphasis>b</emphasis> (See <link linkend="F5-11">Figure <xref linkend="F5-11" remap="5.11"/></link>). Rewriting the hyperplane in canonical form states |<emphasis>b + &#x03C9;<superscript>T</superscript>x</emphasis>| = 1.</para>
<para>The optimal hyperplane is therefore defined as the plane maximizing the margin of the training data. In SVM the data points that lie the closest to the decision surface and therefore the most difficult to classify, are the ones influencing optimality. These points are called support vectors.</para>
<para>The distance between points is defined as <inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg181-1.jpg"/></para>
<para>In a binary classifier, the margin <emphasis>M</emphasis> is defined as twice the distance to the closest examples <inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg181-2.jpg"/></para>
<fig id="F5-11" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-11">Figure <xref linkend="F5-11" remap="5.11"/></link></label>
<caption><para>Linearly separable point can be separated in an infinite number of ways.</para></caption>
<graphic xlink:href="graphics/ch05_fig011.jpg"/>
</fig>
<para>Therefore, the problem of maximizing the margin <emphasis>M</emphasis> is the same as the problem of minimizing the following function <emphasis>L</emphasis> subject to constraints:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-12.jpg"/></para>
<para>Where <emphasis>y<subscript>i</subscript></emphasis> represents each of the labels in the training dataset. This is a constrained optimization problem that is solved by Lagrangian multiplier method as:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-13.jpg"/></para>
<para>By equating the derivatives to 0, the result is <emphasis>w</emphasis> = <superscript><emphasis>P</emphasis></superscript><emphasis>&#x003B1;<subscript>i</subscript>y<subscript>i</subscript>x<subscript>i</subscript></emphasis> and <superscript><emphasis>P</emphasis></superscript><emphasis>&#x003B1;<subscript>i</subscript>y<subscript>i</subscript></emphasis> = 0 (<link linkend="F5-12">Figure <xref linkend="F5-12" remap="5.12"/></link>).</para>
<fig id="F5-12" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-12">Figure <xref linkend="F5-12" remap="5.12"/></link></label>
<caption><para>Optimal hyperplane in a 2D feature vector obtained by maximizing the margin.</para></caption>
<graphic xlink:href="graphics/ch05_fig012.jpg"/>
</fig>
<fig id="F5-13" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-13">Figure <xref linkend="F5-13" remap="5.13"/></link></label>
<caption><para>Slack variables introduced for not linearly separable problem.</para></caption>
<graphic xlink:href="graphics/ch05_fig013.jpg"/>
</fig>
<para>In case the data are not separable by a hyperplane, SVM can use a soft margin, meaning a hyperplane that separates many but not all data points (<link linkend="F5-13">Figure <xref linkend="F5-13" remap="5.13"/></link>). Different formulations are possible for adding soft margin by adding slack variables <emphasis>s<subscript>i</subscript></emphasis> and a penalty parameter C. An example is the L1-norm problem:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-14.jpg"/></para>
<para>Some problems that do not have a simple hyperplane as a separator may require nonlinear transformation with kernels. The idea is to gain linearly separation by mapping the data to a higher dimensional space (<link linkend="F5-14">Figure <xref linkend="F5-14" remap="5.14"/></link>).</para>
<para>SVM are inherently a binary classifier that can be extended to the multiclass problem (e.g., by reducing the multiclass problem into multiple binary classification problem).</para>
<para>In the context of anomaly detection, one-class SVM can be used to detect if the system is behaving normally or not. In order to define the normal behaviour of the system to be monitored, historical data of the system running in normal behaviour is required. The support vector model is hence trained on the anomaly-free data and afterwards each new measurement point is scored by a normalized distance to the decision boundary.</para>
<fig id="F5-14" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-14">Figure <xref linkend="F5-14" remap="5.14"/></link></label>
<caption><para>SVM kernel transformation to gain linearly separation.</para></caption>
<graphic xlink:href="graphics/ch05_fig014.jpg"/>
</fig>
<section class="lev4">
<title>Common applications</title>
<para>SVM was developed by Vapnik and Chervonenkis (1971); Vapnik (1995) and became popular because of its success in handwritten digit recognition. SVMs have demonstrated highly competitive performance in numerous real-world applications such as bioinformatics, text mining, face recognition, and image processing, which has established SVMs as one of the state-of the-art tools for machine learning and data mining.</para>
</section>
<section class="lev4">
<title>Strengths and challenges</title>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Some strengths of SVM are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Training time necessary depends on the number of supports and the kernel chosen, while the execution time is very short so that it is suitable for real-time application;</para></listitem>
<listitem>
<para>SVM do not suffer from the curse of dimensionality;</para></listitem>
<listitem>
<para>SVM is not much affected by local minima in the data;</para></listitem>
<listitem>
<para>Does not require lot of data to be trained, however the data should come from all the possible normal behaviours of the system.</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>On the other hand, it also has some limitations, some of them being listed below:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>The feature space of a SVM is often high dimensional and therefore has many parameters and is hard to estimate;</para></listitem>
<listitem>
<para>Selection of the kernel function is not trivial.</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
</section>
</section>
<section class="lev3" id="sec5-4-2-3">
<title>5.4.2.3 Linear discriminant analysis</title>
<section class="lev4">
<title>General description</title>
<para>Let P(x) be the probability distribution of the input data, and P(y) the class distribution. The goal of a classifier is to assign a sample x to the class y, which corresponds to the largest posterior probability P(y&#x0007C;x) that input x belongs to class y. The LDA classifier relies on the Bayesian theorem</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-15.jpg"/></para>
<para>to compute the posterior using the likelihood P(x&#x0007C;y) of input x being generated by class Y, the prior P(y) and the marginal probability P(x) of input x. The P(y) can be easily estimated from the data by taking the fraction of samples that belong to class Y. Assuming that likelihood P(x&#x0007C;y) is normally distributed, it can be estimated by fitting the mean and variance parameters to the data. See James et al. (2014) for more details, including the formula for the discriminant function used to select the most likely class for input x.</para>
</section>
<section class="lev4">
<title>Common applications</title>
<para>LDA is a general classifier that can be applied in various applications. In the context of failure predictions, it can be applied to classify process measurements as those likely to result in a failure.</para>
</section>
<section class="lev4">
<title>Strengths and challenges</title>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>LDA is applicable in cases when there are few samples, or the classes are overlapping or well separated (in contrast to other classification methods, e.g., logistic regression, which become unstable when classes are well separated). LDA performs best when the classes (i.e., P(x&#x02014;y)) are approximately normally distributed.</para></listitem>
</itemizedlist>
</section>
</section>
<section class="lev3" id="sec5-4-2-4">
<title>5.4.2.4 Pattern mining</title>
<section class="lev4">
<title>General description</title>
<para>Pattern mining is one of the most known methods to extract useful information from the data. By looking for patterns, i.e., closely linked events or items, this technique allows to analyse behaviour and make prediction on future behaviour.</para>
<para>By knowing this behaviour, pattern mining can predict when a failure will occur (by detecting also when the start of the sequence occurs) and can signal the problem before it occurs. This method can be deployed in many various fields, such as maintenance of a factory or a wind mill farm, determining a disease, predicting the next medical prescription, analysing the behaviour of cyclists in a town.</para>
<para>The first pattern mining algorithm, Apriori, was created in 1995 and was only able to find frequent item sets, i.e., to find items that usually occur together. Now pattern-mining methods can also look for sequences of items where the order of the items matters.</para>
</section>
<section class="lev4">
<title>Common applications</title>
<para>The list below only mentions the main topics, but many others exists:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Frequent item sets</para>
<para>A supermarket can analyse the list of goods bought by their customers to find those usually bought together. It can help the supermarket managers to reorganize their shelves or to propose special offers;</para></listitem>
<listitem>
<para>Sequence of events</para>
<para>A logistic company that continuously monitors events occurring in their vehicles, e.g., engine oil level too low or motor temperature too hot. As they also know when the vehicles are break, they can find patterns leading to failures, i.e., the sequence of events that usually precede a failure;</para></listitem>
<listitem>
<para>Multi-level</para>
<para>In multi-level pattern mining, not only the items are considered, but also their hierarchy. Going to the supermarket use case, multi-level pattern mining can output purchases containing apples, pears, kiwis. All of them are considered as distinct items although they are all fruits. Therefore, the supermarket may also be interested to find patterns containing fruits and not the more specific sub-level items, e.g., apples. In addition, by considering the fruits as a whole, it is possible to find patterns that would be hidden if selected independently.</para></listitem>
<listitem>
<para>Multi-domain</para>
<para>The multi-domain pattern mining methods consider multiple properties of the items when looking for patterns. The supermarket could have stores all over a country and would be interested to check if there are different patterns depending on the location of the purchase. These methods could also analyse the patterns depending on gender or income of the customer. Multi-domain methods will answer questions such as which goods are usually bought together by rich women in Brussels.</para></listitem>
<listitem>
<para>Temporal pattern mining</para>
<para>When considering a sequence, the interest property can be not only the next item of the sequence, but also the usual gap time before the occurrence of that item. This is particularly interesting for maintenance purposes. If a pattern leads to a failure or defective product, it can be interesting to know how many times it is necessary to prevent that problem and ensure that there is enough time to do so. Indeed, nobody is interested in a pattern than only allows to predict a car&#x02019;s failure 3 seconds before it happens.</para></listitem>
<listitem>
<para>Constrained pattern mining</para>
<para>This topic covers many different methods that allow to impose constraints to the patterns. A constraint can be defined on the length of the patterns, to only retrieve patterns of more than 5 items. For example, directed pattern can search for patterns containing items of interest, e.g., to find patterns containing fine Belgian chocolate. Aggregate constrained pattern mining imposes a constraint on an aggregate of items, e.g., to find patterns where the average price of all items is above 50 euro.</para></listitem>
</itemizedlist>
<para><emphasis role="strong">Strengths and challenges</emphasis></para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Strengths</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem><para>Insights about the sequences of events;</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>Challenges</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem><para>Computation time on large datasets.</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
</section>
</section>
<section class="lev3" id="sec5-4-2-5">
<title>5.4.2.5 Temporal pattern mining</title>
<section class="lev4">
<title>General description</title>
<para>Temporal abstraction is the process of transforming a point time series into a series of state intervals and is widely used in data mining to aggregate multivariate time series into a representation that is easier to analyse. This higher-level representation is equivalent to a smoothing of the data. Instead of a series of points, the values are discretized using an abstraction alphabet &#x003A3; that represents the set of possible value ranges a variable can assume (e.g., &#x003A3; = &#x0007B;High, Medium, Low&#x0007D;). Each time series is then represented by a series of intervals during which the value is constant. Temporal Pattern Mining then searches for (multivariate) temporal patterns that are common among the input samples, where a pattern encodes the temporal ordering between the intervals present in the sample (e.g., &#x0201C;Sensor1.High before Sensor2.Low and Sensor2.Low concurrent with Sensor3.Medium&#x0201D;). Each sample can then be reduced to a binary vector indicating which temporal patterns it contains. The input dimensionality can be further reduced by keeping track only of those patterns, which are sufficiently long and unique. See Batal et al. [2013] for more details.</para>
</section>
<section class="lev4">
<title>Common applications</title>
<para>Temporal pattern mining is applied for mining symbolic patterns in various domains, including natural language processing, bioinformatics (e.g., finding discriminative patterns, sequence alignment).</para>
</section>
<section class="lev4">
<title>Strengths and challenges</title>
<para>Temporal Pattern Mining can be used as a dimensionality reduction technique for multi-dimensional time series, as a pre-processing step for classifying noisy multivariate time series with irregular time intervals between measurements, different granularity across time series, and missing values.</para>
</section>
</section>
<section class="lev3" id="sec5-4-2-6">
<title>5.4.2.6 Principal component analysis</title>
<section class="lev4">
<title>General description</title>
<para>PCA is a space transformation method used to reduce the dimensional space from a multivariate dataset while retaining most of the information (components). This statistic technique converts a set of observations of possible correlated variables into a set of values of linearly uncorrelated variable called principle components by applying a transformation in such a way that the first principal component has the largest possible variance [Pearson, 1901]. The key idea is that if two signals are highly correlated or dependent, one is enough to carry the information.</para>
<para>Two steps are performed in the PCA transformation [Jolliffe, 2002]:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Find principal components (<link linkend="F5-15">Figure <xref linkend="F5-15" remap="5.15"/></link>);</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Compute the covariance matrix &#x003A3; of the dataset <emphasis>X</emphasis>;</para></listitem>
<listitem>
<para>Find <emphasis>P</emphasis> = (<emphasis>n,n</emphasis>) the <emphasis>n</emphasis> eigenvectors <inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg188-1.jpg"/> of $$$<emphasis>&#x003A3;</emphasis> and the corresponding eigenvalues &#x0039B;<subscript>1</subscript>= &#x0039B;<subscript>2</subscript>= = &#x0039B;<emphasis><subscript>n</subscript></emphasis> where <emphasis>P</emphasis> represents an orthonormal basis.</para></listitem></itemizedlist>
</listitem>
<listitem>
<para>PCA approximation by ignoring components with lower significance.</para></listitem>
</itemizedlist>
<fig id="F5-15" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-15">Figure <xref linkend="F5-15" remap="5.15"/></link></label>
<caption><para>Principal components of a 2-dimensional data.</para></caption>
<graphic xlink:href="graphics/ch05_fig016.jpg"/>
</fig>
<para>Important properties of the PCA transformation are 1) that the principal components are uncorrelated since they are orthogonal to each other; 2) the data can be transformed between the bases without loss of information; 3) the principal components are ordered so that the first retain most of the variation in the data.</para>
<para>The PCA can be used for different applications, one of which is fault detection. This requires having historical data of the system to be monitored in a healthy state <emphasis>Xh</emphasis>. For each new input data <emphasis>Xobs</emphasis>, the projection on the eigenvectors is calculated, and then anti-transformed in order to calculate the normalized reconstruction error, which is a measurement for determining whether the input belong to normal or to abnormal conditions, as follows:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Normalize the data so that the mean is zero and standard deviation is one;</para></listitem>
<listitem>
<para>Calculate the principal components <emphasis>P</emphasis> of <emphasis>Xh</emphasis> and retain only the first <emphasis>n</emphasis> components containing a predefined variance (e.g., 95%);</para></listitem>
<listitem>
<para>Reconstruct the input data Xobs by projecting in the transformed basis, keep only the principal components and anti-transform: <emphasis>X<subscript>rec</subscript> = X<subscript>obs</subscript>PP<superscript>T</superscript></emphasis>;</para></listitem>
<listitem>
<para>If <emphasis>X<subscript>rec</subscript></emphasis> <superscript></superscript>= <emphasis>X<subscript>obs</subscript></emphasis> then the input data belong to the normal condition, otherwise it belongs to the abnormal condition. This distance need to be defined based on threshold values.</para></listitem>
</itemizedlist>
</section>
<section class="lev4">
<title>Common applications</title>
<para>PCA is often used as a technique to reduce the amount of information from an original dataset before a classification or a clustering step is applied. Moreover, it can be used for finding patterns in the data and for data compression.</para>
</section>
<section class="lev4">
<title>Strengths and challenges</title>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Some strengths of PCA are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Training time necessary to find the principal components is proportional to the number of measured signals, usually denoted as <emphasis>n</emphasis>, while the execution time is very short so that it is suitable for real-time applications;</para></listitem>
<listitem>
<para>Reduce the dimensionality of the problem, retaining only the relevant information;</para></listitem>
<listitem>
<para>Sound statistical technique;</para></listitem>
<listitem>
<para>Does not require lots of data to be trained, however the data should come from all the possible normal behaviours of the systems.</para></listitem></itemizedlist>
</listitem>
<listitem>
<para>On the other hand, it also has some limitations, the most prominent of them are listed below:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>PCA has low robustness on highly correlated measurements;</para></listitem>
<listitem>
<para>PCA produce unsatisfactory results for dataset characterized by highly nonlinear relationships.</para></listitem></itemizedlist>
</listitem>
</itemizedlist>
</section>
</section>
<section class="lev3" id="sec5-4-2-7">
<title>5.4.2.7 Hidden Semi-Markov model with Bayes classification</title>
<section class="lev4">
<title>General description</title>
<para>HMM is a form of Markov Mixture Model, which contains a hidden and observable stochastic process. The former is a hidden process representing the system states while the latter is a Markov chain of the output of the system known as observations. A complex system (of which we cannot model its internal operation) can be represented by a HMM. A model can be trained with observation sequences which can (using a proper structure and sufficient number of training sequences) estimate the likelihood distribution of states at each transition, e.g., the most likely state of the system.</para>
<para>The Hidden Semi-Markov Model is an extension of HMM with the duration parameter, e.g., we can use sequences where observations do not follow in constant time.</para>
</section>
<section class="lev4">
<title>Common applications</title>
<para>HMMs have been used in speech recognition for many years, using it for failure prediction is quite new. It is a supervised learning approach, so training data is needed. Multiple models need to be trained, at least one for failure sequences and one for non-failure sequences. Based on observations, sequences are assembled and sequence likelihood is calculated on each model. A Bayes classification algorithm is used to detect a failure-prone case and the probability of failure.</para>
</section>
<section class="lev4">
<title>Strengths and challenges</title>
<para>For the successful outcome of using the HMM approach, a proper model structure needs to be selected. The number of states needs to be selected based on the length of the observation sequences. In addition, this length needs to be the same for each sequence. Shorter sequences need to be padded with synthetic symbols.</para>
</section>
</section>
<section class="lev3" id="sec5-4-2-8">
<title>5.4.2.8 Autoencoders</title>
<section class="lev4">
<title>General description</title>
<para>An autoencoder is a neural network specifically designed for learning representation (encodings) from data. It can be used in an unsupervised setting such as dimensionality reduction. It mays also be used for creating generative models, which were exploited in the MANTIS project.</para>
<para>An autoencoder is a standard feedforward multilayer neural network, however, its topology (structure) is characterized as follows:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>The number of output nodes is the same as the inputs in order to allow for signal reconstruction;</para></listitem>
<listitem>
<para>A central layer with the minimum number of nodes is used to encode the data (minimal representation), denoted as the <emphasis role="strong">z</emphasis> layer;</para></listitem>
<listitem>
<para>The connections from the input nodes through 0 or more hidden layers to the z layer are used to encode the signal;</para></listitem>
<listitem>
<para>The connections from the output nodes through 0 or more hidden layers to the z layer are used to decode the signal;</para></listitem>
<listitem>
<para>The network on left (input) and right (output) are usually symmetric.</para></listitem>
</itemizedlist>
<para>The simplest autoencoder network consists of one input layer, a single hidden z layer and one output layer. Autoencoders are trained to minimize the reconstruction error by measuring the difference between the input and the output. One such loss function can be for example the squared error.</para>
<para>It is important to note that the z layer must have a reduced number of nodes, otherwise it will not be able to learn an encoding and simply &#x0201C;memorize&#x0201D; all of the examples. In effect, the z layer should be small enough to <emphasis>compress</emphasis> the data into a reduced set of useful features, although not too small, otherwise the learned features will not be useful. Several variations of these networks have been developed with the aim of finding such useful features, i.e., capturing essential information and find a richer set of features.</para>
<para>One variant is the <emphasis role="strong">denoising encoder</emphasis> that assumes that the input data consists of a stable set of higher level features that are robust to noise. The aim is to extract those high-level features that represent the data distribution robustly. In this variant, the learning procedure first corrupts the signal and uses that as input to the network. Learning follows the usual procedure of loss evaluation and backpropagation of the error. However, it is important to note that the loss function is between the original uncorrupted signal and the network&#x02019;s output signal.</para>
<para>The <emphasis role="strong">sparse autoencoder</emphasis> works under the assumption that by making the hidden layers sparse, it is possible to better learn the structures that are present within the data. Sparsity is imposed by either manipulating the loss function or deactivating some of the hidden nodes with higher activation values (for example using a L<subscript>1</subscript> regularizer). Intuitively, the network will learn to select a subset of features for a given number of inputs that are characterized by those features only.</para>
<para>The <emphasis role="strong">variational autoencoder</emphasis> has its roots in Bayes inference and graphical models. It considers the hidden nodes as a set of latent variables connected in a regular graph (fully connected layers) and whose parameters (weights) are unknown. Variational Bayesian methods are then used to provide an approximation to the posterior probability (weights) of the unobserved variables (hidden nodes), and base their mechanisms on an iterative procedure similar to expectation maximization to determine the solution. More concretely, they use a loss function with additional penalty terms (function includes a term using Kullback&#x02013;Leibler divergence) and trains the network with the Stochastic Gradient Variational Bayes algorithm. Note the posterior probability that is learned can be seen as two conditional probabilities each with its own set of parameters, one for encoding and another for decoding. Moreover, these are encoded separately in the loss function.</para>
<para>The <emphasis role="strong">contractive autoencoder</emphasis> [Rifai et al., 2011] uses a loss function that promotes learning models that are robust to variations of the input. The loss function uses a Frobenius norm of a Jacobian matrix as a regularizer. The partial derivatives of this norm apply only to the activations of the input nodes. This penalty has the effect of ensuring that only significant changes of the input cause changes in the hidden layers. In other words, if the partial derivatives are zero, then feature learning is reduced. The intuition here is that only important changes will promote significant changes in the learned parameters.</para>
</section>
<section class="lev4">
<title>Common applications</title>
<para>The autoencoders ability to encode data and regenerate the input can be taken advantage of in the following way:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Dimensionality reduction: use the z hidden layer as an output for visualization;</para></listitem>
<listitem>
<para>Feature engineering: use the z hidden layer as an output to generate input for other machine learning algorithms (regression, classification);</para></listitem>
<listitem>
<para>Clustering: use the distance between the z hidden layer&#x02019;s outputs measure similarity and difference between input signals (feature engineering for exploratory analysis). The same technique can be used for classification;</para></listitem>
<listitem>
<para><emphasis>Pre-training</emphasis> of Deep Network: some deep neural networks are difficult to train due to the vanishing and exploding gradients. The deep networks can use autoencoders&#x02019; learning process in an initial stage in order to initialize the parameters (weights) that will facilitate learning in the next stage (the use of ReLU as activation functions has made this largely unnecessary);</para></listitem>
<listitem>
<para>One-Class Classification: the error of the regenerated signal can be used to determine if the input signal has any new features that differentiate it from the training dataset;</para></listitem>
<listitem>
<para>Image denoising: robust autoencoders can regenerate noisy input signals by extracting and using only the most salient features that are resistant to noise;</para></listitem>
<listitem>
<para>Natural language processing (NLP): autoencoders have been used to solve problems in word embedding (converting words or sentences to vectors), build bilingual word and phrase representations, document clustering, sentiment analysis and paraphrase detection.</para></listitem>
</itemizedlist>
</section>
<section class="lev4">
<title>Strengths and challenges</title>
<para>The main difficulty in using the autoencoder was the set up of the hyper-parameters and the topology. A simple grid search was performed. We found that a simple 3 layers network was effective (the input and output layers consisted of 32 nodes). We could successfully detect changes in the normal distribution&#x02019;s standard deviation (z layer with 16 nodes) and the mean and/or standard deviation (z layer with 25 nodes).</para>
<para><emphasis role="strong">Advantages</emphasis></para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>No need to understand the data;</para></listitem>
<listitem>
<para>Automated feature engineering (data transformation);</para></listitem>
<listitem>
<para>Reduced pre-processing required (except for normalization or scaling).</para></listitem>
</itemizedlist>
<para><emphasis role="strong">Disadvantages</emphasis></para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Requires large datasets of relevant high quality data;</para></listitem>
<listitem>
<para>Requires large amounts of CPU time and memory;</para></listitem>
<listitem>
<para>Many hyper-parameters to tune;</para></listitem>
<listitem>
<para>Difficult to select the most appropriate loss function and network topology;</para></listitem>
<listitem>
<para>Model validation is difficult (requires correct data labelling);</para></listitem>
<listitem>
<para>Training is difficult if the data is not representative of the problem (for example, non-failure data contaminated with failures);</para></listitem>
<listitem>
<para>May learn irrelevant features (only detectable during model validation). One further issue regards how to bias the network to learn those important features;</para></listitem>
<listitem>
<para>Features are not interpretable.</para></listitem>
</itemizedlist>
</section>
<section class="lev4">
<title>Use in MANTIS</title>
<para>The technique described above can be used to detect anomalies that may occur in a press brake machine. Signals indicating the movement (velocity and distance) of the press ram and back gauge, oil temperature and vibration can be used to detect failure.</para>
<para>Denoising encoders allows us to encode and regenerate the signals that represent the machine with no failure. To test for failure, the signals is encoded, and it is verified that its reconstruction is within a given bound. If it is not, it can be concluded that the signal signature has not been encountered before (because it cannot be decoded successfully) and it most probably represents a failure. Unlike the statistical test, we are not testing if a given sample belongs to the same population. We are in effect testing if the &#x0201C;curve&#x0201D; that describes the displacement of the ram or back-gauges have the expected &#x0201C;shape&#x0201D;. Samples of non-failures &#x0201C;curves&#x0201D; are the dominant class that were used during an initial training phase. Samples of failures &#x0201C;curves&#x0201D; can then be used to establish the cut-off threshold between the fail and no fail classes during a second phase of learning. This threshold is then used again for model evaluation.</para>
<para>Due to a lack of data, synthetic data were used for the tests, generated using a normal distribution by varying either the mean, standard deviation or both. The autoencoder was trained for datasets with a single set of known statistics (no failure). The autoencoder was then used to reconstruct the signal for data with a different set of statistics (with and without failure labels). During the learning phase, we determined the minimum square error threshold that was necessary to reduce false positives. During a second learning phase, signals with different statistics were used to determine the minimum square error threshold that was necessary to reduce false negatives. The final threshold was then set to the mean of these two separate thresholds.</para>
<para>Test sets were then used to regenerate the input signal and calculate the reconstruction error (containing both failure and non-failure signals). If this error was below the threshold no failure was assumed, otherwise the data was labeled as a failure. The <emphasis>Mathews correlation coefficient</emphasis> was then used to evaluate the performance of the model. The <emphasis>Mathews correlation coefficient</emphasis> emerged from the experiment was evaluated as 1 (perfect score).</para>
</section>
</section>
<section class="lev3" id="sec5-4-2-9">
<title>5.4.2.9 Convolutional neural network with Gramian angular fields</title>
<section class="lev4">
<title>General description</title>
<para>According to a survey from 2014 [Schmidhuber, 2015], one of the currently most successful deep learning methods uses RNNs with LSTM or CNNs with max-pooling. This technique description will focus on the latter one.</para>
<fig id="F5-16" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-16">Figure <xref linkend="F5-16" remap="5.16"/></link></label>
<caption><para>Detection of the presence or absence of a 3 &#x000D7; 3 square in the input image by a Convolutional Neural Network.</para></caption>
<graphic xlink:href="graphics/ch05_fig017.jpg"/>
</fig>
<para>In image classification, the goal is to recognize (classify) what is shown in a given input image. Typically, an image is encoded as a matrix with values between 0 and 1 or 0 and 255, along with the correct label (i.e., what is shown in the image). Colour images are typically encoded as 3-channel (red, green, blue) images. Several benchmark datasets are available in this domain, including MNIST<footnote id="fn5_1" label="1"><para>Y. LeCun, C. Cortes, and C. J.C. Burges. The MNIST database of handwritten digits. http://yann.lecun.com/exdb/mnist/</para></footnote> (grayscale images of handwritten digits), ImageNet [Deng et al., 2009] (colour images of objects, animals and scenes) and CIFAR<footnote id="fn5_2" label="2"><para>A. Krizhevsky. CIFAR-10 and CIFAR-100 datasets. https://www.cs.toronto.edu/kriz/cifar.html</para></footnote> (colour images of vehicles and animals), to name a few.</para>
<para>In essence, CNNs rely on three basic concepts, which are described below and illustrated graphically in <link linkend="F5-16">Figure <xref linkend="F5-16" remap="5.16"/></link>. To simplify discussions, we assume that the input consists of 1-channel images. Note that in a CNN, the input image is treated as an h &#x000D7; h square, with h the number of pixels (or: the height and width of the image). Hence, the input layer of a CNN contains h &#x000D7; h input neurons, where each neuron contains the intensity of one pixel of the image.</para>
<para>Each neuron in the first layer after the input layer, known as the convolutional layer, is only allowed to connect to a small part of the input layer. In particular, each neuron in the latter layer connects to a small square of the input neurons (e.g., 3 &#x000D7; 3), known as the local receptive field. In this work, the size of this square region is denoted as <emphasis>recepsize</emphasis>. Each hidden neuron will learn a weight over the connections from its local receptive field, along with a bias. Intuitively, each neuron in the convolution layer thus learns to analyse the data in its local receptive field.</para>
<para>In <link linkend="F5-16">Figure <xref linkend="F5-16" remap="5.16"/></link>a, a 10 &#x000D7; 10 pixel input image is shown (i.e., h = 10). The neurons of the next convolution layer use local receptive fields of <emphasis>recepsize</emphasis> = 3 to connect to a small square region of the input. The first convolution neuron (top-left neuron first row, light orange circle) connects to the top left receptive field (light orange rectangle) of the input. For the next neuron, the receptive field is shifted by 1 pixel horizontally, obtaining the receptive field for the second convolution neuron (dark orange). This procedure repeats itself, up to the borders of the input image. For the first neuron on the second row of the convolution layer (brown circle), the local receptive field of the first neuron in the first row is shifted by 1 pixel vertically (brown rectangle). The amount of pixels shifted (either horizontally or vertically) to obtain local receptive fields for convolution neurons is known as the stride length, denoted <emphasis>stridelen</emphasis>. Subsequent horizontal or vertical local receptive fields therefore always overlap by <emphasis>recepsize</emphasis> x <emphasis>stridelen</emphasis> pixels from the input.</para>
<para>The number of pixels that remain of an h pixel image after applying convolution with receptive fields of size <emphasis>recepsize</emphasis> and stride length <emphasis>stridelen</emphasis> can be computed according to the following formula:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-16.jpg"/></para>
<para>In <link linkend="F5-16">Figure <xref linkend="F5-16" remap="5.16"/></link>a, h=10, <emphasis>recepsize=3</emphasis> and <emphasis>stridelen=1</emphasis> resulting in a <inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg196-1.jpg"/> pixel image. In order to match this reduced image size, there are 8 &#x000D7; 8 convolution neurons in the subsequent layer.</para>
<para><emphasis>Shared weights and biases</emphasis></para>
<para>In a CNN, all the weights and biases of the neurons in the convolutional layer are shared. This allows the network to detect a particular feature in the input image at any position; this property is also called <emphasis>translation invariance</emphasis>. Note that the term feature in the context of CNNs has a different meaning than the feature from feature selection methods. In the former context, a feature typically represents a certain part of an image (such as a horizontal line). In the latter context, it refers to a piece of information in a dataset (e.g., a column or a combination of columns) that can be used for classification, to distinguish failures from non-failures.</para>
<para>The weights and biases from the input to the convolutional layer are often referred to as a feature map; a CNN usually has multiple feature maps to learn different features in the input image (e.g., horizontal, vertical and diagonal lines). In this work, the number of feature maps is denoted as <emphasis>nrmaps</emphasis>. In <link linkend="F5-16">Figure <xref linkend="F5-16" remap="5.16"/></link>a, three feature maps (<emphasis>nrmaps</emphasis> = 3) are used to analyse the 10 &#x000D7; 10 input image, where each of the neurons in the other feature maps has the same connectivity pattern of the top feature map shown here.</para>
<para><emphasis>Pooling</emphasis></para>
<para>The objective of pooling is to simplify or summarize the information from a convolutional layer in a small square region (e.g., 2 &#x000D7; 2), for a particular feature map. The size of the pooling region is also known as the pooling size, denoted <emphasis>poolsize</emphasis>. Different types of pooling exist, but the most common form is max-pooling, where a pooling unit outputs the maximum activation in its pooling region. Pooling is a nice way to reduce the number of parameters of a CNN, because it reduces the size of the remaining input image (after applying convolution). In particular, if the size of an image after applying convolution is <emphasis>convsize</emphasis>, then this is reduced to a <inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg197-1.jpg"/> image after max-pooling.</para>
<para>Max-pooling is illustrated in <link linkend="F5-16">Figure <xref linkend="F5-16" remap="5.16"/></link>b. Each neuron in the pooling layer uses the convolution layer from <link linkend="F5-16">Figure <xref linkend="F5-16" remap="5.16"/></link>a as input. The first neuron in the pooling layer (light orange circle) connects to a small square pooling region of size 2 (light orange square). For the second neuron (dark orange circle), this square is shifted horizontally by the pooling size, obtaining the pooling region for this neuron (dark orange square). This procedure is repeated up to the borders of the input. For the first pooling neuron on the second row (dark brown circle), the square from the first pooling neuron of the first row (light orange) is shifted vertically by the pooling size to obtain a new pooling region (dark brown square). As in convolution, pooling is done per feature map separately and so there are again 3 feature maps in <link linkend="F5-16">Figure <xref linkend="F5-16" remap="5.16"/></link>b that follow the same connectivity pattern. Unlike convolution however, there is no overlap in subsequent horizontal or vertical (pooling) regions. According to the calculations from above, <emphasis>convsize</emphasis> = 8 in <link linkend="F5-16">Figure <xref linkend="F5-16" remap="5.16"/></link>a after convolution. With <emphasis>poolsize</emphasis> = 2, this image is further reduced to <inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg197-2.jpg"/> after applying pooling; there are therefore 4 &#x000D7; 4 pooling neurons in <link linkend="F5-16">Figure <xref linkend="F5-16" remap="5.16"/></link>b.</para>
<para>Because of these three basic concepts, CNNs are powerful networks that can learn useful spatial representations of images and at the same time require relatively few parameters when compared to other types of networks. Generally speaking, if a neural network has few parameters, it also has a shorter training time. How well a neural network performs is not only determined by its test performance, but also by the time needed to train the network. After all, a network that performs very well (e.g., achieves over 90% accuracy) that takes weeks (or even months) to train is not very useful.</para>
</section>
<section class="lev4">
<title>Common applications</title>
<para>CNNs are typically used to perform classification tasks on images, for instance on the ImageNet dataset [Deng et al., 2009]. Images are not time series however and in some approaches [Wang and Oates, 2015; Bashivan et al., 2015], a transformation technique is employed to transform the raw time series data into a sequence of images, before providing it as data to some kind of CNN. Other approaches make use of EEG, video or log data to do time series classification with CNNs, for instance in the context of making predictions in the medical domain [Hajinoroozi et al., 2016], human action recognition [Ijjina and Chalavadi, 2016], automatic annotation of clinical texts [Hasan et al., 2016] and automatic identification of predatory conversations in chat logs [Ebrahimi et al., 2016].</para>
</section>
<section class="lev4">
<title>Strengths and challenges</title>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Some strengths of CNN are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>CNN is one of the strongest algorithms with a high accuracy for image recognition;</para></listitem>
<listitem>
<para>CNN can automatically extract high-level features from the raw data that are suitable for solving classification or regression problems, without relying on features designed manually by an expert;</para></listitem>
<listitem>
<para>CNN are robust for distortions in the image caused by the camera lens; different lighting conditions, horizontal and vertical shift, etc. [Hijazi et al., 2015];</para></listitem>
<listitem>
<para>Convolving filters over the input vector allow to reduce the number of trained parameters, thus reducing the resource requirements for training and storing CNN models.</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>On the other hand, the usage of CNN gives rise to some challenges:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>The algorithm requires huge amounts of data. Most problems using neural nets require somewhere around 10.000 examples, but preferably more. To collect and label these examples can be very time consuming;</para></listitem>
<listitem>
<para>CNN and artificial neural networks in general are very computationally expensive to train. Due to the advancements in GPUs in the last decade, it is now viable to train neural networks;</para></listitem>
<listitem>
<para>While visualizing the feature maps can provide some insight into the features that a CNN extracts in the intermediate layers, the inclusion of dense layers makes it difficult to interpret how the neural network model produces its prediction and to trust the result. For some domains, e.g., medical and fraud investigation, this is unacceptable;</para></listitem>
<listitem>
<para>The design of CNN architectures for solving a particular machine-learning task remains an art.</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
</section>
</section>
<section class="lev3" id="sec5-4-2-10">
<title>5.4.2.10 Recurrent neural network with long-short-term memory</title>
<section class="lev4">
<title>General description</title>
<para>According to a survey from 2014 [Schmidhuber, 2015], some of the most successful deep learning methods nowadays use Recurrent Neural Networks RNNs with LSTM or CNNs with max-pooling. This technique description focuses on RNN with LSTM.</para>
<para>In a RNN [Zachary et al., 2015], both forward and recurrent edges to neurons are allowed. Recurrent edges may form cycles in a network, including a cycle of length 1 (known as a self-connection or self-loop). Because of the recurrent connections, a certain neuron now not only receives input from the current data at time t, but also from the previous hidden state at time t - 1. The recurrent edges thus introduce a notion of time in RNNs and as such, they are naturally used to model sequences (including time series). An example of a time series classification benchmark dataset is the University of California, Riverside (UCR) repository [Chen et al., 2015]. Although RNNs are very powerful networks, they often suffer from either the vanishing gradients or exploding gradients problem. In the former problem, the gradients (used to update the weights in the various layers) grow exponentially smaller as they are propagated backward through the hidden layers. In the latter problem, the opposite occurs, and the gradients grow exponentially larger. Both problems cause learning to considerably slowdown in RNNs. To overcome this problem, a special type of neuron or memory cell has been introduced, known as a LSTM cell [Gers et al., 2000]. This memory cell provides a form of intermediate storage in between the long-term memory stored in the weights of an RNN (which tend to change slowly over time during training) and the short-term memory stored in the activations passed from one neuron to another (which can change much faster). A simplified overview of an LSTM is shown in <link linkend="F5-17">Figure <xref linkend="F5-17" remap="5.17"/></link>a. The idea of the LSTM cell is to allow a network to hold a value for a longer period of time, using a special internal state that has a self-loop with weight 1. Both the flow in and out of the cell are controlled via gates, as illustrated in <link linkend="F5-17">Figure <xref linkend="F5-17" remap="5.17"/></link>a (gates are shown in gray). If the input gate is set to high (<link linkend="F5-17">Figure <xref linkend="F5-17" remap="5.17"/></link>b), data from the rest of the network (e.g., a value) can flow into the memory cell and change the internal state. If the input gate is set to low (<link linkend="F5-17">Figure <xref linkend="F5-17" remap="5.17"/></link>c), no data can flow into the cell. A similar mechanism is used to control the output of the memory cell, through the use of an output gate (<link linkend="F5-17">Figure <xref linkend="F5-17" remap="5.17"/></link>d and e). The network can also &#x02018;flush&#x02019; (remove) the contents of the memory cell using a special forget gate. In particular, if the forget gate is set to high (<link linkend="F5-17">Figure <xref linkend="F5-17" remap="5.17"/></link>f), the previous state is maintained; if the forget gate is set to low (<link linkend="F5-17">Figure <xref linkend="F5-17" remap="5.17"/></link>g), the content of the memory cell is removed (forgotten). During training, the RNN learns when it is appropriate to remember or forget data using the LSTM cells.</para>
<fig id="F5-17" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-17">Figure <xref linkend="F5-17" remap="5.17"/></link></label>
<caption><para>Illustration of the components of a LSTM cell.</para></caption>
<graphic xlink:href="graphics/ch05_fig018.jpg"/>
</fig>
<para>In an RNN, the neurons in the hidden layers (which typically have <emphasis>sigmoid</emphasis> or <emphasis>tanh</emphasis> activation functions) are replaced by LSTMs, obtaining an entire layer with LSTM cells. Using these cells, an RNN can model long sequences reliably without suffering from the vanishing or exploding gradients problems.</para>
</section>
<section class="lev4">
<title>Common applications</title>
<para>RNNs are naturally used to model time series and have been applied successfully in many application domains, including video description [Yuea et al., 2016], speech recognition [Li and Wu, 2015; Li et al., 2016; Cai and Liu, 2016], computer vision [Chen et al., 2016] and diagnosing medical EEG data [Lipton et al., 2016].</para>
</section>
<section class="lev4">
<title>Strengths and challenges</title>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Some strengths of RNN in combination with LSTM are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Deep learning methods, such as RNN, have the ability to learn and model non-linear and complex relationships, which is a major advantage since in practice many of the relationships between inputs and outputs are non-linear as well as complex;</para></listitem>
<listitem>
<para>Recurrent Neural Networks are suitable for problems that require sequential processing of information;</para></listitem>
<listitem>
<para>LSTM models are powerful to learn and remember the most important past behaviours, and understand whether those past behaviours are important features in making future predictions;</para></listitem>
<listitem>
<para>The recurrent connections effectively allow RNNs to share parameters and thus capture long time dependencies using simpler models than deep feed-forward neural networks, making them suitable for cases with smaller data sets.</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>On the other hand, the usage of RNN with LSTM provides some challenges:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>RNN and artificial neural networks in general are very computationally expensive to train. Moreover, RNN are more difficult to parallelize (e.g., using GPUs) than feed-forward networks, due to the sequential dependencies between the states;</para></listitem>
<listitem>
<para>It is difficult to interpret how the neural network model produces its prediction and to trust the result. For some domains, e.g., medical and fraud investigation, this is unacceptable.</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
</section>
</section>
<section class="lev3" id="sec5-4-2-11">
<title>5.4.2.11 Change detection algorithm</title>
<section class="lev4">
<title>General description</title>
<para>Change detection is the problem of finding abrupt changes in data when the probability distribution of a stochastic process or time series changes in order to determine the fault early enough and to provide an early warning alarm [Aminikhanghahi and Cook, 2017].</para>
<para>Change detection algorithm is therefore not a single algorithm but a family of algorithms that can be categorised as follows [Gustafsson, 2000]:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Methods using one filter, where a whiteness test is applied to the residual</para></listitem>
<listitem>
<para>Methods using two or more filters in parallel, each one following a specific assumption for the change to be detected (<link linkend="F5-18">Figure <xref linkend="F5-18" remap="5.18"/></link>).</para></listitem>
</itemizedlist>
<fig id="F5-18" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-18">Figure <xref linkend="F5-18" remap="5.18"/></link></label>
<caption><para>Filter that map the relation between inputs u<emphasis><subscript>t</subscript></emphasis> and output y<emphasis><subscript>t</subscript></emphasis> and generate residuals &#x1D700;<subscript><emphasis>t</emphasis></subscript>.</para></caption>
<graphic xlink:href="graphics/ch05_fig019.jpg"/>
</fig>
<para>A filter in system identification is a mathematical model that maps the relation between the input and output of the system and returns a sequence of residuals. Which filter to be used is a research topic by itself and it depends on the problem under study [Ljung, 1999] and does not matter from a change detection perspective [Gustafsson, 2000]. Examples of filters can be Kalman filter, kernel, parametric models, state space models and physical models.</para>
<para>The basic idea of a change detection system is that if there is no change in the system and the model is correct, given the input and output variables, the result is a sequence of independent stochastic variables with zero mean and know variance. When a change occurs in the system, the model cannot capture anymore the behaviour of the system and this affects the residuals. The essential step now is to define a rule of thumb to indicate whether the residuals are too large and an alarm needs to be generated or not.</para>
<para>This can be formulated into the problem of deciding between two hypotheses where a stopping rule is achieved by low-pass filtering of <emphasis>s<subscript>t</subscript></emphasis> and comparing the value with a threshold [Gustafsson, 2000]:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-17.jpg"/></para>
<para>Below two examples of common rules are given [Gustafsson, 2000]:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>CUMSUM is a sequential analysis technique [Barnard, 1959]. It implies calculating the cumulative sum from the residuals of the filter as follows:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-18.jpg"/></para>
</listitem>
</itemizedlist>
<para>When the value of S exceeds a certain threshold value, a change in the timeseries has been found where <emphasis>v</emphasis> represents the likely-hood function.</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Geometric Moving Average:</para>
<para>The Geometric Moving Average calculates the geometric mean of the previous N values in the time series as follows:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq5-19.jpg"/></para>
<para>Where <emphasis>&#x003BB;</emphasis> introduces a forgetting factor to account for the slow timevarying behaviour (i.e., degradation) of the system. When the value of S exceeds a certain threshold value, a change in the time-series has been detected.</para></listitem>
</itemizedlist>
<para>Therefore, a change detection algorithm based on time-series requires the definition of the model to represent a system and a rule to define when the model does not match anymore the behaviour of the system and therefore an alarm must be triggered.</para>
</section>
<section class="lev4">
<title>Common applications</title>
<para>Change detection has been used in a number of applications [Aminikhanghahi and Cook, 2017] from stock analysis, to condition based monitoring, to speech detection and image analysis. Moreover, it has been used for detecting climate change and human activity analysis. Generally, it can be used for any process in which time series are available and a model representing the system can be defined.</para>
</section>
<section class="lev4">
<title>Strengths and challenges</title>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Some strengths of change detection algorithms are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Expert knowledge can be included in the model;</para></listitem>
<listitem>
<para>Fast technique that can be used also on embedded devices in real-time;</para></listitem>
<listitem>
<para>Do not require historical data;</para></listitem>
<listitem>
<para>Can track dynamic change of the behaviour of the system if timevarying parameters are used.</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>On the other hand, it also has the limitations listed below:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Requires defining or identifying a model of the system of interest;</para></listitem>
<listitem>
<para>Selection of the stopping rule is not trivial and can vary depending on the system of interest.</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
</section>
</section>
<section class="lev3" id="sec5-4-2-12">
<title>5.4.2.12 Fisher&#x02019;s exact test</title>
<section class="lev4">
<title>General description</title>
<para>Fisher&#x02019;s exact test<footnote id="fn5_3" label="3"><para>https://en.wikipedia.org/wiki/Fisher%27s_exact test</para></footnote> is a statistical significance test used in the analysis of contingency tables. It is in particular suitable for small sample sizes, where an <emphasis>&#x003C7;</emphasis>-squared test is less appropriate or would yield unreliable results. For a given contingency table, its <emphasis>p</emphasis>-value is given by the sum of all probabilities that are at most the probability of the given contingency table of any other contingency table that is realizable with the original row and column sums in the given contingency table [Fisher, 1922, 1954; Agresti, 1992]</para>
</section>
<section class="lev4">
<title>Common applications</title>
<para>The test is useful for categorical data that result from classifying objects in two different ways. As such, it has a broad applicability in many domains. A typical example is where one classification constitutes a so-called <emphasis>ground truth</emphasis>, of which the validity is beyond doubt, and the other consists of the output of a machine-learning algorithm. In this case, one normally speaks of a confusion matrix instead of a contingency table.</para>
</section>
<section class="lev4">
<title>Strengths and challenges</title>
<para>Fisher&#x02019;s exact test is particularly suitable when the sample size is small. Also, the fact that it is a nonparametric test is a strength, as no assumptions have to made about the shape of the distributions. It is, however, based on the assumption that the row- and sum-totals of the contingency table are fixed, which may be problematic in practice. In the current context, this assumption could be defended. Where, in the social sciences, <emphasis>p</emphasis>-values in the order of 0.01 or 0.001 are common, in the current context, values in the order of 10<superscript>-7</superscript>and even orders of magnitude smaller were often encountered. This has to do with the fact that, in a given dataset, there are many calls where neither a given part is replaced, nor the log pattern occurs. The latter, by the way, demanded an implementation of the binomial coefficients using logarithms and additions instead of multiplications.</para>
</section>
</section>
<section class="lev3" id="sec5-4-2-13">
<title>5.4.2.13 Bonferroni correction</title>
<section class="lev4">
<title>General description</title>
<para>Statistical hypothesis testing is based on rejecting a null hypothesis if the likelihood of the observed data under the null hypotheses is low. If multiple hypotheses are tested, the chance of a rare event increases, and therefore, the likelihood of incorrectly rejecting a null hypothesis (i.e., making a Type I error) increases. The Bonferroni correction provides a remedy for this. For a given confidence level <emphasis>&#x003B1;</emphasis>and <emphasis>m</emphasis> hypotheses, by rejecting a hypothesis if its <emphasis>p</emphasis>value exceeds <emphasis>&#x003B1;/m</emphasis>, the so-called familywise error rate, i.e., the probability of making at least one type-I error is bounded from above by <emphasis>&#x003B1;</emphasis> [Mittelhammer et al., 2000] (https://en.wikipedia.org/wiki/Bonferroni correction).</para>
</section>
<section class="lev4">
<title>Common applications</title>
<para>As this technique is quite general, it is broadly applicable to cases where statistical hypothesis testing is done repeatedly. In the context of alerting and predicting failures, it is one of the key methods by which log patterns are selected.</para>
</section>
<section class="lev4">
<title>Strengths and challenges</title>
<para>Its broad applicability and simplicity as well as being non-parametric is an important strength. One of its weaknesses is that it may be overly conservative if there is a large number of tests. The correction comes at the cost of increasing the probability of accepting false hypotheses, i.e., missing good log patterns.</para>
</section>
</section>
<section class="lev3" id="sec5-4-2-14">
<title>5.4.2.14 Hypothesis testing using univariate parametric statistics</title>
<section class="lev4">
<title>General description</title>
<para>Statistical hypothesis testing allows one to compare two processes by comparing the distributions generated by the random variables that describe those processes [Stuart et al., 1999]. A distribution of a given process may be described by one or more random variables. Univariate statistics refers to the use of statistical models that use only one variable to describe the process. Several univariate statistics can be extended to cater for multi-variate statistics that use two or more random variables (for example, the 2 variable Gaussian processes using the correlation matrix instead of the mean and deviation).</para>
<para>There are two basic methods used in statistical inference. The first establishes a null hypothesis that stipulates the no differences between two processes (distributions) exist. Inference therefore consists in testing the hypothesis and rejecting the null hypothesis. For example, given the heights of men and women we can test whether or not the mean heights are different. The null hypothesis states that no such difference exists. If we reject this null hypothesis, then the alternative hypothesis, that states that there is a difference, is assumed to be true.</para>
<para>The second consists of modelling each process (distribution) with a set of possible statistical models. Each model is a parametric mathematical equation that describes a possible hypothesis (given a set of random input variables, it outputs a frequency). Inference then consists of evaluating and selecting the most appropriate model. The models are selected according to a specific criterion such as the Akaike information criterion or the Bayes factor [Burnham and Anderson, 2002].</para>
<para><emphasis role="strong">Parametric statistics</emphasis> refers to the use of parametric mathematical models of distributions. Examples of these distributions include both continuous (Uniform, Gaussian, Beta, Exponential, Weibull, etc.) and discrete distributions (Uniform, Binomial, Poisson, Bernoulli, etc.). Statistical inference requires that we estimate the parameters and use this directly or indirectly to compare the distributions.</para>
<para>When performing statistical inference, statistical significance is used to determine how likely it is to assume that the null hypothesis is not rejected under the current circumstances. The statistical significance level establishes a threshold, under which a null hypothesis rejection cannot be accepted. Note that, in addition to this threshold, additional limits on the type 1 (incorrect rejection of a true null hypothesis &#x02013; false positive) and type 2 errors (incorrect retaining a false null hypothesis &#x02013; false negative) can be set.</para>
<para>As a title of example, let us consider the p-value and establish the significance level of 0.05 (could be as low as 0.02 or lower). If the p-value of the statistical test is below the threshold, it is highly unlikely (less than 5% of the time) that such a distribution can be observed, and the null hypothesis is true. In this case, the null hypothesis can be rejected. Note however that if it fails, there is not enough information to reject the null hypothesis. Therefore, the test is inconclusive.</para>
<para>The <emphasis role="strong">testing process</emphasis> is as follows:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Select the null and alternative hypothesis;</para></listitem>
<listitem>
<para>Establish the assumption under which the test will be made (type of distribution);</para></listitem>
<listitem>
<para>Select the significance level to use as the threshold;</para></listitem>
<listitem>
<para>Select the test statistic to use and calculate it (single sample, two samples, pared test, etc.);</para></listitem>
<listitem>
<para>Compare the &#x0201C;t&#x0201D; statistics with the p-value threshold and decide if it can be rejected or not.</para></listitem>
</itemizedlist>
<para>Note that test statistic &#x0201C;t&#x0201D; can be calculated from the observations and compared with the p-value, or it is possible to use the statistical tests to compare the distributions&#x02019; samples directly. In either case a &#x0201C;t&#x0201D; statistic that is comparable to the p-value is obtained.</para>
<para>The naive Gaussian parametric tests that are used are of two types: one checks for differences in the distribution mean and another one is used to test differences in variance. The variance tests allow us to check that a sensor signal is within a given threshold of a constant (zero for example).</para>
<para>In this case, the t-Test was performed. More concretely, the <emphasis role="strong">t-test</emphasis> was performed by means of a two-sample location test of the null hypothesis that states that the means of the populations are equal. The independent two sample t-test calculates the t statistic based on the number of samples and the means and variances of each population. The t statistic led to a 2-sided p value that is then used to compare against the threshold that allows either accepting or rejecting the null hypothesis.</para>
</section>
<section class="lev4">
<title>Common applications</title>
<para>Statistical tests allow to analyse data collections when no prior theory exists that can be used to predict relationships. In such cases, one can design and implement experiments to evaluate if a hypothesis is likely to be true. In these experiments, data is collected under the conditions of the null hypothesis and the alternate hypothesis under test. For example, to test the effectiveness of a certain pharmacological treatment two data sets are collected, one where no treatment is applied and one wherein it is (note that in the clinical cases usually a third dataset that uses placebos is also used). The effectiveness of the treatment can then be established by testing if there is a statistically significant difference between the two distributions.</para>
<para>Other examples include [Larsen and Stroup, 1976]:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Determining the effect of the full moon on behaviour (null hypothesis: no full moon);</para></listitem>
<listitem>
<para>Establishing the range at which bats can detect an insect by echo (null hypothesis is the mean detection range);</para></listitem>
<listitem>
<para>Deciding whether or not carpeting results in more infections or not (null hypothesis: no carpets).</para></listitem>
</itemizedlist>
<para>Note that depending on the hypothesis that is under test and the experiment set-up, it is either possible to use a single distribution (for example the bat&#x02019;s echo location range) or two (for example testing whether or not carpets cause infections).</para>
</section>
<section class="lev4">
<title>Strengths and challenges</title>
<para>The following is a list of general advantages and disadvantages of the parametric tests.</para>
<para><emphasis role="strong">Advantages:</emphasis></para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Simple process;</para></listitem>
<listitem>
<para>Easy to calculate;</para></listitem>
<listitem>
<para>Provides additional information (confidence intervals);</para></listitem>
<listitem>
<para>Some models can be learned (updated) and tested as streams of data (Gaussian model);</para></listitem>
<listitem>
<para>Can deal with skewed data;</para></listitem>
<listitem>
<para>Greater statistical power;</para></listitem>
<listitem>
<para>Robust to samples with different variability (2-sample t-test or one-way ANOVA).</para></listitem>
</itemizedlist>
<para><emphasis role="strong">Disadvantages:</emphasis></para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Assumes the type of distribution is known and does not change with time;</para></listitem>
<listitem>
<para>If the null hypothesis is not rejected, no conclusion can be reached;</para></listitem>
<listitem>
<para>Requires large data sets;</para></listitem>
<listitem>
<para>Lack of robustness (sensitive to outliers, sensitive to comparison point<footnote id="fn5_4" label="4"><para>For example, when comparing two Gaussian curves, because the tails fall off at an exponential rate, the difference in distributions is small when close to the mean but very large close to the tail ends.</para></footnote>);</para></listitem>
<listitem>
<para>Assume all samples are i.i.d (Independent and identically distributed random variable);</para></listitem>
<listitem>
<para>Experiments may produce unexpected effects resulting in failed tests (type 1 and 2);</para></listitem>
<listitem>
<para>Each test has its own assumption (example t-Test of independent means assumes populations of equal variance).</para></listitem>
</itemizedlist>
</section>
<section class="lev4">
<title>Use in MANTIS</title>
<para>The technique described above was used to detect anomalies that may occur in a press brake machine. Signals indicating the velocity and distance of the press ram and back gauge, oil temperature and vibration were used as the univariate variables.</para>
<para>The parametrical statistical tests were used to determine whether new samples conform to the same distribution of previously sampled machine states that are known to be failure free. If the distributions are not equivalent, (null hypothesis rejected) then it was assumed that a failure occurred and alerts are generated.</para>
<para>In the case of the formal statistical tests, if the p-value does not allow us to reject the null hypothesis, it is not possible to infer that the machine has no failure (type 2 error). However, for simplicity, it was assumed that it was true and alerts were only sent when the null hypothesis was successfully rejected.</para>
<para>In the case of a test for a constant (for example position), the signal is normalized prior modelling and testing (a machine may stop its ram at any position, and the interest regards only knowing if it is stopped, and not were). In addition to this, two signals (for example the rams speed at the two different pneumatic pistons) are subtracted one from another. The result is a single sample whose means should be 0 (no need for normalization). The statistical test is then performed as is done for the first case.</para>
<para>Parametric Gaussian tests were performed on the mean (<emphasis role="strong">t-Test</emphasis>) and deviation (directly compare deviations, F-test, Bonett&#x02019;s test and Levene&#x02019;s test).</para>
<para>The test for a difference in the mean and deviation were also done using the naive statistical test. In this case, an online Gaussian model is obtained via the calculation of a mean and variance. These means and variances are then directly compared to the continually sampled signals from a working machine. If significant divergence is found, alerts are generated. Due to the high false positive and false negative rates (type 1 and type 2 respectively), an additional multiplicative threshold (in respect to one standard deviation) is used when comparing deviations. The initial values of this threshold are set automatically by selecting the lowest possible threshold that reduces type 1 errors.</para>
<para>Due to lack of data, initial tests were performed with artificial data. For large samples the effectiveness of the t-test appeared to be reduced (increased number of false positives and false negatives), and it was also concluded that the naive statistical tests using simple means and variance are very effective. The naive statistical tests were used on a sample of real data first before committing to the more complex t-test. The latest tests on the real data seem to show that the t-test is ineffective for some of the signals. Work in progress is still trying to ascertain why this is so (noise, incorrect labelling of failure, non-normal distribution, variance does not follow the Chi-square distribution, etc).</para>
</section>
</section>
<section class="lev3" id="sec5-4-2-15">
<title>5.4.2.15 Hypothesis testing using univariate non-parametric statistics</title>
<section class="lev4">
<title>General description</title>
<para>As with the case of parametric statistics (see hypothesis testing using univariate parametric statistics in Section 5.4.2.14), the process of defining the hypothesis, sampling data and establishing a significance level as a threshold are the same. However, not all of the non-parametric methods provide a p-value for a significance level comparison. This section describes three types of statistical tests that were used in project MANTIS: the Kolmogorov-Smirnov test (K-S test), the Man-Whitney U test and the use of a Kernel Density Estimation. In the first case, the U statistic (as performed by the software library) is approximated by a normal distribution and a p-value is available. In this case, the p-value is used to establish a threshold to accept or reject the null hypothesis (distributions are the same).</para>
<para>In the case of the KDE, an online algorithm was used that generates a configurable number of (Gaussian) kernels. The kernels and respective parameters of two different distributions cannot be directly compared. Experimentation shows that the estimated densities may be visually very similar, but the kernels themselves differ significantly. However, since the kernel can be used to sample the underlying estimated distribution, the statistical tests were used to compare the samples (for both the parametric cases and nonparametric cases using the Kolmogorov-Smirnov test and Mann-Whitney U test). Another option could have been the use of alternate algorithms such as the earth mover&#x02019;s distance, but they were not applied to the case at hand because the naive parametric tests seemed to be working well [Levina and Bickel, 2001].</para>
</section>
<section class="lev4">
<title>Kolmogorov-Smirnoff test</title>
<para>This test compares an empirical distribution function (estimate of the CDF) of a sample to a cumulative distribution of a reference distribution. The two sample K-S test can detect differences in both the location and shape of the empirical CDFs of both populations. Given two CDFs (the reference population distribution being a known function or sample), the Kolmogorov&#x02013;Smirnov statistic is the maximum difference between these two CDFs. The two sample S-K test can be used to compare two arbitrary distribution functions of a single random variable. Here the Kolmogorov&#x02013; Smirnov statistic D<emphasis><subscript>n,m</subscript></emphasis> is the supremum of only two samples (with <emphasis>n</emphasis> and <emphasis>m</emphasis> being the number of samples in each population). This supremum is used to reject the null hypothesis. The rejection threshold is calculated by applying a significance level (<emphasis>&#x003B1;</emphasis>) to a known function (c(<emphasis>&#x003B1;</emphasis>) <superscript><emphasis>p</emphasis></superscript><emphasis>n</emphasis> + <emphasis>m/nm</emphasis>)). Note that this expression is an approximation. The critical values D<emphasis><subscript>&#x003B1;</subscript></emphasis> are usually generated via other methods and can be made available in a table form [Facchinetti, 2009].</para>
</section>
<section class="lev4">
<title>Man-Whitney U-Test [Mann and Whitney, 1947]</title>
<para>This test allows to check if a random variable is stochastically larger than another one. The test is done by calculating the U statistic, which has a known distribution. Usually the U statistic is calculated directly (see next paragraph) and not from the distribution itself. For large datasets however, the U statistic can be approximated by a normal distribution. The large U statistic associated to a given dataset indicates that the dataset&#x02019;s random variable is larger.</para>
<para>Two methods exist to calculate the U statistic: The first can be used in small data sets. In this case all pairs are compared, to sum the number of times the first sample is greater than the second (ties are assigned a value of 0.5). This sum is the U statistic U1. The process is repeated by swapping the pairs that are compared. This U statistic is U2. In the second case, both datasets are joined and ranked. The ranks of all elements from the first dataset are the summed to obtain the U1. U2 can now be calculated based on the total number elements in each dataset. In either case, the large U value indicates which dataset is stochastically higher.</para>
<para>The advantages of this test compared to the t-test is that it can be used for ordinal data, is robust against outlier (based on rank) and is a better estimator (higher efficiency) for large datasets that are not normally distributed (for normal distributions its efficiency is 5% below that of the t-test). Note that this test can also be used when comparing two populations of different distributions.</para>
</section>
<section class="lev4">
<title>Kernel density estimation [Rosenblatt, 1954; Parzen, 1962]</title>
<para>Also known as Parzen&#x02013;Rosenblatt window method, this is a data smoothing technique that estimates the shape of a function of unknown density.</para>
<para>The kernel density estimator is a sum of scaled kernel functions. A kernel function is a function that has an integral of 1. Density estimation consists of identifying the type and number of kernels to use and determining the kernels and smoothing parameters (also referred to as the bandwidth). These parameters are selected in order to reduce an error, usually the MISE. This process of estimating the density has similarities to that of manifold learning, which includes among other techniques auto-encoders (see category Artificial Neural Network, technique auto-encoders in Section 5.4.2.8).</para>
<para>Several kernels exist, such as the uniform, triangular, Epanechnikov and normal functions, among others. Usually a normal (Gaussian) kernel is used for mathematical convenience. Its efficiency is close to that of the Epanechnikov, which is optimal with respect to MISE.</para>
<para>The selection of the parameters, based on the MISE, is not possible because it requires prior knowledge of the density function that is being estimated. Several techniques have been developed [Xu et al., 2015], which include the most effective plug-in and cross validation selectors. However, bandwidth selection for heavy-tailed distributions is difficult.</para>
<para>Besides the various techniques used to estimate the various parameters such as the bandwidth, kernel density estimators also differ in the number of random variables modelled (uni- or multi variate), the use of constant or variable bandwidth (variable kernel density estimation) and the use of a single kernel per data-point or a variable (optimal) number of kernels.</para>
</section>
<section class="lev4">
<title>Common applications</title>
<para>Function smoothing in general has many uses that include:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Kernel smoothing allows to estimate underlying function of a noisy sample;</para></listitem>
<listitem>
<para>Kernel regression uses the learned function for prediction (non-linear regression);</para></listitem>
<listitem>
<para>Conditional probability density estimates can be obtained from the data by estimating the density functions of the data that are selected according the desired conditionals (Bayes theorem);</para></listitem>
<listitem>
<para>The maxima can be used for data analysis and machine learning such as clustering and image processing (multivariate kernel density estimation).</para></listitem>
</itemizedlist>
</section>
<section class="lev4">
<title>Strengths and challenges</title>
<para><emphasis role="strong">Advantages</emphasis></para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Does not require the process to be Gaussian;</para></listitem>
<listitem>
<para>Can be used in small datasets;</para></listitem>
<listitem>
<para>Some tests have simple calculations and are easy to understand. MannWhitney U test is based on the more robust median (instead of the mean of the parametric tests, therefore not so sensitive to outliers);</para></listitem>
<listitem>
<para>Mann-Whitney U test can analyse ordinal and ranked data (for example, data using the Likert scale use the de Winter and Dodou).</para></listitem>
</itemizedlist>
<para><emphasis role="strong">Disadvantages</emphasis></para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Lower probability of detecting an effect as compared to parametric methods;</para></listitem>
<listitem>
<para>May have additional assumptions that are required of the dataset (for example the Mann-Whitney U-Test requires that the distribution be continuous and that the alternative hypothesis being tested represent the distribution is stochastically greater than that of the null hypothesis.);</para></listitem>
<listitem>
<para>Still require the estimation of parameters for model generation;</para></listitem>
<listitem>
<para>These tests are usually less efficient than the parameterized versions;</para></listitem>
<listitem>
<para>CPU intensive (KDE);</para></listitem>
<listitem>
<para>The K-S test requires large data sets to correctly reject the null hypothesis.</para></listitem>
</itemizedlist>
</section>
<section class="lev4">
<title>Use in MANTIS</title>
<para>The technique described above was used to detect anomalies that may occur in a press brake machine. Signals indicating the velocity and distance of the press ram and back gauge, oil temperature and vibration were used as the univariate variables.</para>
<para>The non-parametric statistical tests were used to determine whether new samples conform to the same distribution of previously sampled machine states that are known to be failure free. If the distributions are not equivalent, (null hypothesis rejected) then it can be assumed that a failure occurred and alerts are generated.</para>
<para>When testing for a constant (for example position), the signal is normalized prior modelling and testing (a machine may stop its ram at any position, we are only interested in knowing if it is stopped, but not where it happened). In addition to this, two signals (for example the rams speed at the two different pneumatic pistons) are subtracted one from another. The result is a single sample whose means should be 0 (no need for normalization). The statistical test is then performed as it is done for the first case.</para>
<para>The Mann Whitney U test was applied directly to the previously recorded samples of a correctly functioning machine and new samples were obtained from a working machine. As with the parametric case, a p-value was set to reduce the type 1 errors. In the case of the KDE, a model was generated for a sample of the correctly function machine. This model was used to generate a reference sample. The reference sample was then compared to the new working machine samples. Once again, Mann Whitney U test was applied as previously described.</para>
<para>Of all the parametric and non-parametric tests, the Mann-Whitney seemed to be the most robust and effective method. The KDE was tested by applying the t-Test, the Kolmogorov-Smirnov test and the Mann-Whitney U test. The result was no better than applying the Mann-Whitney U test directly. The Mann-Whitney U test is used were the naive statistical tests fails (note that it is necessary to use alternate tests when comparing variability because all the tests discussed here only deal with the mean).</para>
</section>
</section>
<section class="lev3" id="sec5-4-2-16">
<title>5.4.2.16 Mean, thresholds, normality tests</title>
<section class="lev4">
<title>General description</title>
<para>Descriptive statistics [Mann, 1995] is often used to provide a basic analysis of an unknown dataset and identify the basic properties of the data that can be used to select proper techniques to perform a deeper investigation and build statistical models. Moreover, such analyses are useful to identify the characteristics that allow the use of specific statistical approaches. The most common are the following:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis>Mean:</emphasis> it measures the average value of a dataset;</para></listitem>
<listitem>
<para><emphasis>Median:</emphasis> it measures the central value of a dataset (half of the dataset is greater than the median value and the other half is smaller). In many cases, it is preferred compared to the mean since it is less affected by outliers;</para></listitem>
<listitem>
<para><emphasis>Quartiles:</emphasis> they provide a basic understanding on how the values are distributed. Usually, they are calculated to identify where the 25%, 50%, 75% of the data reside;</para></listitem>
<listitem>
<para><emphasis>Variance:</emphasis> it measures the variability and the spread of the data.</para></listitem>
</itemizedlist>
<para>After this preliminary set of analyses, it is required to investigate more deeply the distribution of the data. In particular, it is required to understand if the distribution is normal (or Gaussian) since many statistical approaches can be used only if the data distribution is of this kind.</para>
<para>To test normality, there are many different approaches but the most used is the Shapiro-Wilk test [Shapiro and Wilk, 1965] (another popular but less powerful test is the Kolmogorov&#x02013;Smirnov one [Wayne, 1990]). After verifying whether a dataset follows a normal distribution, it is possible to select properly the statistical tools to perform further investigations.</para>
<para>To define &#x0201C;normal&#x0201D; and &#x0201C;abnormal&#x0201D; values for a variable in a dataset, it is required to define thresholds, and it requires a dataset that includes a large number of repeated measures of the variable. Such measures should be labelled as &#x0201C;normal&#x0201D; or &#x0201C;abnormal&#x0201D; since the threshold can be built using the &#x0201C;normal&#x0201D; measures and verified against both &#x0201C;normal&#x0201D; and &#x0201C;abnormal&#x0201D; ones to build a good descriptive model. In addition, in this case there are several approaches and one that is very popular when dealing with data not distributed normally is the Tukey&#x02019;s range test [Tukey, 1949]. This technique allows the definition of thresholds for outliers, such as the values outside the range are candidate to be considered &#x0201C;abnormal&#x0201D;.</para>
</section>
<section class="lev4">
<title>Common applications</title>
<para>Descriptive statistics is used as a preliminary investigation in any analysis of an unknown dataset regardless of the application domain. They include basic measures that are useful to characterize the data and decide further actions (e.g., subsequent analysis, split/merge of datasets, apply transformations, etc.).</para>
<para>After the extraction of descriptive statistics, testing the data for normality distribution is the second most common activity in any kind of application domains if the distribution of the data is not known. It is also possible to test the compliance with other distributions, but several statistical methods require the normality distribution of the data to be applicable. Therefore, this kind of tests is performed as a pre-requisite for the identification of the proper statistical tool to apply.</para>
<para>Thresholds definition using different kinds of approaches is frequently used to perform statistical process control, to define statistical models, and to evaluate their performances using only data from the field (e.g., historical data) and perform online checks of the behaviour of a system to rise alerts in case of violations. The developed models can also be used to make predictions of violations in conjunction with other statistical techniques that are able to identify trends in data (e.g., using time series analysis techniques).</para>
</section>
<section class="lev4">
<title>Strengths and challenges</title>
<para>Strengths</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Simple mathematical approach;</para></listitem>
<listitem>
<para>The definition of the model requires a limited amount of data;</para></listitem>
<listitem>
<para>The model can be generated on-the-fly and adapted to different working conditions;</para></listitem>
<listitem>
<para>Models can be generated using limited computational resources.</para></listitem>
</itemizedlist>
<para>Challenges</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Can be sensitive to noise;</para></listitem>
<listitem>
<para>Requires fully labelled data.</para></listitem>
</itemizedlist>
</section>
<section class="lev4">
<title>Use in MANTIS</title>
<para>The described techniques have been used for the characterization of the datasets coming from railway switches and to define their expected behaviour based on available data. In particular, the used data regarded the adsorbed current of the motor and the position of the switch over time to define the profiles of correct behaviour through the analysis of the variability of the data at each time instant. The distribution of the samples from different movements of the switch at each time instant was analysed, followed by the definition of thresholds of acceptable values based on an analysis of the outliers. Moreover, since the collected data are affected by the temperature of the environment, the adopted approach can continuously adapt itself using the data from the latest movements that were considered correct. Finally, the work on the project also proved that abnormal behaviours can be easily detected since the correct profiles have a low percentage of samples that are outside the threshold defined (less than 5%) while the abnormal ones have a much higher percentage and/or the duration of the movement is very different from the expected one.</para>
</section>
</section>
</section>
</section>
<section class="lev1" id="sec5-5">
<title>5.5 Examples</title>
<para>This section entails some example use cases of the unsupervised algorithms and the recurrent neural networks for analysing and predicting temporal message log respectively, and investigates the use of convolutional neural network architectures for predicting textures of the metal surface, which is very useful for maintenance.</para>
<section class="lev2" id="sec5-5-1">
<title>5.5.1 Usage Patterns/k-means</title>
<para>In the context of the MANTIS project, a small pilot experiment was performed by Liebher on a dataset containing log files with messages and time stamps for about 900 machines of two types. Assume a machine is a stochastic message generator, assume for the time being that the order is not important (in other words, that the order is of secondary importance if compared to the presence vs absence of a message). The message patterns emitted by the machine are a function of machine type, part characteristics, usage conditions, possibly by the conditions of repair (quality of the servicer/mechanic). This research attempted to answer the following questions.</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Are there groups (clusters) of message pattern generators?</para></listitem>
<listitem>
<para>Are there machines that have a pattern of message generation?</para></listitem>
<listitem>
<para>How many of such groups do exist in the data? Is there a relation to the service mechanic? Is there a relation to the customer? Is there a relation to the type of the machine?</para></listitem>
</itemizedlist>
<section class="lev3" id="sec5-5-1-1">
<title>5.5.1.1 Data analysis</title>
<para>Data were preprocessed in the following sequence;</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Extracting machine-specific information;</para></listitem>
<listitem>
<para>Computation of message histogram row counts and message probabilities per machine;</para>
<fig id="F5-19" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-19">Figure <xref linkend="F5-19" remap="5.19"/></link></label>
<caption><para>Error curve for clustering results as a function of the expected number of clusters k.</para></caption>
<graphic xlink:href="graphics/ch05_fig020.jpg"/>
</fig>
</listitem>
<listitem>
<para>Use of own K-means clustering script for computing k centroid vectors for the message probabilities to identify patterns in the machine generators;</para></listitem>
<listitem>
<para>Computation of an optimal number of clusters (machine groups). <link linkend="F5-19">Figure <xref linkend="F5-19" remap="5.19"/></link> shows the Euclidean distance between cluster centroids and the input data. The figure represents the error curve for clustering results as a function of the expected number of customer groups. Average Euclidean distance between message histograms of k-means centroids and machine instances, respectively, was used. Each clustering attempt for a value k was performed 20 times and the mean curve as well as the standard-deviation band are shown. There is no &#x02018;knee&#x02019; or convincing discontinuity indicating the optimal number of clusters. The variants of this method showed a similar behaviour (silhouette method, gap statistic). Hence, a chi-square was experimented with, which has, to our knowledge, not been used in this manner before. The rationale is that if there is an underlying distribution of k clusters in the data, there should be a distribution of instances over clusters that is deviating from a uniform or expected distribution. Conversely, if there is no underlying cluster structure, the assignment of instances to clusters is random over several Monte-Carlo clustering attempts and a statistical frequency test should yield non-significant results;</para></listitem>
<listitem>
<para>Using <emphasis>&#x003C7;</emphasis> 2 for testing cluster independence;</para></listitem>
<listitem>
<para>Computation of factors such as : machine type, customer ID, servicerID;</para></listitem>
<listitem>
<para>Random trials of k-means estimations for a range of <emphasis>k</emphasis>;</para></listitem>
<listitem>
<para>Analysis of contingency table significance of interesting factor interactions. For this step, the research question was: what will be the average estimate of the significance of the count independence in the selected table (cls x cust or cls x serv)? In order to get reliable results, the average p values are calculated over one hundred (100) independent k-means model estimations for each of the <emphasis>k</emphasis> = [2,&#x02026;,25] values. For instance, <link linkend="T5-4">Table <xref linkend="T5-4" remap="5.4"/></link> reports the ClusterID vs MachTyp: cls x typ (the other, cust and srv tables would be too big horizontally to fit here);</para>
<table-wrap position="float" id="T5-4">
<label><link linkend="T5-4">Table <xref linkend="T5-4" remap="5.4"/></link></label>
<caption><para>ClusterID vs MachTyp contingency table</para></caption>
<table cellspacing="5" cellpadding="5" frame="above" rules="groups">
<thead>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="left">typ2</td>
<td valign="top" align="left">typ1</td>
<td valign="top" align="left">Totals</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">c1</td>
<td valign="top" align="left">115</td>
<td valign="top" align="left">0</td>
<td valign="top" align="left">115</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">c2</td>
<td valign="top" align="left">267</td>
<td valign="top" align="left">9</td>
<td valign="top" align="left">276</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">c3</td>
<td valign="top" align="left">103</td>
<td valign="top" align="left">0</td>
<td valign="top" align="left">103</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">c4</td>
<td valign="top" align="left">4</td>
<td valign="top" align="left">0</td>
<td valign="top" align="left">4</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">c5</td>
<td valign="top" align="left">27</td>
<td valign="top" align="left">6</td>
<td valign="top" align="left">33</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">c6</td>
<td valign="top" align="left">53</td>
<td valign="top" align="left">0</td>
<td valign="top" align="left">53</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">c7</td>
<td valign="top" align="left">4</td>
<td valign="top" align="left">0</td>
<td valign="top" align="left">4</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">c8</td>
<td valign="top" align="left">52</td>
<td valign="top" align="left">1</td>
<td valign="top" align="left">53</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">c9</td>
<td valign="top" align="left">142</td>
<td valign="top" align="left">34</td>
<td valign="top" align="left">176</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">c10</td>
<td valign="top" align="left">38</td>
<td valign="top" align="left">67</td>
<td valign="top" align="left">105</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">Totals</td>
<td valign="top" align="left">805</td>
<td valign="top" align="left">117</td>
<td valign="top" align="left">922</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">Analysis</td>
<td valign="top" align="left">for cls</td>
<td valign="top" align="left">x typ:</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">chisq</td>
<td valign="top" align="left"></td>
<td>323.812158</td>
<td>df</td>
<td>9</td>
<td>P 0.000000</td>
</tr>
<tr>
<td valign="top" align="left">Cramer&#x02019;s</td>
<td valign="top" align="left"></td>
<td valign="top" align="left">V</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left">0.592627</td>
</tr>
</tbody>
<table-wrap-foot>
<tr>
<td valign="top" align="left" colspan="6"><emphasis>Source</emphasis>: cls typ</td>
</tr>
</table-wrap-foot>
</table>
</table-wrap>
</listitem>
<listitem>
<para>The p value was extracted from the contab log file and was averaged over the 100 random retries of the kmeans estimation. In <link linkend="T5-4">Table <xref linkend="T5-4" remap="5.4"/></link>, p equals &#x02018;zero&#x02019;, because of the trivial fact that the messages for MachTyp=1 and MachTyp=2 are significantly different: These are physically different machines, each type sending messages with its own pattern statistics;</para></listitem>
<listitem>
<para>Average expected significance probability: compute the averages for p values, their standard deviation (sd) and standard error (se). The results obtained are discussed in the next section.</para></listitem>
</itemizedlist>
</section>
<section class="lev3" id="sec5-5-1-2">
<title>5.5.1.2 Results</title>
<para>Average significance of clustering results for different k, for interaction &#x02018;cluster x customer&#x02019; in <link linkend="T5-5">Table <xref linkend="T5-5" remap="5.5"/></link>.</para>
<table-wrap position="float" id="T5-5">
<label><link linkend="T5-5">Table <xref linkend="T5-5" remap="5.5"/></link></label>
<caption><para>Results of multiple Monte Carlo experiments (N = 100), for different k. A low average p-significance level indicates an underlying structure for this value of k</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="center">k</td>
<td valign="top" align="center">N</td>
<td valign="top" align="center">Avg_p_signif</td>
<td valign="top" align="center">sd</td>
<td valign="top" align="center">se</td>
<td></td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="center">2</td>
<td valign="top" align="center">100</td>
<td valign="top" align="center">0.2674</td>
<td valign="top" align="center">0.3294</td>
<td>0.0329</td>
<td></td></tr>
<tr>
<td valign="top" align="center">3</td>
<td valign="top" align="center">100</td>
<td valign="top" align="center">0.1627</td>
<td valign="top" align="center">0.2307</td>
<td>0.0231</td>
<td></td>
</tr>
<tr>
<td valign="top" align="center">4</td>
<td valign="top" align="center">100</td>
<td valign="top" align="center">0.1763</td>
<td valign="top" align="center">0.2902</td>
<td>0.0290</td>
<td></td></tr>
<tr>
<td valign="top" align="center">5</td>
<td valign="top" align="center">100</td>
<td valign="top" align="center">0.1340</td>
<td valign="top" align="center">0.2811</td>
<td>0.0281</td>
<td></td>
</tr>
<tr>
<td valign="top" align="center">6</td>
<td valign="top" align="center">100</td>
<td valign="top" align="center">0.1245</td>
<td valign="top" align="center">0.2708</td>
<td valign="top" align="center">0.0271</td>
<td></td>
</tr>
<tr>
<td valign="top" align="center">7</td>
<td valign="top" align="center">100</td>
<td valign="top" align="center">0.0588</td>
<td valign="top" align="center">0.1426</td>
<td valign="top" align="center">0.0143</td>
<td valign="top" align="center">&#x0003C;==</td>
</tr>
<tr>
<td valign="top" align="center">8</td>
<td valign="top" align="center">100</td>
<td valign="top" align="center">0.1543</td>
<td valign="top" align="center">0.2932</td>
<td valign="top" align="center">0.0293</td>
<td></td>
</tr>
<tr>
<td valign="top" align="center">9</td>
<td valign="top" align="center">100</td>
<td valign="top" align="center">0.2029</td>
<td valign="top" align="center">0.3608</td>
<td valign="top" align="center">0.0361</td>
<td></td>
</tr>
<tr>
<td valign="top" align="center">&#x02026;</td>
<td valign="top" align="center">&#x02026;</td>
<td valign="top" align="center">&#x02026;</td>
<td valign="top" align="center">&#x02026;</td>
<td valign="top" align="center">&#x02026;</td>
</tr>
<tr>
<td valign="top" align="center">25</td>
<td valign="top" align="center">100</td>
<td valign="top" align="center">0.2630</td>
<td valign="top" align="center">0.3748</td>
<td valign="top" align="center">0.0375</td>
<td></td></tr>
<tr>
<td valign="top" align="center" colspan="6"></td></tr>
</tbody>
</table>
</table-wrap>
<section class="lev4" id="sec5-5-1-2-1">
<title>5.5.1.2.1 Plotting</title>
<para>The data information from column 1 and 3 in <link linkend="T5-5">Table <xref linkend="T5-5" remap="5.5"/></link> are plotted in a histogram as shown in <link linkend="F5-20">Figure <xref linkend="F5-20" remap="5.20"/></link>, with k on the x axis and avg p signif on the y axis. The left plot of <link linkend="F5-20">Figure <xref linkend="F5-20" remap="5.20"/></link> is the average p-significance levels as a function of k (number of groups) for the interaction (clusters x customers). There is a clear minimum at k = 7, indicating that for this number of groups the distribution of machines over (clusters x customers) is not likely to be random, in case of seven groups of prototypical message generators. Although the minimum value (p = 0.0588) is above the commonly maintained &#x02018;0.05&#x02019;, the result is sufficiently different from the other values to be actually meaningful (error = 0.0143). The right plot is the average p-significance levels as a function of k (number of groups), for the interaction (clusters x servicers) There is no tendency for non-random distributions, thus it is highly unlikely to find significant differences in the distribution of machines over bins defined by (clusters x servicers).</para>
</section>
<section class="lev4" id="sec5-5-1-2-2">
<title>5.5.1.2.2 Replicability of results</title>
<para>Since the k-means algorithm entails random initialisation and the data scattering is noisy by itself, the question is warranted what a repetition of the whole procedure would yield. When repeating the complete procedure 24 times, the average value for k was 6.6 with a standard deviation of 1.5 (see <link linkend="T5-6">Table <xref linkend="T5-6" remap="5.6"/></link>).</para>
<fig id="F5-20" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-20">Figure <xref linkend="F5-20" remap="5.20"/></link></label>
<caption><para>Average p-significance levels as a function of number of groups for interactions (clusters &#x000D7; customers) on the left, and (clusters &#x000D7; servicers) on the right.</para></caption>
<graphic xlink:href="graphics/ch05_fig021.jpg"/>
</fig>
<table-wrap position="float" id="T5-6">
<label><link linkend="T5-6">Table <xref linkend="T5-6" remap="5.6"/></link></label>
<caption><para>Results of k-means algorithm repetition</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="center">k</td>
<td valign="top" align="center">Freq</td>
<td valign="top" align="left"></td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="center">4</td>
<td valign="top" align="center">1</td>
<td valign="top" align="left">&#x0002A;</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="center">5</td>
<td valign="top" align="center">2</td>
<td valign="top" align="left">&#x0002A;&#x0002A;</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="center">6</td>
<td valign="top" align="center">12</td>
<td valign="top" align="left">&#x0002A;&#x0002A;&#x0002A;&#x0002A;&#x0002A;&#x0002A;&#x0002A;&#x0002A;&#x0002A;&#x0002A;&#x0002A;&#x0002A;</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="center">7</td>
<td valign="top" align="center">4</td>
<td valign="top" align="left">&#x0002A;&#x0002A;&#x0002A;&#x0002A;</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="center">8</td>
<td valign="top" align="center">2</td>
<td valign="top" align="left">&#x0002A;&#x0002A;</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="center">9</td>
<td valign="top" align="center">2</td>
<td valign="top" align="left">&#x0002A;&#x0002A;</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="center">10</td>
<td valign="top" align="center">0</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="center">11</td>
<td valign="top" align="center">1</td>
<td valign="top" align="left">&#x0002A;</td></tr>
</tbody>
</table>
</table-wrap>
<para>However, apart from the &#x02018;best&#x02019; number of clusters, the most interesting finding is the contrast between the &#x02018;Customers&#x02019; and &#x02018;Servicers&#x02019; factors, where the latter do not seem to induce a clear clustering of message patterns generated by the machine instances.</para>
</section>
<section class="lev4" id="sec5-5-1-2-3">
<title>5.5.1.2.3 Summary of results</title>
<para>There is a very strong relation <emphasis>p &#x0003C;</emphasis> 0<emphasis>.</emphasis>000001 between cls x typ (cluster-ID x machine type) i.e., the algorithm detects that machine type 1 and 2 generate different message patterns, in all possible cluster forms, with k in the range [2, 25]. There are no different messaging patterns for the machines in the clusters with respect to Servicer-ID. There is a clear significance if the counts in the clusters are based on Customer-ID. This occurs when the data is segmented into <emphasis>k</emphasis> = 6 &#x000B1; 1 clusters (see <link linkend="T5-6">Table <xref linkend="T5-6" remap="5.6"/></link>).</para>
<para>Since there are only two machine types, a grouping into six or seven prototypical message patterns cannot be the result of machine type alone. <emphasis role="strong">Customers</emphasis> may be using machines differently, leading to different message patterns for each of the user groups. On the other hand, for the <emphasis role="strong">Servicers</emphasis>, no indications are found for a cluster structure with different machine-messaging patterns for different clusters. Clustering results varies slightly over several estimations, but not drastically. Repeated runs (i.e., again a full set of 100 tests per value of k from 2&#x02013;25) yielded comparable results.</para>
</section>
</section>
</section>
<section class="lev2" id="sec5-5-2">
<title>5.5.2 Message Log Prediction Using LSTM</title>
<para>An efficient approach to time-series prediction of message code sequences generated from the Litronic machine (for Liebherr) is proposed. For this aim, the choice fell on an Integrated Dimensionality-Reduction LSTM (ID-LSTM) that mainly consist of one encoder LSTM, a chain decoder with three LSTMs, one repeat-vector that connects the encoding and decoding sections of the network architecture, and a time-distributed wrapper. Preliminary experiments were carried out using separate single layer predictive models (SL-LSTM or SL-GRU) and ID-LSTM for predicting two or more forms of the message code representations. The best approach from this earlier investigation is used to analyse all the message codes by considering ten unique subsets of the same size that sum up to the entire message codes. The results show that ID-LSTM using one-hot-encoding data-representation yields a performance that surpasses all other approaches on the small sample of the data. Additionally, the average prediction accuracy for the ten unique subsets of the entire message codes during training and testing phases of the proposed method is very good.</para>
<para>There exist groups of message pattern generator, and there is a need to know how well a neural network can predict this pattern. The resulting questions to be answered regard how we should encode arbitrary nominal machine codes to be useable in a vectorial processing paradigm, and whether we can predict error codes or not.</para>
<section class="lev3" id="sec5-5-2-1">
<title>5.5.2.1 Data interpretation and representation</title>
<para>This section describes the Litronic dataset, and provides a brief description for each of the message code data-representations passed as input to the different predictive models.</para>
<section class="lev4" id="sec5-5-2-1-1">
<title>5.5.2.1.1 Litronic dataset</title>
<para>This dataset contains several machine code generator. Each row within the dataset describes nine unique data fields that provide the machine ID, message code, time information and some other industrial properties. The total number of timestamps present in the dataset is 15, 461, 890 (i.e., 15M) which also corresponds to the total-time counts.</para>
<para>In the experiments, we considered two forms of data sizes:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>A small number of samples: this data size contains one machine ID (mach4545) over a short duration of 10,000 time-counts;</para></listitem>
<listitem>
<para>A more significant number of samples: this data size involves a collection of different machine IDs with unique message codes for ten different subsets but with the same amount of data samples. Note that each subset uses the same number of time-counts 1<emphasis>.</emphasis>546&#x000D7;10<superscript>6</superscript>. This sample size is convenient for a memory-based model estimation. We remark that the sum of each subset time-counts yields the total time-counts of 15.46 &#x000D7; 10<superscript>6</superscript>, which is ten times as large.</para></listitem>
</itemizedlist>
</section>
<section class="lev4" id="sec5-5-2-1-2">
<title>5.5.2.1.2 Data representation</title>
<para>We considered three forms of data-representation, individually passed as input to the predictive models.</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Raw Code: only the raw message code is used for the time-series prediction with no transformation or pre-processing on the raw code. Some of the unique sets of message code include: &#x0007B;464, 465, 466, ..., 946, ...., 31662&#x0007D;;</para></listitem>
<listitem>
<para>OHE Code: these codes are obtained by transforming the RC to an index encoded integers; then the output is further processed to a categorical representation of the message codes. Note the final versions of these message codes contain binary values &#x0007B;0, 1&#x0007D;. Note that the dimensionality of the feature vector of the encoding is dependent on the amount of the index codes for a given number of time counts. Note that for the entire 15,461,890 samples of codes, the vector matrix of the OHE is <emphasis>R</emphasis><superscript>15,461,890&#x000D7;1304</superscript>, that is, there are 1304 codes (unique index);</para></listitem>
<listitem>
<para>PCA Code: these codes are obtained by using a linear dimensionality reduction (DR) algorithm to reduce the feature dimensionality of the OHE codes into a reduced feature space that contains the most important information, based on the covariance structure of the data.</para></listitem>
</itemizedlist>
</section>
</section>
<section class="lev3" id="sec5-5-2-2">
<title>5.5.2.2 Predictive models</title>
<para>We employed Single Layer Predictive Models, namely LSTMs and GRU, which are mainly designed to avoid the long-term dependency problem.</para>
<para>An approach to sequence prediction using an integrated dimensionality reduction LSTM model was devised. The model is set-up using one encoder LSTM interconnected to a decoder section (that comprises three LSTMs) using a repeat-vector algorithm. The first-two LSTMs in the decoder section help in the reconstruction process of the squeezed feature vector output from the encoding section into a lower dimensional feature space. Later on, the final feature <emphasis>x</emphasis> from the last LSTM is wrapped with a time distributed algorithm that presents the reproduced data in a sequential series; this was achieved using a fully connected layer <emphasis>f</emphasis> (<emphasis>x</emphasis>) = <emphasis>Wx</emphasis> + <emphasis>b</emphasis>, that is activated using a rectified linear units, or RELU, activation function <emphasis>max</emphasis>(0<emphasis>,f</emphasis>(<emphasis>x</emphasis>)). Note that all the LSTMs use a return sequence (that is set to TRUE). Moreover, each of the LSTM uses 10 output nodes except for the last LSTM that contains output nodes that corresponds to the exact number of the input feature dimension, and is dependent on the input data-representation if they exist as either OHE or PCA codes. <link linkend="F5-21">Figure <xref linkend="F5-21" remap="5.21"/></link> shows the block diagram of the integrated dimensionality-reduction LSTM architecture. The bold lines indicate the network connection while the dotted lines show the non-network connection, which represent data copying.</para>
<fig id="F5-21" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-21">Figure <xref linkend="F5-21" remap="5.21"/></link></label>
<caption><para>Block diagram describing the predictive recurrent neural network using an Integrated Dimensionality LSTM architecture applied on Litronic message codes.</para></caption>
<graphic xlink:href="graphics/ch05_fig022.jpg"/>
</fig>
<para>The network is compiled using an Adam optimizer [Kingma and Jimmy, 2014], while training for specific conditions depending on the data representation and data-size:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Small sample data represented in OHE codes: trained for 100 epochs and setting the batch size to 10;</para></listitem>
<listitem>
<para>Small sample data represented in PCA codes: trained for either 100 or 200 epochs and setting the batch size to 10;</para></listitem>
<listitem>
<para>Large sample data represented in OHE codes: trained for 100 epochs and setting batch size to 10,000, presents a significant computational benefit as it computes faster using higher batch size compared to lower batch size (which require more extended training time).</para></listitem>
</itemizedlist>
<para>Note that training the proposed network on the PCA codes presents varying trainable parameters, and this is due to the variation in the PCA feature dimensions &#x0007B;1, 2, 3, 4, 5, 10, 20, 40&#x0007D; on the small data samples. In this method, we consider 70% samples of the data-distribution for training and the remaining 30% for testing, for each of the used data sizes (small or large).</para>
</section>
<section class="lev3" id="sec5-5-2-3">
<title>5.5.2.3 Results</title>
<para>This section reports on performances of the various predictive models with respect to the used data sizes.</para>
<section class="lev4" id="sec5-5-2-3-1">
<title>5.5.2.3.1 Evaluation of predictive models on small number of samples</title>
<para>The performances (accuracies and output-target plots) for the proposed predictive method on OHE codes are represented graphically in <link linkend="F5-22">Figure <xref linkend="F5-22" remap="5.22"/></link>. The experiments considered a small number of samples, and featured 10 K time samples, of 64 possible message codes. The upper left graph (<link linkend="F5-22">Figure <xref linkend="F5-22" remap="5.22"/></link>a) shows the ID-LSTM prediction accuracy for both training and testing phases while training the network for 100 epochs. The upper right graph (<link linkend="F5-22">Figure <xref linkend="F5-22" remap="5.22"/></link>b) shows a plot of the index prediction in the training and testing phases to their corresponding target values. The bottom left graph (<link linkend="F5-22">Figure <xref linkend="F5-22" remap="5.22"/></link>c) and bottom right graph (<link linkend="F5-22">Figure <xref linkend="F5-22" remap="5.22"/></link>d) represent the confusion matrix, by showing the plot of the output predictions against their target values for both training and testing phases respectively. Additionally, the several predictive model&#x02019;s evaluations on the different data representations are reported in <link linkend="T5-7">Table <xref linkend="T5-7" remap="5.7"/></link>.</para>
<para><emphasis role="strong"><emphasis>ID-LSTM Evaluation on OHE Codes</emphasis></emphasis></para>
<para>As can be observed from <link linkend="T5-7">Table <xref linkend="T5-7" remap="5.7"/></link>, the use of ID-LSTM on OHE yields a performance that surpasses all other approaches: the method shows a high level of precision in the predictive values compared to their corresponding target values. Other interesting techniques involve the use of ID-LSTM examined on high order dimensional PCA codes, which persistently outperform both ID-LSTM or single layer predictive models when applied on one-dimensional PCA versions or raw codes.</para>
<fig id="F5-22" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-22">Figure <xref linkend="F5-22" remap="5.22"/></link></label>
<caption><para>Result on small pilot test 10K time samples, 64 possible message codes.</para></caption>
<graphic xlink:href="graphics/ch05_fig023.jpg"/>
</fig>
<para>The second best approach is the ID-LSTM on 20-dimensional PCA codes; this is due to their performance in the testing phase. There exist some failure in the prediction for few epoch instances in the testing phase. An exception to this is the ID-LSTM on 20 or 1-dimensional PCA codes. Based on this observation, it can be deduced that it is essential to optimize PCA codes with different feature dimensionality to solve a large-scale sequence prediction problem; this will aid to provide a predictive model that will be free from prediction failure in the testing phase. Moreover, this allows to retain useful covariance information that can provide an alternative basis for good temporal prediction. Even the external cases of applying PCA and only using largest eigenvector yielded a reasonable performance of 93%.</para>
<para>Hence the experimental result indicates that the use of OHE or PCA codes with their respective index provide a better basis for temporal prediction. Due to the performance of the best approach, ID-LSTM only was employed to examine OHE codes from the more significant amount of the message sequence.</para>
<table-wrap position="float" id="T5-7">
<label><link linkend="T5-7">Table <xref linkend="T5-7" remap="5.7"/></link></label>
<caption><para>Training performance metrics on the different approaches using a batch size of 10</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left">Method</td>
<td valign="top" align="left">Train Accuracy</td>
<td valign="top" align="left">Test Accuracy</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">ID-LSTM-I-OHE-Codes</td>
<td valign="top" align="left">0.9957</td>
<td valign="top" align="left">0.9920</td></tr>
<tr>
<td valign="top" align="left">ID-LSTM-I-20-DIM-PCA-Codes</td>
<td valign="top" align="left">0.9763</td>
<td valign="top" align="left">0.9843</td>
</tr>
<tr>
<td valign="top" align="left">ID-LSTM-I-40-DIM-PCA-Codes</td>
<td valign="top" align="left">0.9760</td>
<td valign="top" align="left">0.9733</td></tr>
<tr>
<td valign="top" align="left">ID-LSTM-I-10-DIM-PCA-Codes</td>
<td valign="top" align="left">0.9316</td>
<td valign="top" align="left">0.9727</td>
</tr>
<tr>
<td valign="top" align="left">ID-LSTM-I-5-DIM-PCA-Codes</td>
<td valign="top" align="left">0.9139</td>
<td valign="top" align="left">0.9593</td></tr>
<tr>
<td valign="top" align="left">ID-LSTM-I-4-DIM-PCA-Codes</td>
<td valign="top" align="left">0.9424</td>
<td valign="top" align="left">0.9410</td>
</tr>
<tr>
<td valign="top" align="left">ID-LSTM-I-3-DIM-PCA-Codes</td>
<td valign="top" align="left">0.9463</td>
<td valign="top" align="left">0.9593</td></tr>
<tr>
<td valign="top" align="left">ID-LSTM-I-2-DIM-PCA-Codes</td>
<td valign="top" align="left">0.9424</td>
<td valign="top" align="left">0.9590</td>
</tr>
<tr>
<td valign="top" align="left">ID-LSTM-I-1-DIM-PCA-Codes</td>
<td valign="top" align="left">0.8729</td>
<td valign="top" align="left">0.9340</td>
</tr>
<tr>
<td valign="top" align="left">SL-LSTM-I-1-DIM-PCA-Codes</td>
<td valign="top" align="left">0.8757</td>
<td valign="top" align="left">0.9340</td>
</tr>
<tr>
<td valign="top" align="left">SL-GRU-MSE-SI-1-DIM-PCA-Codes</td>
<td valign="top" align="left">0.8715</td>
<td valign="top" align="left">0.9316</td>
</tr>
<tr>
<td valign="top" align="left">SL-GRU-MAE-SI-1-DIM-PCA-Codes</td>
<td valign="top" align="left">0.8715</td>
<td valign="top" align="left">0.9316</td>
</tr>
<tr>
<td valign="top" align="left">SL-LSTM-MAE-SI-1-DIM-PCA-Codes</td>
<td valign="top" align="left">0.8715</td>
<td valign="top" align="left">0.9316</td>
</tr>
<tr>
<td valign="top" align="left">SL-LSTM-MSE-SI-1-DIM-PCA-Codes</td>
<td valign="top" align="left">0.8715</td>
<td valign="top" align="left">0.9316</td>
</tr>
<tr>
<td valign="top" align="left">SL-LSTM-I-Raw-Codes</td>
<td valign="top" align="left">1.429 &#x000D7; 10&#x002C6;-4</td>
<td valign="top" align="left">0.0000</td>
</tr>
<tr>
<td valign="top" align="left">SL-GRU-MSE-SI-Raw-Codes</td>
<td valign="top" align="left">2.858 &#x000D7; 10&#x002C6;-4</td>
<td valign="top" align="left">0.0000</td>
</tr>
<tr>
<td valign="top" align="left">SL-GRU-MAE-SI-Raw-Codes</td>
<td valign="top" align="left">2.858 &#x000D7; 10&#x002C6;-4</td>
<td valign="top" align="left">0.0000</td>
</tr>
<tr>
<td valign="top" align="left">SL-LSTM-MSE-SI-Raw-Codes</td>
<td valign="top" align="left">2.858 &#x000D7; 10&#x002C6;-4</td>
<td valign="top" align="left">0.0000</td>
</tr>
<tr>
<td valign="top" align="left">SL-LSTM-MAE-SI-Raw-Codes</td>
<td valign="top" align="left">1.429 &#x000D7; 10&#x002C6;-4</td>
<td valign="top" align="left">0.0000</td></tr>
</tbody>
</table>
</table-wrap>
<fig id="F5-23" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-23">Figure <xref linkend="F5-23" remap="5.23"/></link></label>
<caption><para>ID-LSTM prediction accuracy for both training and testing phases while training the network on subset 9 for 100 epochs.</para></caption>
<graphic xlink:href="graphics/ch05_fig024.jpg"/>
</fig>
</section>
<section class="lev4" id="sec5-5-2-3-2">
<title>5.5.2.3.2 Evaluation of the ID-LSTM on OHE codes for more significant number of samples</title>
<para>In this subsection, the proposed network was trained on ten different subsets that sum up to the entire message codes. The ID-LSTM evaluation on 196 machines from subset 9, is shown in <link linkend="F5-23">Figure <xref linkend="F5-23" remap="5.23"/></link> with subfigures that describe the prediction accuracy curve, index prediction, and the confusion matrix plot in both training and testing phases respectively. In particular, the upper left graph (<link linkend="F5-23">Figure <xref linkend="F5-23" remap="5.23"/></link>a) shows the ID-LSTM prediction accuracy for both training and testing phases while training the network on subset 9 for 100 epochs. The upper right graph (<link linkend="F5-23">Figure <xref linkend="F5-23" remap="5.23"/></link>b) shows a plot of the index prediction in the training and testing phases to their corresponding target values. The bottom left graph (<link linkend="F5-23">Figure <xref linkend="F5-23" remap="5.23"/></link>c) and bottom right graph (<link linkend="F5-23">Figure <xref linkend="F5-23" remap="5.23"/></link>d) represent the confusion matrix, by showing the plot of the output predictions against their target values for both training and testing phases respectively. The performance metrics obtained after 100 epochs for the different subsets are reported in <link linkend="T5-8">Table <xref linkend="T5-8" remap="5.8"/></link>. In the training phase each of the models performed very well with an average accuracy of 0.9844, while the average test performance is 0.9506. The proposed predictive model can handle very complex sequential problem irrespective of the size of data (time-counts). Please note that the lower performance in the test phase may have arisen due to variation in the codes that were used during training of the proposed network. Both the training and testing phases show outliers. Some of the outlier points represent few instances of absolute average accuracy error in prediction is <emphasis>&#x0003C;</emphasis> 5% for both phases. This performance measure indicates that the proposed method yields a good result.</para>
</section>
</section>
<section class="lev3" id="sec5-5-2-4">
<title>5.5.2.4 Discussion</title>
<para>This research has addressed the two questions in Section 5.5.2: The research demonstrated the transformation of arbitrary machine codes (raw data) into a useable vector representation, with the objective to identify correlated patterns. To deal with this problem, OHE was used to convert the nominal codes to a vector representation. However, the vector dimensionality is high, and correlations are to be expected between message patterns. To reduce the dimensionality to a lower feature space, PCA was employed to select the most informative dimensions. The best result was obtained on OHE code, however with the risk of overfitting depending on the number of message sample and lexicon sizes. Additionally, the results obtained with PCA transformed data are very good, as the use of only 20-dimensional PCA yielded a prediction accuracy of 98%. However, the use of raw codes with no transformation yields the worst result; this implies that the raw codes are not suitable or sensible to the neural networks.</para>
<table-wrap position="float" id="T5-8">
<label><link linkend="T5-8">Table <xref linkend="T5-8" remap="5.8"/></link></label>
<caption><para>Train and Test performance assessment on the different subsets using a batch size set to 10,000</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left">Data Type</td>
<td valign="top" align="left">Time counts</td>
<td valign="top" align="left">No. of Machines</td>
<td valign="top" align="left">No. of Index (class label)</td>
<td valign="top" align="left">Train Accuracy</td>
<td valign="top" align="left">Test Accuracy</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Subset 1</td>
<td valign="top" align="left">0.00-1.54M</td>
<td valign="top" align="left">20</td>
<td valign="top" align="left">948</td>
<td valign="top" align="left">0.9826</td>
<td valign="top" align="left">0.9751</td></tr>
<tr>
<td valign="top" align="left">Subset 2</td>
<td valign="top" align="left">1.54-3.09M</td>
<td valign="top" align="left">30</td>
<td valign="top" align="left">606</td>
<td valign="top" align="left">0.9979</td>
<td valign="top" align="left">0.9695</td>
</tr>
<tr>
<td valign="top" align="left">Subset 3</td>
<td valign="top" align="left">3.09-4.63M</td>
<td valign="top" align="left">36</td>
<td valign="top" align="left">535</td>
<td valign="top" align="left">0.9886</td>
<td valign="top" align="left">0.9624</td></tr>
<tr>
<td valign="top" align="left">Subset 4</td>
<td valign="top" align="left">4.63-6.18M</td>
<td valign="top" align="left">48</td>
<td valign="top" align="left">619</td>
<td valign="top" align="left">0.9961</td>
<td valign="top" align="left">0.9021</td>
</tr>
<tr>
<td valign="top" align="left">Subset 5</td>
<td valign="top" align="left">6.18-7.73M</td>
<td valign="top" align="left">62</td>
<td valign="top" align="left">620</td>
<td valign="top" align="left">0.9837</td>
<td valign="top" align="left">0.9806</td>
</tr>
<tr>
<td valign="top" align="left">Subset 6</td>
<td valign="top" align="left">7.73-9.27M</td>
<td valign="top" align="left">109</td>
<td valign="top" align="left">675</td>
<td valign="top" align="left">0.9962</td>
<td valign="top" align="left">0.9347</td>
</tr>
<tr>
<td valign="top" align="left">Subset 7</td>
<td valign="top" align="left">9.27-10.8M</td>
<td valign="top" align="left">64</td>
<td valign="top" align="left">648</td>
<td valign="top" align="left">0.9205</td>
<td valign="top" align="left">0.9293</td>
</tr>
<tr>
<td valign="top" align="left">Subset 8</td>
<td valign="top" align="left">10.8-12.3M</td>
<td valign="top" align="left">95</td>
<td valign="top" align="left">679</td>
<td valign="top" align="left">0.9973</td>
<td valign="top" align="left">0.9576</td>
</tr>
<tr>
<td valign="top" align="left">Subset 9</td>
<td valign="top" align="left">12.3-13.9M</td>
<td valign="top" align="left">196</td>
<td valign="top" align="left">717</td>
<td valign="top" align="left">0.9943</td>
<td valign="top" align="left">0.9681</td>
</tr>
<tr>
<td valign="top" align="left">Subset 10</td>
<td valign="top" align="left">13.9-15.4M</td>
<td valign="top" align="left">263</td>
<td valign="top" align="left">624</td>
<td valign="top" align="left">0.9871</td>
<td valign="top" align="left">0.9268</td>
</tr>
<tr>
<td valign="top" align="left">Average</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left">0.9844</td>
<td valign="top" align="left">0.9506</td></tr>
</tbody>
</table>
</table-wrap>
<para>Furthermore, we investigated whether predictive recurrent neural networks can be trained and tested on datasets of varying samples sizes, message lexicon size, and underlying machines. For the experiment led to an integrated dimensionality reduction LSTM (ID-LSTM) and external LSTM (SL-LSTM or SL-GRU) for predicting the discussed data. The results show that in the pilot experiment on the small samples, the use of the proposed method (ID-LSTM on OHE codes) yields performance that surpasses all other approaches. ID-LSTM also obtain better performance on higher order dimensional PCA codes than on one-dimensional PCA or raw codes. It is possible to conclude that, for a large data sample with PCA code, there is the is a need to optimize the dimensionality to obtain good performance. Moreover, the ID-LSTM on OHE codes obtained <emphasis>&#x0003C;</emphasis> 5% error on the predicted codes in a realistically large dataset.</para>
<para>The research suggests that it may be possible to combine the proposed model with an early anomaly detection algorithm to allow continuous prediction of physical problems in the machines generating the message logs. Finally, with advance knowledge on the message log data, it is possible to develop a precise system that can handle detection of an anomaly in the message log, and hence providing a comprehensive basis for the root-cause analysis.</para>
</section>
</section>
<section class="lev2" id="sec5-5-3">
<title>5.5.3 Metal-defect Classification</title>
<para>This section describes using deep learning techniques for predicting textures of metal surface, which is very useful for maintenance. The defects on the surface of, e.g., a cutting tool in a production line are detrimental to the quality of the end product produced by the tool. Therefore, it is important to inspect the cutting tool regularly to maintain the quality of the product, which is usually done manually by human experts. However, manual inspection of cutting tools on production line is costly and difficult whereby the entire process slows down. Some defects are hard to measure by digital sensors, but might be easily spotted by the naked eye. For example, the pollution of tool surface due to scraped metal particles can be detected by a camera. Visual inspection is a promising way to automatically find out defects in the cutting surface in real-time in production lines. Images of metal surfaces of cutting tools can be obtained with sensors, such as the digital microscope. Different defects on metal surfaces result in different types of textures and recognizing the textural surface helps to reveal original (root) causes. For example, if a burnt texture is found on the metal surface, the possible reason is that the temperature is high and the cooling system is failing. The defect textures of metal surfaces are usually different on different metal materials. However, there are some generally known and well-documented defect textures in metal surfaces, which provide a good indication of the root causes. In this pilot project, CNNs are used for metal texture classification. Since data sets with labels are hard to collect from the real-word production line, a public metal surface defect data set was used and later a new data set from metal images with different textures was collected.</para>
<section class="lev3" id="sec5-5-3-1">
<title>5.5.3.1 Data collection</title>
<para>The NEU surface defect database is a public data set<footnote id="fn5_5" label="5"><para>http://faculty.neu.edu.cn/yunhyan/NEU surface defect database.html</para></footnote>, which contains six kinds of typical surface defects of the hot-rolled steel strip. It has 300 images per surface defect type and 1800 greyscale images in total, where the size of each image is 200x200 pixels. This image size is very suitable for the application of CNNs. Examples of images are shown in <link linkend="F5-24">Figure <xref linkend="F5-24" remap="5.24"/></link> (left). A new data set named MTEXTURE was also collected. Example images are shown in <link linkend="F5-24">Figure <xref linkend="F5-24" remap="5.24"/></link> (right).This data set has seven texture attributes, which are related to defect textures on metal surface. The size of images is 100 &#x000D7; 100 with greyscale. <link linkend="T5-9">Table <xref linkend="T5-9" remap="5.9"/></link> below shows the attributes of each data set.</para>
<fig id="F5-24" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-24">Figure <xref linkend="F5-24" remap="5.24"/></link></label>
<caption><para>Example of images in the NEU data set (left) and the newly collected MTEXTURE data set (right).</para></caption>
<graphic xlink:href="graphics/ch05_fig025.jpg"/>
</fig>
<table-wrap position="float" id="T5-9">
<label><link linkend="T5-9">Table <xref linkend="T5-9" remap="5.9"/></link></label>
<caption><para>Attributes in the two data sets</para></caption>
<table cellspacing="5" cellpadding="5" frame="above" rules="groups">
<thead>
<tr style="border-bottom: 0.1em solid">
<td valign="middle" align="center" rowspan="2">Database</td>
<td valign="top" align="center" colspan="7">Attribute</td>
</tr>
<tr>
<td valign="top" align="center">1</td>
<td valign="top" align="center">2</td>
<td valign="top" align="center">3</td>
<td valign="top" align="center">4</td>
<td valign="top" align="center">5</td>
<td valign="top" align="center">6</td>
<td valign="top" align="center">7</td></tr>
</thead>
<tbody>
<tr>
<td valign="top" align="center">NEU</td>
<td valign="top" align="center">Rolled-in scale</td>
<td valign="top" align="center">Patches</td>
<td valign="top" align="center">Crazing</td>
<td valign="top" align="center">Pitted</td>
<td valign="top" align="center">Inclusion</td>
<td valign="top" align="center">scratches</td>
<td valign="top" align="center"></td>
</tr>
<tr>
<td valign="top" align="center">MTEXTURE</td>
<td valign="top" align="center">Worn</td>
<td valign="top" align="center">Debris</td>
<td valign="top" align="center">Dented</td>
<td valign="top" align="center">Rough</td>
<td valign="top" align="center">Rust</td>
<td valign="top" align="center">Scratched</td>
<td valign="top" align="center">Crumpled</td>
</tr>
</tbody>
<table-wrap-foot>
<tr>
<td valign="top" align="left" colspan="8">Attribute Database</td></tr>
</table-wrap-foot>
</table>
</table-wrap>
</section>
<section class="lev3" id="sec5-5-3-2">
<title>5.5.3.2 Experiments</title>
<para>The AlexNet [Krizhevsky et al., 2012] CNN has five convolutional layers with Leakly-ReLU as activation function. The Leakly-ReLU is defined as: y = max(x<emphasis>,&#x003BB;</emphasis> x), where x is the input, y is the output and <emphasis>&#x003BB;</emphasis> is a constant parameter which is set to 0.25. After each convolutional layer, a max-pooling layer is used with kernel size 2x2 and stride 2. Three fully-connected layers are applied on the output and the cross-entropy softmax loss is used to train the model. The total number of trained parameters for AlexNet is 26.4M. VGGNet [Simonyan and Zisserman, 2014] is deeper than AlexNet. It consists of sixteen convolutional layers, five max-pooling layers and three fully-connected layers. The training procedure is the same as AlexNet. There are 50.8M parameters in VGGNet. The performance of AlexNet and VGGNet on the two datasets with different epochs is shown in <link linkend="F5-25">Figure <xref linkend="F5-25" remap="5.25"/></link>. The performance looks promising and the precision of metal texture recognition is higher than 95%. The performance of AlexNet and VGGNet has no significant difference, but the number of parameters of AlexNet is much smaller and therefore it is easier to be trained and applied in real-world product lines. The defect textures in metal surface can be detected by a sliding-window strategy. Each patch on the metal image is classified into different defect textures. Once the texture attributes are detected, the related root causes may be found, such as an external (metal) element, which is affecting the surface of the tool. This is an exploratory result and more wear and tear textures can be added if available.</para>
<fig id="F5-25" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F5-25">Figure <xref linkend="F5-25" remap="5.25"/></link></label>
<caption><para>Performance of VGGnet and AlexNet on NEU<sup>6</sup><?lb?>(left) and the collect set (right) with different epochs.</para></caption>
<graphic xlink:href="graphics/ch05_fig026.jpg"/>
</fig>
</section>
<section class="lev3" id="sec5-5-3-3">
<title>5.5.3.3 Discussion</title>
<para>Hunting for root causes of metal surface defects can be performed by an analysis of the textural attributes of metal surfaces. Textural attributes can be learned from an annotated data set. This experiment gives initial results useful for RCA where damage of metal surfaces is involved. From the initial results of two small data sets, it is clear that textural analysis on metal surface can provide information for maintenance of product lines with classification of 95%.</para>
<para>Several issues should be solved to apply this method in the real-word:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Real data set collection and annotation. The challenge of using deep learning in general on metal surfaces concerns the labelled data set. The number of parameters in the network is very large and thus the system provides higher performance when more data is available or the network is pre-trained on other large-scale databases;</para></listitem>
<listitem>
<para>Running the complex deep neural network takes computing time and memory. Networks that are more efficient might be investigated, such as quantizing the parameters of the network;</para></listitem>
<listitem>
<para>Continuous learning is also very important because unseen causes are coming in over time. The system should learn new things when new labels are added and operators may need to perform this additional task before valid results can be obtained.</para></listitem>
</itemizedlist>
</section>
</section>
</section>
<section class="lev1" id="sec5-6">
<title>References</title>
<orderedlist numeration="arabic" continuation="restarts" spacing="normal">
<listitem>
<para>Aghasaryan, A., Fabre, E., Benveniste, A., Boubour, R., and Jard, C. (1997) &#x02018;A Petri net approach to fault detection and diagnosis in distributed systems,&#x02019; in 36th IEEE Conference on Decision and Control, IEEE CDC&#x02019;97, San Diego, California, USA. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Aghasaryan%2C+A%2E%2C+Fabre%2C+E%2E%2C+Benveniste%2C+A%2E%2C+Boubour%2C+R%2E%2C+and+Jard%2C+C%2E+%281997%29+%27A+Petri+net+approach+to+fault+detection+and+diagnosis+in+distributed+systems%2C%27+in+36th+IEEE+Conference+on+Decision+and+Control%2C+IEEE+CDC%2797%2C+San+Diego%2C+California%2C+USA%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Agresti, A. (1992) &#x02018;A survey of exact inference for contingency tables,&#x02019; <emphasis>Statistical Science</emphasis> 7(1), pp. 131&#x02013;153. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Agresti%2C+A%2E+%281992%29+%27A+survey+of+exact+inference+for+contingency+tables%2C%27+Statistical+Science+7%281%29%2C+pp%2E+131-153%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Ali, J. B., Chebel-Morello, B., Saidi, L., Malinowski, S. and Fnaiech, F. (2015) &#x02018;Accurate bearing remaining useful life prediction based on Weibull distribution and artificial neural network,&#x02019; Mechanical Systems and Signal Processing, 56&#x02013;57, pp. 150&#x02013;172. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ali%2C+J%2E+B%2E%2C+Chebel-Morello%2C+B%2E%2C+Saidi%2C+L%2E%2C+Malinowski%2C+S%2E+and+Fnaiech%2C+F%2E+%282015%29+%27Accurate+bearing+remaining+useful+life+prediction+based+on+Weibull+distribution+and+artificial+neural+network%2C%27+Mechanical+Systems+and+Signal+Processing%2C+56-57%2C+pp%2E+150-172%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Aminikhanghahi, S. and Cook, D. J. (2017) &#x02018;A survey of methods for time series change point detection,&#x02019; Knowledge and information systems 51.2, 339&#x02013;367. PMC. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Aminikhanghahi%2C+S%2E+and+Cook%2C+D%2E+J%2E+%282017%29+%27A+survey+of+methods+for+time+series+change+point+detection%2C%27+Knowledge+and+information+systems+51%2E2%2C+339-367%2E+PMC%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Andrews, R., Diederich, J., and Tickle, A. B. (1995) &#x02018;Survey and critique of techniques for extracting rules from trained artificial neural networks,&#x02019; Knowledge-Based Systems, 8(6), pp. 373&#x02013;389. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Andrews%2C+R%2E%2C+Diederich%2C+J%2E%2C+and+Tickle%2C+A%2E+B%2E+%281995%29+%27Survey+and+critique+of+techniques+for+extracting+rules+from+trained+artificial+neural+networks%2C%27+Knowledge-Based+Systems%2C+8%286%29%2C+pp%2E+373-389%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Barnard, G. A. (1959) &#x02018;Control charts and stochastic processes,&#x02019; Journal of the Royal Statistical Society. B (Methodological) 21(2), pp. 239&#x02013;271. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Barnard%2C+G%2E+A%2E+%281959%29+%27Control+charts+and+stochastic+processes%2C%27+Journal+of+the+Royal+Statistical+Society%2E+B+%28Methodological%29+21%282%29%2C+pp%2E+239-271%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Bashivan, P., Rish, I., Yeasin, M., and Codella, N. (2015) &#x02018;Learning representations from EEG with deep recurrent-convolutional neural networks,&#x02019; International Conference on Learning Representations. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Bashivan%2C+P%2E%2C+Rish%2C+I%2E%2C+Yeasin%2C+M%2E%2C+and+Codella%2C+N%2E+%282015%29+%27Learning+representations+from+EEG+with+deep+recurrent-convolutional+neural+networks%2C%27+International+Conference+on+Learning+Representations%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Batal, I., Valizadegan, H., Cooper, G. F., and Hauskrecht, M. (2013) &#x0201C;A temporal pattern mining approach for classifying electronic health record data.&#x0201D; ACM Trans. Intell. Syst. Technol., 4(4), pp. 63:1&#x02013;63:22. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Batal%2C+I%2E%2C+Valizadegan%2C+H%2E%2C+Cooper%2C+G%2E+F%2E%2C+and+Hauskrecht%2C+M%2E+%282013%29+%22A+temporal+pattern+mining+approach+for+classifying+electronic+health+record+data%2E%22+ACM+Trans%2E+Intell%2E+Syst%2E+Technol%2E%2C+4%284%29%2C+pp%2E+63%3A1-63%3A22%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Baxt, W. G. (1990) &#x02018;Use of an artificial neural network for data analysis in clinical decision-making: The diagnosis of acute coronary occlusion,&#x02019; Neural Computation, 2, pp. 480&#x02013;489. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Baxt%2C+W%2E+G%2E+%281990%29+%27Use+of+an+artificial+neural+network+for+data+analysis+in+clinical+decision-making%3A+The+diagnosis+of+acute+coronary+occlusion%2C%27+Neural+Computation%2C+2%2C+pp%2E+480-489%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Burnham, K. P. and Anderson, D. R. (2002) Model Selection and Multimodel Inference: A Practical InformationTheoretic Approach (2nd ed.). Springer-Verlag.</para></listitem>
<listitem>
<para>Cai, M. and Liu, J. (2016) &#x02018;Maxout neurons for deep convolutional and LSTM neural networks in speech recognition,&#x02019; Speech Communication, 77, p. 53. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Cai%2C+M%2E+and+Liu%2C+J%2E+%282016%29+%27Maxout+neurons+for+deep+convolutional+and+LSTM+neural+networks+in+speech+recognition%2C%27+Speech+Communication%2C+77%2C+p%2E+53%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Caruana, A. and Niculescu-Mizil, A. (2005) An Empirical Comparison of Supervised Learning Algorithms Using Different Performance Metrics. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Caruana%2C+A%2E+and+Niculescu-Mizil%2C+A%2E+%282005%29+An+Empirical+Comparison+of+Supervised+Learning+Algorithms+Using+Different+Performance+Metrics%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Chen, Y., Keogh, E., Hu, B., Begum, N., Bagnall, A., Mueen, A., and Batista, G. (2015) The ucr time series classification archive. www.cs.ucr.edu/eamonn/time series data/. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Chen%2C+Y%2E%2C+Keogh%2C+E%2E%2C+Hu%2C+B%2E%2C+Begum%2C+N%2E%2C+Bagnall%2C+A%2E%2C+Mueen%2C+A%2E%2C+and+Batista%2C+G%2E+%282015%29+The+ucr+time+series+classification+archive%2E+www%2Ecs%2Eucr%2Eedu%2Feamonn%2Ftime+series+data%2F%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Chen, Y., Yang, J., Qian, J. (2016) &#x02018;Recurrent neural network for facial landmark detection,&#x02019; Neurocomputing. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Chen%2C+Y%2E%2C+Yang%2C+J%2E%2C+Qian%2C+J%2E+%282016%29+%27Recurrent+neural+network+for+facial+landmark+detection%2C%27+Neurocomputing%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Chilimbi, T. M., Suzue, Y., Apacible, J. and Kalyanaraman, K. (2014) &#x02018;Project Adam: Building an efficient and scalable deep learning training system,&#x02019; in 11th USENIX Symposium on Operating Systems Design and Implementation, Broomfield, CO, USA. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Chilimbi%2C+T%2E+M%2E%2C+Suzue%2C+Y%2E%2C+Apacible%2C+J%2E+and+Kalyanaraman%2C+K%2E+%282014%29+%27Project+Adam%3A+Building+an+efficient+and+scalable+deep+learning+training+system%2C%27+in+11th+USENIX+Symposium+on+Operating+Systems+Design+and+Implementation%2C+Broomfield%2C+CO%2C+USA%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Cho, K., van Merrienboer, B., Gulcehre, C., Bahdanau, D., Bougares, F., Schwenk, H., and Bengio, Y. (2014) &#x02018;Learning phrase representations using RNN encoder-decoder for statistical,&#x02019; in Proceedings of the 2014 Conference on Empirical Methods in Natural, Doha, Qatar. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Cho%2C+K%2E%2C+van+Merrienboer%2C+B%2E%2C+Gulcehre%2C+C%2E%2C+Bahdanau%2C+D%2E%2C+Bougares%2C+F%2E%2C+Schwenk%2C+H%2E%2C+and+Bengio%2C+Y%2E+%282014%29+%27Learning+phrase+representations+using+RNN+encoder-decoder+for+statistical%2C%27+in+Proceedings+of+the+2014+Conference+on+Empirical+Methods+in+Natural%2C+Doha%2C+Qatar%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Collett, D. (2003) Modelling Survival Data in Medical Research, Second Edition, Taylor Francis. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Collett%2C+D%2E+%282003%29+Modelling+Survival+Data+in+Medical+Research%2C+Second+Edition%2C+Taylor+Francis%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Collobert, R. and Weston, J. (2008) &#x02018;A unified architecture for natural language processing: Deep neural networks with multitask learning,&#x02019; in Proceedings of the 25th International Conference on Machine Learning, Helsinki, Finland. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Collobert%2C+R%2E+and+Weston%2C+J%2E+%282008%29+%27A+unified+architecture+for+natural+language+processing%3A+Deep+neural+networks+with+multitask+learning%2C%27+in+Proceedings+of+the+25th+International+Conference+on+Machine+Learning%2C+Helsinki%2C+Finland%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Daniel, Wayne W. (1990) &#x02018;Kolmogorov&#x02013;Smirnov onesample test,&#x02019; Applied Nonparametric Statistics (2nd ed.). Boston: PWSKent. pp. 319&#x02013;330. ISBN 0-534-91976-6.</para></listitem>
<listitem>
<para>Deng, J., Dong, W., Socher, R., Li, L., Li, K., and Fei-Fei, L. (2009) &#x02018;ImageNet: A large-scale hierarchical image database,&#x02019; IEEE Conference on Computer Vision and Pattern Recognition. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Deng%2C+J%2E%2C+Dong%2C+W%2E%2C+Socher%2C+R%2E%2C+Li%2C+L%2E%2C+Li%2C+K%2E%2C+and+Fei-Fei%2C+L%2E+%282009%29+%27ImageNet%3A+A+large-scale+hierarchical+image+database%2C%27+IEEE+Conference+on+Computer+Vision+and+Pattern+Recognition%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Domhan, T., Springenberg, J. T. and Hutter, F. (2015) &#x02018;Speeding up automatic hyperparameter optimization of deep neural networks by extrapolation of learning curves,&#x02019; in Proceedings of the 24th International Conference on Artificial Intelligence, Buenos Aires, Argentina. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Domhan%2C+T%2E%2C+Springenberg%2C+J%2E+T%2E+and+Hutter%2C+F%2E+%282015%29+%27Speeding+up+automatic+hyperparameter+optimization+of+deep+neural+networks+by+extrapolation+of+learning+curves%2C%27+in+Proceedings+of+the+24th+International+Conference+on+Artificial+Intelligence%2C+Buenos+Aires%2C+Argentina%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Ebrahimi, M., Suen, C. Y., and Ormandjieva, O. (2016) &#x02018;Detecting predatory conversations in social media by deep convolutional neural networks,&#x02019; Digital Investigation, 18, p. 33. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ebrahimi%2C+M%2E%2C+Suen%2C+C%2E+Y%2E%2C+and+Ormandjieva%2C+O%2E+%282016%29+%27Detecting+predatory+conversations+in+social+media+by+deep+convolutional+neural+networks%2C%27+Digital+Investigation%2C+18%2C+p%2E+33%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Egmont-Petersen, M., deRidder, D., and Handels, H. (2002) &#x02018;Image processing with neural networks-a review,&#x02019; Pattern Recognition, 35, pp. 2279&#x02013;2301. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Egmont-Petersen%2C+M%2E%2C+deRidder%2C+D%2E%2C+and+Handels%2C+H%2E+%282002%29+%27Image+processing+with+neural+networks-a+review%2C%27+Pattern+Recognition%2C+35%2C+pp%2E+2279-2301%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Facchinetti, S. (2009) &#x02018;A procedure to find exact critical values Of Kolmogorov-Smirnov test,&#x02019; Statistica Applicata &#x02013; Italian Journal of Applied Statistics 21(3&#x02013;4), pp. 337&#x02013;359. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Facchinetti%2C+S%2E+%282009%29+%27A+procedure+to+find+exact+critical+values+Of+Kolmogorov-Smirnov+test%2C%27+Statistica+Applicata+-+Italian+Journal+of+Applied+Statistics+21%283-4%29%2C+pp%2E+337-359%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Fisher, R. A. (1922) &#x02018;On the interpretation of <emphasis>&#x003C7;</emphasis>2 from contingency tables, and the calculation of P,&#x02019; <emphasis>Journal of the Royal Statistical Society</emphasis>, 85(1), pp. 87&#x02013;94, doi:10.2307/2340521. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Fisher%2C+R%2E+A%2E+%281922%29+%27On+the+interpretation+of+&#x03C7;2+from+contingency+tables%2C+and+the+calculation+of+P%2C%27+Journal+of+the+Royal+Statistical+Society%2C+85%281%29%2C+pp%2E+87-94%2C+doi%3A10%2E2307%2F2340521%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Fisher, R. A. (1954) <emphasis>Statistical Methods for Research Workers</emphasis>, Oliver and Boyd. ISBN 0-05-002170-2.</para></listitem>
<listitem>
<para>Fisher, L. D. and Lin, D. Y. (1999) &#x02018;Time-dependent covariates in the Cox proportional-hazards regression model,&#x02019; Annual Review of Public Health, 20, pp. 145&#x02013;157. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Fisher%2C+L%2E+D%2E+and+Lin%2C+D%2E+Y%2E+%281999%29+%27Time-dependent+covariates+in+the+Cox+proportional-hazards+regression+model%2C%27+Annual+Review+of+Public+Health%2C+20%2C+pp%2E+145-157%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Fox, J. and Weisberg, S. (2010) Appendix to An R Companion to Applied Regression Appendix to An R Companion to Applied Regression, SAGE Publications. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Fox%2C+J%2E+and+Weisberg%2C+S%2E+%282010%29+Appendix+to+An+R+Companion+to+Applied+Regression+Appendix+to+An+R+Companion+to+Applied+Regression%2C+SAGE+Publications%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Friedman, J. (1999) &#x02018;Stochastic gradient boosting,&#x02019; Computational Statistics and Data Analysis, 38, pp. 367&#x02013;378. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Friedman%2C+J%2E+%281999%29+%27Stochastic+gradient+boosting%2C%27+Computational+Statistics+and+Data+Analysis%2C+38%2C+pp%2E+367-378%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Gebraeel, N. Z. (2003) &#x02018;Real-time degradation modelling and residual life prediction for component maintenance and replacement,&#x02019; Ph.D. Dissertation. School of Industrial Engineering. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Gebraeel%2C+N%2E+Z%2E+%282003%29+%27Real-time+degradation+modelling+and+residual+life+prediction+for+component+maintenance+and+replacement%2C%27+Ph%2ED%2E+Dissertation%2E+School+of+Industrial+Engineering%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Gebraeel, N., Lawley, M. A., Li, R., and Ryan, J. (2005) &#x02018;Residuallife distributions from component degradation signals: A Bayesian approach,&#x02019; <emphasis>IIE Transactions 37</emphasis>, 37, pp. 543&#x02013;557. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Gebraeel%2C+N%2E%2C+Lawley%2C+M%2E+A%2E%2C+Li%2C+R%2E%2C+and+Ryan%2C+J%2E+%282005%29+%27Residuallife+distributions+from+component+degradation+signals%3A+A+Bayesian+approach%2C%27+IIE+Transactions+37%2C+37%2C+pp%2E+543-557%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Gers, F. A., Schmidhuber, J., Cummins, F. (2000) &#x02018;Learning to forget: Continual prediction with LSTM,&#x02019; Neural Computation, 12(10). <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Gers%2C+F%2E+A%2E%2C+Schmidhuber%2C+J%2E%2C+Cummins%2C+F%2E+%282000%29+%27Learning+to+forget%3A+Continual+prediction+with+LSTM%2C%27+Neural+Computation%2C+12%2810%29%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Guo, C. and Morris, S. A. (2017) &#x02018;Engineering cell identity: Establishing new gene regulatory and chromatin landscapes,&#x02019; Current opinion in genetics &#x00026; development, 46, pp. 50&#x02013;57. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Guo%2C+C%2E+and+Morris%2C+S%2E+A%2E+%282017%29+%27Engineering+cell+identity%3A+Establishing+new+gene+regulatory+and+chromatin+landscapes%2C%27+Current+opinion+in+genetics+%26+development%2C+46%2C+pp%2E+50-57%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Gustafsson, F. (2000) Adaptive Filtering and Change Detection. Wiley. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Gustafsson%2C+F%2E+%282000%29+Adaptive+Filtering+and+Change+Detection%2E+Wiley%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Haeusser, P. (2017) &#x02018;How computers learn to understand our world,&#x02019; [Online]. Available: http://www.in.tum.de/forschung/forschung shighlights/how-computers-learn-to-understand-our-world.html.</para></listitem>
<listitem>
<para>Hajinoroozi, M., Mao, Z., Jung, T., Lin, C., Huang, Y. (2016) &#x02018;EEG-based prediction of driver&#x02019;s cognitive performance by deep convolutional neural network,&#x02019; Signal Processing: Image Communication, 47, p. 549. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Hajinoroozi%2C+M%2E%2C+Mao%2C+Z%2E%2C+Jung%2C+T%2E%2C+Lin%2C+C%2E%2C+Huang%2C+Y%2E+%282016%29+%27EEG-based+prediction+of+driver%27s+cognitive+performance+by+deep+convolutional+neural+network%2C%27+Signal+Processing%3A+Image+Communication%2C+47%2C+p%2E+549%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Hasan, M., Kotov, A., Carcone, A. I., Dong, M., Naar, S., Hartlieb, K. B. (2016) &#x02018;A study of the effectiveness of machine learning methods for classification of clinical interview fragments into a large number of categories,&#x02019; Journal of Biomedical Informatics, 62, p. 21. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Hasan%2C+M%2E%2C+Kotov%2C+A%2E%2C+Carcone%2C+A%2E+I%2E%2C+Dong%2C+M%2E%2C+Naar%2C+S%2E%2C+Hartlieb%2C+K%2E+B%2E+%282016%29+%27A+study+of+the+effectiveness+of+machine+learning+methods+for+classification+of+clinical+interview+fragments+into+a+large+number+of+categories%2C%27+Journal+of+Biomedical+Informatics%2C+62%2C+p%2E+21%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Hijazi, S., Kumar, R., and Rowen, C. (2015) Using Convolutional Neural Networks for Image Recognition, Cadence. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Hijazi%2C+S%2E%2C+Kumar%2C+R%2E%2C+and+Rowen%2C+C%2E+%282015%29+Using+Convolutional+Neural+Networks+for+Image+Recognition%2C+Cadence%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Horster, H., Kauer, E., and Lechner, W. (1971) &#x02018;The Burn-out mechanism of incandescent lapms,&#x02019; Philips tech Rev, 32(6), pp. 155&#x02013;164. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Horster%2C+H%2E%2C+Kauer%2C+E%2E%2C+and+Lechner%2C+W%2E+%281971%29+%27The+Burn-out+mechanism+of+incandescent+lapms%2C%27+Philips+tech+Rev%2C+32%286%29%2C+pp%2E+155-164%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>International Institute of Welding. (2008) Recommendations For Fatigue Design Of Welded Joints And Components.</para></listitem>
<listitem>
<para>Ijjina, E. P., Chalavadi, K. M. (2016) &#x02018;Human action recognition using genetic algorithms and convolutional neural networks,&#x02019; Pattern Recognition, 59, p. 199. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ijjina%2C+E%2E+P%2E%2C+Chalavadi%2C+K%2E+M%2E+%282016%29+%27Human+action+recognition+using+genetic+algorithms+and+convolutional+neural+networks%2C%27+Pattern+Recognition%2C+59%2C+p%2E+199%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>ISO 13381-1, &#x0201C;Condition monitoring and diagnostics of machines-prognostics-part 1: General guidelines,&#x02019; International Standards Organization. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=ISO+13381-1%2C+%22Condition+monitoring+and+diagnostics+of+machines-prognostics-part+1%3A+General+guidelines%2C%27+International+Standards+Organization%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>James, G., Witten, D., Hastie, T., Tibshirani, R. (2014) An Introduction to Statistical Learning: With Applications in R, Springer. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=James%2C+G%2E%2C+Witten%2C+D%2E%2C+Hastie%2C+T%2E%2C+Tibshirani%2C+R%2E+%282014%29+An+Introduction+to+Statistical+Learning%3A+With+Applications+in+R%2C+Springer%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Jolliffe, I. T. (2002) Principal Component Analysis, 2nd edition, Springer.</para></listitem>
<listitem>
<para>Kaushik, M. and Mathur, B. (2014) &#x02018;Comparative study of K-means and hierarchical clustering techniques,&#x02019; International Journal of Software and Hardware Research in Engineering, pp. 93&#x02013;98. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Kaushik%2C+M%2E+and+Mathur%2C+B%2E+%282014%29+%27Comparative+study+of+K-means+and+hierarchical+clustering+techniques%2C%27+International+Journal+of+Software+and+Hardware+Research+in+Engineering%2C+pp%2E+93-98%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Kingma, D. P. and Jimmy, B. (2014) &#x02018;Adam: A method for stochastic optimization,&#x02019; <emphasis>arXiv preprint arXiv:1412.6980</emphasis>. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Kingma%2C+D%2E+P%2E+and+Jimmy%2C+B%2E+%282014%29+%27Adam%3A+A+method+for+stochastic+optimization%2C%27+arXiv+preprint+arXiv%3A1412%2E6980%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2012) &#x02018;Imagenet classification with deep convolutional neural networks,&#x02019; <emphasis>Advances in neural information processing systems</emphasis>, pp. 1097&#x02013;1105. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Krizhevsky%2C+A%2E%2C+Sutskever%2C+I%2E%2C+and+Hinton%2C+G%2E+E%2E+%282012%29+%27Imagenet+classification+with+deep+convolutional+neural+networks%2C%27+Advances+in+neural+information+processing+systems%2C+pp%2E+1097-1105%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Larsen, R. J. and Stroup, D. F. (1976) Statistics in the Real World: A Book of Examples, Macmillan. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Larsen%2C+R%2E+J%2E+and+Stroup%2C+D%2E+F%2E+%281976%29+Statistics+in+the+Real+World%3A+A+Book+of+Examples%2C+Macmillan%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>LeCun, Y., Bengio, Y., and Hinton, G. (2015) &#x02018;Deep learning,&#x02019; Nature, 521, pp. 436&#x02013;444. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=LeCun%2C+Y%2E%2C+Bengio%2C+Y%2E%2C+and+Hinton%2C+G%2E+%282015%29+%27Deep+learning%2C%27+Nature%2C+521%2C+pp%2E+436-444%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Lee, H., Largman, Y., Pham, P., and Ng, A. Y. (2009) &#x0201C;Unsupervised feature learning for audio classification using convolutional deep belief networks,&#x0201D; in 22nd International Conference on Neural Information Processing Systems (NIPS&#x02019;09), USA. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Lee%2C+H%2E%2C+Largman%2C+Y%2E%2C+Pham%2C+P%2E%2C+and+Ng%2C+A%2E+Y%2E+%282009%29+%22Unsupervised+feature+learning+for+audio+classification+using+convolutional+deep+belief+networks%2C%22+in+22nd+International+Conference+on+Neural+Information+Processing+Systems+%28NIPS%2709%29%2C+USA%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Levina, E. and Bickel, P. (2001). &#x0201C;The earthmover&#x02019;s distance is the mallows distance: Some insights from statistics,&#x02019; Proceedings of ICCV 2001, Vancouver, Canada, pp. 251&#x02013;256. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Levina%2C+E%2E+and+Bickel%2C+P%2E+%282001%29%2E+%22The+earthmover%27s+distance+is+the+mallows+distance%3A+Some+insights+from+statistics%2C%27+Proceedings+of+ICCV+2001%2C+Vancouver%2C+Canada%2C+pp%2E+251-256%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Li, J., Deng, L., Gong, Y., Haeb-Umbach, R. (2016) &#x02018;Robust automatic speech recognition &#x02013; a bridge to practical applications. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Li%2C+J%2E%2C+Deng%2C+L%2E%2C+Gong%2C+Y%2E%2C+Haeb-Umbach%2C+R%2E+%282016%29+%27Robust+automatic+speech+recognition+-+a+bridge+to+practical+applications%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Li, X. and Wu, X. (2015) &#x02018;Constructing long short-term memory based deep recurrent neural networks for large vocabulary speech recognition,&#x02019; IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Li%2C+X%2E+and+Wu%2C+X%2E+%282015%29+%27Constructing+long+short-term+memory+based+deep+recurrent+neural+networks+for+large+vocabulary+speech+recognition%2C%27+IEEE+International+Conference+on+Acoustics%2C+Speech+and+Signal+Processing+%28ICASSP%29%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Lipton, Z. C., Kale, D. C., Elkan, C., and Wetzell, R. (2016) &#x02018;Learning to diagnose with LSTM recurrent neural networks,&#x02019; International Conference on Learning Representations. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Lipton%2C+Z%2E+C%2E%2C+Kale%2C+D%2E+C%2E%2C+Elkan%2C+C%2E%2C+and+Wetzell%2C+R%2E+%282016%29+%27Learning+to+diagnose+with+LSTM+recurrent+neural+networks%2C%27+International+Conference+on+Learning+Representations%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Ljung, L. (1999) System Identification: Theory for the User, Prentice Hall, 2 edition. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ljung%2C+L%2E+%281999%29+System+Identification%3A+Theory+for+the+User%2C+Prentice+Hall%2C+2+edition%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>van der Maaten, L. and Hinton, G. (2008) &#x02018;Visualizing high-dimensional data using t-SNE,&#x02019; Journal of Machine Learning Research, 9, pp. 2579&#x02013;2605. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=van+der+Maaten%2C+L%2E+and+Hinton%2C+G%2E+%282008%29+%27Visualizing+high-dimensional+data+using+t-SNE%2C%27+Journal+of+Machine+Learning+Research%2C+9%2C+pp%2E+2579-2605%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Maier, H. R. and Dandy, G. C. (2000) &#x02018;Neural networks for the prediction and forecasting of water resources variables: A review of modelling issues and applications,&#x02019; Environmental Modelling Software, 15, pp. 101&#x02013;124. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Maier%2C+H%2E+R%2E+and+Dandy%2C+G%2E+C%2E+%282000%29+%27Neural+networks+for+the+prediction+and+forecasting+of+water+resources+variables%3A+A+review+of+modelling+issues+and+applications%2C%27+Environmental+Modelling+Software%2C+15%2C+pp%2E+101-124%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Mann, H. B. and Whitney, D. R. (1947) &#x02018;On a test of whether one of two random variables is stochastically larger than the other,&#x02019; Annals of Mathematical Statistics, 18(1), pp. 50&#x02013;60. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Mann%2C+H%2E+B%2E+and+Whitney%2C+D%2E+R%2E+%281947%29+%27On+a+test+of+whether+one+of+two+random+variables+is+stochastically+larger+than+the+other%2C%27+Annals+of+Mathematical+Statistics%2C+18%281%29%2C+pp%2E+50-60%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Mann, P. S. (1995). Introductory Statistics (2nd ed.). Wiley. ISBN 0-471-31009-3.</para></listitem>
<listitem>
<para>Mittelhammer, R. C., Judge, G. G., Miller, J., and Douglas, J. (2000). <emphasis>Econometric Foundations</emphasis>, Cambridge University Press. pp. 73&#x02013;74. ISBN 0-521-62394-4. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Mittelhammer%2C+R%2E+C%2E%2C+Judge%2C+G%2E+G%2E%2C+Miller%2C+J%2E%2C+and+Douglas%2C+J%2E+%282000%29%2E+Econometric+Foundations%2C+Cambridge+University+Press%2E+pp%2E+73-74%2E+ISBN+0-521-62394-4%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Moghaddam, A. H., Moghaddam, M. H., and Esfandyari, M. (2016) &#x02018;Stock market index prediction using artificial neural network,&#x02019; Journal of Economics, Finance and Administrative Science, 21, pp. 89-93. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Moghaddam%2C+A%2E+H%2E%2C+Moghaddam%2C+M%2E+H%2E%2C+and+Esfandyari%2C+M%2E+%282016%29+%27Stock+market+index+prediction+using+artificial+neural+network%2C%27+Journal+of+Economics%2C+Finance+and+Administrative+Science%2C+21%2C+pp%2E+89-93%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Nielsen, M. (2017) &#x02018;Using neural nets to recognize handwritten digits,&#x02019; Available: http://neuralnetworksand deeplearning.com/chap1.html.</para></listitem>
<listitem>
<para>Parzen, E. (1962) &#x02018;On estimation of a probability density function and mode,&#x02019; The Annals of Mathematical Statistics 33(3), p. 1065. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Parzen%2C+E%2E+%281962%29+%27On+estimation+of+a+probability+density+function+and+mode%2C%27+The+Annals+of+Mathematical+Statistics+33%283%29%2C+p%2E+1065%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Pearson, K. (1901) &#x0201C;On lines and planes of closest fit to systems of points in space,&#x0201D; Philosophical Magazine, 2(11), pp. 559&#x02013;572. doi:10.1080/14786440109462720. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Pearson%2C+K%2E+%281901%29+%22On+lines+and+planes+of+closest+fit+to+systems+of+points+in+space%2C%22+Philosophical+Magazine%2C+2%2811%29%2C+pp%2E+559-572%2E+doi%3A10%2E1080%2F14786440109462720%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Refenes, A.-P. N., Zapranis, A., and Francis, G. (1994) &#x02018;Stock performance modeling using neural networks: A comparative study with regression models,&#x02019; Neural Networks, 7, pp. 375&#x02013;388. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Refenes%2C+A%2E-P%2E+N%2E%2C+Zapranis%2C+A%2E%2C+and+Francis%2C+G%2E+%281994%29+%27Stock+performance+modeling+using+neural+networks%3A+A+comparative+study+with+regression+models%2C%27+Neural+Networks%2C+7%2C+pp%2E+375-388%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Rice, R. (1997) SAE Fatigue Design Handbook, Warrendale, Penn.: Society of Automotive Engineers.</para></listitem>
<listitem>
<para>Rifai, S. and Vincent, P., Muller, X., Glorot, X., and Bengio, Y. (2011) &#x02018;Contractive auto-encoders: explicit invariance during feature extraction&#x02019;, Proceedings of the 28th International Conference on Machine Learning. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Rifai%2C+S%2E+and+Vincent%2C+P%2E%2C+Muller%2C+X%2E%2C+Glorot%2C+X%2E%2C+and+Bengio%2C+Y%2E+%282011%29+%27Contractive+auto-encoders%3A+explicit+invariance+during+feature+extraction%27%2C+Proceedings+of+the+28th+International+Conference+on+Machine+Learning%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Rosenblatt, M. (1956) &#x02018;Remarks on some nonparametric estimates of a density function,&#x02019; The Annals of Mathematical Statistics, 27(3), p. 832. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Rosenblatt%2C+M%2E+%281956%29+%27Remarks+on+some+nonparametric+estimates+of+a+density+function%2C%27+The+Annals+of+Mathematical+Statistics%2C+27%283%29%2C+p%2E+832%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Schmidhuber, J. (2015) &#x02018;Deep learning in neural networks: An overview,&#x02019; Neural Networks, 61, pp. 85&#x02013;117. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Schmidhuber%2C+J%2E+%282015%29+%27Deep+learning+in+neural+networks%3A+An+overview%2C%27+Neural+Networks%2C+61%2C+pp%2E+85-117%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Shapiro, S. S. and Wilk, M. B. (1965) &#x02018;An analysis of variance test for normality (complete samples),&#x02019; Biometrika, 52(3&#x02013;4), pp. 591&#x02013;611. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Shapiro%2C+S%2E+S%2E+and+Wilk%2C+M%2E+B%2E+%281965%29+%27An+analysis+of+variance+test+for+normality+%28complete+samples%29%2C%27+Biometrika%2C+52%283-4%29%2C+pp%2E+591-611%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Si, X.-S., Wang, W., Hua, C.-H., and Zhou, D.-H. (2011) &#x02018;Remaining useful life estimation &#x02013; A review on the statistical data driven approaches,&#x02019; <emphasis>European Journal of Operational Research</emphasis>, 213, pp. 1&#x02013;14. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Si%2C+X%2E-S%2E%2C+Wang%2C+W%2E%2C+Hua%2C+C%2E-H%2E%2C+and+Zhou%2C+D%2E-H%2E+%282011%29+%27Remaining+useful+life+estimation+-+A+review+on+the+statistical+data+driven+approaches%2C%27+European+Journal+of+Operational+Research%2C+213%2C+pp%2E+1-14%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Sikorska, J., Hodkiewicz, M., and Ma, L. (2011) &#x02018;Prognostic modelling options for remaining useful life estimation by industry,&#x02019; <emphasis>Mechanical Systems and Signal Processing</emphasis> 25(5), pp. 1803&#x02013;1836. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Sikorska%2C+J%2E%2C+Hodkiewicz%2C+M%2E%2C+and+Ma%2C+L%2E+%282011%29+%27Prognostic+modelling+options+for+remaining+useful+life+estimation+by+industry%2C%27+Mechanical+Systems+and+Signal+Processing+25%285%29%2C+pp%2E+1803-1836%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Simonyan, K. and Zisserman, A. (2014) &#x02018;Very deep convolutional networks for large-scale image recognition,&#x02019; <emphasis>arXiv preprint arXiv:1409.1556</emphasis>. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Simonyan%2C+K%2E+and+Zisserman%2C+A%2E+%282014%29+%27Very+deep+convolutional+networks+for+large-scale+image+recognition%2C%27+arXiv+preprint+arXiv%3A1409%2E1556%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Simpson, J. W. and Sheppard, W. (1998) &#x02018;Inducing diagnostic inference models from case data,&#x02019; in Frontiers in Electronic Testing, vol. 13, Kluwer. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Simpson%2C+J%2E+W%2E+and+Sheppard%2C+W%2E+%281998%29+%27Inducing+diagnostic+inference+models+from+case+data%2C%27+in+Frontiers+in+Electronic+Testing%2C+vol%2E+13%2C+Kluwer%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Stuart, A., Ord, K., and Arnold, S. (1999) Kendall&#x02019;s Advanced Theory of Statistics: Volume 2A&#x02014;Classical Inference &#x00026; the Linear Mode. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Stuart%2C+A%2E%2C+Ord%2C+K%2E%2C+and+Arnold%2C+S%2E+%281999%29+Kendall%27s+Advanced+Theory+of+Statistics%3A+Volume+2A-Classical+Inference+%26+the+Linear+Mode%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Tamilselvan, P. and Wang, P. (2013) &#x02018;Failure diagnosis using deep belief learning based health state classification,&#x02019; Reliability Engineering System Safety, 115, pp. 124&#x02013;135. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Tamilselvan%2C+P%2E+and+Wang%2C+P%2E+%282013%29+%27Failure+diagnosis+using+deep+belief+learning+based+health+state+classification%2C%27+Reliability+Engineering+System+Safety%2C+115%2C+pp%2E+124-135%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Team, R. C. (2014) &#x02018;R: A language and environment for statistical computing,&#x02019; R Foundation for Statistical Computing, Vienna, Austria. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Team%2C+R%2E+C%2E+%282014%29+%27R%3A+A+language+and+environment+for+statistical+computing%2C%27+R+Foundation+for+Statistical+Computing%2C+Vienna%2C+Austria%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Therneau, T. M. (2015) &#x02018;A Package for Survival Analysis in S.&#x02019; <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Therneau%2C+T%2E+M%2E+%282015%29+%27A+Package+for+Survival+Analysis+in+S%2E%27" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Therneau, T. C. C. and Atkinson, E. (2017) &#x02018;Using time dependent covariates and time dependent coefficients in the cox model,&#x02019; Survival Vignettes. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Therneau%2C+T%2E+C%2E+C%2E+and+Atkinson%2C+E%2E+%282017%29+%27Using+time+dependent+covariates+and+time+dependent+coefficients+in+the+cox+model%2C%27+Survival+Vignettes%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Thomas, L. and Reyes, E. M. (2014) &#x02018;Tutorial: Survival estimation for cox regression models with time-varying coefficients using SAS and R,&#x02019; Journal of Statistical Software, 61. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Thomas%2C+L%2E+and+Reyes%2C+E%2E+M%2E+%282014%29+%27Tutorial%3A+Survival+estimation+for+cox+regression+models+with+time-varying+coefficients+using+SAS+and+R%2C%27+Journal+of+Statistical+Software%2C+61%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Tran, T. D., Khoa Nguyen, D., Tran, T. A. X., Nguyen, Q. C., and Nguyen, H. B. (2011) &#x02018;Speech enhancement using combination of dereverberation and noise reduction for robust speech recognition,&#x02019; in Proceedings of the Second Symposium on Information and Communication Technology, Hanoi, Viet Nam. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Tran%2C+T%2E+D%2E%2C+Khoa+Nguyen%2C+D%2E%2C+Tran%2C+T%2E+A%2E+X%2E%2C+Nguyen%2C+Q%2E+C%2E%2C+and+Nguyen%2C+H%2E+B%2E+%282011%29+%27Speech+enhancement+using+combination+of+dereverberation+and+noise+reduction+for+robust+speech+recognition%2C%27+in+Proceedings+of+the+Second+Symposium+on+Information+and+Communication+Technology%2C+Hanoi%2C+Viet+Nam%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Tu, J. V. (1996) &#x02018;Advantages and disadvantages of using artificial neural networks versus logistic regression for predicting medical outcomes,&#x02019; Journal of Clinical Epidemiology, 49, pp. 1225&#x02013;1231. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Tu%2C+J%2E+V%2E+%281996%29+%27Advantages+and+disadvantages+of+using+artificial+neural+networks+versus+logistic+regression+for+predicting+medical+outcomes%2C%27+Journal+of+Clinical+Epidemiology%2C+49%2C+pp%2E+1225-1231%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Tukey, J. (1949) &#x02018;Comparing individual means in the analysis of variance,&#x02019; Biometrics 5(2), pp. 99&#x02013;114. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Tukey%2C+J%2E+%281949%29+%27Comparing+individual+means+in+the+analysis+of+variance%2C%27+Biometrics+5%282%29%2C+pp%2E+99-114%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Vapnik, V. N. (1995) The Nature of Statistical Learning Theory, Springer-Verlag. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Vapnik%2C+V%2E+N%2E+%281995%29+The+Nature+of+Statistical+Learning+Theory%2C+Springer-Verlag%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Vapnik, V. N. and Chervonenkis, A. Y. (1971) &#x0201C;On the uniform convergence of relative frequencies of events to their probabilities,&#x02019; Theory of Probability &#x00026; Its Applications, 16(2), p. 264. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Vapnik%2C+V%2E+N%2E+and+Chervonenkis%2C+A%2E+Y%2E+%281971%29+%22On+the+uniform+convergence+of+relative+frequencies+of+events+to+their+probabilities%2C%27+Theory+of+Probability+%26+Its+Applications%2C+16%282%29%2C+p%2E+264%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Wang, Z. and Oates, T. (2015) &#x02018;Spatially encoding temporal correlations to classify temporal data using convolutional neural networks,&#x02019; CoRR. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Wang%2C+Z%2E+and+Oates%2C+T%2E+%282015%29+%27Spatially+encoding+temporal+correlations+to+classify+temporal+data+using+convolutional+neural+networks%2C%27+CoRR%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Webers, G., Boosten, M., and Sprong, H. (2016) &#x02018;Mantis D3.1 Report on sensors selected and to develop - Appendix 9,&#x02019; ECSEL.</para></listitem>
<listitem>
<para>Welte, T. M. and Wang, K. (2014) &#x02018;Models for lifetime estimation: An overview with focus on applications to wind turbines,&#x0201D; <emphasis>Advances in Manufacturing</emphasis> 2(1), pp. 79&#x02013;87. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Welte%2C+T%2E+M%2E+and+Wang%2C+K%2E+%282014%29+%27Models+for+lifetime+estimation%3A+An+overview+with+focus+on+applications+to+wind+turbines%2C%22+Advances+in+Manufacturing+2%281%29%2C+pp%2E+79-87%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Widrow, B., Rumelhart, D. E., and Lehr, M. A. &#x02018;Neural networks: Applications in industry, business and science,&#x02019; Commun. ACM, 37, pp. 93&#x02013;105. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Widrow%2C+B%2E%2C+Rumelhart%2C+D%2E+E%2E%2C+and+Lehr%2C+M%2E+A%2E+%27Neural+networks%3A+Applications+in+industry%2C+business+and+science%2C%27+Commun%2E+ACM%2C+37%2C+pp%2E+93-105%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Xu, X., Yan, Z., and Xu, S. (2015). &#x02018;Estimating wind speed probability distribution by diffusion-based kernel density method,&#x02019; Electric Power Systems Research, 121, pp. 28&#x02013;37. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Xu%2C+X%2E%2C+Yan%2C+Z%2E%2C+and+Xu%2C+S%2E+%282015%29%2E+%27Estimating+wind+speed+probability+distribution+by+diffusion-based+kernel+density+method%2C%27+Electric+Power+Systems+Research%2C+121%2C+pp%2E+28-37%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Yuea, W., Xiaojiea, W., and Yuzhao, M. (2016) &#x02018;First-feed LSTM model for video description,&#x02019; The Journal of China Universities of Posts and Telecommunications, 23(3), p. 89. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Yuea%2C+W%2E%2C+Xiaojiea%2C+W%2E%2C+and+Yuzhao%2C+M%2E+%282016%29+%27First-feed+LSTM+model+for+video+description%2C%27+The+Journal+of+China+Universities+of+Posts+and+Telecommunications%2C+23%283%29%2C+p%2E+89%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Zachary, J. B., Lipton, C., and Elkan, C. (2015) A Critical Review of Recurrent Neural Networks for Sequence Learning, CoRR. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Zachary%2C+J%2E+B%2E%2C+Lipton%2C+C%2E%2C+and+Elkan%2C+C%2E+%282015%29+A+Critical+Review+of+Recurrent+Neural+Networks+for+Sequence+Learning%2C+CoRR%2E" target="_blank">Google Scholar</ulink></para></listitem>
</orderedlist>
</section>
</chapter>
<chapter class="chapter" id="ch06" label="6" xreflabel="6">
<title>From KPI Dashboards to Advanced Visualization</title>
<para><emphasis role="strong">Goreti Marreiros<superscript><emphasis role="strong">1</emphasis></superscript>, Peter Craamer<superscript><emphasis role="strong">2</emphasis></superscript>, I&#x000F1;aki Garitano<superscript><emphasis role="strong">3</emphasis></superscript>, Roberto Gonz&#x000E1;lez<superscript><emphasis role="strong">4</emphasis></superscript>, Manja Gorenc Novak<superscript><emphasis role="strong">5</emphasis></superscript>, Ale&#x00161; Kancilija<superscript><emphasis role="strong">5</emphasis></superscript>, Quinten Marcelis<superscript><emphasis role="strong">6</emphasis></superscript>, Diogo Martinho<superscript><emphasis role="strong">1</emphasis></superscript>, Antti Niemel&#x000E4;<superscript><emphasis role="strong">7</emphasis></superscript>, Franc Novak<superscript><emphasis role="strong">8</emphasis></superscript>, Gregor Papa<superscript><emphasis role="strong">8</emphasis></superscript>, &#x00160;pela Poklukar<superscript><emphasis role="strong">8</emphasis></superscript>, Isabel Pra&#x000E7;a<superscript><emphasis role="strong">1</emphasis></superscript>, Ville Rauhala<superscript><emphasis role="strong">7</emphasis></superscript>, Daniel Reguera<superscript><emphasis role="strong">3</emphasis></superscript>, Marjan &#x00160;terk<superscript><emphasis role="strong">5</emphasis></superscript>, Gorka Urchegui<superscript><emphasis role="strong">2</emphasis></superscript>, Roberto Uribeetxeberria<superscript><emphasis role="strong">3</emphasis></superscript>, Juha Valtonen<superscript><emphasis role="strong">7</emphasis></superscript>, and Anja Vidmar<superscript><emphasis role="strong">5</emphasis></superscript></emphasis></para>
<para><superscript>1</superscript>ISEP, Polytechnic Institute of Porto, Porto, Portugal</para>
<para><superscript>2</superscript>Mondragon Sistemas De Informacion, Spain</para>
<para><superscript>3</superscript>Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain</para>
<para><superscript>4</superscript>Tekniker, Spain</para>
<para><superscript>5</superscript>XLAB, Ljubljana, Slovenia</para>
<para><superscript>6</superscript>Ilias Solutions, Belgium</para>
<para><superscript>7</superscript>Lapland University of Applied Sciences Ltd, Finland</para>
<para><superscript>8</superscript>Jo&#x0017E;ef Stefan Institute, Slovenia</para>
<para>New technologies are being developed towards Industry 4.0 such as the establishment of smart factories, smart products and smart services embedded in an internet of things and of services. As a result, the development of prognostic and collaborative technologies have become a necessity. New technological solutions that can make the best use of existing physical assets while following company&#x02019;s important metrics to provide a quick overview of business performance are now starting to be developed. Examples of such solutions include the development of dashboards that can show the Key Performance Indicators (KPI) for processes involved in the production site and Human-Machine Interfaces (HMI) that allow the visualization of data collected directly from the machine in different formats such as graphic, table or even through augmented and virtual reality, while their adaptive interfaces can present relevant information according to the user and the context.</para>
<para>This chapter explores these new technological solutions in the context of the MANTIS project and consists of four main sections. The first section presents the HMI technology and the specifications, design principles and recommendations, requirements, and modelling of the HMI that guides and follows the principles of the MANTIS Reference Architecture (see <link linkend="ch02">Chapter <xref linkend="ch02" remap="2"/></link>). The second section discusses the concept of adaptive interfaces and the two main approaches that were considered (Context-aware and Interaction based/driven). The third section discusses advanced data visualization methods for HMIs and presents different scenarios according to the different use cases defined in MANTIS. In the last section, the usability testing methodology considered for industrial HMIs is discussed.</para>
<section class="lev1" id="sec6-1">
<title>6.1 HMI Functional Specifications and Interaction Model</title>
<para>Human-machine interaction denotes real-time interaction and communication between human users and a machine via a human-machine interface [Techopedia, 2011]. Hereby, the term &#x0201C;machine&#x0201D; indicates any kind of dynamic technical system and it relates to different technical and production processes in diverse application domains. Beside traditional functionalities of HMI such as presentation and processing of information, advanced features include explanation and adaptability based on user and application models and knowledge-based systems for decision support. HMI for proactive maintenance should therefore contribute to:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>enhanced monitoring of shop-floor conditions (i.e., the machines health condition, the efficiency of the production lines, and the safety of workers);</para></listitem>
<listitem>
<para>automatic self-adaptation of control strategies based on the context related to user status, machine status, general environment and time;</para></listitem>
<listitem>
<para>user-friendly, ergonomic and intuitive interaction between workers and machines, and consequently positive user motivation leading to higher efficiency and safety.</para></listitem>
</itemizedlist>
<para>While MANTIS strongly emphasises autonomy, self-testing and self-adaptation, human role remains one of the important factors in system operation. The human role is twofold: controlling, which comprises continuous and discrete tasks of open- and closed-loop activities, and problem solving which includes the higher cognitive tasks of fault management and planning. The increased degree of automation in control of dynamic technical systems does not replace the human users, but rather modifies the interaction between both. Appropriate matching of both leads to a user-centred design.</para>
<para>To propose a user-centred design in the project comprising eleven use cases from four different sectors transpired to be a challenging task. Diversity of MANTIS use cases resulted in a wide range of requirements that could hardly fit a common MANTIS HMI structure. Design and development of a common MANTIS HMI is not only difficult but also most likely to result in a poor usability of the products. One of the goals of the MANTIS project is therefore to offer a common ground for designing user-centred, usable and use case specific HMIs. The goal has been achieved by developing the human-machine proactive maintenance-based interaction model that on one hand covers all proactive maintenance related requirements of every use case and stays general enough to be applicable to every MANTIS and potential future use case. The MANTIS HMI model has been conceived to provide means that would help to identify the HMI content elements and their relationships of a given use case. Together with the functional specification, described later in this section, it may serve as a reference point for writing use case specific requirements specifications and for designing the user interaction.</para>
<section class="lev2" id="sec6-1-1">
<title>6.1.1 HMI Design Principle Followed in the MANTIS Project</title>
<para>MANTIS followed Scenario-Based Design (SBD) [MANTIS Consortium, 2016] which is an established approach for describing the use of a system at an early point in the development process. Narrative descriptions of the envisioned scenarios help to guide the development of the system and serve, among others, as a basis for setting efficient human-machine interaction. Use case owners described some typical problem scenarios and refined them through the activity, information and interaction phase and as a result to provide the scenarios that would include sufficient details for HMI prototyping [MANTIS Consortium, 2016].</para>
<para>Scenarios gathered in this document are the result of iterative process of the SBD phase, which is reflected in their common structure:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>situation that describes the circumstances in which the scenario occurs, focused on the perspective as seen by the user;</para></listitem>
<listitem>
<para>device which holds the interaction, most suitable for the nature of the user&#x02019;s activity;</para></listitem>
<listitem>
<para>information, available for the user (<emphasis>What options are available in the interface</emphasis>);</para></listitem>
<listitem>
<para>the way user interacts with the interface (<emphasis>How is it done</emphasis>);</para></listitem>
<listitem>
<para>current implementation.</para></listitem>
</itemizedlist>
<para>The activity scenarios describe pure functionality of HMI. They have been refined to information scenarios through the <emphasis>What options are available in the interface</emphasis> section of each scenario and further particularized to interaction scenarios through the <emphasis>How is it done</emphasis> section. This way the scenarios are elaborated to the point where they provide the details of user action and feedback.</para>
<para>A common structure of each scenario intended to unify the diversity of requirements, imposed by the wide range of distinct use cases, and to gather functional specifications, common to all use cases and specific for proactive maintenance.</para>
</section>
<section class="lev2" id="sec6-1-2">
<title>6.1.2 MANTIS HMI Specifications</title>
<para>To provide the right information, in the right modality and in the best way for users when needed, the user interface should be highly personalized and adapted to each specific user or user role. Any unification of the HMI design might impose the constraints that could result in an HMI with a poor usability.</para>
<para>The approach, adopted in the MANTIS project, therefore focuses on the requirements, common to most of the use cases and specifics for proactive and collaborative maintenance. In the following section, a generic MANTIS HMI is specified to the extent that does not introduce any constraints for the use cases, but at the same time describes the most important features of the MANTIS HMI that should be considered when designing the HMI in individual use cases.</para>
<section class="lev3" id="sec6-1-2-1">
<title>6.1.2.1 Functional specifications</title>
<para>The specifications provided in this section, are the result of refinement of scenarios, provided by industrial partners. Functional specifications describe the HMI functionalities present in most use cases and abstracted from the specific situation of every single use case. They are not meant as a replacement of MANTIS HMI requirements specifications for a separate use case but may serve as a reference point when writing ones.</para>
<para><emphasis role="strong">Scheduling of Maintenance Tasks</emphasis></para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>MANTIS HMI allows the user to see all relevant maintenance tasks together with some additional information such as description of the task (including suggested time schedule), relevant asset related information (e.g., sensor logs, maintenance history and statistics), guides, manuals or instructions for maintenance task, task progress information and client information;</para></listitem>
<listitem>
<para>MANTIS HMI allows adding of task related information, such as task acceptance or rejection, task progress (e.g., start and stop indication), assigning resources (e.g., necessary time and equipment);</para></listitem>
<listitem>
<para>MANTIS HMI allows adding of asset related information such as asset status, image of the failure (in case of failure), and feedback to the system (e.g., identification of the failure root cause, estimation of the actual wear, &#x02026;);</para></listitem>
<listitem>
<para>MANTIS HMI allows spare parts managing. This may include the inquiry of spare part availability, ordering spare parts and vendors contact information;</para></listitem>
<listitem>
<para>MANTIS HMI allows maintenance tasks rescheduling (automatically, based on MANTIS maintenance optimization and manually by the user);</para></listitem>
<listitem>
<para>Maintenance tasks display enables filtering and sorting;</para></listitem>
<listitem>
<para>Maintenance tasks display is updated immediately after a new maintenance task is scheduled;</para></listitem>
<listitem>
<para>MANTIS HMI is able to automatically generate reports on maintenance activities and to transfer the maintenance related data to other users;</para></listitem>
<listitem>
<para>MANTIS HMI displays an alert in case of asset failure, or if the spare parts required for scheduled maintenance task are not available. The alert contains additional information, such as description of the failure, asset status, or the additional information on spare parts.</para></listitem>
</itemizedlist>
<para><emphasis role="strong">Monitoring Assets</emphasis></para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>MANTIS HMI displays current, historical and predicted parameter values of monitored assets and expected range of these parameters;</para></listitem>
<listitem>
<para>MANTIS HMI displays comparison between actual and estimated asset wear and/or predicted remaining useful life;</para></listitem>
<listitem>
<para>MANTIS HMI displays various statistics of historical parameter values of monitored assets;</para></listitem>
<listitem>
<para>MANTIS HMI displays possible failures of assets together with some additional description, such as current and historical parameter values related to the faulty asset and possible feedback from other;</para></listitem>
<listitem>
<para>MANTIS HMI allows the user to sort and filter monitored assets, select different data sources, and to select time range of monitored parameters;</para></listitem>
<listitem>
<para>MANTIS HMI allows the user to select and flag the data;</para></listitem>
<listitem>
<para>MANTIS HMI is able to automatically generate reports on monitored parameters and to transfer the monitored data to other users;</para></listitem>
<listitem>
<para>MANTIS HMI displays data in real time;</para></listitem>
<listitem>
<para>MANTIS HMI displays an alert if the monitored asset parameter is out of a predefined range. The alert carries additional information on the monitored asset parameter such as historical values of the parameter.</para></listitem>
</itemizedlist>
<para><emphasis role="strong">Data Analysis</emphasis></para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>MANTIS HMI is able to display remaining useful life of the assets, predicted future values of monitored parameters, comparison between predicted and actual parameter values and feedback from other users in textual as well as graphical form;</para></listitem>
<listitem>
<para>MANTIS HMI allows the user to manage prediction models. This includes model inspection, activation or deactivation of the model, updating, generating and evaluating predictions;</para></listitem>
<listitem>
<para>MANTIS HMI displays an alert if the prediction performance of MANTIS system is below the predefined threshold. The alert will carry additional information on prediction performance.</para></listitem>
</itemizedlist>
<para><emphasis role="strong">Reporting</emphasis></para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>MANTIS HMI is able to generate automatic reports in pdf or html format;</para></listitem>
<listitem>
<para>MANTIS HMI is able to process spoken reports;</para></listitem>
<listitem>
<para>MANTIS HMI allows the user to manually generate reports. This includes information input (textual and graphical) and data export.</para></listitem>
</itemizedlist>
<para><emphasis role="strong">Communication</emphasis></para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>MANTIS HMI supports the textual, visual and audial communication among the users;</para></listitem>
<listitem>
<para>MANTIS HMI enables the transfer of different data sources, images, videos, and documents among the users as well as to and from the MANTIS platform.</para></listitem>
</itemizedlist>
</section>
<section class="lev3" id="sec6-1-2-2">
<title>6.1.2.2 General requirements</title>
<para>The following general requirements have also been identified by industrial partners and must require that:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>MANTIS HMI performs just-in-time;</para></listitem>
<listitem>
<para>MANTIS HMI supports different types of devices, including mobile phones, tablet computers, laptop computers, network computers and machine displays;</para></listitem>
<listitem>
<para>MANTIS HMI supports at least the most commonly used operating systems such as Windows, Android, and IOS;</para></listitem>
<listitem>
<para>MANTIS HMI supports multiple levels of alerts;</para></listitem>
<listitem>
<para>MANTIS HMI allows user&#x02019;s input such as confirmation of the alert and providing feedback to the intelligent MANTIS features;</para></listitem>
<listitem>
<para>MANTIS HMI is able to automatically generate reports on alerting activities and to transfer the maintenance related data to other users.</para></listitem>
</itemizedlist>
</section>
</section>
<section class="lev2" id="sec6-1-3">
<title>6.1.3 MANTIS HMI Model</title>
<para>As stated in [Techopedia, 2011], human-machine interface (HMI) is a component of certain devices that are capable of handling human-machine interactions [Boy, 2011]. The interface consists of hardware and software that allow user inputs to be translated as signals for machines that, in turn, provide the required result to the user. Since HMI technology is ubiquitous, the interfaces involved can include motion sensors, keyboards and similar peripheral devices, speech-recognition interfaces and any other interaction in which information is exchanged using sight, sound, heat and other cognitive and physical modes are considered to be part of HMIs.</para>
<para>The initial phase of scenario-based design approach commonly applied for all MANTIS use cases resulted in an extensive set of divergent scenarios, which required considerable additional activities to get the descriptions suitable for HMI design.</para>
<para>In the following we describe a generic static model that can be used together with the requirement specifications of each individual use case to formalize the structure of the target HMI implementation. The model has been conceived, in particular, with two ideas in mind: to provide means that would help to identify the HMI content elements and their relationships of a given use case and to unify (as much as possible) the HMI design of different use cases, which would be useful for comparison of implementations and exchange of good practices. When setting up the model structure we follow the concepts of descriptive models applied in task analysis [Diaper, et al., 2004] and add specifics of MANTIS gathered maintenance scenarios, denoted as MANTIS high level tasks. For each of these high level tasks we provide a list of functionalities supporting the given high level task. In addition to the model, a Requirements Specification template was used to identify the HMI content elements and their functionalities supporting the high level tasks of a given use case.</para>
<fig id="F6-1" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-1">Figure <xref linkend="F6-1" remap="6.1"/></link></label>
<caption><para>MANTIS human-machine interaction.</para></caption>
<graphic xlink:href="graphics/ch06_fig001.jpg"/>
</fig>
<para>MANTIS human-machine interaction comprises five main elements (<link linkend="F6-1">Figure <xref linkend="F6-1" remap="6.1"/></link>):</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>user interfaces;</para></listitem>
<listitem>
<para>users;</para></listitem>
<listitem>
<para>MANTIS platform;</para></listitem>
<listitem>
<para>assets;</para></listitem>
<listitem>
<para>environment.</para></listitem>
</itemizedlist>
<para>MANTIS platform allows communication with several different users through their user interface. Interaction between the users and the platform is bidirectional; users can not only access the information retrieved from assets and stored in the platform, but can also provide an input to the MANTIS system. In case of proactive maintenance, user feedback is especially valuable since it is providing additional input to the prediction algorithms. Users can also initiate an operation which is then carried out by the platform, such as rescheduling maintenance task, or respond to a system triggered operation such as handling an alarm. Another aspect of the user interaction is the communication between different users through the MANTIS platform. In addition to the straightforward communication in terms of the textual or video chat functions, the users can also communicate via established workflows or shared widgets.</para>
<para>Although environment cannot be treated neither as a direct link between the user and the system nor as a part of communication between the users, it can significantly improve the human-machine interaction in terms of context-aware functionalities.</para>
<para>Proactive and collaborative maintenance oriented human-machine interaction within the MANTIS system supports five main high-level user tasks:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>monitoring assets;</para></listitem>
<listitem>
<para>data analysis;</para></listitem>
<listitem>
<para>maintenance tasks scheduling;</para></listitem>
<listitem>
<para>reporting;</para></listitem>
<listitem>
<para>communication.</para></listitem>
</itemizedlist>
<para>These tasks were identified as the key user tasks in the initial process of user-centred design. Monitoring the assets, data analysis and maintenance task scheduling proved to be vital for proactive maintenance, while reporting and communication allow collaboration between different user roles. Tasks can be carried out using several MANTIS specific functionalities that can be classified as user input, system output, user- or system- triggered operation. Functionalities, described later in this section, cover all the main aspects of MANTIS human-machine interaction and are general enough to be applicable to any MANTIS or potential future use case.</para>
<section class="lev3" id="sec6-1-3-1">
<title>6.1.3.1 Functionalities supporting high level tasks</title>
<para>A more detailed description of each functionality previously identified is now provided including the expected Output, Input and/or related Operations.</para>
<para><emphasis role="strong">Scheduling Maintenance Tasks</emphasis></para>
<para>Output:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Maintenance tasks schedule together with some additional information for each task, such as:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>description of the task (including suggested time schedule);</para></listitem>
<listitem>
<para>relevant asset related information (e.g., sensor logs, maintenance history and statistics);</para></listitem>
<listitem>
<para>task progress information;</para></listitem>
<listitem>
<para>client information;</para></listitem>
<listitem>
<para>spare parts related information (availability, vendor information).</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
<para>Input:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Input of task related information, such as:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>task acceptance/rejection;</para></listitem>
<listitem>
<para>task progress (e.g., start and stop indication);</para></listitem>
<listitem>
<para>assigned resources (e.g., necessary time and equipment).</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>Input of asset related information:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>asset state information (e.g., description, image, failure description);</para></listitem>
<listitem>
<para>feedback to the system (e.g., identification of the failure root cause, estimation of the actual wear).</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
<para>Operations:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>User-triggered operations:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>task rescheduling;</para></listitem>
<listitem>
<para>filtering and sorting maintenance tasks;</para></listitem>
<listitem>
<para>report generation (automatic or manual);</para></listitem>
<listitem>
<para>spare parts management.</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>System-triggered operations:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>update when new task is scheduled;</para></listitem>
<listitem>
<para>alert (when new task is scheduled, spare parts are not available, etc.).</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
<para>In scheduling maintenance tasks, the display of maintenance tasks schedule, produced as a result of MANTIS system intelligent functions, is the most important functionality. It offers the overview of all relevant maintenance tasks and provides them with additional, task related information such as task description (including suggested time schedule), relevant asset related information (e.g., sensor logs, maintenance history and statistics), maintenance personnel support such as guides, manuals or instructions for a maintenance task, task progress information and client information. Information, displayed in the schedule is, together with the user input data, one of the main sources for automatic report generation.</para>
<para>MANTIS HMI allows the user to input some of the task and critical asset related information, such as task acceptance or rejection, task progress (e.g., start and stop indication), assigning resources (e.g., necessary time and equipment), critical asset status, or image of the failure (in case of failure). Such input is important for monitoring the maintenance activities&#x02019; progress and especially for providing feedback to the system (e.g., identification of the failure root cause, estimation of the actual wear, etc.), which may have a considerable impact on improvement of predictive algorithms. The user feedback may be taken into account in two ways. It can serve as a direct input to predictive models, or it can be used indirectly as a domain expert knowledge that can provide an important insight in the quality of predictive models.</para>
<para>Schedule updating operation can be triggered automatically by the system when a new maintenance task is scheduled according to the maintenance tasks scheduling algorithms. If the newly scheduled maintenance task is considered critical for the production process or for the health of assets, the system may trigger an alert as well. These two system-triggered operations can affect the display of the maintenance task schedule by changing the schedule or/and in case of alert by modifying the graphical display of the schedule, which happens mostly in the case of a critical maintenance task.</para>
<para>Manual rescheduling, filtering and sorting maintenance tasks, and spare parts managing are user-initiated operations and can affect the display of the maintenance task schedule. In addition to these operations, users can trigger the automatic report generation. In response to this user action the system gathers tasks related information and the user input to generate a report in any desired format.</para>
<para><emphasis role="strong">Monitoring</emphasis></para>
<para>Output:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Current, historical and predicted parameter values of monitored assets, together with the expected range of monitored parameters;</para></listitem>
<listitem>
<para>Comparison between actual and estimated wear or predicted remaining useful life;</para></listitem>
<listitem>
<para>Possible failures of assets together with some additional description (e.g., current and historical parameter values related to the faulty asset, possible feedback from other users).</para></listitem>
</itemizedlist>
<para>Input:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Flagging the data.</para></listitem>
</itemizedlist>
<para>Operations:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>User-triggered operations:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>sort and filter monitored assets;</para></listitem>
<listitem>
<para>select different data sources;</para></listitem>
<listitem>
<para>select time range of monitored parameters;</para></listitem>
<listitem>
<para>generate reports on monitored parameters;</para></listitem>
<listitem>
<para>transfer the monitored data to other users.</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>System-triggered operations:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>continuous updating of the monitored parameters;</para></listitem>
<listitem>
<para>alert when monitored asset parameter out of predefined range.</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
<para>The most common assets monitoring related functionality is definitively the real time display of parameter values, measured by multiple sensors in the MANTIS system. Amounts, displayed on the user interface, vary from the actual current and historical parameter values to the predicted future parameter values. In case of abnormal values of these parameters, the interface can adapt the display to alert the users. It is often required to display the expected (normal) range of the parameter values, the comparison between the predicted and actual parameter values or remaining useful life of the asset, and various statistics of historical parameter values of the monitored asset.</para>
<para>Other monitoring features include the display of the possible assets failures together with some additional description, such as current and historical parameter values related to the faulty asset and possibly the feedback from other users.</para>
<para>Although the user input is typically not required for monitoring itself, MANTIS HMI should allow the user to flag, label and comment the data. In this way, the users can provide the additional data that might not be captured by the sensors.</para>
<para>Real time display of information is often vital for an efficient maintenance process which means that the MANTIS HMI should be able to frequently update the parameter values. Also, the display of alerts if the monitored asset parameters are out of predefined range is another important system-triggered operation. It is often helpful if the alert carries some additional information related to the monitored asset parameter such as historical values of the parameter. Both operations have influence on the display of parameter values and the display of possible failures. While an update of the monitored parameter changes the values of the parameter itself, alarms or alerts have influence only on the display of the parameter values.</para>
<para>MANTIS HMI should allow the users to sort and filter monitored assets to advance the navigation among different assets and monitored parameters. To make the monitoring more flexible and tailored to the users&#x02019; current needs, the interface should allow the selection of different data sources and the time range of the monitored parameters.</para>
<para>Finally, the interface should be able to produce automatically generated reports that include various information about the monitored parameters and to transfer the monitored data to other users.</para>
<para><emphasis role="strong">Data Analysis</emphasis></para>
<para>Output:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Assets wear;</para></listitem>
<listitem>
<para>Remaining useful life of the assets;</para></listitem>
<listitem>
<para>Predicted future values of monitored parameters;</para></listitem>
<listitem>
<para>Comparison between predicted and actual parameter values;</para></listitem>
<listitem>
<para>Feedback from other users.</para></listitem>
</itemizedlist>
<para>Operations:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>User-triggered operations:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Prediction models management:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>model inspection;</para></listitem>
<listitem>
<para>activation or deactivation of the model;</para></listitem>
<listitem>
<para>updating, generating and evaluating predictions.</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>Report generation.</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>System-triggered operations:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Alert when prediction performance of MANTIS system is below the predefined threshold.</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
<para>Since the data analysis is one of the key tasks in proactive maintenance, it is important that it is supported by the MANTIS HMI. In most of the MANTIS use cases, data analysts are already using various software. However, in order to reduce the time of frequent tasks it might still be useful to have an additional user interface. Such interface can also allow the users that are not specialised in data analysis to perform some basic data analysis operations such as displaying some basic statistics or choosing between predefined models.</para>
<para>Displaying the production assets wear, remaining useful life of the assets, or predicted future values of monitored parameters can therefore represent valuable features of the MANTIS HMI. Also, the of comparison between predicted and actual parameter values and feedback from other users in textual and/or graphical form can aid in evaluating the performance of the predictive algorithms.</para>
<para>To some extent, the users can also be able to manage the prediction models. These functionalities are usually limited to model inspection, activation or deactivation of the model, and updating, generating and evaluating predictions. Manipulation of the prediction models influences not only on the display of different parameter values in scope of data analysis, but does also have a significant impact on every aspect of the proactive maintenance. Designers of such interfaces should pay a special attention to the automatic update of the predicted parameter values and estimated remaining useful life of the assets in case of applying a new model. Usually, the maintenance tasks should be rescheduled as well. If the new estimation of the remaining useful life of the asset is lower than the previous one, this might also trigger some possibly indispensable alarms.</para>
<para>The results of the model management should be reported with the use of automatic report generation feature of the MANTIS HMI. The report should contain the information, displayed on the data analyst&#x02019;s user interface, and optionally the description and interpretation of the used models.</para>
<para>The prediction performance of the MANTIS system can be estimated from the comparison of the predicted and actual parameter values or the feedback from the users working on the field. If the performance is below the predefined threshold, MANTIS HMI should display an alert with additional information on prediction performance and send the relevant data to other users whose work is influenced by these models.</para>
<para><emphasis role="strong">Reporting</emphasis></para>
<para>Output:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Report in pdf or html format.</para></listitem>
</itemizedlist>
<para>Input:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Textual information input;</para></listitem>
<listitem>
<para>Graphical information input;</para></listitem>
<listitem>
<para>Spoken report.</para></listitem>
</itemizedlist>
<para>Operations:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>User-triggered operations: data importing and exporting;</para></listitem>
<listitem>
<para>System-triggered operations: processing spoken reports.</para></listitem>
</itemizedlist>
<para>Generating reports can be triggered manually by the user or automatically by the system on regular bases. If a user triggers the report generation, the system should be able to produce the report in any required format, most commonly: pdf or html. The report should contain all the relevant information related to the maintenance process, assets and the input that the user has provided.</para>
<para>In addition to the content that is automatically generated by the system, the MANTIS HMI should allow the users to input any additional information, either by means of importing the data from different data sources or manually input textual or graphical information.</para>
<para>Although the reporting is more of a by-product than a vital part of the maintenance process, it can significantly reduce the workers&#x02019; time and the effort dedicated to this task. Some advanced reporting features include process the spoken reports, which could especially benefit the maintenance technicians on the field.</para>
<para><emphasis role="strong">Communication</emphasis></para>
<para>Operations:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>User-triggered operations:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>textual, visual and audial communication between the users;</para></listitem>
<listitem>
<para>transfer of different data sources, images, videos, and documents among the users;</para></listitem>
<listitem>
<para>transfer of different data sources, images, videos, and documents to and from the MANTIS platform.</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
<para>Communication is an important aspect of proactive and collaborative maintenance. Communication is not present only between the user and the system but also between different users. Enhanced communication not only boosts the working productivity but also helps to avoid the human mistakes caused by misunderstanding. MANTIS HMI should therefore support both direct messaging in form of chat or video and indirect communications via shared widgets or established workflows.</para>
<para>For each of these high level tasks we provide a list of functionalities supporting given high level task (<link linkend="F6-2">Figure <xref linkend="F6-2" remap="6.2"/></link>). In addition to the model, we provide a Requirements Specification template, which can serve for identification of the HMI content elements and their functionalities supporting the high level tasks of a given use case. More details are also available in [Poklukar et al., 2017].</para>
<fig id="F6-2" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-2">Figure <xref linkend="F6-2" remap="6.2"/></link></label>
<caption><para>MANTIS HMI model.</para></caption>
<graphic xlink:href="graphics/ch06_fig002.jpg"/>
</fig>
</section>
</section>
<section class="lev2" id="sec6-1-4">
<title>6.1.4 HMI Design Recommendations</title>
<para>Functional specification and thereby derived interaction model, described in the previous section, cover the functional aspect of the MANTIS HMI. In order for resulting interfaces to be intuitive and easy to use, the established design principles should be applied. Regardless of technical implementation choices, design recommendations and guidelines can be provided that facilitate HMI implementations to fulfil the requirements of maintenance-based user interaction and offer a good user experience.</para>
<para>The first step towards a personalised interface that enables the user&#x02019;s maximum efficiency is the definition and design of different HMI representation types, associated with different scenarios, users (workers, administrators, services, managers, etc.), and platforms. The selection of the HMI representation types was accomplished by extracting the requirements of interaction scenarios from the MANTIS use cases. All use cases require the use of PC interfaces (desktop/laptop) and most require mobile devices (smartphone/tablet). A few need to use industrial PC consoles or specialized external devices as well.</para>
<para>A central feature of most use cases is displaying and responding to alarms and monitoring processes by displaying data in the forms of tables, charts or graphs. Some scenarios also require the interface to display maintenance plans and guides on how to perform tasks, to exchange messages with other users, and/or to input reports on the performed maintenance tasks. Non-textual input is required by some use cases, such as sound and video recording and text-to-speech and speech-to-text.</para>
<para>Based on user interface design best practices, but focussed on collected MANTIS requirements, provided design guidelines strive to find a balance of being specific enough to add value over generic UI design literature, while still being applicable to most MANTIS use cases. They are also applicable to industrial maintenance use cases outside of MANTIS. The guidelines relate to the structure of the interface (a levels-based screen hierarchy and navigation between the screens), layout of the screen elements, visual design (general appearance of the interface, colour selection, etc.), and data representation. Furthermore, interactions and collaboration between users are addressed, for example managing alarms and events, creating reports, and communicating via messaging and chat.</para>
<para>The aspect of supporting different HMI platforms is taken into account by analysing their specifics and differences. We argue that for most interaction scenarios all platforms can be supported adequately by implementing the HMI as a web application. Responsive design should be used to adapt automatically to the screen size, input capabilities, and other specifics of the particular device a user chooses to employ. To this end, we recommend following the model-view-controller architecture for web application development. We also briefly list some of the popular web development back- and front-end frameworks that could serve as a basis for implementations. Finally, some suitable libraries for data visualisation are also listed, e.g., for graphs and charts.</para>
</section>
<section class="lev2" id="sec6-1-5">
<title>6.1.5 MANTIS Platform Interface Requirements</title>
<para>The definition and design of different HMI representation types is associated with different scenarios and different users, e.g., workers, administrators, services, managers, etc. This includes defining the modalities of adaptive user interfaces in order to setup a context sensitive monitoring environment as well as taking the aspect of supporting different HMI platforms into account, e.g., web, mobile. The maintenance personnel should get mobile and easy-to-use extensions to the existing industrial dashboards (usually) on fixed-position screens. Maintenance procedures are presented via user-friendly, ergonomic and intuitive human-machine interactions, which might include the use of monitors, cameras and other HMI-specific sensors.</para>
<para>The devices that can be used for human-machine interaction range from general-purpose input- and output-capable devices, either static (desktop PCs), portable (laptops), or mobile (tablets, smart phones). Certain scenarios also include special-purpose PCs integrated into industrial equipment, as well as special purpose output (line monitors) and input devices (e.g., GoPro cameras).</para>
<para>Preferably each MANTIS HMI implementation should be usable on PCs and mobile devices, rather than requiring two separate implementations. The two most prominent interface elements are alarms and guides. Alarms play an important role in cyber-physical systems in general and in their maintenance in particular. Guides need to be taken into account when the operators need to follow the prescribed maintenance processes or machine failure interventions. The other elements are more generic; however, numerous design guidelines and examples can be provided to facilitate the implementation of HMI that can fulfil the objectives of MANTIS.</para>
<section class="lev3" id="sec6-1-5-1">
<title>6.1.5.1 Analysis of different interface types</title>
<para>User interfaces can be divided into many types based on the input and output devices that a user can interact with (Wikipedia). The most traditional is the combination of keyboard and mouse input with screen output. Touchscreens serve as both input and output. Additional types of interaction can be facilitated with audio (microphones for voice recording or recognition, speakers or headphones for voice synthesis or playback of pre-made sounds) or video devices (cameras). Finally, pointing devices (mouse, touchscreen, trackballs, user&#x02019;s hand recognized by a camera, etc.) can be used to recognize user&#x02019;s gestures rather than as pointing to a particular object on the screen.</para>
<para>Types of user interfaces were strictly divided in the past, but nowadays trends in design are more oriented into combining different existing types, exploring new ones, to enable most natural interaction between user and machine (natural user interfaces) [Wigdor, et al., 2011].</para>
<para>Smartphones and tablets have become two most commonly used types of mobile smart devices in everyday life. Most of the users are therefore familiar with their UI and should not need any device-specific training.</para>
<para>Smartphones and tablets run various operating systems (most commonly, Android, iOS or Windows platforms are used) with a graphical user interface. Mobile devices offer a wide variety of additional hardware characteristics, which can be used as a part of application HMI, e.g., playing sound for alarms. The camera has become an almost essential part of mobile devices. Every device is also equipped with a microphone. Mobile devices connect to networks wirelessly, either to local networks via Wi-Fi or to the internet via mobile networks. Because of the size and hardware characteristic, mobile devices can be brought almost everywhere [Dunlop, et al., 2002].</para>
<para>The characteristics of the smartphones and tablets enable the use of different HMI types. The basic type of a mobile HMI is the graphical user interface, most commonly using a touchscreen as a combined input and output device. A first type of advanced HMI is the gesture interface, which enables inputs made in a form of hand or stylus gestures. To improve user experience, some additional features, such as photo capturing and video or voice recording can be added to the HMI, as an example, for error reporting.</para>
<para>The second type of advanced HMI that is discussed in this context, is voice user interface, where the input is made by voice commands. Voice recordings can be added to error reports for later playback to other users or provide real-time interaction between the machine and its user. An example is voice dialling on mobile phones. Similarly, generated voice output is meant to be interpreted and acted upon by the user, such as in the case of voice commands of car navigation devices.</para>
<section class="lev4">
<title>Display Size</title>
<para>The main difference between a smartphone and a tablet is the display size. While smartphones are smaller and &#x0201C;pocket-sized&#x0201D;, the tablet is able to leverage on a larger diaplay to show more information. From an analysis on smartphone commercialized by market leaders, it was found that display diagonal size ranges between 3.5&#x02032;&#x02032; and 6&#x02032;&#x02032;, which means HMI should be adjusted to hardware restrictions. In this kind of scenario, the user interface should provide only basic information, and be menu based to enable easy access to application features.</para>
<para>Tablets display diagonal sizes ranges between 7&#x02032;&#x02032; and 12&#x02032;&#x02032;. Larger display provides easier access and interaction. The larger display is also crucial when more urgent information needs to be available, so there are little actions needed to navigate to them.</para>
</section>
<section class="lev4">
<title>Controls</title>
<para>The controls are one of the most notable aspects when user is selecting a device for specific MANTIS task. Mobile HMI is usually controlled by means of a touch screen display with hand or stylus. Graphical user interface is combined with menu user interface, which means there are visual buttons on the display to lead interaction with certain actions. Most of the mobile devices have some physical buttons (depends on the manufacturer), where basic functions of devices can be applied, for instance on/off button, camera shortcut etc. Finally, smartphones and certain tablets can provide simple haptic feedback to the user through vibration.</para>
<para>Other possibilities of controlling tablets or smartphones are with external controls, such as an external keyboard or access through an external device, for instance, PC-smartphone connection, Bluetooth connection, and remote control.</para>
</section>
<section class="lev4">
<title>User Role</title>
<para>The mobile HMI is typically used by users who only need to see a limited amount of information and alerts.</para>
<para>Examples of roles in MANTIS use case scenarios interacting with mobile HMI are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>An operator, who is a person usually dealing is usually dealing directly with the technical process. Due to the type of work, operators may have wet, dirty hands or wear gloves, which means they are incapable of using a keyboard or mouse. Devices are therefore touch-screen based and should work with gloves as well as hands. Some scenarios require them to also be mobile, while in other scenarios the cyber-physical system being worked on contains fixed operator consoles. In case of production line noise, voice user interfaces, excluding alarms, can only be used with caution or not at all;</para></listitem>
<listitem>
<para>Other user roles, such as maintenance team member, 3rd level development support, service technician, who are working both on the terrain and in the office but only need to see certain information, alerts and notifications, are using mobile HMI as well.</para></listitem>
</itemizedlist>
</section>
<section class="lev4">
<title>Additional Sensors and Context Sensitivity</title>
<para>Mobile devices can also include other sensors, such as a GPS receiver, air temperature and pressure sensor, accelerometer, compass, fingerprint reader, etc. Some of those may be useless in industrial scenarios. For example, in production plants where ambient temperature is important, highly accurate temperature sensors will be permanently installed, rather than relying on measurements from a worker&#x02019;s tablet. The compass may not work well in vicinity of large metal structures. Furthermore, the MANTIS use case scenarios do not explicitly mention human interaction with data from such sensors; therefore these aspects are not further considered in this chapter.</para>
<para>On the other hand, the MANTIS deliverable on context-awareness explored when, where and how the human-machine interaction can be supplemented by implicitly taking into account the context of the task being performed. For example, if the mobile user&#x02019;s location can be determined accurately enough, the HMI could automatically switch to the screen relevant to the machine closest to her.</para>
<para>The context is not limited to the sensors on the mobile device, but rather includes any data related to the task but not explicitly being handled (typed in, read) by the user. Some of the goals of the MANTIS project are measuring, recording, and statistically processing huge amounts of data, all of which provide additional context for HMI tasks. Moreover, the latest global trend of making industrial machines connected and intelligent shows that context-sensitive HMI is an important topic and not just limited to MANTIS.</para>
</section>
</section>
<section class="lev3" id="sec6-1-5-2">
<title>6.1.5.2 PC HMI</title>
<para>PC platforms are more commonly used professionally, for instance in industry, production and design. In industry, it is usually used as a set of desktops in the control room, or as industrial PC located on the production site (extreme environment). The majority of average users are able to manage PC devices.</para>
<para>PCs, used in MANTIS scenarios, can be divided into three groups, desktop PC, laptop and industrial PC. All of them have some advantages, which can be used in different environments, with different user types. They are typically connected into local wired (e.g., Ethernet) networks, which are secured more easily than the wireless networks used by mobile devices.</para>
<para>PCs are highly customizable. Their operating systems are different than those on mobile devices (different versions of Windows, Linux, MacOS, etc.) and they usually, but not always, use a graphical user interface. Some older or specialized software still uses the command-line interface (CLI). Desktop PCs are dependent on peripherals (external components), such as a display, keyboard, mouse, to enable access to the HMI. Through the use of external components, PCs are much more customizable than smartphones and tablets. Except for laptops, PCs are not mobile, which means they are engaged in the certain environment (office, production site, etc.).</para>
<para>Interactions on PCs are usually restricted to keyboard and mouse, although hybrid laptops with touchscreens are also popular in some circles. External components, such as microphone and cameras, can be used where needed for additional interaction modes. However, there is less need for voice-based interactions (voice commands and voice generation) than with mobile devices, because with PCs there is rarely a need for hands-free interaction and because the keyboard is much more usable for entering text then the on-screen emulated keyboards of mobile devices.</para>
<section class="lev4">
<title>Display Size</title>
<para>The currently common display diagonal size of external monitors can range from 19&#x02032;&#x02032; to 30&#x02032;&#x02032; and 13&#x02032;&#x02032;&#x02013;17&#x02032;&#x02032; for laptop screens. Wide-screen ratio 16:9 and 16:10 are common formats on the market, though more information can be shown on traditional 4:3 screens. Some external monitors can be switched from landscape to portrait orientation. In any case, MANTIS HMI design should be adjustable to one format or another, to avoid vital information not being available. Display size, compared to mobile HMIs, is larger and therefore can display greater amount of information on the screen. Users like data analysts, in need of having to look at plenty of data at once, need large displays; mobile display sizes do not meet the requirements.</para>
</section>
<section class="lev4">
<title>Controls</title>
<para>Desktop PC can only be controlled by external input and output modalities.<footnote id="fn6_1" label="1"><para>A modality is a path of communication employed by the user to carry input and output.</para></footnote> Common modalities are based on vision (screen), sound (audio outputs) and sense of touch (vibrations, movements).</para>
<para>Considering users&#x02019; needs, different components are used for input. The computer keyboard allows the user to enter typed text and the mouse allows the user to input spatial data to a computer. On a desktop computer, a virtual keyboard might provide an alternative input mechanism for users with disabilities who cannot use a conventional keyboard, or for bi- or multi-lingual users who switch frequently between different character sets or alphabets. There are many variations of pointing devices, such as 3D mice, joysticks, etc. Some other devices, such as a digital pen, digitizing tablet, high-degree of freedom input and composite devices are available on the market. Most common input devices for imaging, which are used to digitalize video or image to the computer are for instance digital camera, webcam, fingerprint scanner etc. Audio input allows the user to capture sound.</para>
<para>The output is another aspect of HMI. Displays visually represent text, graphic, and other video material. The visual material can be also printed on paper by other output devices, such as printers, 3D printers, etc. Audio can be heard through speakers or headphones. More uncommon is the haptic technology, which provides tactile feedback using the sense of touch, vibrations, the motion of the user.</para>
</section>
<section class="lev4">
<title>User Role</title>
<para>PC HMI is usually used by users based in one place, such as an office, control room or production site. PC users need to manage a larger amount of data and therefore need a more powerful device. There are usually no restrictions for keyboard or mouse use.</para>
<para>Examples of users from MANTIS use case scenarios interacting with PC HMI:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Maintenance managers are responsible for long-term analysis of tool usage. They communicate with the business manager. For a better overview, a maintenance manager uses PC HMI, where all available data can be shown. Important notifications can be also received via an e-mail;</para></listitem>
<listitem>
<para>The data analyst is responsible for analysing the results from prediction/learning algorithms. His work is connected to a large amount of information, which can only be represented on PC display. He or she can access all live streaming data and historical data from several data sources. Best HMI for this user role is the graphical user interface as there is many data to be displayed;</para></listitem>
<listitem>
<para>Production manager, qualifier, maintenance planner, maintenance planner unit, plant operator need to use a PC HMI due to a larger amount of data, better access to other applications and permanent workspace.</para></listitem>
</itemizedlist>
</section>
</section>
</section>
<section class="lev2" id="sec6-1-6">
<title>6.1.6 Recommendations for Platform Selection</title>
<para>The main factors defining which type of HMI to be used in certain situation are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>users (needs, preferences, capabilities);</para></listitem>
<listitem>
<para>user tasks, interactions and goals;</para></listitem>
<listitem>
<para>platform (hardware, software constraints);</para></listitem>
<listitem>
<para>environment (noise, lighting, dirt, vibrations, etc.).</para></listitem>
</itemizedlist>
<para>MANTIS scenarios present common situations in terms of maintenance of the system and trouble-shooting:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>monitoring the system (e.g., monitor the information on screen, reading reports);</para></listitem>
<listitem>
<para>simple interactions;</para></listitem>
<listitem>
<para>analysis;</para></listitem>
<listitem>
<para>usage of the machine and trouble-shooting;</para></listitem>
<listitem>
<para>communication.</para></listitem>
</itemizedlist>
<para>The selection of HMI to be used for certain situation can be summarized as follows:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>For monitoring the machines and generated data, both mobile and PC HMIs are required. The preferred selection of mobile device is tablet as it can present more data on screen due to larger display size. The users, performing monitoring tasks, are diverse &#x02013; from machine operator, to maintenance engineer and data analyst. The selection of the HMI, based on the user is therefore more dependent on the location of the user rather than their role. The users with fixed workplace location will prefer the PC HMI with larger display size, while users working on the terrain will prefer mobile HMIs due to ease of accessibility. Mobile HMI can be used as a remote extension to PC HMI as well;</para></listitem>
<listitem>
<para>For simple interactions with HMI (confirmations, calculations, ratings etc.), buttons and menus are mainly used as interface elements. As trivial interface elements to use and implement, there is no limitation for usage to either HMI. The main difference is the control function; normally touch for mobile HMI and keyboard with mouse for PC HMI;</para></listitem>
<listitem>
<para>For analysis of system data, generated reports or financial aspects, responsible by data analyst, PC HMI is preferred in all MANTIS scenarios. For component analysis and usage of various analysis tools by some maintenance managers, tablet is required as a HMI;</para></listitem>
<listitem>
<para>For usage of the machine, troubleshooting and following guidelines and instructions when repairing the machine, both mobile and PC HMIs are required. The preferred selection of mobile device is tablet, while the PC HMI is mainly presented as industrial PC. The users of PC HMI can be defined as machine operators and the users of tablet are defined as service technicians, support team or maintenance engineers, without fixed workplace, moving from site to site, or working remotely;</para></listitem>
<listitem>
<para>For communication between users (e.g., text, audio or video chat), both mobile and PC HMIs are required. Communication is not limited to any user role and device, it could be used on various devices (smartphone, tablet, desktop PC and laptop PC) by various users. It can be assumed that the selection of HMI by specific user is affected by other factors or tasks rather than chat requirements.</para></listitem>
</itemizedlist>
<para>All the user roles enumerated for mobile and PC HMIs can be supported well with a web-based HMI. To run well on the type of device each user role typically uses and to ensure good user experience, the web-based HMI has to be implemented carefully so that it is able to adapt to different devices and display sizes.</para>
<para>A web application intended for both mobile and PC users must be able to adapt to screen sizes from 4&#x02032;&#x02032; to 30&#x02032;&#x02032; in both portrait and landscape orientations. However, individual screens of the web HMI only have to adapt to the typical screen sizes of the devices that will be used for tasks that the screen is part of. The screens intended for mobile devices will still be usable on PCs. The opposite, using a mobile device for a task intended for PC, will be possible as a workaround (e.g., the PC broke down) but not recommended for extended periods.</para>
<para>The same can be said for the difference in controls. The tasks intended for mobile devices should require at most a minimal amount of typing, thus the inputs are limited to multiple-choice buttons and alternative modalities (sound, camera). The workflow should lead the user through well-defined procedures using linear navigation. No high-precision pointing should be required unless a stylus is used.</para>
<para>The tasks intended for PCs, however, can offer richer navigation, can use typing as the default input method and should offer keyboard shortcuts for buttons, menus, and navigation. Obviously, multi-touch gestures must not be required.</para>
<para>LED-based projection technology is currently getting build into smartphones and tablets, giving them the potential of projecting a big (touch) screen on any surface. Although a very promising solution to the current screen size limitations, this technology is still in its infancy and there are many obstacles to overcome before it will become mainstream.</para>
<section class="lev3" id="sec6-1-6-1">
<title>6.1.6.1 Web-based HMI</title>
<para>Web applications are increasingly being used in all domains where the application is not an isolated island, but has to also interact with other users and/or data residing elsewhere. A well-known example are office suites (Google Drive, MS Office Online), where web applications enable collaboration in the sense of multiple users simultaneously editing the same document. On the other hand, they are less feature-rich than their desktop counterparts, the interaction is less responsive in certain cases and might require a continuous web connection to work.</para>
<para>In case of MANTIS maintenance HMI, most use cases require both PC and mobile HMI. The operating systems and consequently the native software development stacks (programming languages, integrated development environments, libraries and frameworks for user interface creation, debuggers, etc.) are completely different in these two cases, requiring significant duplicated effort. Windows platforms do strive to unify mobile and desktop development to a degree. However, only supporting Windows as a mobile platform is currently limited and has low availability of hardware. Web applications, on the other hand, run out of the box on all PC and mobile operating systems, making it much easier to support both types of devices.</para>
<para>Another advantage of web applications is that they can be deployed and managed centrally and are easier to ensure that all users use the same (latest) version [Miller, 2008]. They need continuous access through a network with a (local) server to work. Nowadays, all devices are already suitable for connections, either through Wi-Fi or mobile network. Since most of MANTIS scenarios involve cooperation of multiple users. This is a requirement regardless of whether web or native applications are used.</para>
<para>The upcoming evolution in web applications is the ability to run and store data when no connection to a server is present. In this case, data is managed on the local browser and once the connection restores, all data will be synchronized. This is already the case in some progressive applications and receives increasing supported from browser vendors.</para>
</section>
<section class="lev3" id="sec6-1-6-2">
<title>6.1.6.2 Responsive design</title>
<para>The user interface of the MANTIS platform has to be flexible and modular, to easily adapt layout, content and appearance to different screen sizes. In this context responsive design approach proposes an efficient and suitable method to solve the challenges of modularization, flexibility and scalability and to optimize user experience across devices of varying sizes and capabilities.</para>
<para>Responsive design allows a page to adapt layout and content to viewing contexts across a spectrum of digital devices. Responsive design approach supports the adjustment of device&#x02019;s resolution, size, and layout, from smartphones to desktop PCs. Devices such as tablets and smartphones also support orientation changing, providing two possible screen widths. Developers now have new web standards like Hypertext Markup Language version 5 (HTML5), and Cascading Style Sheets version 3 (CSS3), enabling them to design and build user-sensitive sites that respond to a range of contexts and device capabilities [Gardner, 2011].</para>
<para>&#x0201C;Mobile first&#x0201D; is a more recent design paradigm whereby web-based user interfaces are optimized for mobile use first and have a graceful fall back when being used on devices with larger screens instead of the other way around. This is done by gradually offering more information and options once the real estate allows it. Starting from the essential must-haves on small screens to additional nice-to-haves on larger screens.</para>
<para>Marcotte [Marcotte, 2010] outlined a method for creating fluid layouts that are screen-resolution agnostic and capable of dynamically changing according to user context. He describes responsive design as having three elements:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>a fluid layout that uses a flexible grid, which in turn ensures that a website can scale to a browser&#x02019;s full width;</para></listitem>
<listitem>
<para>images that work in a flexible context, whether fluid themselves or perhaps controlled through overflow mechanisms;</para></listitem>
<listitem>
<para>media queries, which optimize the design for different viewing contexts and spot-fix bugs that occur at different resolution ranges.</para></listitem>
</itemizedlist>
</section>
</section>
<section class="lev2" id="sec6-1-7">
<title>6.1.7 Interface Design Recommendations for MANTIS Platform</title>
<para>Although MANTIS use cases come from different domains (production assets, vehicles, energy production and health equipment), common recommendations can be provided that apply to most or all of them. The objective was to provide a set of guidelines that, if followed by the various MANTIS HMI designs, will ensure a consistent interaction between the users and the system with the common pitfalls avoided. The user will be provided with simple, intuitive, legible displays that are suitable for the intended purpose. These recommendations can be applied to any potential interface, emphasizing maintenance tasks, alarm management, pro-activeness and collaboration.</para>
<para>Based on the understanding of the use case scenarios and a design philosophy, a set of recommendations for designing interfaces for MANTIS platform was established and refined:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>The <emphasis role="strong">structure</emphasis> of the interface has to be defined (a four-level screen hierarchy and organization of screen flows) and the implemented logic of navigation controls and menus that allow the user to navigate more easily through the HMI;</para></listitem>
<listitem>
<para>Overall <emphasis role="strong">screen layout</emphasis> has to be defined including a consistent arrangement of interface elements and distribution of information on the screen to enable a fast orientation of the user;</para></listitem>
<listitem>
<para>The <emphasis role="strong">visual design</emphasis> section has to describe the general appearance of the interface and specify the properties of the interface elements, such as colour selection and display text;</para></listitem>
<listitem>
<para>Decision has to be made about the most effective way of <emphasis role="strong">data presentation</emphasis> using components such as tables, charts, bar charts and line graphs;</para></listitem>
<listitem>
<para><emphasis role="strong">Interaction</emphasis> and <emphasis role="strong">collaboration</emphasis> betweenusers have to be defined, for example managing alarms and events, creating reports, following step-by-step guides and communicating via messaging and chat.</para></listitem>
</itemizedlist>
</section>
</section>
<section class="lev1" id="sec6-2">
<title>6.2 Adaptive Interfaces</title>
<para>A great challenge in human-machine interaction is to ensure that the information presented on the interface is meaningful and relevant, properly represented, location-aware and targeting the appropriate person. It is also important to bring the attention of the user at the most suitable moment, instead of disrupting him/her with information overload.</para>
<para>In a proactive and collaborative maintenance platform, gathering and combining condition monitoring data with contextual information can provide numerous benefits such as a better productivity, improved decision-making, accuracy of predictions and process optimisation, together with an enhanced usability and personalisation on the HMI.</para>
<section class="lev2" id="sec6-2-1">
<title>6.2.1 Context-awareness Approach</title>
<para>Context-aware computing is a paradigm where applications and services use environmental information acquired by sensors (such as user and device location, state, time, nearby places, people and devices, etc.) to provide relevant information and/or services to the user.</para>
<para>The concept of context has been researched over the last decades in software engineering, especially in areas such as Natural Language Processing, and more generally in Human-Computer Interaction.</para>
<para>Context makes the interaction with computers easier by adapting the information to the user, discriminating what is relevant and what is not, so that the user in human-machine interactions can focus on high-level tasks, which is very important in scenarios of information overload, especially as we move towards a world of ubiquitous and pervasive computing and the Internet of Things.</para>
<section class="lev3" id="sec6-2-1-1">
<title>6.2.1.1 Context and context awareness fundamentals</title>
<para>The term &#x0201C;context&#x0201D; has been addressed by some authors such as Abowd and Dey [Abowd, et al., 1999], which identified the two main types of context: primary and secondary context. These contexts differ through four main categories which are location, identity, activity and time.</para>
<para>Following this study, [Perera, et al., 2014] referred to the main difference between each context as:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Primary context:</emphasis> Any information retrieved without using existing context and without performing any kind of sensor data fusion operations;</para></listitem>
<listitem>
<para><emphasis role="strong">Secondary context:</emphasis> Any information that can be computed using primary context by using sensor data fusion operations or data retrieval operations.</para></listitem>
</itemizedlist>
<para><link linkend="F6-3">Figure <xref linkend="F6-3" remap="6.3"/></link> shows some examples of context categorization from a conceptual and operational perspective:</para>
<fig id="F6-3" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-3">Figure <xref linkend="F6-3" remap="6.3"/></link></label>
<caption><para>Categories of Context [Perera, et al., 2014].</para></caption>
<graphic xlink:href="graphics/ch06_fig003.jpg"/>
</fig>
<para>The term &#x0201C;context awareness&#x0201D; refers to the ability of computing systems to acquire and reason about the context information and subsequently adapt the corresponding applications accordingly.</para>
<para>This term was first introduced in the research field of pervasive and ubiquitous computing by Schilit and Theimer in 1994 in a paper entitled &#x0201C;Context Aware Computing Applications&#x0201D; [Schilit, et al., 1994], describing software which adapts according to its location of use, the collection of nearby people and objects, as well as changes to those objects over time.</para>
<para>Abowd and Dey [Abowd, et al., 1999], provided the most widely accepted categorizations of context aware features which are the presentation of information and services, automatic execution of services an tagging of context.</para>
<para>Another interesting study is the one done by Barkhuus and Dey [Barkhuus, et al., 2003] where they identified three levels of interactivity in context awareness based on the user interaction which are personalization (based on the preferences and expectations of the user), passive context awareness (updated contextual information is presented to the user) and active context awareness (application autonomously changes its behaviour according to the sensed information).</para>
</section>
<section class="lev3" id="sec6-2-1-2">
<title>6.2.1.2 Context lifecycle in context-aware applications</title>
<para>Perera et al. [2014] selected ten popular data lifecycles [Hynes, et al., 2009; Chantzara, et al., 2005; Ferscha, et al., 2001; Wrona, et al., 2006] to analyse them in their survey. After reviewing these works, the authors stated that applications use typically four phases when processing context, from the moment it is acquired from sensors in raw format, to the moment it is consumed by the end-user application:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Context Acquisition:</emphasis> Contextual data is captured from the environment using sensors;</para></listitem>
<listitem>
<para><emphasis role="strong">Context Modelling:</emphasis> The collected data needs to be represented in a meaningful manner through a context model;</para></listitem>
<listitem>
<para><emphasis role="strong">Context Reasoning:</emphasis> Modelled data needs to be processed to derive high-level context information from low-level raw sensor data;</para></listitem>
<listitem>
<para><emphasis role="strong">Context Dissemination:</emphasis> High-level and low-level context need to be distributed to the consumers who are interested in context.</para></listitem>
</itemizedlist>
</section>
<section class="lev3" id="sec6-2-1-3">
<title>6.2.1.3 Adaptive and intelligent HMIs</title>
<para>Information overload, variety of heterogeneous users and cognitive overload for decision making are different problems to deal with the process of HMI design and development. Different research areas such as Intelligent User Interfaces (IUI) and Adaptive User Interfaces (AUI) face with those problems applying intelligence during the process, investigating new algorithms and promising techniques for user, context and content adaptation.</para>
<section class="lev4">
<title>Intelligent User Interfaces (IUI)</title>
<para>Intelligent User Interfaces (IUI)<footnote id="fn6_2" label="2"><para>IUIs can be described as interfaces that, using intelligent technology, can improve the communication between the machine and the end user.</para></footnote> is a multidisciplinary area inside the Human-Computer Interaction (HCI) research field that aims to improve human-computer interaction by applying technology to those interfaces [Ehlert, 2003].</para>
<para>Over the years, many researchers from different fields (as shown in <link linkend="F6-4">Figure <xref linkend="F6-4" remap="6.4"/></link>) have influenced and made improvements on areas related to IUI, for example in psychology (advances in cognitive sciences or human perception), in artificial intelligence (improvements in user modelling, or in machine learning to predict user behaviours), and in the HCI field (new visualization and interface evaluation techniques to have a better user experience and usability).</para>
<fig id="F6-4" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-4">Figure <xref linkend="F6-4" remap="6.4"/></link></label>
<caption><para>Research Fields Influencing the Development of Intelligent User Interfaces.</para></caption>
<graphic xlink:href="graphics/ch06_fig004.jpg"/>
</fig>
<para>Intelligent interfaces can <emphasis role="strong">adapt to user, context and situation,</emphasis> they have the <emphasis role="strong">ability to communicate</emphasis> and they have the <emphasis role="strong">ability to solve different problems</emphasis>, improving the usability, flexibility [Maybury, et al., 1998] and user experience adding Artificial Intelligence.</para>
</section>
<section class="lev4">
<title>Adaptive User Interfaces (AUI)</title>
<para>Adaptive User Interfaces is a subtype of IUIs that improve the interaction with the user with knowledge taken from this user [Langley, 1997]. We can divide them in two main groups according to the type of feedback that the user must provide:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Informative Interfaces.</emphasis> For example, recommender systems or information filtering, where the interfaces select or filter information for the end user to improve the user experience;</para></listitem>
<listitem>
<para><emphasis role="strong">Generative Interfaces.</emphasis> For example, systems for planning or document preparation, where the generation of a useful structure is needed.</para></listitem>
</itemizedlist>
<para>AUI was recognized as very promising and challenging area [Norcio, et al., 1989]. The purpose of this field is to provide a user experience that is automatically created via machine-learned processes. We can define AUI as the intersection between HMI and Machine Learning.</para>
<para>A user model is the description and knowledge of the user maintained by the system. User modelling is concentrated on individual users&#x02019; knowledge, goals, plans, emotions, personality, ability etc. [Kobsa, 2011]. Early researches of adaptive interface models are based on user data.</para>
<para>Context-aware user interfaces play an important role in many adaptive human-computer interaction tasks of location-based services. Examples of works in this area is the context-aware adaptive models for mobile location proposed by [Feng et al., 2015], the personalized traveler information system (ATIS) presented by LBS, [Lathia, et al., 2012] or the AUI based on various possible contexts such as handicap user profile proposed by [Zouhaier, et al., 2013].</para>
</section>
</section>
<section class="lev3" id="sec6-2-1-4">
<title>6.2.1.4 Context awareness for fault prediction and maintenance optimisation</title>
<para>The use of context information within prediction and maintenance-related processes is one application of context awareness mechanisms. Here, the focus is to enhance the reasoning/modelling systems (e.g., diagnostics, prognostics, maintenance scheduling) to improve decision support.</para>
<para>Concerning predictive analytics, the predictive ability of a system is enhanced by contextual information present in the environment [Kiseleva, 2013]. Besides that, context can also be used to improve the Remaining Useful Life (RUL) prediction, as expressed in [Ahmadzadeh, et al., 2012; Thaduri, et al., 2014]. Different approaches can be used to obtain an accurate RUL using operational context, there could be better ways to make this prediction.</para>
<para>An example of an approach would be to predict the degradation of an equipment (e.g., machine tool) based on how it is being used using &#x0201C;Fingerprint&#x0201D; is the recorded data obtained periodically when monitoring a sensorised machine doing the same set of predefined operations (see <link linkend="F6-5">Figure <xref linkend="F6-5" remap="6.5"/></link>).</para>
<fig id="F6-5" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-5">Figure <xref linkend="F6-5" remap="6.5"/></link></label>
<caption><para>Context-Dependent Data Integration in Maintenance Scenarios (POWER-OM, Lulea Technology Univ., 2015).</para></caption>
<graphic xlink:href="graphics/ch06_fig005.jpg"/>
</fig>
<para>Another approach is to calculate a more accurate RUL, where context information concerning the future operational conditions have an impact in the final prognosis scenario. This approach was applied in a study done in [Ferreiro, et al., 2012] to predict the RUL of aircraft brake wear.</para>
</section>
<section class="lev3" id="sec6-2-1-5">
<title>6.2.1.5 Context awareness for maintenance personalisation and decision-making</title>
<para>Personalization has become an important aspect in many areas, e.g., personalization of car interfaces [Endres, et al., 2010; Garzon, et al., 2011], in smart home environments [Ma, et al., 2005], automatic profile selection [Coutand, et al., 2006], healthcare [Zhang, et al., 2005; Hashiguchi, et al., 2003; Koutkias, et al., 2001; Koutkias, in drugi, 2001].</para>
<para>Finding proper patterns and accurately predicting the results could provide better personalization and adaptation. Here, context has an important role, because the behaviour of persons/users may be related on the context (i.e., location, time, access device).</para>
<para>Context awareness allows the development of personalized services that automatically adapt to the user&#x02019;s situation, and in this sense, context management mechanisms can complement prediction models so that predictive analytics decisions can be more accurate.</para>
<para>One of the main challenges in this area is to construct the mechanisms which would detect what context is, how to integrate context into the prediction models, or monitoring the stream of contextual data over time to detect anomalies.</para>
<para>Results could be displayed in various human-machine interface applications. One of the main characteristics of modern intelligent user interfaces is the integration of multiple users with diverse needs and requirements. Benefits of such personalization include improved safety, added comfort, increased efficiency, or enabling access for users with special needs.</para>
<para>Personalization in maintenance could be reached by using different artificial intelligence concepts to predict next steps and help maintenance workers in decision-making tasks. These tasks could consist of fault detection and diagnosis, detecting anomalies, scheduling suggestions, choosing proper maintenance concepts, optimizing energy consumption during the operations, informing operators about actions, or planning a repair action, among other things, which could be displayed through different HMIs. HMIs in maintenance are customized to the worker&#x02019;s need and depends on the type and environment of industry.</para>
<para>To make accurate predictions in maintenance, qualitative data (context information) is needed which is usually historical data of maintenance activities; large amounts of sensor measurements, history of user interactions, anomalies, faults, etc.</para>
<para>Wearable devices are important in the plant maintenance, because they allow the user&#x02019;s hands to remain free to do the work and currently rely on voice recognition and voice response [Nagamatsu, et al., 2003; Nicolai, 2005; Nicolai, et al., 2006; Stiefmeier, et al., 2008].</para>
<para>As stated by Lee in the article &#x0201C;Cyber physical systems: Design challenges&#x0201D; [Lee, 2008] cyber-physical systems (CPS) are integrations of computation and physical processes, where physical processes affect the computation and vice versa. The addition of context information to the monitoring and prediction in maintenance activities can contribute to improve maintenance approaches, enhancing the cost, time and quality of the processes.</para>
<para>The large amount of data collected using sensors can be used for detecting and analysing anomalies and faults in large and complex systems. Data-driven approaches leverage on this large amount of data which is collected by CPS and is used to learn the necessary models automatically; recognize unusual situations, optimize energy consumption during the operations and inform operators who use this information to modify system processes, or plan for repair or maintenance. System&#x02019;s engineers and experts can use this information to take further actions (e.g., update operations procedures or redesign the system).</para>
<para>The data-driven prognostic approach [Swanson, et al., 2000; Niggemann, et al., 2015; Krueger, et al., 2014; J&#x000E4;ms&#x000E4;-Jounela, et al., 2013; Zhang, et al., 2015] could be used to determine the fault and predict the amount of time before it reaches a predetermined threshold level.</para>
</section>
<section class="lev3" id="sec6-2-1-6">
<title>6.2.1.6 Context awareness approaches in a proactive collaborative maintenance platform</title>
<para>Here are three generic context awareness approaches that could be incorporated in a collaborative maintenance platform to provide some kind of benefit to its users:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Adaptation to scenarios</emphasis></para>
<para>From alerting and warning situations, to special events in the state of the production process, or changes in the location of a user, these are examples of scenarios where the use of context awareness could be relevant to deliver the right information, at the right moment, in the right format, to the right person by means of an adaptive and intelligent user interface;</para></listitem>
<listitem>
<para><emphasis role="strong">Enhanced reasoning algorithms</emphasis></para>
<para>The goal here is to use operational information as a contextual extra input to reasoning algorithms (diagnostics, prognostics, scheduling, etc.) to optimize their results;</para></listitem>
<listitem>
<para><emphasis role="strong">Personalised maintenance suggestions</emphasis></para>
<para>Context could be used to make personalized maintenance suggestions to users when performing everyday processes in the system to improve their productivity and overall experience on the HMI.</para></listitem>
</itemizedlist>
<para><link linkend="T6-1">Table <xref linkend="T6-1" remap="6.1"/></link> summarizes the main benefits of the approaches described above.</para>
<table-wrap position="float" id="T6-1">
<label><link linkend="T6-1">Table <xref linkend="T6-1" remap="6.1"/></link></label>
<caption><para>Context Awareness Approaches in a Proactive Collaborative Maintenance Platform</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left">Context Awareness Approach</td>
<td valign="top" align="left">Benefits</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Adaptation to scenarios</td>
<td valign="top" align="left">&#x02219; Personalisation</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="left">&#x02219; Usability</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="left">&#x02219; Maintenance optimisation</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="left">&#x02219; Better ergonomics</td>
</tr>
<tr>
<td valign="top" align="left">Enhanced reasoning algorithms</td>
<td valign="top" align="left">&#x02219; Improved decision-making</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="left">&#x02219; Better maintenance planning</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="left">&#x02219; Accuracy of predictions</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="left">&#x02219; Cost savings</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="left">&#x02219; Maintenance optimisation</td>
</tr>
<tr>
<td valign="top" align="left">Personalized maintenance suggestions</td>
<td valign="top" align="left">&#x02219; Productivity</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="left">&#x02219; User experience</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="left">&#x02219; Enhanced asset management</td></tr>
</tbody>
</table>
</table-wrap>
</section>
</section>
<section class="lev2" id="sec6-2-2">
<title>6.2.2 Interaction Based/Driven Approach</title>
<para>Nowadays interaction analysis (e.g., clickstream analysis) is one of the most frequently used techniques to understand user behaviour while he is using the interface. This understanding can help not only in the interface design and development process but also providing some inputs to carry on intelligent adaptation. This interface adaptation, commonly called adaptive user interface, can enhance usability and user experience.</para>
<para>User interaction data have detailed information of how users perform actions with the different elements of the interface such as visualization elements, buttons or icons, and also how the user navigate between them. Interaction history gathered in these datasets can be described as a collection of different timestamped actions performed by a single user whilst is using the interface [Nguyen, et al., 2017].</para>
<para>Analysing these datasets can give us information about user sequences or recurrent patterns [Soh, et al., 2017] Furthermore, this analysis can be used as relevant information to perform different automatic adaptation in the interface [Dev, et al., 2017].</para>
<para>Therefore, it is necessary to have different tools which allow to track and store all the interaction between user and the interface. In the context of MANTIS project this interaction driven approach will be focused on defining a methodology to track and store navigation and actions performed by the user with the interface.</para>
<section class="lev3" id="sec6-2-2-1">
<title>6.2.2.1 Introduction</title>
<para>When the user accesses a web interface and interacts with it, a digital <emphasis>fingerprint</emphasis> is recorded. This digital fingerprint can be defined as the record of actions and steps performed by the user in a time slot. These actions have been captured and stored automatically with a timestamp key that allow us user tracking.</para>
<para>The issue has been approached from two different perspectives at it is shown in <link linkend="F6-6">Figure <xref linkend="F6-6" remap="6.6"/></link>:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Navigation between the different interfaces. Extracted and parsed from the web server logs. Analysing this data can give us information about the common paths and navigation flows;</para></listitem>
<listitem>
<para>The interaction carried out in the different interfaces. Extracted automatically from the interface via JavaScript. Analysing this data can give us information about the common actions or sequences.</para></listitem>
</itemizedlist>
<para>For the user interaction capture and storage system Elastic<footnote id="fn6_3" label="3"><para>http://elastic.co</para></footnote> technological framework has been proposed. Elastic is a tool which is often used in Log Analytics use cases.</para>
<para>Elastic is an open source platform based on Lucene<footnote id="fn6_4" label="4"><para>https://lucene.apache.org/core/</para></footnote> that allows save, search and display information stored in different indexes. Elastic provides different tools such as Logstash<footnote id="fn6_5" label="5"><para>http://elastic.co/products/logstash</para></footnote> for data parsing, ElasticSearch<footnote id="fn6_6" label="6"><para>http://elastic.co/products/elasticsearch</para></footnote> for data storage and Kibana<footnote id="fn6_7" label="7"><para>http://elastic.co/products/kibana</para></footnote> for data visualization. The main point to be considered is that the information should be indexed by time (<emphasis>Timestamped data</emphasis>) and ElasticSearch is a good option for that purpose. ElasticSearch is a powerful tool with many advantages such as:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>speed: ElasticSearch uses distributed inverted indexes;</para></listitem>
<listitem>
<para>APIs: ElasticSearch offers a REST API and uses JSON schema;</para></listitem>
<listitem>
<para>variety of plugins: Logstash for data parsing and processing or Kibana for developing different dashboards and interactive visualizations;</para></listitem>
<listitem>
<para>real time index updates: Very important when you are monitoring a scenario;</para></listitem>
<listitem>
<para>available for different languages: Python, Java, Ruby or Node.js.</para></listitem>
</itemizedlist>
<fig id="F6-6" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-6">Figure <xref linkend="F6-6" remap="6.6"/></link></label>
<caption><para>Interaction driven approach for visualization recommender.</para></caption>
<graphic xlink:href="graphics/ch06_fig006.jpg"/>
</fig>
</section>
<section class="lev3" id="sec6-2-2-2">
<title>6.2.2.2 Navigation tracking and storage</title>
<para>When a user is navigating among different interfaces, this navigation information can be extracted from the access log files. A log file is a collection of events and actions that is stored in the web server. In the context of the MANTIS project, common log file format will be considered. This format, NCSA Common format<footnote id="fn6_8" label="8"><para>https://httpd.apache.org/docs/trunk/logs.html#common</para></footnote>, is an standardized text file that is generated by different web servers.</para>
<para>Access log files can have the following structure:</para>
<para>192.168.2.22 - -[12/Feb/2018:15:20:02 +0200]&#x0201D;GET /mantis? P=1 HTTP/1.1 &#x0201D; 200 136 &#x0201D; http://mantis.com/true &#x0201D;&#x0201D; Mozilla/5.0 (Windows NT 6.1; rv: 24.0) Gecko/20100101 Firefox/24.0 &#x0201D;.</para>
<para>Parsing this file can give us the following information:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>user IP address;</para></listitem>
<listitem>
<para>date time. Timestamp value that will be used for index;</para></listitem>
<listitem>
<para>method. (GET or POST);</para></listitem>
<listitem>
<para>HTTP protocol version;</para></listitem>
<listitem>
<para>response status code: 10x Informative response, 20x successful response, 30x Redirection, 40x Client error or 50x Server errors;</para></listitem>
<listitem>
<para>system information: Operative system and browser.</para></listitem>
</itemizedlist>
<para>Accessing the server log files and parsing them with Logstash, raw information can be converted into a format that allows traceability. This information will be stored in an ElasticSearch index.</para>
<para>Indexing by &#x02018;datetime&#x02019; let us to analyse the navigation flow and detect the most common paths while using the interfaces. This information can be used for instance to improve the process by reducing the number of clicks performing some action.</para>
<para>As it is shown in <link linkend="F6-7">Figure <xref linkend="F6-7" remap="6.7"/></link>, this navigation capture and storage can be done in an automatic and non-intrusive way. Other frameworks or solutions have been found in the literature [Hashemi, et al., 2016; Atterer, et al., 2006] but they are not as flexible as this approach. Each navigation will add a new register in the access log file and in turn will be recorded in the ElasticSearch index.</para>
</section>
<section class="lev3" id="sec6-2-2-3">
<title>6.2.2.3 Action logs</title>
<para>Other important component for the interaction analysis is the register of different actions performed by the user with the interface. Those actions will be executed through different devices such us mouse, keyboards or tactile interfaces. One of the main advantages of using web based interfaces is that JavaScript can be used for adding functionalities to the interface.</para>
<para>JavaScript<footnote id="fn6_9" label="9"><para>https://developer.mozilla.org/en-US/docs/Web/JavaScript</para></footnote> is an event-driven, dynamic and multi-paradigm programming language that allows to add functionality not only to the client side but also to the server side. JavaScript is proposed as the programming language to track and send the interaction to the ElasticSearch index.</para>
<fig id="F6-7" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-7">Figure <xref linkend="F6-7" remap="6.7"/></link></label>
<caption><para>Navigation parse and index on ElasticSearch.</para></caption>
<graphic xlink:href="graphics/ch06_fig007.jpg"/>
</fig>
<para>To track and store all the interactions with the interface performed by the user, different triggers must be defined and implemented. These triggers will record insert new record into the ElasticSearch index. The triggers will be associated to different interaction events.</para>
<para>In the context of MANTIS project only will be considered six kind of interaction events: mouse events, keyboard events, focus, drag and drop events, clipboard events and view events. Once the user performs an action on the different interface elements, the trigger will be fired and automatically upload new record to ElasticSearch index via JavaScript. <link linkend="F6-8">Figure <xref linkend="F6-8" remap="6.8"/></link> shows the process of event triggering and storing into ElasticSearch index.</para>
<para>Apart from the system &#x02018;datetime&#x02019; date and time other important information must be stored, for instance: which interface the user is using, with which element the user has interacted or what kind of action was triggered by the user. This way, we can trace interactions performed on the different interfaces and in turn we can perform an analysis for instance to detect patterns. Note that this should only be done on a temporarily basis for study purposes and with clear and written consent of the monitored user.</para>
<fig id="F6-8" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-8">Figure <xref linkend="F6-8" remap="6.8"/></link></label>
<caption><para>Interaction capture and store in ElasticSearch index.</para></caption>
<graphic xlink:href="graphics/ch06_fig008.jpg"/>
</fig>
<fig id="F6-9" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-9">Figure <xref linkend="F6-9" remap="6.9"/></link></label>
<caption><para>Index structure created in ElasticSearch to capture the interaction of the user with the interface.</para></caption>
<graphic xlink:href="graphics/ch06_fig009.jpg"/>
</fig>
<para>As we can see in <link linkend="F6-9">Figure <xref linkend="F6-9" remap="6.9"/></link>, for each of the interactions carried out in the interface the following information is stored in the ElasticSearch index:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>type of event that the user has made: Click, mouse movement... These events are predefined by JavaScript;</para></listitem>
<listitem>
<para>timestamp of when the action was taken. With this information the traceability can be done;</para></listitem>
<listitem>
<para>interaction element;</para></listitem>
<listitem>
<para>which interface;</para></listitem>
<listitem>
<para>ElasticSearch index name.</para></listitem>
</itemizedlist>
<para>This methodology of capturing navigation and interaction allows storing the interaction with the interface in a non-intrusive and transparent way. On one hand, we have information about navigation flows indexed by time and on the other hand, we have the information about different performed actions between user and interface.</para>
<para>This dataset can be used as a powerful set of information to perform different analysis techniques in order to detect patters, provide recommendation or develop automatic adaptation engines.</para>
</section>
</section>
</section>
<section class="lev1" id="sec6-3">
<title>6.3 Advanced Data Visualizations for HMIs</title>
<para>The visualization requirements of a data analyst or a highly skilled maintenance expert go beyond the basic graphs and other widgets typically present on industrial dashboards. Most importantly, such experts need to be able to choose what data to visualize and how in order to gain further insight.</para>
<section class="lev2" id="sec6-3-1">
<title>6.3.1 Visualization of Raw Data</title>
<para>The specific requirements in the different maintenance use cases, in MANTIS and beyond, are so diverse that no single visualization tool can cover them all. On one hand, raw data can be visualized by various methods in order for a human expert to explore relations between data. On the other, visualizing the results of analysis and decision-making algorithms, such as those presented in <link linkend="ch05">Chapter <xref linkend="ch05" remap="5"/></link>, can provide insights into their operation and help improve the algorithms or validate their results. Thus, an overview of some data visualization tools is given here. Most tools require specific input formats and thus need to import the data rather than use it directly from the data store already set up for maintenance data. One of them, Kibana, was chosen as applicable to the widest array of maintenance use cases and therefore the process of its integration is also described.</para>
<section class="lev3" id="sec6-3-1-1">
<title>6.3.1.1 Visualisation tools overview</title>
<para>There is an abundance of utilities for producing the basic kinds of graphs, from office software to web graphing libraries to the built-in capabilities of scientific computing tools. The latter (e.g., Matlab, Octave, R) are most suitable for data analysts and already routinely used for such purposes.</para>
<para>Certain tools stand out due to innovativeness of the visualizations they offer. Plot.ly<footnote id="fn6_10" label="10"><para>https://plot.ly/</para></footnote> is a tool with libraries for JavaScript, Python, and R. It can, for example, intuitively illustrate how various machine learning algorithms work on given datasets<footnote id="fn6_11" label="11"><para>See the on-line example at https://plot.ly/&#x002DC;jackp/16209/machine-learning-classifier-comparison.embed.</para></footnote>. Inner workings of neural networks can also be visualized in interesting, interactive ways<footnote id="fn6_12" label="12"><para>A. W. Harley, &#x0201D;An Interactive Node-Link Visualization of Convolutional Neural Networks,&#x0201D; in ISVC, pages 867-877, 2015.</para></footnote>. This can be particularly engaging if the problem being solved by the neural network is also visual in its nature, such as optical character recognition<footnote id="fn6_13" label="13"><para>See interactive example at http://scs.ryerson.ca/&#x002DC;aharley/vis/conv/.</para></footnote> or, for a maintenance-related example, optical recognition of worn-out machine components.</para>
<fig id="F6-10" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-10">Figure <xref linkend="F6-10" remap="6.10"/></link></label>
<caption><para>Data analysis and visualization workflow in Orange.</para></caption>
<graphic xlink:href="graphics/ch06_fig010.jpg"/>
</fig>
<para>Finally, specialized machine learning tools, such as Orange<footnote id="fn6_14" label="14"><para>https://orange.biolab.si/</para></footnote>, typically also include visualization capabilities. Orange enables data analysis, machine learning, and visualization via interactive workflows, as shown in <link linkend="F6-10">Figure <xref linkend="F6-10" remap="6.10"/></link>. The upper right part shows a workflow where an example data set is clustered using the k-Means method and then shown as a scatter plot (upper branch of workflow). The raw data is shown as mosaic (lower branch). The actual mosaic visualization is shown in the lower part. A powerful feature is the &#x0201C;Find informative &#x02026;&#x0201D; button included in most Orange visualizations. As shown in the lower right corner, this suggests which attributes of the data set to base the visualization on for most informative results. The left part of <link linkend="F6-10">Figure <xref linkend="F6-10" remap="6.10"/></link> is the Orange visualization toolbox, listing all available types of visualizations.</para>
<para>Orange can be recommended as the most universally applicable tool with a shallow learning curve, while Matlab, R, and similar tools offer more flexibility and possibilities of automation of processes at the expense of having to learn the respective programming languages. However, neither of them scales well to truly large datasets of gigabytes or more.</para>
</section>
<section class="lev3" id="sec6-3-1-2">
<title>6.3.1.2 Scenario 1: Kibana</title>
<para>Kibana<footnote id="fn6_15" label="15"><para>https://www.elastic.co/products/kibana</para></footnote> is a visualization plugin for ElasticSearch<footnote id="fn6_16" label="16"><para>https://www.elastic.co/</para></footnote>, and the latter is a distributed, highly scalable indexing and search engine. ElasticSearch can store a huge number of schema-free JSON documents. The fields of the documents are automatically indexed and can be searched for using a powerful query syntax. The documents of different types can either be stored separately (in different indices in ElasticSearch terminology) or in the same index, such that the search query determines whether all documents or just those with certain attributes present should be searched for. Architecturally, ElasticSearch is a service based on Apache Lucene indexing/search library, accessed via a REST/HTTP(S) API.</para>
<fig id="F6-11" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-11">Figure <xref linkend="F6-11" remap="6.11"/></link></label>
<caption><para>MANTIS data visualization in Kibana.</para></caption>
<graphic xlink:href="graphics/ch06_fig011.jpg"/>
</fig>
<para>Kibana is implemented as an interactive web portal into ElasticSearch. It provides a user interface for performing queries on the stored data, but more interestingly, it can produce various visualizations. It is particularly powerful for viewing datasets that include a time dimension and allows interactive selection of time ranges, scales, filters, and aggregations. Together with Elastic search&#x02019;s support for large datasets, it is well suited to the maintenance use cases that continuously monitor processes and assets and thus, over longer time, invariable end up with large datasets.</para>
<para>The first step of integration of Kibana into a maintenance MANTIS prototype consists of development of a service that continuously requests new data from the data store, such as a MIMOSA database, and adds it to the ElasticSearch index. ElasticSearch also provides libraries for common programming languages, which are more straightforward to use than the HTTP-based REST API. Within the MANTIS project, a MIMOSA-to-ElasticSearch import service was developed in Python. The service simply adds the JSON objects returned by the MIMOSA REST API as ElasticSearch documents. Full contents of the (chosen subset of) MIMOSA tables are thus available for search and visualization. On the other hand, the data is raw and thus requires knowledge of MIMOSA data model (or, in general, the specific data model used). This can be seen in <link linkend="F6-11">Figure <xref linkend="F6-11" remap="6.11"/></link>, which shows a Kibana dashboard with three different graphs from a MANTIS use case<footnote id="fn6_17" label="17"><para>Please note that the time range includes periods when the MIMOSA database contained test data rather than real sensor measurements, therefore, the values shown here may not be realistic.</para></footnote> &#x02013; alarm types and environmental measurements are labelled by raw MIMOSA field values, because that is how they are stored in ElasticSearch.</para>
<para>Kibana individual visualizations or whole dashboards can be embedded in any HTML document using the &#x0003C;iframe> tag. <link linkend="F6-12">Figure <xref linkend="F6-12" remap="6.12"/></link> shows a single visualization inside the MANTIS generic HMI prototype, based on the data from a MANTIS use case. Note that the embedded graph is interactive and allows the user to change the time range shown or apply custom filter, as shown in this example. In cases when this is not desired, static snapshots of visualizations can be embedded instead.</para>
<fig id="F6-12" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-12">Figure <xref linkend="F6-12" remap="6.12"/></link></label>
<caption><para>A Kibana graph embedded in the MANTIS HMI prototype.</para></caption>
<graphic xlink:href="graphics/ch06_fig012.jpg"/>
</fig>
<para>To conclude, Kibana is particularly useful for (but not limited to) exploring large datasets of time-based data, such as sensor measurements and alarms. It is being widely used for monitoring assets in cloud computing and should thus perform equally well in industrial asset monitoring.</para>
</section>
<section class="lev3" id="sec6-3-1-3">
<title>6.3.1.3 Scenario 2: Textual and graphical data representation</title>
<para>Data on the machine is collected by means of sensors that are part the machine&#x02019;s control systems or from sensors which were added specifically for maintenance purposes. In the MANTIS sheet metal industry use case [Ferreira et al., 2017], raw data visualization includes the visualization of both historical and streamed data.</para>
<section class="lev4">
<title>Historical Data</title>
<para>Historical data is related with the values detected by sensors installed in the machine and is stored in the centralized database. It can then be consulted in a table or graphical format.</para>
<fig id="F6-13" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-13">Figure <xref linkend="F6-13" remap="6.13"/></link></label>
<caption><para>Machine Data Visualization (Table).</para></caption>
<graphic xlink:href="graphics/ch06_fig013.jpg"/>
</fig>
<para><link linkend="F6-13">Figure <xref linkend="F6-13" remap="6.13"/></link> shows the data collected from the machine in a table format. This data includes all the variable values that were measured during the machine operation.</para>
</section>
<section class="lev4">
<title>Streamed Data</title>
<para>Streamed data is also related with sensor values detected however the data is received directly from the message broker in the communication middleware and is presented to the user immediately after being consumed. For this the HMI subscribes to the appropriate machine queue and as soon as new data is loaded into that queue, it is transmitted to the HMI and is converted to a graphical format.</para>
<para><link linkend="F6-14">Figure <xref linkend="F6-14" remap="6.14"/></link> shows the data collected from the machine in a graphic format. After the user selects the data interval and the variables to be analysed the corresponding plot will be generated with all the values recorded from the machine within the selected interval. The user may then perform different analysis operations such as variables comparison, zoom in/out, define offsets, etc.</para>
<fig id="F6-14" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-14">Figure <xref linkend="F6-14" remap="6.14"/></link></label>
<caption><para>Graphic Visualization of Machine Data.</para></caption>
<graphic xlink:href="graphics/ch06_fig014.jpg"/>
</fig>
<fig id="F6-15" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-15">Figure <xref linkend="F6-15" remap="6.15"/></link></label>
<caption><para>Maintenance Sensor Data Visualization.</para></caption>
<graphic xlink:href="graphics/ch06_fig015.jpg"/>
</fig>
<para><link linkend="F6-15">Figure <xref linkend="F6-15" remap="6.15"/></link> shows data obtained from Maintenance Sensors which are placed on the machine, usually communicating over an independent channel (e.g., a wireless network). These sensors only acquire specific maintenance related information. In this case the sensor is acquiring the values regarding machine&#x02019;s moving parts.</para>
</section>
</section>
</section>
<section class="lev2" id="sec6-3-2">
<title>6.3.2 Augmented and Virtual Reality</title>
<para>Virtual reality and augmented reality were chosen as the technologies to be used in the advanced HMI approaches for the Finnish conventional energy production use case. They are an emerging market especially on consumer side and will most likely have an impact on maintenance in the future in one form or another. There is also a lot of innovation potential in these technologies.</para>
<para>The distinction between industrial maintenance related usage of VR and AR approaches can be roughly defined between factory-floor and back-office, where AR is more applicable for factory-floor and on the field maintenance tasks and guidance. VR is inherently more suited for back-office and other at office activities such as training and planning. VR&#x02019;s reliance on raw graphical computing power and, depending on the hardware solution used, external location and position hardware eliminates any possibility of it being mobile.</para>
<para>HTC Vive and its direct competitor Oculus, both rely on external hardware for position and location functionalities and are considered to be outside-in tracking solutions, meaning that the location and position data for the headset on the user&#x02019;s head comes from external beacons. However, hardware solutions such as the Microsoft Mixed Reality platform, which uses cameras attached to the headset itself to orient themselves and the controllers, are considered inside-out tracking solutions. This negates the need for external beacons and makes the VR more portable. The Microsoft MR platform also enables mixing in live, real-world stereo video using stereo cameras for a through-camera AR like approach, hence the name mixed reality.</para>
<para>Initially interest in the use case was placed on the AR approach, as it was more suitable for use in maintenance monitoring on the field or factory floor. The AR approach was done on the Google Tango platform that consists of a comprehensive Unity compatible AR SDK and a special hardware platform that consists of an IR dot matrix projector and a special camera capable of measuring the time-of-flight of the independent dots projected onto a shape. The combination of the SDK and the hardware platform is used to mitigate inherent drift in any pure IMU based positioning solution.</para>
<para>The AR application named AHMI (Advanced HMI) would enable users to create a virtual representation of the flue gas blower in the J&#x000E4;rvenp&#x000E4;&#x000E4; plant (see <link linkend="F6-16">Figure <xref linkend="F6-16" remap="6.16"/></link>) and retrieve real-world measurement data onto the measurement points attached to the 3D model. It also supports adding virtual measurement points to real-world objects using real-world measurement data retrieved from MIMOSA (see <link linkend="F6-17">Figure <xref linkend="F6-17" remap="6.17"/></link>). It also had additional features such as disassembling of 3D models (where the model allows it) and virtual post-it notes for leaving virtual messages onto the factory floor. The texts contained on the notes are stored in MIMOSA and thus could be used to gather tacit knowledge.</para>
<para>Development of the AR version of AHMI was halted as it became apparent that Google was halting support and development of the Tango platform. With the emergence of Apple&#x02019;s own proprietary augmented reality ARKit SDK, Google announced their own, mostly hardware-independent AR SDK the AR Core, which then quickly superseded the Tango platform that relied on specific hardware to be present on the device. On 15th of December 2017 Google announced that the Google Tango platform will be deprecated on 1st of March 2018<footnote id="fn6_18" label="18"><para>https://twitter.com/projecttango/status/941730801791549440</para></footnote>, and that finalized the cessation of AR development for the MANTIS platform as AR Core was still not ready for release.</para>
<fig id="F6-16" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-16">Figure <xref linkend="F6-16" remap="6.16"/></link></label>
<caption><para>3D model of flue gas blower placed at meeting room table.</para></caption>
<graphic xlink:href="graphics/ch06_fig016.jpg"/>
</fig>
<fig id="F6-17" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-17">Figure <xref linkend="F6-17" remap="6.17"/></link></label>
<caption><para>Marker object representing data from sensor.</para></caption>
<graphic xlink:href="graphics/ch06_fig017.jpg"/>
</fig>
<para>As the fate of the Tango Platform was already quite apparent during development, it was decided to develop a VR solution based on the work done with AR. In fact, the VR application was very similar to the AHMI one as it re-used most of code, as both were developed in Unity and were therefore mostly compatible. Some additional features were introduced such as more in-depth measurement windows and the possibility of viewing spectrums and even listening to the raw vibration measurement data. The VR AHMI application was built solely on the HTC Vive platform, however it could be transferred to other VR hardware platforms such as Oculus, possibly even the Microsoft mixed reality platform.</para>
<fig id="F6-18" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-18">Figure <xref linkend="F6-18" remap="6.18"/></link></label>
<caption><para>Screen capture of the VR demonstrator displaying real measurement data.</para></caption>
<graphic xlink:href="graphics/ch06_fig018.jpg"/>
</fig>
<para><link linkend="F6-18">Figure <xref linkend="F6-18" remap="6.18"/></link> shows the windows displaying real-world data from the J&#x000E4;rvenp&#x000E4;&#x000E4; plant. The users can open and reposition these windows to their own preferences. They also have a snap functionality that allows for neat alignment of all windows. It is possible to load FFTs and raw vibration data for any data point by moving a red indicator. It is also possible to move the indicator on all windows at the same time by enabling the indicator lock visible on the right hand corner of the window. <link linkend="F6-19">Figure <xref linkend="F6-19" remap="6.19"/></link> shows these measurement windows opened to a selected data point, where the raw measurements are shown in the top windows and the FFT are shown in the lowest windows. <link linkend="F6-20">Figure <xref linkend="F6-20" remap="6.20"/></link> shows closer inspection of the FFT data using gesture control enabled by Leap Motion.</para>
<fig id="F6-19" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-19">Figure <xref linkend="F6-19" remap="6.19"/></link></label>
<caption><para>FFT and raw vibration data opened for a measurement.</para></caption>
<graphic xlink:href="graphics/ch06_fig019.jpg"/>
</fig>
<fig id="F6-20" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-20">Figure <xref linkend="F6-20" remap="6.20"/></link></label>
<caption><para>FFT data inspection with Leap Motion controls.</para></caption>
<graphic xlink:href="graphics/ch06_fig020.jpg"/>
</fig>
<para>Both VR and AR solutions could be utilized as a part of collaborative decision-making. The VR could be used to convene and observe anomalies online over wide geographic distances. Experts around the world could communicate with each other using avatars in a 3D space. AR could be used, for instance locally to observe machinery status. It would allow the users to load the 3D model onto a conference room table and it could be visible to all users with AR capable devices. This could allow for new business opportunities in maintenance related decision support and collaboration. A separate &#x0201C;control room&#x0201D; was created as a part of the VR application. It&#x02019;s an empty VR space where collaborators can &#x0201C;send&#x0201D; measurement windows from the VR factory floor. This enables collaborativeness in the VR world between experts.</para>
<para>During testing and demonstration of the VR system, utilizing the HTC Vive included controllers, observations were made that the controllers themselves makes the VR less approachable and less intuitive. The controllers were then replaced with Leap Motion controller, which is a structured light based IR projection camera system intended for recognizing hand positions and gestures. This was incorporated into the VR demo to replace the hindrance of the controllers allowing for an immediately more intuitive approach using the user&#x02019;s hands as control tools. The Leap Motion unit came with a cradle that allowed it to be attached to the HTC Vive headset. The application was then modified to support the Leap Motion SDK and functionalities.</para>
<para>In Use Case 3.3 two separate scenarios related to HMIs were identified in the requirement specifications; automated vibration monitoring and condition and incoming maintenance alert for plant operators. Each of these scenarios were implemented in the Finnish consortium partners&#x02019; own HMI approaches. These scenarios were also implemented in the advanced visualization approaches, as described in the following subsections.</para>
<section class="lev3" id="sec6-3-2-1">
<title>6.3.2.1 Scenario 1: Automated vibration monitoring</title>
<para>In the VR implementation, the system level is represented by the virtual models of the equipment. The component levels are represented by the virtual sensor nodes. Drilling down to the measurements is enabled by opening the individual sensor positions and accessing the measurements by picking the point in time the operator would like to examine. At its current state the VR implementation does not allow for automated opening of measurement levels, it depends on user interaction to drill down to the FFT and raw data.</para>
<para>All the measurement data displayed within the VR environment is obtained from the REST interface and the MIMOSA database. Due to the way the measurement data is inserted into the database, it is not real-time. However, were the data stored in the database in real-time, it could be displayed near real-time within the VR environment.</para>
</section>
<section class="lev3" id="sec6-3-2-2">
<title>6.3.2.2 Scenario 2: Condition and incoming maintenance alert for plant operators</title>
<para>The VR demonstrator supports alarms via MIMOSA&#x02019;s alarm related tables. They are implemented in a manner that makes the alarms visible to the people using the VR application and the headset. The alarms appear as hovering items in the field of view of the operator and does not go away until the operator has confirmed the alarm by interacting with it with either the HTC Vive controllers or by hands.</para>
<para>The alarm item does also show the measurement that has caused the alarm and can provide the operator a virtual representation of the equipment or measurement point that has caused the alarm. From there the operator can observe the cause of the alarm using the normal VR demonstrator tools. <link linkend="F6-21">Figure <xref linkend="F6-21" remap="6.21"/></link> shows two active alarms and the measurements that has caused the alarm. Using the VR controller to click and dismiss the alarm will teleport the user to the equipment the measurement point is attached to, which is visible in <link linkend="F6-21">Figure <xref linkend="F6-21" remap="6.21"/></link> as the electric motor with the red dot attached. There are plans to incorporate user recognition to direct the alarms at correct personnel.</para>
<fig id="F6-21" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-21">Figure <xref linkend="F6-21" remap="6.21"/></link></label>
<caption><para>Screenshot showing a super-imposed alarm and the synthetic data that has caused it.</para></caption>
<graphic xlink:href="graphics/ch06_fig021.jpg"/>
</fig>
</section>
</section>
</section>
<section class="lev1" id="sec6-4">
<title>6.4 Usability Testing Methodology for Industrial HMIs</title>
<para>Usability aspects are becoming increasingly important and a common criterion associated with performance is the quality of service. Quality of service can be studied from both the system point of view and the user point of view. In this context, quality of service from the system point of view means, that the system is capable of offering and doing required functionalities which users are observing or commanding through the HMI. Quality of service from the user point of view means, that the users are capable of reading or doing those required functionalities through the HMI. Even if there are divergences for the evaluation approaches of these services we shall follow the best practice in individual aspects such as input performance, interpretation performance, context appropriateness (at the system side) and perceptual effort, cognitive workload, physical effort on the user side. In addition, quality of experience covering user perception and satisfaction issues is also gaining importance [MANTIS Consortium, 2015].</para>
<para>Before proceeding to the usability evaluation of specific HMI implementation, first refer to the issue of usability as defined by the standards (ISO) dealing with human computer interaction (HCI), focusing on ergonomics of human-system interaction and associated product quality.</para>
<para>The website of the UsabilityNet, a project funded by the European Union to promote usability and user centred design (Usability Net), provides an extensive list of standards related to HCI and usability. They are categorised into four groups, as those primarily concerned with:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>the use of the product (effectiveness, efficiency and satisfaction in a particular context of use);</para></listitem>
<listitem>
<para>the user interface and interaction;</para></listitem>
<listitem>
<para>the process used to develop the product;</para></listitem>
<listitem>
<para>the capability of an organization to apply user centred design.</para></listitem>
</itemizedlist>
<section class="lev2" id="sec6-4-1">
<title>6.4.1 Human-system Interaction &#x02013; Usability Standards</title>
<para>As regards usability definition, ISO 9241-11, ISO 9241-210, ISO/IEC 9126 standards and ISO/TR 16982 technical report are exposed. In the following, we borrow some parts of their descriptions in order to reveal their specific features for the purposes of the MANTIS project.</para>
</section>
<section class="lev2" id="sec6-4-2">
<title>ISO 9241-11: Guidance on Usability (1998)</title>
<para>In this standard, usability is defined as the &#x0201C;<emphasis>extent to which a product can be used by specified users to achieve specified goals with effectiveness, efficiency, and satisfaction in a specified context of use</emphasis>&#x0201D;.</para>
<para>Hereby:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>effectiveness denotes the accuracy and completeness with which the users achieves specified goals;</para></listitem>
<listitem>
<para>efficiency measures the resources spent in relation to the accuracy and completeness with which users achieve their goals;</para></listitem>
<listitem>
<para>satisfaction designates the freedom from discomfort, and positive attitudes towards the use of the product;</para></listitem>
<listitem>
<para>context of use includes a description of users, their tasks, employed equipment and physical/social environment.</para></listitem>
</itemizedlist>
<para>In order to specify or measure usability, it is necessary to identify the goals and to decompose effectiveness, efficiency and satisfaction and the components of the context of use into sub-components with measurable and verifiable attributes.</para>
<para>Guidance is given on how to describe the context of use of the product and the measures of usability in an explicit way. The document also explains how measures of user performance and satisfaction can be used to measure how any component of a work system affects the quality of the whole work system in use [ISO, 1998].</para>
</section>
<section class="lev2" id="sec6-4-3">
<title>ISO 9241-210: Ergonomics of Human-system Interaction &#x02013; Part 210: Human-centred Design for Interactive Systems (2010)</title>
<para>As the title suggests, this standard deals with human-centred design for interactive systems and explains the activities required for user centred design. It is intended to be used by those managing design processes.</para>
<para>Principles of human-centred design are listed and described. Next, the main steps of planning human-centred design are surveyed. Once the planning is done, the following design activities are foreseen and described in more details:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>understanding and specifying the context of use;</para></listitem>
<listitem>
<para>specifying the user requirements;</para></listitem>
<listitem>
<para>producing design solutions;</para></listitem>
<listitem>
<para>evaluating the design.</para></listitem>
</itemizedlist>
<para>User-centred evaluation should involve:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>properly allocating resources, such as properly selected users, description of product functionalities and context of use (in an early stage) to obtain feedback that can be used to improve or redesign the product and (later) to determine whether the requirements have been met;</para></listitem>
<listitem>
<para>planning the user-centred evaluation;</para></listitem>
<listitem>
<para>carrying out comprehensive testing to provide meaningful results for the system as a whole;</para></listitem>
<listitem>
<para>analysing the results;</para></listitem>
<listitem>
<para>communicating the results so that they can be used by the design team.</para></listitem>
</itemizedlist>
<para>Two widely used approaches to user-centred evaluation are</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>user-based testing;</para></listitem>
<listitem>
<para>inspection-based evaluation.</para></listitem>
</itemizedlist>
<para>Both are described at the appropriate level of details considering the fact that the standard is intended primarily for managers and not for product developers [ISO, 2010].</para>
</section>
<section class="lev2" id="sec6-4-4">
<title>ISO/IEC 9126: Software Product Evaluation &#x02013; Quality Characteristics and Guidelines for their Use (1991)</title>
<para>This standard was developed separately as a software engineering standard. It defined usability as one relatively independent contribution to software quality associated with the design and evaluation of the user interface and interaction. This standard has been replaced by a new four-part standard ISO/IEC FDIS 9126 (2000) [ISO, 1991].</para>
</section>
<section class="lev2" id="sec6-4-5">
<title>ISO/IEC FDIS 9126-1: Software Engineering &#x02013; Product Quality &#x02013; Part 1: Quality Model (2000)</title>
<para>This is the first of the four parts of ISO/IEC FDIS 9126 (2000) and describes a two-part model for software product quality:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>internal quality and external quality;</para></listitem>
<listitem>
<para>quality in use.</para></listitem>
</itemizedlist>
<para>Internal Quality Requirements specify the level of required quality from the internal view of the product and address implementation issues (such as employed models, source code, etc.). Internal quality represents the internal characteristics of the software product and can be evaluated against the internal quality requirements. External Quality Requirements specify the required level of quality from the external view and are derived from user quality needs. External quality thus represents the characteristics of the software product from an external view.</para>
<para>The quality model of the internal quality and external quality specifies six categories of software quality that are relevant during product development: functionality, reliability, usability, efficiency, maintainability and portability. Each of them is further divided in sub-categories.</para>
<para>Quality in use is the user&#x02019;s view of quality. It is specified in four categories: effectiveness, productivity, safety and satisfaction. Hereby the effectiveness, productivity and satisfaction somehow correspond to the notions of effectiveness, efficiency, and satisfaction defined in ISO 9241-11 [ISO, 2001].</para>
</section>
<section class="lev2" id="sec6-4-6">
<title>ISO/IEC FDIS 9126-2: Software Engineering &#x02013; Product Quality &#x02013; Part 2: External Metrics and ISO/IEC FDIS 9126-3: Software Engineering &#x02013; Product Quality &#x02013; Part 3: Internal Metrics</title>
<para>As the title suggests, these two parts of the standard define metrics for quantitative measuring software quality. Each of the two parts contains an explanation of how to apply software quality metrics and a basic set of metrics for each sub-characteristic of the stated six categories. Additional explanations, such as considerations when using metrics (i.e., interpretation of measures, validation of metrics, etc.) and explanation of metric scale types and measurement types are given in informative annexes [ISO, 2003; ISO, 2003].</para>
</section>
<section class="lev2" id="sec6-4-7">
<title>SO/IEC FDIS 9126-4: Software Engineering &#x02013; Product Quality &#x02013; Part 4: Quality in Use Metrics</title>
<para>Similar to the above two, this part of the standard defines metrics for quantitative measuring of software quality related to the four categories defined by the quality model for quality in use [ISO, 2004].</para>
</section>
<section class="lev2" id="sec6-4-8">
<title>SO/TR 16982: Ergonomics of Human-system Interaction &#x02013; Usability Methods Supporting Human-centred Design</title>
<para>In reference to ISO 9241-11 and ISO 9241-210 this technical report provides an overview of existing usability methods which can be used on their own or in combination to support design and evaluation. Each method is described with its advantages, disadvantages and other factors relevant to its selection and use. The purpose of this technical report is to help project managers make informed decisions about the choice of usability methods to support human-centred design principles as described in ISO 9241-210.</para>
<para>In order to incorporate usability requirements, the following four human-centred design activities are suggested:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>understanding and specifying the context of use;</para></listitem>
<listitem>
<para>specifying user requirements;</para></listitem>
<listitem>
<para>producing designs and prototypes;</para></listitem>
<listitem>
<para>performing user-based assessment.</para></listitem>
</itemizedlist>
<para>The described usability methods are focused either on design or evaluation. In the first case data-gathering techniques that are applied early in the design phase and are used to guide the design. The second case refers to the evaluation of design such as assessment of interface features, expected task completion time, expected use pattern, etc. The methods that are presented in this technical report are those that are most frequently used. The methods are listed in a table and divided into two broad categories:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>methods that imply the direct involvement of users;</para></listitem>
<listitem>
<para>methods that imply the indirect involvement of users, which are used either when it is not possible to gather usage data due or where they provide complementary data and information.</para></listitem>
</itemizedlist>
<para>After the description of individual usability methods, the preferred choice of usability methods some general guidelines for their application during the different design phases are given. The last part is devoted to the choice of usability methods depending on life-cycle process, constraints of project environment, user characteristics, characteristics of the task to be performed, the product used and the abilities required for the designer or evaluator [ISO, 2002].</para>
<para>While ISO 9241-11 and ISO/IEC FDIS 9126-1 address the usability in a slightly different way, the approach described in ISO 9241-11 is more closely related to the issues of the MANTIS project and we shall adopt it as the starting point of usability testing. According to ISO 9241-11, the following information is needed when measuring usability:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>a description of intended goals;</para></listitem>
<listitem>
<para>a description of the components of the context of use including users, tasks, equipment and environments;</para></listitem>
<listitem>
<para>target or actual values of effectiveness, efficiency, and satisfaction for the intended contexts.</para></listitem>
</itemizedlist>
<para>According to ISO 9241-11, measures of usability should be based on data, which reflect the result of user interaction with the system. Effectiveness, efficiency and satisfaction can be measured as follows:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Effectiveness is defined as accuracy and completeness with which the users achieve specified goals. Hereby, the accuracy can be measured to which extend the quality of the implemented HMI corresponds to the specified criteria. For example, how consistent are the implemented functionalities. The completeness can be measured as the extent of the achieved target quantity. For example, how many of the specified functionalities have actually been implemented;</para></listitem>
<listitem>
<para>Efficiency is measured by relating the level of effectiveness achieved to the resources used. For example, for a given functionality, how long does it take to perform given task and achieve the result complying to the stated goal;</para></listitem>
<listitem>
<para>Satisfaction can be assessed by objective and subjective measures. Objective measures are based on observation of the behaviour of the user. Subjective measures comprise data expressing user&#x02019;s opinions, attitudes and reactions. These data can be obtained by asking users to express their feeling when performing specific task, or by using an attitude scale based on a questionnaire.</para></listitem>
</itemizedlist>
<para>In order to properly measure effectiveness, efficiency and satisfaction, appropriate metrics should be applied. In this regard, metrics for software product quality described in the four-part standard ISO/IEC FDIS 9126 (2000) can serve as an example or a guideline. In particular, metrics described in ISO/IEC FDIS 9126-2 and ISO/IEC FDIS 9126-3 can be applied to measuring effectiveness and efficiency of the implemented HMI; and metrics described in ISO/IEC FDIS 9126-4 for measuring user satisfaction when performing tasks on the implemented HMI, respectively.</para>
<para>Usability tests differ in the way that usability measurements are performed, depending on the phase of the design process in which they are applied. In ISO/TR 16982, advantages and disadvantages of each type of usability methods are described.</para>
<para>While all standards described previously deal with usability issues, none of them gives explicit guidelines for performing usability tests. To some minor extent, the usability testing is described in ISO 9241-210. Two widely used approaches to user-centred evaluation are: user-based testing and inspection-based testing.</para>
<para>User-based testing can be undertaken at any stage in the design. At a very early stage, users can be presented with models, scenarios or sketches of the design concepts and asked to evaluate them in relation to a real context. Such early testing can provide valuable feedback on the acceptability of the proposed design. At a later stage in the development, user-based testing can be carried out to assess whether usability objectives, including measurable usability performance and satisfaction criteria, have been met in the intended context.</para>
<para>Inspection-based evaluation complements user testing. It can be used to eliminate major issues before user testing and hence make user testing more cost-effective. Usually two to three analysts evaluate the system with reference to established guidelines or principles, noting down their observations and often ranking them in order of severity. The analysts are usually experts in human factors or HCI. However, inspection does not always find the same problems that would be found in user-based testing. The greater the difference between the knowledge and experience of the inspectors and the real users, the less reliable are the results.</para>
<para>For detailed description, how to perform usability testing one should rather rely upon the information and guidelines provided in (Usability Net) (Usability.gov), or in [Rubin, et al., 2008]. Based on the above resources we describe in the next section the usability testing methods.</para>
</section>
<section class="lev2" id="sec6-4-9">
<title>6.4.2 Usability Testing Methodology for MANTIS</title>
<para>In the case of MANTIS HMI, usability testing was performed primarily as user-based testing. Usability test is performed by representative users under guidance of a skilled moderator. Users are observed when performing given tasks (e.g., the most frequent or the most critical). The collected qualitative and quantitative data related to the performed tasks serve for the improvement of the product.</para>
<para>Basic elements of usability testing can be summarised as follows:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>development of research questions or test objectives;</para></listitem>
<listitem>
<para>use of a representative sample of end users;</para></listitem>
<listitem>
<para>presentation of the actual work environment and usability tasks;</para></listitem>
<listitem>
<para>observation of users who use the product in order to perform given task;</para></listitem>
<listitem>
<para>interviewing the users by the test moderator;</para></listitem>
<listitem>
<para>collection of quantitative and qualitative performance and preference measures;</para></listitem>
<listitem>
<para>recommendation of improvements to the design of the product.</para></listitem>
</itemizedlist>
<para>As described in [Rubin, et al., 2008], three types of usability testing are distinguished depending on the phase of a development cycle:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>exploratory study;</para></listitem>
<listitem>
<para>assessment or summative test;</para></listitem>
<listitem>
<para>validation or verification test.</para></listitem>
</itemizedlist>
<para>The exploratory study is conducted early in the development cycle, when a product is still in the preliminary stages of design when its basic concept and functionalities are being defined. Some typical user-oriented questions that an exploratory study would attempt to answer might include: what do users think about using the product, are product&#x02019;s functionalities useful to the user, how easily can the user learn to use the product, etc.</para>
<para>The main objective of the exploratory study is to examine the effectiveness of preliminary design concepts. Exploratory tests are characterised by extensive interaction between the participant and the test moderator. Since the product is in the early design phase, only preliminary versions (in a form of prototype) are available for user&#x02019;s evaluation. During the test of such a prototype, the user would attempt to perform representative tasks or just simply express his feeling about the product under test. The user is encouraged to &#x0201C;think aloud&#x0201D; and his comments and remarks are collected for subsequent analysis and improvement of the product. Collected data is qualitative. Its benefits are twofold:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Potential usability problems can be detected at an early stage before development is complete;</para></listitem>
<listitem>
<para>A deeper understanding of the users&#x02019; expectations and impressions of the system.</para></listitem>
</itemizedlist>
<para>Assessment tests are performed early or midway into the product development cycle, usually after the fundamental concept and functionalities of the product have been established. The product is developed with specified functionalities but probably requires optimization and polishing. The purpose of the assessment test is to examine how effectively the concept has been implemented. Assessment test thus reveals how well a user can actually perform a typical realistic task and identifies possible usability deficiencies that manifest during the task completion. In contrast to exploratory study:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>the user will always perform tasks rather than simply walking through and commenting;</para></listitem>
<listitem>
<para>the communication with the test moderator is less concerned about user&#x02019;s feeling and comments and more focused on the actual task execution.</para></listitem>
</itemizedlist>
<para>Qualitative and quantitative data are collected. Its benefits are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>identified deficiencies can be improved since the product is still in a development phase, with all development tools and development team available;</para></listitem>
<listitem>
<para>possible missing functionalities can be implemented.</para></listitem>
</itemizedlist>
<para>The validation test, also referred to as the verification test, is usually conducted late in the development cycle and, as the name suggests, it should confirm that the problems identified in the earlier phases have been solved and that the product under test operates without faults. It may also be used to measure usability of a product against established benchmarks. This test typically takes place close to the release of the product.</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Usability goals are stated in terms of performance criteria. The effectiveness and efficiency metrics are defined in accordance with the stated goals;</para></listitem>
<listitem>
<para>Participants perform tasks without (or very little) interaction with the test moderator;</para></listitem>
<listitem>
<para>Quantitative data related to the stated effectiveness and efficiency metrics is collected.</para></listitem>
</itemizedlist>
<para>Exploratory test will be made in an early design phase. The main purpose of exploratory test is to help HMI developers to improve their HMI already in the design phase. The advantage of the exploratory test is that the shortcomings are noticed early enough before the first version of HMI has actually been built. The exploratory test can be performed independently for every HMI related use case.</para>
<para>Assessment test will be made when the basic HMI functionalities are implemented in use-case but not yet optimized. Some general guidelines how to perform assessment test for the first version of implemented HMI are given in the following section. The purpose is to help use-case owners and HMI developing partners to improve their implemented HMIs towards the final version and validation testing. The same way as exploratory testing, the assessment test can be made independently among use-case owners and HMI developing partners based on the given guidelines.</para>
<para>Validation is the final testing phase. The purpose of the validation test is to provide valuable feedback how to improve individual HMIs in use-cases. The main idea is that validation tests are done in each use-case according to the common guideline. In this way, the results of different use-cases can be compared and possible deviations and deficiencies removed.</para>
<para>The following guidelines have been adopted from (Usability Net).</para>
<para>Planning:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Define goals of the performed usability testing;</para></listitem>
<listitem>
<para>Define metrics that will be used to assess or measure to what extent the goals have been achieved;</para></listitem>
<listitem>
<para>Select representative users. The same users can be employed in all stages of usability tests;</para></listitem>
<listitem>
<para>Select the most important tasks (e.g., the most frequent or the most critical) related to the stated goals;</para></listitem>
<listitem>
<para>Produce task scenarios and input data and write instructions for the user (tell the user what to achieve, not how to do it);</para></listitem>
<listitem>
<para>Plan sessions allowing time for giving instructions, running the test, and a post-test interview;</para></listitem>
<listitem>
<para>Invite developers to observe the sessions if possible. An alternative is to videotape the sessions, and show developers edited clips of the main issues.</para></listitem>
</itemizedlist>
<para>Running sessions:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Welcome the user, and give the task instructions. Ask for their consent do gather, keep and use the collected data (according to GDPR rules);</para></listitem>
<listitem>
<para>Ask user to perform the task and record and/or measure task implementation (e.g., record user&#x02019;s comments and questions, measure task completion time);</para></listitem>
<listitem>
<para>Do not give any hints or assistance unless the user is unable to complete the task;</para></listitem>
<listitem>
<para>Observe the interaction and note any problems encountered;</para></listitem>
<listitem>
<para>During exploratory study and in some cases also during assessment test, the user may be prompted for their impressions of the design, what they think different elements may do, and what they expect the result of their next action to be. The user may also be asked to suggest how individual elements could be improved;</para></listitem>
<listitem>
<para>Interview the user to gain general opinions, and to ask about specific problems encountered.</para></listitem>
</itemizedlist>
<para>Output:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Report of the conducted usability tests, including employed metrics and achieved results;</para></listitem>
<listitem>
<para>Based on the above results produce a list of usability problems, categorised by importance and an overview of the types of problems encountered;</para></listitem>
<listitem>
<para>Arrange a meeting with the designers to discuss whether and how each problem can be fixed.</para></listitem>
</itemizedlist>
<para>While the above guidelines are quite general and could be applicable to any type of usability testing, some additional commends and recommendations specific to exploratory study, assessment test, and validation or verification test in the case of MANTIS HMI given in the following paragraphs might be helpful.</para>
<para>Since the initial steps toward HMI design in individual use cases have already started and each use case addresses specific user requirements, the approaches in exploratory studies and assessment tests are likely to be divergent to some extent. Nevertheless, it would be prudent to collect and document the acquired usability test data for possible exchange of &#x0201C;lessons learned&#x0201D; among use cases and for reporting.</para>
<para>In the early design phase, preliminary concepts are evaluated with representative users. The focus group research employs simultaneous involvement of more than one user and deals with the target product in a very preliminary form (i.e., paper drawing, screen-based prototypes) and explores users&#x02019; judgements and feelings. In this way, basic functionalities of the target product are explored, and possible missing items can be identified. Such proof of concept may include also surveys and questionnaires employed in order to understand the preferences of a broader base of users. For these procedures no formal templates are foreseen in the frame of MANTIS project but use case owners are advised to document the data in an appropriate way.</para>
<para>Assessment test and verification test differ in the goals and the corresponding metrics. Assessment test typically establishes if the functionalities stated in the requirement specifications have been implemented. On the other hand, verification test checks the efficiency of the implemented product and typically measures the time and resources required to accomplish given tasks. For example, the assessment test can be employed to check the functional implementation completeness in the following way: count the number of missing functions detected in evaluation and compare with the number of functions described in the requirement specifications. The goal of a verification test could be, for example, to verify that 70 percent of the users could meet the established successful completion criteria for each major task scenario.</para>
<para>When conducting usability testing it is good to have a validation plan. In MANTIS one of the methods to plan this activity was to fill in a two-page, interactive dashboard-like planning template. This offers a concise yet rather comprehensive way to plan the usability testing procedure. It is based on David Travis&#x02019; single page usability testing dashboard document. The original document was expanded and modified slightly with additional space for objectives, test tasks and measureable usability goals descriptions. This was done in order to make it more suitable for MANTIS and industrial HMI testing purposes. The idea was to provide a comprehensive dashboard for testing procedures.</para>
<fig id="F6-22" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-22">Figure <xref linkend="F6-22" remap="6.22"/></link></label>
<caption><para>An example of a filled Usability Validation Plan document page 1.</para></caption>
<graphic xlink:href="graphics/ch06_fig022.jpg"/>
</fig>
<fig id="F6-23" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F6-23">Figure <xref linkend="F6-23" remap="6.23"/></link></label>
<caption><para>An example of a filled Usability Validation Plan document page 2.</para></caption>
<graphic xlink:href="graphics/ch06_fig023.jpg"/>
</fig>
<para><link linkend="F6-22">Figures <xref linkend="F6-22" remap="6.22"/></link> and <link linkend="F6-23"><xref linkend="F6-23" remap="6.23"/></link> are screen captures for a validation plan done for a VR based advanced HMI approach developed by Lapland University of Applied Sciences. This document, together with the report template, was used to conduct the usability testing of the developed VR HMI. The same validation plan templates were also used in other use cases successfully.</para>
<para>The usability validation plan was done in a dashboard form factor. This was then released as a PDF document template with two pages and multiple prepared form fields that allow for insertion of text. Each text field also had a tooltip that would guide how to fill each field accordingly. The tooltip was a mouse-over based event and would appear once the user moves the mouse over the form field similar to those used in webpages. Each usability validation plan document was then saved as its own PDF document for the use case usability testing.</para>
<para>The document is divided into 12 separate sections that further explain the testing procedure and requirements. There are also MANTIS specific fields that tie the planned test to the use case and related content or HMI element. Text intensive fields such as the test objectives, tasks and measurable usability goals have been extended to the second page in order to fit more content. It is advised however to try to keep the content as short as possible to keep the plan as concise and easy to approach as possible.</para>
<para>The ultimate goal of MANTIS HMI is to provide interaction facilities that comply to the established usability goals: effectiveness, efficiency, and satisfaction. In order to help use case owners to perform usability tests and report the results in a way that would allow easy comparison and exchange of good practices.</para>
</section>
</section>
<section class="lev1" id="sec6-5">
<title>References</title>
<orderedlist numeration="arabic" continuation="restarts" spacing="normal">
<listitem>
<para>Abowd G. D. [et al.] (1999) Towards a Better Understanding of Context and Context-Awareness, Handheld and ubiquitous computing (pp. 304&#x02013;307). Springer Berlin Heidelberg. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Abowd+G%2E+D%2E+%5Bet+al%2E%5D+%281999%29+Towards+a+Better+Understanding+of+Context+and+Context-Awareness%2C+Handheld+and+ubiquitous+computing+%28pp%2E+304-307%29%2E+Springer+Berlin+Heidelberg%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Ahmadzadeh F., Ghodrati, B. and Kumar U. (2012) &#x02018;Mean Residual Life Estimation Considering Operating Environment International Conference on Quality, Reliability, Infocom Technology and Industrial Technology Management. Delhi, India. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ahmadzadeh+F%2E%2C+Ghodrati%2C+B%2E+and+Kumar+U%2E+%282012%29+%27Mean+Residual+Life+Estimation+Considering+Operating+Environment+International+Conference+on+Quality%2C+Reliability%2C+Infocom+Technology+and+Industrial+Technology+Management%2E+Delhi%2C+India%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Atterer Richard, Wnuk Monika and Schmidt Albrecht (2006) &#x02018;Knowing the user&#x02019;s every move: User activity tracking for website usability evaluation and implicit interaction,&#x02019; Proceedings of the 15th international conference on World Wide Web, pp. 203&#x02013;212. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Atterer+Richard%2C+Wnuk+Monika+and+Schmidt+Albrecht+%282006%29+%27Knowing+the+user%27s+every+move%3A+User+activity+tracking+for+website+usability+evaluation+and+implicit+interaction%2C%27+Proceedings+of+the+15th+international+conference+on+World+Wide+Web%2C+pp%2E+203-212%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Barkhuus L. and Dey A. (2003) &#x02018;Is context-aware computing taking control away from the user? Three Levels of Interactivity Examined,&#x02019; Proceedings of the Fifth Annual Conference on Ubiquitous Computing (UBICOMP 2003), Springer Berlin Heidelberg, pp. 149&#x02013;156. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Barkhuus+L%2E+and+Dey+A%2E+%282003%29+%27Is+context-aware+computing+taking+control+away+from+the+user%B4+Three+Levels+of+Interactivity+Examined%2C%27+Proceedings+of+the+Fifth+Annual+Conference+on+Ubiquitous+Computing+%28UBICOMP+2003%29%2C+Springer+Berlin+Heidelberg%2C+pp%2E+149-156%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Boy G. A. (2011) The Handbook of human&#x02013;machine interaction: A human-centered approach [Book]. [s.l.]: Aldershot, 2011. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Boy+G%2E+A%2E+%282011%29+The+Handbook+of+human-machine+interaction%3A+A+human-centered+approach+%5BBook%5D%2E+%5Bs%2El%2E%5D%3A+Aldershot%2C+2011%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Chantzara M. and Anagnostou M. (2005) Evaluation and selection of context information, In Second International Workshop on Modeling and Retrieval of Context. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Chantzara+M%2E+and+Anagnostou+M%2E+%282005%29+Evaluation+and+selection+of+context+information%2C+In+Second+International+Workshop+on+Modeling+and+Retrieval+of+Context%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Coutand O. [et al.] (2006) A Case-based Reasoning Approach for Personalizing Location-aware Services Workshop on Case-based Reasoning and Context Awareness. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Coutand+O%2E+%5Bet+al%2E%5D+%282006%29+A+Case-based+Reasoning+Approach+for+Personalizing+Location-aware+Services+Workshop+on+Case-based+Reasoning+and+Context+Awareness%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Dev Himel and Zhicheng Liu (2017) &#x02018;Identifying Frequent User Tasks from Application Logs,&#x02019; Proceedings of the 22nd International Conference on Intelligent User Interfaces, pp. 263&#x02013;273. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Dev+Himel+and+Zhicheng+Liu+%282017%29+%27Identifying+Frequent+User+Tasks+from+Application+Logs%2C%27+Proceedings+of+the+22nd+International+Conference+on+Intelligent+User+Interfaces%2C+pp%2E+263-273%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Diaper Dan and Stanton Neville (2004) The Handbook of Task Analysis for Human-Computer Interaction. London: Lawrence Erlbaum Associates, Publishers. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Diaper+Dan+and+Stanton+Neville+%282004%29+The+Handbook+of+Task+Analysis+for+Human-Computer+Interaction%2E+London%3A+Lawrence+Erlbaum+Associates%2C+Publishers%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Dunlop Mark and Brewster Stephen (2002) The challenge of mobile devices for human computer interaction, Personal and Ubiquitous Computing, pp. 235&#x02013;236.</para></listitem>
<listitem>
<para>Ehlert Patrick (2003) Intelligent user interfaces. Introduction and survey. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ehlert+Patrick+%282003%29+Intelligent+user+interfaces%2E+Introduction+and+survey%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Endres C. [et al.] (2010) Cinematic analysis of automotive personalization, In Proceedings of the International Workshop on Multimodal Interfaces for Automotive Applications (MIAA&#x02019;10) in Conjunction with IUI, pp. 1&#x02013;6. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Endres+C%2E+%5Bet+al%2E%5D+%282010%29+Cinematic+analysis+of+automotive+personalization%2C+In+Proceedings+of+the+International+Workshop+on+Multimodal+Interfaces+for+Automotive+Applications+%28MIAA%2710%29+in+Conjunction+with+IUI%2C+pp%2E+1-6%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Feng J. and Liu Y. (2015) Intelligent context-aware and adaptive interface for mobile LBS, Computational Intelligence and Neuroscience. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Feng+J%2E+and+Liu+Y%2E+%282015%29+Intelligent+context-aware+and+adaptive+interface+for+mobile+LBS%2C+Computational+Intelligence+and+Neuroscience%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Ferreira, L. L., Albano, M., Silva, J., Martinho, D., Marreiros, G., di Orio, G., Mal&#x000F3;, P., and Ferreira, H. (2017) &#x02018;A Pilot for Proactive Maintenance in Industry 4.0,&#x02019; In <emphasis>13th IEEE International Workshop on Factory Communication Systems (WFCS 2017)</emphasis>. Trondheim, Norway. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ferreira%2C+L%2E+L%2E%2C+Albano%2C+M%2E%2C+Silva%2C+J%2E%2C+Martinho%2C+D%2E%2C+Marreiros%2C+G%2E%2C+di+Orio%2C+G%2E%2C+Mal%F3%2C+P%2E%2C+and+Ferreira%2C+H%2E+%282017%29+%27A+Pilot+for+Proactive+Maintenance+in+Industry+4%2E0%2C%27+In+13th+IEEE+International+Workshop+on+Factory+Communication+Systems+%28WFCS+2017%29%2E+Trondheim%2C+Norway%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Ferreiro Susana [et al.] (2012) &#x02018;Application of Bayesian networks in prognostics for a new Integrated Vehicle Health Management concept,&#x02019; Expert Systems with Applications, pp. 6402&#x02013;6418. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ferreiro+Susana+%5Bet+al%2E%5D+%282012%29+%27Application+of+Bayesian+networks+in+prognostics+for+a+new+Integrated+Vehicle+Health+Management+concept%2C%27+Expert+Systems+with+Applications%2C+pp%2E+6402-6418%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Ferscha A., Vogl S. and Beer W. (2001) &#x02018;Context sensing, aggregation, representation and exploitation in wireless networks,&#x02019; Scalable Computing: Practice and Experience. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ferscha+A%2E%2C+Vogl+S%2E+and+Beer+W%2E+%282001%29+%27Context+sensing%2C+aggregation%2C+representation+and+exploitation+in+wireless+networks%2C%27+Scalable+Computing%3A+Practice+and+Experience%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Gardner Brett S. (2011) &#x02018;Responsive web design: Enriching the user experience,&#x02019; Sigma Journal: Inside the Digital Ecosystem, pp. 13&#x02013;19. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Gardner+Brett+S%2E+%282011%29+%27Responsive+web+design%3A+Enriching+the+user+experience%2C%27+Sigma+Journal%3A+Inside+the+Digital+Ecosystem%2C+pp%2E+13-19%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Garzon S. R. and Poguntke M. (2011) The personal adaptive in-car HMI: Integration of external applications for personalized use, In International Conference on User Modeling, Adaptation, and Personalization. Springer Berlin Heidelberg, pp. 35&#x02013;46. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Garzon+S%2E+R%2E+and+Poguntke+M%2E+%282011%29+The+personal+adaptive+in-car+HMI%3A+Integration+of+external+applications+for+personalized+use%2C+In+International+Conference+on+User+Modeling%2C+Adaptation%2C+and+Personalization%2E+Springer+Berlin+Heidelberg%2C+pp%2E+35-46%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Hashemi Mohammad and Herbert John (2016) User interaction monitoring and analysis framework, Proceedings of the International Conference on Mobile Software Engineering and Systems, pp. 7&#x02013;8.</para></listitem>
<listitem>
<para>Hashiguchi T., Matsuo H. and Hashizume A. (2003) &#x02018;Healthcare dynamics informatics for personalized healthcare,&#x02019; Hitachi Review, pp. 183&#x02013;188. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Hashiguchi+T%2E%2C+Matsuo+H%2E+and+Hashizume+A%2E+%282003%29+%27Healthcare+dynamics+informatics+for+personalized+healthcare%2C%27+Hitachi+Review%2C+pp%2E+183-188%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Hynes G., Reynolds V. and Hauswirth M. (2009) &#x02018;A Context lifecycle for web-based context management services,&#x02019; Smart Sensing and Contextser. Lecture Notes in Computer Science, Heidelberg: Springer Berlin, pp. 51&#x02013;65. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Hynes+G%2E%2C+Reynolds+V%2E+and+Hauswirth+M%2E+%282009%29+%27A+Context+lifecycle+for+web-based+context+management+services%2C%27+Smart+Sensing+and+Contextser%2E+Lecture+Notes+in+Computer+Science%2C+Heidelberg%3A+Springer+Berlin%2C+pp%2E+51-65%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>ISO ISO 9241-11:1998 (1998) Ergonomic requirements for office work with visual display terminals (VDTs) Part 11: Guidance on usability. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=ISO+ISO+9241-11%3A1998+%281998%29+Ergonomic+requirements+for+office+work+with+visual+display+terminals+%28VDTs%29+Part+11%3A+Guidance+on+usability%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>ISO ISO 9241-210: (2010) Ergonomics of human-system interaction - Part 210: Human-centred design for interactive systems. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=ISO+ISO+9241-210%3A+%282010%29+Ergonomics+of+human-system+interaction+-+Part+210%3A+Human-centred+design+for+interactive+systems%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>ISO ISO International Organization for Standardization. 10 April 2017. https://www.iso.org/standards.html.</para></listitem>
<listitem>
<para>ISO ISO/IEC 9126: Software product evaluation &#x02013; Quality characteristics and guidelines for their use. 1991.</para></listitem>
<listitem>
<para>ISO ISO/IEC 9126-1:2001 Software engineering &#x02013; Product quality &#x02013; Part 1: Quality model. 2001. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=ISO+ISO%2FIEC+9126-1%3A2001+Software+engineering+-+Product+quality+-+Part+1%3A+Quality+model%2E+2001%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>ISO ISO/IEC TR 9126-2:2003 Software engineering &#x02013; Product quality &#x02013; Part 2: External metrics. 2003. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=ISO+ISO%2FIEC+TR+9126-2%3A2003+Software+engineering+-+Product+quality+-+Part+2%3A+External+metrics%2E+2003%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>ISO ISO/IEC TR 9126-3:2003 Software engineering &#x02013; Product quality &#x02013; Part 3: Internal metrics. 2003. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=ISO+ISO%2FIEC+TR+9126-3%3A2003+Software+engineering+-+Product+quality+-+Part+3%3A+Internal+metrics%2E+2003%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>ISO ISO/IEC TR 9126-4:2004 Software engineering &#x02013; Product quality &#x02013; Part 4: Quality in use metrics. 2004.</para></listitem>
<listitem>
<para>ISO ISO/TR 16982:2002 Ergonomics of human-system interaction &#x02013; Usability methods supporting human-centred design. 2002.</para></listitem>
<listitem>
<para>J&#x000E4;ms&#x000E4;-Jounela S. L. [et al.] (2013) &#x02018;Outline of a fault diagnosis system for a large-scale board machine,&#x02019; The International Journal of Advanced Manufacturing Technology. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=J%E4ms%E4-Jounela+S%2E+L%2E+%5Bet+al%2E%5D+%282013%29+%27Outline+of+a+fault+diagnosis+system+for+a+large-scale+board+machine%2C%27+The+International+Journal+of+Advanced+Manufacturing+Technology%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Kiseleva Julia (2013) &#x02018;Context mining and integration into predictive web analytics,&#x02019; Proceedings of the 22nd International Conference on World Wide Web., ACM, pp. 383&#x02013;388. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Kiseleva+Julia+%282013%29+%27Context+mining+and+integration+into+predictive+web+analytics%2C%27+Proceedings+of+the+22nd+International+Conference+on+World+Wide+Web%2E%2C+ACM%2C+pp%2E+383-388%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Kobsa A. (2011) &#x02018;Generic user modeling systems,&#x02019; User modeling and user-adapted interaction, pp. 49&#x02013;63. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Kobsa+A%2E+%282011%29+%27Generic+user+modeling+systems%2C%27+User+modeling+and+user-adapted+interaction%2C+pp%2E+49-63%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Koutkias V. G., Meletiadis S. L., and Maglaveras N. (2001) &#x02018;WAP-based personalized healthcare systems&#x02019; Health Informatics Journal, pp. 183&#x02013;189. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Koutkias+V%2E+G%2E%2C+Meletiadis+S%2E+L%2E%2C+and+Maglaveras+N%2E+%282001%29+%27WAP-based+personalized+healthcare+systems%27+Health+Informatics+Journal%2C+pp%2E+183-189%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Krueger M. [et al.] (2014) &#x02018;A data-driven maintenance support system for wind energy conversion systems, IFAC Proceedings Volumes, pp. 11470&#x02013;11475. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Krueger+M%2E+%5Bet+al%2E%5D+%282014%29+%27A+data-driven+maintenance+support+system+for+wind+energy+conversion+systems%2C+IFAC+Proceedings+Volumes%2C+pp%2E+11470-11475%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Langley Pat (1997) &#x02018;Machine learning for adaptive user interfaces,&#x02019; Annual Conference on Artificial Intelligence, Springer, pp. 53&#x02013;62. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Langley+Pat+%281997%29+%27Machine+learning+for+adaptive+user+interfaces%2C%27+Annual+Conference+on+Artificial+Intelligence%2C+Springer%2C+pp%2E+53-62%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Lathia N. [et al.] (2012) &#x02018;Personalizing mobile travel information services,&#x02019; Procedia-Social and Behavioral Sciences, pp. 1195&#x02013;1204. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Lathia+N%2E+%5Bet+al%2E%5D+%282012%29+%27Personalizing+mobile+travel+information+services%2C%27+Procedia-Social+and+Behavioral+Sciences%2C+pp%2E+1195-1204%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Lee E. A. (2008) &#x02018;Cyber physical systems: Design challenges,&#x02019; In 2008 11th IEEE International Symposium on Object and Component-Oriented Real-Time Distributed Computing (ISORC), IEEE, pp. 363&#x02013;369. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Lee+E%2E+A%2E+%282008%29+%27Cyber+physical+systems%3A+Design+challenges%2C%27+In+2008+11th+IEEE+International+Symposium+on+Object+and+Component-Oriented+Real-Time+Distributed+Computing+%28ISORC%29%2C+IEEE%2C+pp%2E+363-369%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Ma T. [et al.] (2005) &#x02018;Context-aware implementation based on CBR for smart home,&#x02019; International Conference on Wireless And Mobile Computing, Networking And Communications, IEEE. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ma+T%2E+%5Bet+al%2E%5D+%282005%29+%27Context-aware+implementation+based+on+CBR+for+smart+home%2C%27+International+Conference+on+Wireless+And+Mobile+Computing%2C+Networking+And+Communications%2C+IEEE%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>MANTIS Consortium D1.5 MANTIS Platform Requirements V2. 2016. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=MANTIS+Consortium+D1%2E5+MANTIS+Platform+Requirements+V2%2E+2016%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>MANTIS Consortium D5.1 List and Description of Different Maintenance Scenarios. 2016. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=MANTIS+Consortium+D5%2E1+List+and+Description+of+Different+Maintenance+Scenarios%2E+2016%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>MANTIS Consortium MANTIS - Cyber Physical System based Proactive Collaborative Maintenance Full Project Proposal (FPP). 2015.</para></listitem>
<listitem>
<para>Marcotte Ethan On Being &#x0201C;Responsive&#x0201D; [Online]. 14 09 2010: http://unstoppablerobotninja.com/entry/on-being-responsive.</para></listitem>
<listitem>
<para>Market Share Statistics for Internet Technologies [Online]. 2016. 17 08 2016. https://www.netmarketshare.com/operating-system-market-share.aspx?qprid=8&#x00026;qpcustomd=1&#x00026;qptimeframe=Y. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Market+Share+Statistics+for+Internet+Technologies+%5BOnline%5D%2E+2016%2E+17+08+2016%2E+https%3A%2F%2Fwww%2Enetmarketshare%2Ecom%2Foperating-system-market-share%2Easpx%B4qprid%3D8%26qpcustomd%3D1%26qptimeframe%3DY%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Maybury Mark T. and Wahlster Wolfgang (1998) Readings in Intelligent User Interfaces, Morgan Kaufmann. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Maybury+Mark+T%2E+and+Wahlster+Wolfgang+%281998%29+Readings+in+Intelligent+User+Interfaces%2C+Morgan+Kaufmann%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Miller Michael (2008) Cloud computing: Web-based Applications that Change the Way you Work and Collaborate Online, Que publishing. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Miller+Michael+%282008%29+Cloud+computing%3A+Web-based+Applications+that+Change+the+Way+you+Work+and+Collaborate+Online%2C+Que+publishing%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Nagamatsu T. [et al.] (2003) &#x02018;Information support for annual maintenance with wearable device,&#x02019; In Proceedings of HCI International, pp. 1253&#x02013;1257. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Nagamatsu+T%2E+%5Bet+al%2E%5D+%282003%29+%27Information+support+for+annual+maintenance+with+wearable+device%2C%27+In+Proceedings+of+HCI+International%2C+pp%2E+1253-1257%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Nguyen Phong [et al.] (2017) A Visual Analytics Approach for User Behaviour Understanding through Action Sequence Analysis. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Nguyen+Phong+%5Bet+al%2E%5D+%282017%29+A+Visual+Analytics+Approach+for+User+Behaviour+Understanding+through+Action+Sequence+Analysis%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Nicolai T. [et al.] (2006) &#x02018;Wearable computing for aircraft maintenance: Simplifying the user interface, In Applied Wearable Computing (IFAWC) 3rd International Forum, VDE, pp. 1&#x02013;12. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Nicolai+T%2E+%5Bet+al%2E%5D+%282006%29+%27Wearable+computing+for+aircraft+maintenance%3A+Simplifying+the+user+interface%2C+In+Applied+Wearable+Computing+%28IFAWC%29+3rd+International+Forum%2C+VDE%2C+pp%2E+1-12%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Nicolai T., Sindt, T., Kenn, H., &#x00026; Witt, H. (2005) &#x02018;Case study of wearable computing for aircraft maintenance,&#x02019; In IFAWC-International Forum on Applied Wearable Computing, VDE VERLAG GmbH.</para></listitem>
<listitem>
<para>Niggemann O. [et al.] (2015) &#x02018;Data-driven monitoring of cyber-physical systems leveraging on big data and the internet-of-things for diagnosis and control, 26th International Workshop on Principles of Diagnosis, Paris. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Niggemann+O%2E+%5Bet+al%2E%5D+%282015%29+%27Data-driven+monitoring+of+cyber-physical+systems+leveraging+on+big+data+and+the+internet-of-things+for+diagnosis+and+control%2C+26th+International+Workshop+on+Principles+of+Diagnosis%2C+Paris%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Norcio A. F. and Stanley J. (1989) &#x02018;Adaptive human-computer interfaces: A literature survey and perspective,&#x02019; IEEE Transactions on Systems, Man, and Cybernetics, pp. 399&#x02013;408. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Norcio+A%2E+F%2E+and+Stanley+J%2E+%281989%29+%27Adaptive+human-computer+interfaces%3A+A+literature+survey+and+perspective%2C%27+IEEE+Transactions+on+Systems%2C+Man%2C+and+Cybernetics%2C+pp%2E+399-408%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Perera C., Zaslavsky, A., Christen P. and Georgakopoulos D. (2014) &#x02018;Context aware computing for the internet of things: A survey,&#x02019; IEEE Communications Surveys &#x00026; Tutorials, pp. 414&#x02013;454. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Perera+C%2E%2C+Zaslavsky%2C+A%2E%2C+Christen+P%2E+and+Georgakopoulos+D%2E+%282014%29+%27Context+aware+computing+for+the+internet+of+things%3A+A+survey%2C%27+IEEE+Communications+Surveys+%26+Tutorials%2C+pp%2E+414-454%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Poklukar, &#x00160;pela, Papa, G., and Novak, F. (2018) &#x02018;A formal framework of human machine interaction in proactive maintenance &#x02013; MANTIS experience,&#x02019; Automatika, 58(4), pp. 450&#x02013;459. DOI: 10.1080/00051144.2018.1465226 <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Poklukar%2C+&#x0160;pela%2C+Papa%2C+G%2E%2C+and+Novak%2C+F%2E+%282018%29+%27A+formal+framework+of+human+machine+interaction+in+proactive+maintenance+-+MANTIS+experience%2C%27+Automatika%2C+58%284%29%2C+pp%2E+450-459%2E+DOI%3A+10%2E1080%2F00051144%2E2018%2E1465226" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>POWER-OM, Lulea Technology Univ. Product Reliability Evaluation Tools(Power consumption driven Reliability, Operation and Maintenance optimisation) Internal deliverable report D7.3 [Report]. 2015.</para></listitem>
<listitem>
<para>Rubin Jeffrey and Chisnell Dana (2008) Handbook of Usability Testing, Willey Publishing Inc. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Rubin+Jeffrey+and+Chisnell+Dana+%282008%29+Handbook+of+Usability+Testing%2C+Willey+Publishing+Inc%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Schilit B., Adams N., and Want R. (1994) &#x02018;Context-aware computing applications,&#x02019; In Mobile Computing Systems and Applications, WMCSA 1994. First Workshop on IEEE, pp. 85&#x02013;90. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Schilit+B%2E%2C+Adams+N%2E%2C+and+Want+R%2E+%281994%29+%27Context-aware+computing+applications%2C%27+In+Mobile+Computing+Systems+and+Applications%2C+WMCSA+1994%2E+First+Workshop+on+IEEE%2C+pp%2E+85-90%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Soh Harold [et al.] (2017) &#x02018;Deep sequential recommendation for personalized adaptive user interfaces,&#x02019; Proceedings of the 22nd International Conference on Intelligent User Interfaces, pp. 589&#x02013;593. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Soh+Harold+%5Bet+al%2E%5D+%282017%29+%27Deep+sequential+recommendation+for+personalized+adaptive+user+interfaces%2C%27+Proceedings+of+the+22nd+International+Conference+on+Intelligent+User+Interfaces%2C+pp%2E+589-593%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Stiefmeier T. [et al.] (2008) &#x02018;Wearable activity tracking in car manufacturing,&#x02019; IEEE Pervasive Computing, pp. 42&#x02013;50. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Stiefmeier+T%2E+%5Bet+al%2E%5D+%282008%29+%27Wearable+activity+tracking+in+car+manufacturing%2C%27+IEEE+Pervasive+Computing%2C+pp%2E+42-50%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Swanson D. C., Spencer J. M., and Arzoumanian S. H. (2000) &#x02018;Prognostic modelling of crack growth in a tensioned steel band,&#x02019; Mechanical systems and signal processing, pp. 789&#x02013;803. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Swanson+D%2E+C%2E%2C+Spencer+J%2E+M%2E%2C+and+Arzoumanian+S%2E+H%2E+%282000%29+%27Prognostic+modelling+of+crack+growth+in+a+tensioned+steel+band%2C%27+Mechanical+systems+and+signal+processing%2C+pp%2E+789-803%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Techopedia Techopedia [Online]. 2011. 10 January 2017. https://www.techopedia.com/definition/12829/human-machine-interface-hmi.</para></listitem>
<listitem>
<para>Thaduri A., Kumar U., and Verma A. K. (2014) &#x02018;Computational intelligence framework for context-aware decision making,&#x02019; International Journal of System Assurance Engineering and Management, pp. 1&#x02013;12. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Thaduri+A%2E%2C+Kumar+U%2E%2C+and+Verma+A%2E+K%2E+%282014%29+%27Computational+intelligence+framework+for+context-aware+decision+making%2C%27+International+Journal+of+System+Assurance+Engineering+and+Management%2C+pp%2E+1-12%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Usability Net Usability Net [Online] // Usability Net. 10 April 2017. http://www.usabilitynet.org/home.htm.</para></listitem>
<listitem>
<para>Usability.gov Usability.gov [Online]. 10 April 2017. https://www.us ability.gov/.</para></listitem>
<listitem>
<para>Wigdor Daniel and Wixon Dennis (2011) Brave NUI World: Designing Natural User Interfaces for Touch and Gesture, Elsevier. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Wigdor+Daniel+and+Wixon+Dennis+%282011%29+Brave+NUI+World%3A+Designing+Natural+User+Interfaces+for+Touch+and+Gesture%2C+Elsevier%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Wikipedia User Interface [Online]. 2017. https://en.wikipedia.org/wiki/User_interface.</para></listitem>
<listitem>
<para>Wrona K. and Gomez L. (2006) &#x02018;Context-aware security and secure context-awareness in ubiquitous computing environments,&#x02019; In Annales UMCS, Informatica, pp. 332&#x02013;348. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Wrona+K%2E+and+Gomez+L%2E+%282006%29+%27Context-aware+security+and+secure+context-awareness+in+ubiquitous+computing+environments%2C%27+In+Annales+UMCS%2C+Informatica%2C+pp%2E+332-348%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Zhang D., Yu Z., and Chin C. Y. (2005) &#x02018;Context-aware infrastructure for personalized healthcare,&#x02019; Studies in health technology and informatics, pp. 154&#x02013;163. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Zhang+D%2E%2C+Yu+Z%2E%2C+and+Chin+C%2E+Y%2E+%282005%29+%27Context-aware+infrastructure+for+personalized+healthcare%2C%27+Studies+in+health+technology+and+informatics%2C+pp%2E+154-163%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Zhang Z., He X., and Kusiak A. (2015) &#x02018;Data-driven minimization of pump operating and maintenance cost,&#x02019; Engineering Applications of Artificial Intelligence, pp. 37&#x02013;46. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Zhang+Z%2E%2C+He+X%2E%2C+and+Kusiak+A%2E+%282015%29+%27Data-driven+minimization+of+pump+operating+and+maintenance+cost%2C%27+Engineering+Applications+of+Artificial+Intelligence%2C+pp%2E+37-46%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Zouhaier L., Hlaoui Y. B., and Ayed L. J. B. (2013) &#x02018;Building adaptive accessible context-aware for user interface tailored to disable users,&#x02019; In 2013 IEEE 37th Annual Computer Software and Applications Conference Workshops. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Zouhaier+L%2E%2C+Hlaoui+Y%2E+B%2E%2C+and+Ayed+L%2E+J%2E+B%2E+%282013%29+%27Building+adaptive+accessible+context-aware+for+user+interface+tailored+to+disable+users%2C%27+In+2013+IEEE+37th+Annual+Computer+Software+and+Applications+Conference+Workshops%2E" target="_blank">Google Scholar</ulink></para></listitem></orderedlist>
</section>
</chapter>
<chapter class="chapter" id="ch07" label="7" xreflabel="7">
<title>Success Stories on Real Pilots</title>
<para><emphasis role="strong">Rafael Socorro<superscript>1</superscript>, Mar&#x000ED;a Aguirregabiria<superscript>2</superscript>, Alp Ak&#x000E7;ay<superscript>3</superscript>, Michele Albano<superscript>4</superscript>, Mikel Anasagasti<superscript>5</superscript>, Andoitz Aranburu<superscript>6</superscript>, Mauro Barbieri<superscript>7</superscript>, Iban Barrutia<superscript>8</superscript>, Ansgar Bergmann<superscript>9</superscript>, Karel De Brabandere<superscript>10</superscript>, Marcel Boosten<superscript>11</superscript>, Rui Casais<superscript>12</superscript>, David Chico<superscript>6</superscript>, Paolo Ciancarini<superscript>13</superscript>, Paulien Dam<superscript>14</superscript>, Giovanni DiOrio<superscript>15</superscript>, Karel Eerland<superscript>11</superscript>, Xabier Eguiluz<superscript>2</superscript>, Salvatore Esposito<superscript>16</superscript>, Catarina F&#x000E9;lix<superscript>17</superscript>, Javier Fernandez-Anakabe<superscript>8</superscript>, Hugo Ferreira<superscript>17</superscript>, Luis Lino Ferreira<superscript>4</superscript>, Attila Frank&#x000F3;<superscript>18</superscript>, Iosu Gabilondo<superscript>2</superscript>, Raquel Garc&#x000ED;a<superscript>1</superscript>, Jeroen Gijsbers<superscript>11</superscript>, Mathias Gr&#x000E4;dler<superscript>19</superscript>, Csaba Heged&#x00171;s<superscript>18</superscript>, Silvia Hern&#x000E1;ndez<superscript>1</superscript>, Petri Helo<superscript>19</superscript>, Mike Holenderski<superscript>3</superscript>, Erkki Jantunen<superscript>20</superscript>, Matti Kaija<superscript>21</superscript>, Ale&#x00161; Kancilija<superscript>22</superscript>, F&#x000E9;lix Larrinaga Barrenechea<superscript>8</superscript>, Pedro Mal&#x000F3;<superscript>15</superscript>, Goreti Marreiros<superscript>4</superscript>, Eva Mart&#x000ED;nez<superscript>1</superscript>, Diogo Martinho<superscript>4</superscript>, Asif Mohammed<superscript>12</superscript>, Mikel Mondragon<superscript>5</superscript>, Istv&#x000E1;n Moldov&#x000E1;n<superscript>23</superscript>, Antti Niemel&#x000E4;<superscript>24</superscript>, Jon Olaizola<superscript>8</superscript>, Gregor Papa<superscript>25</superscript>, &#x00160;pela Poklukar<superscript>25</superscript>, Isabel Pra&#x000E7;a<superscript>4</superscript>, Stefano Primi<superscript>1</superscript>, Verus Pronk<superscript>7</superscript>, Ville Rauhala<superscript>24</superscript>, Mario Riccardi<superscript>16</superscript>, Rafael Rocha<superscript>4</superscript>, Jon Rodriguez<superscript>26</superscript>, Ricardo Romero<superscript>27</superscript>, Antonio Ruggieri<superscript>16</superscript>, Oier Sarasua<superscript>6</superscript>, Eduardo Saiz<superscript>2</superscript>, Veli-Pekka Salo<superscript>19</superscript>, M&#x000F3;nica S&#x000E1;nchez<superscript>1</superscript>, Paolo Sannino<superscript>16</superscript>, Babacar Sarr<superscript>10</superscript>, Alberto Sillitti<superscript>28</superscript>, Carlos Soares<superscript>17</superscript>, Hans Sprong<superscript>11</superscript>, Daan Terwee<superscript>14</superscript>, Bas Tijsma<superscript>14</superscript>, Tom Tourw&#x000E9;<superscript>29</superscript>, Nayra Uranga<superscript>1</superscript>, Lauri V&#x000E4;limaa<superscript>27</superscript>,Juha Valtonen<superscript>24</superscript>, P&#x000E1;l Varga<superscript>18</superscript>, Alejandro Veiga<superscript>1</superscript>, Mikel Viguera<superscript>26</superscript>, Jaapvander Voet<superscript>11</superscript>, Godfried Webers<superscript>11</superscript>, Achim Woyte<superscript>10</superscript>, Kees Wouters<superscript>7</superscript>, Ekhi Zugasti<superscript>8</superscript> and Urko Zurutuza<superscript>8</superscript></emphasis></para>
<para><superscript>1</superscript>Acciona Construcci&#x000F3;n S.A., Spain<?lb?><superscript>2</superscript>IK4-Ikerlan, Arrasate-Mondrag&#x000F3;n, Spain<?lb?><superscript>3</superscript>Technische Universiteit Eindhoven, The Netherlands<?lb?><superscript>4</superscript>ISEP, Polytechnic Institute of Porto, Porto, Portugal<?lb?><superscript>5</superscript>Goizper, Spain<?lb?><superscript>6</superscript>FAGOR ARRASATE, Arrasate-Mondrag&#x000F3;n, Spain<?lb?><superscript>7</superscript>Philips Electronics Nederland B.V., The Netherlands<?lb?><superscript>8</superscript>Mondragon Unibertsitatea, Arrasate-Mondrag&#x000F3;n, Spain<?lb?><superscript>9</superscript>STILL GmbH, Germany<?lb?><superscript>10</superscript>3e, Belgium<?lb?><superscript>11</superscript>Philips Medical Systems Nederland B.V., The Netherlands<?lb?><superscript>12</superscript>ADIRA Metal-Forming Solutions SA, Portugal<?lb?><superscript>13</superscript>Consorzio Interuniversitario Nazionale per l&#x02019; Informatica, Italy <?lb?><superscript>14</superscript>Philips Consumer Lifestyle B.V., The Netherlands<?lb?><superscript>15</superscript>FCT-UNL, UNINOVA-CTS, Caparica, Portugal<?lb?><superscript>16</superscript>Ansaldo STS, Italy<?lb?><superscript>17</superscript>Instituto de Engenharia de Sistemas e Computadores do Porto, Portugal<?lb?><superscript>18</superscript>AITIA International Inc., Hungary <?lb?><superscript>19</superscript>Wapice, Finland<?lb?><superscript>20</superscript>VTT Technical Research Centre of Finland Ltd, Finland<?lb?><superscript>21</superscript>Fortum Power and Heat Oy, Finland<?lb?><superscript>22</superscript>XLAB, Ljubljana, Slovenia<?lb?><superscript>23</superscript>Budapest University of Technology and Economics, Hungary<?lb?><superscript>24</superscript>Lapland University of Applied Sciences Ltd, Finland<?lb?><superscript>25</superscript>Jo&#x0017E;ef Stefan Institute, Slovenia<?lb?><superscript>26</superscript>KONIKER, Arrasate-Mondrag&#x000F3;n, Spain<?lb?><superscript>27</superscript>Tekniker, Spain<?lb?><superscript>28</superscript>Innopolis University, Russian Federation<?lb?><superscript>29</superscript>Sirris, Belgium</para>
<para>This chapter describes success stories. The MANTIS architecture (<link linkend="ch03">Chapter <xref linkend="ch03" remap="3"/></link>) was implemented for a number of use cases on real pilots, and the techniques described in Chapters 4, 5, and 6 were experimented with in real settings. Results on the techniques were already presented in previous chapters. This chapter, on the other hand, describes the pilots, from their objectives and context to the system integration efforts to the attained results. One of the results of the application of the techniques was the enhanced Technology Readiness Level of the techniques, which is summarized in <link linkend="F7-1">Figure <xref linkend="F7-1" remap="7.1"/></link>.</para>
<fig id="F7-1" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-1">Figure <xref linkend="F7-1" remap="7.1"/></link></label>
<caption><para>Maturity Level before and after the MANTIS project for the techniques in the use cases.</para></caption>
<graphic xlink:href="graphics/ch07_fig001.jpg"/>
</fig>
<para>Each section takes care of providing details for a different use case, and as a whole the chapter proves the large breadth of the applicability of the MANTIS approach.</para>
<section class="lev1" id="sec7-1">
<title>7.1 Shaver Production Plant</title>
<blockquote role="flushleft">
<para><emphasis>Contributors: Bas Tijsma, Paulien Dam and Daan Terwee</emphasis></para></blockquote>
<para>The goal of the shaver production plant use case is to increase the predictability of the maintenance actions through smart use of data. By actively utilizing various data sources in an automated manner, it is expected to reduce the cost of the tooling, and reduce the amount of unexpected downtime, thereby reducing the total cost of production. This section provides an overview of the practical application of the several elements developed and applied to this use case. The rationale of the approach is represented in <link linkend="F7-2">Figure <xref linkend="F7-2" remap="7.2"/></link>.</para>
<fig id="F7-2" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-2">Figure <xref linkend="F7-2" remap="7.2"/></link></label>
<caption><para>Goals and vision for the shaver manufacturing site.</para></caption>
<graphic xlink:href="graphics/ch07_fig002.jpg"/>
</fig>
<section class="lev2" id="sec7-1-1">
<title>7.1.1 Introduction to the Shaver Manufacturing Plant</title>
<para>The shaver manufacturing plant is one of the largest manufacturing plants of Philips. For mass production of certain parts of the shaver an advanced machining technique is used as manufacturing technology. The tools used in this manufacturing process are the focus area for this use case.</para>
<para>Large amounts of data are gathered in the manufacturing plant about the products and processes. These data is mostly used for manual, after-the-fact analysis of process disruptions, machine failures and quality issues.</para>
<para>It is expected that these data (and where necessary additional data) can be used to make predictions about product quality, process disruptions and impending maintenance actions. By actively utilizing the data in an automated manner, it is expected to reduce the cost of our tooling and maintenance. The objective is to demonstrate that impending failures can be accurately predicted by mining large amounts of data from heterogeneous databases, such that tooling maintenance can be timely scheduled to prevent unexpected downtime of the production lines and maximize tooling lifetime.</para>
</section>
<section class="lev2" id="sec7-1-2">
<title>7.1.2 Scope and Logic</title>
<para>For the complete project, a full project scope was made. This chapter focusses on a subset shown in <link linkend="F7-3">Figure <xref linkend="F7-3" remap="7.3"/></link>, and is largely based on practical experience and domain expertise. The main focus, the prediction of tool failure, is put in the center of the picture.</para>
<para>There are three main influencing factors regarding the life-time of a tool:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Process behavior</para>
<para>The combined behavior of the process (measured by many sensors, see Section 7.1.3) during the discrete manufacturing processes, as well as process error behavior over time. Both the process sensor measures behaviors, as well as process errors which may cause damage to the tool and influence the state of the tool;</para></listitem>
<listitem>
<para>Quality Status</para>
<para>It entails the geometrical measurements of the products made by the aforementioned process, which need to comply with product specifications. Many quality deviations on the products can be related to a change (damage, wear) in the tooling geometry, which might indicate an (upcoming) failure;</para></listitem>
<listitem>
<para>Tooling status</para>
<para>The current status of the tools with respect to wear, small damage, etc., over time. These cannot be measured by the process sensors and the quality status, and usually imply measurements performed over longer time periods.</para></listitem>
</itemizedlist>
<para>The main goal is to use these three data sources in a combined model to predict tool failures.</para>
<para>With the ability to model tool failure behavior, the output can be used as an input to optimize Philips&#x02019; current maintenance policy and strategy. For example, the amount of spare stock can be regulated much better, if the future failures of tools can be predicted.</para>
<fig id="F7-3" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-3">Figure <xref linkend="F7-3" remap="7.3"/></link></label>
<caption><para>Basic interactions.</para></caption>
<graphic xlink:href="graphics/ch07_fig003.jpg"/>
</fig>
</section>
<section class="lev2" id="sec7-1-3">
<title>7.1.3 Data Platform and Sensors</title>
<para>In this use case, the existing proprietary platform of the manufacturing site is used as much as possible, to be able to focus more on the analysis and application part of the project. Most of the manufacturing machines are connected to a legacy data platform, known as the Factory Information System (FIS). It consists of various relational databases, to which the machines are connected by custom developed drivers. This was custom built over the course of several decades by the internal IT department.</para>
<para>Despite to recent developments in the so-called industry 4.0 revolution, where much progress has been made into generic data exchange protocols (e.g., OPC-UA) and storage architectures (e.g., Mimosa), the equipment in scope is legacy and cannot be changed easily. During the course of the project, it became clear that performance of the existing platform was sufficient. However, in the future, increasing amounts of data requests might hamper overall performance.</para>
<para>The main data source in the platform is the process data. The machines in scope are equipped with a wide variety of sensors. The output of the sensors is collected by a machine controller to perform several pre-processing steps, like filtering and aggregation, before the data is sent to the FIS platform. In general, a set of data is collected and sent once per production cycle (in one cycle one product is made). Each cycle contains over 100 parameters. The data is externally accessible via the FIS platform.</para>
<para>Tooling information is also automatically stored in a separate database. A digital log is kept on the lifecycle of each individual tool, for example on which machines it was placed, the amount of products the tool has made and maintenance actions taken by the tooling maintenance department.</para>
<para>The last data source is the product quality metrics. Quality data is gathered by taking o&#x0FB04;ine product samples on dedicated measuring devices. These measurements are inputs for the quality system, which is also part of the FIS platform, meaning they can also be accessed externally. All measurements are geometrical, like form accuracy and thickness. Usually these data are aggregated values of a larger set of measurement points, like average, standard deviations, etc.</para>
</section>
<section class="lev2" id="sec7-1-4">
<title>7.1.4 Data Analytics and Maintenance Optimization</title>
<para>The manufacturing process consists of several physical elements. Electrical, chemical and mechanical elements are working together in order to manu- facture the products, making it a highly complex process where interactions between different signals can be easily overlooked when just monitoring every signal individually. A prediction model (soft sensor) that combines all different signals and processing them together gives better insight in these interaction effects via computational intelligence. This sensor fusion deals with disparate sources that do not have to originate from identical sensors.</para>
<section class="lev3" id="sec7-1-4-1">
<title>7.1.4.1 Physical models and background</title>
<para>Before being able to successfully analyze process and manufacturing data, domain knowledge is required, which can be provided by process engineers. Without domain knowledge, it is very hard to understand the data, do the analysis and validate the results.</para>
<para>The machining process in scope is electrochemical machining (ECM). This is an unconventional electrochemical manufacturing process, but it is well established in niche applications like turbine blades and medical implants [McGeough, 1988; De Barr and Oliver, 1975]. This process removes material at the anode (work piece) using current controlled electrochemical process. By feeding a shaped cathode (tool or electrode) towards the work piece, the reverse shape of the tool is copied to the work piece.</para>
<para>It is a complicated process incorporating a number of physical phenomena interacting with each other. Common problems with this process are related to the variations of the material composition, variations in chemical conditions, as well as the influence of the geometry of the tools. All these effects may, in some form or another, change or damage tooling geometry, which in effect will lead to quality issues.</para>
</section>
<section class="lev3" id="sec7-1-4-2">
<title>7.1.4.2 Process monitoring with Principal Component Analysis &#x00026; Hotelling&#x02019;s T<superscript><emphasis role="strong">2</emphasis></superscript></title>
<para>The Principal Component Analysis algorithm in combination with the Hotelling&#x02019;s T<superscript>2</superscript> score is used to get insight in the interaction effect of all different process parameters. To train the model, the data is extracted from the FIS platform and analyzed to make sure that the historic dataset consists of data that indicates only normal process behavior, with no deviations or outliers. This is an important step, since this data will serve as a reference for future predictions.</para>
<para>The PCA algorithm is trained on the historical data where the dimensionality of the entire dataset is reduced, while retaining as much as possible of the variation present in the data set. This is done by transforming the data to a new set of variables called the Principle Components (PCs), which are, by definition, uncorrelated. By definition the PCs are ordered in such a way that the first few PCs contain most of the variation present in the original variables (see <link linkend="F7-4">Figure <xref linkend="F7-4" remap="7.4"/></link>). In this example, the red line indicates that 5 PCs explain more than 90% of the original variance thus reducing the dimensionality from 15 to 5 parameters. The PCA model transforms every observation of the dataset into a set of scores of the same size as the number of PCs.</para>
<fig id="F7-4" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-4">Figure <xref linkend="F7-4" remap="7.4"/></link></label>
<caption><para>Cumulative explained variance of the principal components.</para></caption>
<graphic xlink:href="graphics/ch07_fig004.jpg"/>
</fig>
<para>For real-time calculations the trained PCA model is deployed on a server and data from the PLC-PMAC system is fed into this model in order to obtain the new weighted scores that indicate how close new observations are related to the historic dataset. Because the PCA algorithm is a &#x02018;white box&#x02019; algorithm, it has self-diagnostics abilities. The model can be used to determine the root causes for fluctuations, trends and outliers.</para>
<para>The resulting PCA model allows for real time streaming data to be transformed onto the principal component space defined during the training phase. Assuming the training dataset is a good representative of normal machine behavior, any significant deviations seen with the live streaming data can be interpreted as signs of a defective machine behavior.</para>
<para>The Hotelling&#x02019;s T<superscript>2</superscript> value can be used as a measure for how close the transformed live streaming data are to the training set. It is reasonable to start from the simpler univariate case t-test that is defined as follows:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq7-1.jpg"/></para>
<para>where the null hypothesis H<subscript>0</subscript> (the historic dataset) provides a mean &#x003BC; = &#x003BC;<subscript>0</subscript> and a sample of n can be acquired from the population with mean <emphasis><emphasis role="overline">x</emphasis></emphasis> and standard deviation s. The statistical interpretation is as follows: on average the difference between the sample and null hypothesis will fall within s if normal statistical variation can explain these differences. The overall t-score is weighted by <inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg318-1.jpg"/> since any differences in the numerator become more significant the larger the sample size becomes.</para>
<para>To generalize this result, it is possible to square the expression for the t-test to obtain:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq7-2.jpg"/></para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq7-3.jpg"/></para>
<para>This is the same as an F-distributed random variable with 1 and n-1 degrees of freedom. We can replace the difference between the sample mean <emphasis role="overline"><emphasis>x</emphasis></emphasis> and &#x003BC;<subscript>0</subscript> with the difference between the sample mean vector X <superscript>&#x021C0;</superscript> and the hypothesized mean vector &#x003BC;<subscript>o</subscript> <superscript>&#x021C0;</superscript>. The inverse of the sample variance is replaced be the inverse of the sample variance-covariance matrix S:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq7-4.jpg"/></para>
<para>Using the above expression for <emphasis>T<superscript>2</superscript></emphasis> it is possible to compress a multivariate dataset into one scalar metric that can quantify the status of a given manufacturing machine. The use of a single parameter also facilitates machine status visualizations with the monitoring of a single quantity, and the calculated <emphasis>T<superscript>2</superscript></emphasis> value can be compared with predetermined confidence limits to trigger warnings or alarms, should human interaction be needed to return the given machine to a nominal status (see <link linkend="F7-5">Figure <xref linkend="F7-5" remap="7.5"/></link>).</para>
<fig id="F7-5" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-5">Figure <xref linkend="F7-5" remap="7.5"/></link></label>
<caption><para>Control chart showing the Hotelling&#x02019;s T<superscript>2</superscript> score and tolerances to identify trends and outliers in the multivariate data.</para></caption>
<graphic xlink:href="graphics/ch07_fig005.jpg"/>
</fig>
</section>
<section class="lev3" id="sec7-1-4-3">
<title>7.1.4.3 Product quality prediction with partial least squares regression</title>
<para>As described in Section 7.1.4.1, the product geometry is a negative copy of the tool. Therefore, the product geometry is a suitable metric for detecting deviations in the tool geometry such as damages or wear. In production, the product geometry is measured on a sample basis, once every 4 hours. This is not frequent enough for monitoring potential tooling issues. However, by relating the process data as described in the previous section with the product quality data, a predictive model can be trained to estimate the quality of every product made.</para>
<para>For this predictive model, the process data, enriched with a few extra variables of the tooling is used as predictors (X). This process data has already been pre-processed, and consists of an observation for every product made. Each observation, in turn, consists of a collection of data points coming from a variety of different sensors in the production machine. After careful research and bootstrap modelling, 12 predictor variables were selected as input for the modelling algorithm.</para>
<para>As response data (Y), product geometry data is used, which consists of four Y-variables, which are sub selection of all the quality indicators for products made by the manufacturing process. Both data sets have a common product identifier which makes a good join between the two possible data sets. Since product quality data is only measured on a sample basis, the time interval between observations differs greatly from the process data. For this reason, the data set is filtered, where only those observations that consist of both process and quality data are kept.</para>
<para>The last step in the data preparation is outlier removal. The outliers (see <link linkend="F7-5">Figure <xref linkend="F7-5" remap="7.5"/></link>) due to short term sensor failures or miscalculations during ETL are removed from the dataset. Outliers due to physical events in the process itself remain in the dataset, since they can hold valuable information for modelling the relationships between parameters. Finally, the data is ready to serve as an input for model training.</para>
<para>When the number of predictors (X) is too large (10) the more traditional regression methods, such as linear regression, are not adequately effective. Furthermore, in many cases manufacturing and/or sensory data have a correlated nature. This causes the sample covariance matrix to be ill-conditioned, because it becomes almost singular, which is a problem for the more traditional regression methods. This can be solved by using linear projection methods such as Partial Least Squares Regression [Wold, 1975].</para>
<para>Mathematically, there are quite some advantages for PLS compared to traditional regression algorithms. Because PLS is a linear projection method, it decomposes the covariance matrix that settles the singularity problem. This gives PLS the ability to handle multicollinearity among the predictors (X). Furthermore, these linear projection methods have the advantage that they can handle missing data points in the data set. For example, if a particular sensor has a short term failure or certain data points are removed from the data set during outlier analysis, then the whole observation does not have to be discarded, but can still serve as input to the algorithm. Moreover, PLS has the extra advantage that it can incorporate multiple Y-variables (or responses) in one statistical model. Finally, PLS is suitable for modelling and monitoring larger number of variables simultaneously.</para>
<para>The trained PLS model gives promising results. In <link linkend="F7-6">Figure <xref linkend="F7-6" remap="7.6"/></link>, the results of the PLS predictions are plotted against the actual (observed) results of a specific product quality parameter. We see that there is only a small deviation between the two trend lines; the Root Mean Square Error (RMSE) is 1.86, and the R<superscript>2</superscript> is 71.2%, which are good results given the acceptable range of the quality parameter and the complexity of the production process itself.</para>
<fig id="F7-6" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-6">Figure <xref linkend="F7-6" remap="7.6"/></link></label>
<caption><para>Graph of the predicted and the observed quality with PLS regression.</para></caption>
<graphic xlink:href="graphics/ch07_fig006.jpg"/>
</fig>
</section>
<section class="lev3" id="sec7-1-4-4">
<title>7.1.4.4 Computational trust</title>
<para>Every production line has incidents. These incidents can be related to, for example, machine failures or process errors. Due to the complexity of the manufacturing process, quite many process errors are generated in time. Some of these errors are critical, as they may cause additional damage. Other errors have minor impact. The impact level and the frequency of occurrence are both important factors in calculating the current state of a particular process, but more advanced interactions are only possible, in which a particular order or process errors can be critical.</para>
<para>Therefore, the concept of computational trust is researched, in order to quantify the current &#x02018;trust&#x02019; in the machine being in a &#x02018;good&#x02019; state, or in a &#x02018;bad&#x02019; state. This allows for errors to be specified (impact-level, fall-off level) and to be combined in a specific &#x02018;trust in good machine&#x02019;-metric, as input for the overall tool failure predictions. For example, refer to the particular case presented in <link linkend="F7-7">Figure <xref linkend="F7-7" remap="7.7"/></link>, which reports generated error codes over time. Values under the threshold of 0.6 are considered as a &#x02018;bad&#x02019; state.</para>
<fig id="F7-7" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-7">Figure <xref linkend="F7-7" remap="7.7"/></link></label>
<caption><para>Error codes over time, and corresponding trust values.</para></caption>
<graphic xlink:href="graphics/ch07_fig007.jpg"/>
</fig>
</section>
</section>
<section class="lev2" id="sec7-1-5">
<title>7.1.5 Visualization and HMI</title>
<para>For visualization of the analytics toolset, a prototype interactive dashboard was developed and tested in production. It is inspired by a principle component analysis (PCA) score plot, but redesigned to be better usable in a production environment.</para>
<para>The goal is to give operators direct insight to the current status of the production process in terms of overall process stability, but also enable more detailed insight by allowing to drill down to the specific cause of potential deviations. This is a major advantage of the methods described in Sections 7.1.4.2 and 7.1.4.3, since it allows to relate aggregated scores (like Hotelling&#x02019;s T<superscript>2</superscript>) with individual sensor values.</para>
<fig id="F7-8" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-8">Figure <xref linkend="F7-8" remap="7.8"/></link></label>
<caption><para>Main screen for production operators, showing process performance for eight similar machines. The green circles indicate &#x02018;good&#x02019; behaviour.</para></caption>
<graphic xlink:href="graphics/ch07_fig008.jpg"/>
</fig>
<para>The main screen (<link linkend="F7-8">Figure <xref linkend="F7-8" remap="7.8"/></link>) for giving direct insight consists of eight plots, corresponding to eight machines performing the same process in parallel. The individual graphs consist of a scatterplot, where the first and second principal components are plotted. The green ellipse indicates the 95% confidence limit, in which the process can be considered stable. Each point represents an individual product manufactured on that specific machine (the yellow point is the latest product).</para>
<para>When the outcome of the process is too critical, the top bar will change to either orange or red, indicating potential problems. The operators are expected to respond to the alarm, and perform pre-specified actions.</para>
<fig id="F7-9" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-9">Figure <xref linkend="F7-9" remap="7.9"/></link></label>
<caption><para>Analysis screen of operator dashboard.</para></caption>
<graphic xlink:href="graphics/ch07_fig009.jpg"/>
</fig>
<para>By clicking on one of the eight machines, the dashboard will show more information about this particular machine (see <link linkend="F7-9">Figure <xref linkend="F7-9" remap="7.9"/></link>). This screen provides additional details, such as time-series graphs, to get a better understanding of the problem. The operator (or engineer) can select individual points, can show individual plots of sensor values and can colour the scores based on a selected sensor. These are all experimental tools to help find root-causes and solve production issues much faster.</para>
</section>
<section class="lev2" id="sec7-1-6">
<title>7.1.6 Maintenance and Inventory Optimization Results</title>
<para>Within the MANTIS project, research has been performed on both the prediction of machine process errors and failures of tools. So far, remaining useful life estimations for the tooling are still uncertain, but moderate results have been achieved with classification methods. Hence, it is investigated what the added value is by using imperfect predictions of tool failures to decide when to perform maintenance actions (e.g., tool inspections and replacements) and when to place orders for new tools. Therefore, it is assumed that a classification model (and its approximate performance) will be the predictive input to the designed policy.</para>
<para>Furthermore, an analysis is carried out into the amount of products that a tool can produce before being discarded. It turns out that the chance that a tool is replaced due to a defect decreases when a tool reaches a longer tool life (see <link linkend="F7-10">Figure <xref linkend="F7-10" remap="7.10"/></link>). In other words, tools that have a longer lifetime are more likely to be replaced due to the age of the tool rather than defects caused by the machine.</para>
<fig id="F7-10" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-10">Figure <xref linkend="F7-10" remap="7.10"/></link></label>
<caption><para>Time to failure (in amount of products) for the three main tool types.</para></caption>
<graphic xlink:href="graphics/ch07_fig010.jpg"/>
</fig>
<para>For the proposed policy, predictions of upcoming tool failures are generated for every predefined time period (8-hour shift) within a prediction horizon. With a multi-period prediction horizon, the predictions thus overlap. The aggregate of the overlapping predictions is compared to an &#x02018;inspection threshold&#x02019; to decide whether to inspect a tool in the upcoming shift. This threshold is optimized by explicitly modeling the imperfectness in predictions. The predictions are also added as a data-driven component to a (R, s, Q) inventory control policy. The expected amount of tools is added to <emphasis>not</emphasis> fail in the coming review period (based on imperfect predictions) to the conventional inventory position (IP). This altered IP is compared to the reorder level <emphasis>s</emphasis> at each review period <emphasis>R</emphasis> to decide on whether to order <emphasis>Q</emphasis> tools.</para>
<para>The predictions thus allow for <emphasis>postponing</emphasis> orders to the next review period if enough tools are still expected to be operational by then. Implementing this joint policy for the use case study will lead to an estimated annual savings that consists of inventory savings, maintenance savings and a decrease in tools ordered. There are also one-time savings in tooling purchasing costs from lowering the current average inventory on hand level to the optimal level.</para>
</section>
<section class="lev2" id="sec7-1-7">
<title>7.1.7 Conclusions</title>
<para>For the shaver production plant use case, quite a lot of effort has been spent on the analytics parts during the MANTIS project. This was mainly possible due to the already existing data acquisition platform, which provided a solid starting point for analytics. Some very promising results have been found for this use case.</para>
<para>It is proven that it is possible to use process data of the machines to make good estimations of product quality. As a result, insights into current quality performance has increased significantly, as well as a reduced reliability on (slow) o&#x0FB04;ine quality measurements of the products. On top of this, it can also be used as an input for estimation on tooling status, since there is a known relationship between the quality (shape) of the product and the shape of the tool.</para>
<para>Another result is that with some creative thinking, concepts from the academic world can be translated to real-world use cases. One such an example is the computational trust-modeling, which looks quite promising to quantify machine performance with respect to error behavior over time. Again, this quantification is important, as it can be used as an input to the tooling status model.</para>
<para>Image recognition techniques applied to the tools has proven to be difficult, especially on complex shapes and glossy surfaces. Several attempts have been made to use image recognition techniques to calculate tool wear, which can also be used as an input for the wear model. None of them, so far, have given any usable results.</para>
<para>Last, we also looked at the promises of predictive maintenance on tooling, while keeping in mind the future applications to other production tooling assets as well. One challenge is the predictive performance of models, which is hard to estimate when the real output of the models is yet unknown. Data driven assumptions and simulation modeling can still provide good insights in potential (financial) benefits. From a business point of view this is important, as businesses typically demand clear business cases in order to provide resources for further enhancements of these concepts.</para>
</section>
</section>
<section class="lev1" id="sec7-2">
<title>7.2 Deploying an User Friendly Monitoring System for Pultrusion Line Production</title>
<blockquote role="flushleft">
<para><emphasis>Contributors: Rafael Socorro, Raquel Garc&#x000ED;a, Nayra Uranga, Silvia Hern&#x000E1;ndez, M&#x000F3;nica S&#x000E1;nchez, Alejandro Veiga, Eva Mart&#x000ED;nez, Stefano Primi</emphasis></para></blockquote>
<para>This use case aims to design and develop a reliable monitoring system locating different sensors in key locations for gathering relevant data to improve the preventive maintenance for different processes included in the pultrusion process. Moreover, it allows to create an historical storage of all the data collected to identify patterns in the future, and contribute to better proactive maintenance.</para>
<section class="lev2" id="sec7-2-1">
<title>7.2.1 Introduction to the Pultrusion Use Case</title>
<para>ACCIONA operates one manufacturing plant (with two production lines) in Alcobendas (Madrid) for production of composite structures through a pultrusion process. This process has been widely used for manufacturing highly strengthened and continuous composite structures with low weight, elevated mechanical and chemical resistance, and electrical and thermal insulation. For example, they were used for the Pajares tunnels in Asturias (<link linkend="F7-11">Figure <xref linkend="F7-11" remap="7.11"/></link>), for Valencia Lighthouse (<link linkend="F7-12">Figure <xref linkend="F7-12" remap="7.12"/></link>), and for the pedestrian bridge in Madrid (<link linkend="F7-13">Figure <xref linkend="F7-13" remap="7.13"/></link>). The properties of the composite structures are the main reason why this method has become essential in the development of ACCIONA&#x02019;s highly differentiated construction projects.</para>
<para>This process is very challenging in terms of production and maintenance of the equipment involved as it is a continuous process and the machines are running 24 hours a day, so it is necessary to avoid production stops or unexpected delays.</para>
</section>
<section class="lev2" id="sec7-2-2">
<title>7.2.2 Scope and Logic</title>
<para>The production line, shown in <link linkend="F7-14">Figure <xref linkend="F7-14" remap="7.14"/></link>, is continuous, it stops only when the part model being produced changes. A new product to be produced entails a new configuration of the machine. The current maintenance policy for the pultrusion line is preventive (some tasks are performed every time unit periodically), reactive corrective (if a failure is detected) and opportunistic (if the line is stopped and the deadline for the maintenance task is close), which means that once the production line stops to perform maintenance tasks, units can be replaced because they were detected as defective but also because it is an opportunity to change it because the line is stopped. Ideally the line will not stop until the type of product being produced changes.</para>
<fig id="F7-11" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-11">Figure <xref linkend="F7-11" remap="7.11"/></link></label>
<caption><para>Waterproofing the Pajares tunnel in Asturias, Spain.</para></caption>
<graphic xlink:href="graphics/ch07_fig011.jpg"/>
</fig>
<fig id="F7-12" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-12">Figure <xref linkend="F7-12" remap="7.12"/></link></label>
<caption><para>Valencia lightouse (Spain).</para></caption>
<graphic xlink:href="graphics/ch07_fig012.jpg"/>
</fig>
<fig id="F7-13" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-13">Figure <xref linkend="F7-13" remap="7.13"/></link></label>
<caption><para>Pedestrian bridge in Madrid (Spain).</para></caption>
<graphic xlink:href="graphics/ch07_fig013.jpg"/>
</fig>
<fig id="F7-14" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-14">Figure <xref linkend="F7-14" remap="7.14"/></link></label>
<caption><para>Pultrusion machine at ACCIONA workshop.</para></caption>
<graphic xlink:href="graphics/ch07_fig014.jpg"/>
</fig>
<para>There exist three roles involved in the production site of pultrusion line from ACCIONA; production manager, process engineer and operator. Each of them has the following responsibilities and objectives in the framework of the pultrusion process:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="underline"><emphasis>Production Manager responsibilities:</emphasis></emphasis></para>
<para>Study how to improve the overall process for achieve a higher level of efficiency, adapt the process to new kinds of work requests, transmit the outputs obtained from the aforementioned evaluations to the process engineer, and along with the process engineer make appropriate decisions about how to solve deviations caused by repetitive failures;</para></listitem>
</itemizedlist>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="underline"><emphasis>Process Engineer responsibilities:</emphasis></emphasis></para>
<para>General maintenance of the machine, designing the incoming updates from the production manager, deploying the aforementioned updates and test them, leave the equipment ready for use, develop technical instruction for the operators to explain how to proceed with the process, give training sessions to the operators involved in the process, decide how to proceed when a deviation occurs;</para></listitem>
</itemizedlist>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="underline"><emphasis>Operator responsibilities</emphasis>:</emphasis></para>
<para>Daily use of the machine, acquire a high level of autonomy to perform the work requests without supervision, in case of any nonconformity in the process the operator must report the problem to the process engineer.</para></listitem>
</itemizedlist>
<para>Focusing on maintenance tasks, the process engineer is the person who makes important decisions related to several aspects, such as machine maintenance or in the case of any deviation occurs. He is responsible for the equipment to be ready for use and this requires to develop technical instruction for the operators explaining how to proceed with the maintenance tasks. Furthermore, he is in charge of giving training sessions to the operators involved in the maintenance process oriented to apply knowledge in a practical way from the technical instruction.</para>
<para>Maintenance tasks within pultrusion line are manual processes based on visual checks or manual tasks scheduled from time to time. For this reason, these processes must be optimized in order to achieve a reliable analysis of maintenance tasks, foresee potential failures in the systems, decrease production delays and assure proper machine functioning.</para>
</section>
<section class="lev2" id="sec7-2-3">
<title>7.2.3 Data Platform and Sensors</title>
<para>The maintenance tasks related to this machine do not only involve the machine itself. It is necessary to monitor the workshop&#x02019;s environmental conditions.</para>
<para>Data to be collected from the workshop.</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis><emphasis role="underline">Environmental parameters:</emphasis></emphasis></para>
<para>Temperature, humidity and luminosity. These parameters concern to possible change of machine configuration, which can affect the maintenance tasks. The obtained information through temperature and humidity sensors will be an indication of workshop status with a direct implication on the workshop maintenance tasks. For example, the resin used during the production process becomes solid at different points depending on temperature and humidity. The proper mixture of the components should be carried out in a controlled environment that influences on the reaction rate and the proper maintenance of these substances to ensure good criteria of quality. In turn, environmental conditions affect the cleanup and purge tasks of the machine;</para></listitem>
<listitem>
<para><emphasis><emphasis role="underline">Workshop air extraction capacity:</emphasis></emphasis></para>
<para>The maintenance of the ventilation system can vary significantly depending on its use and the outside environment. It is an important system to be taking into account due to the kind of substances that are used during the pultrusion process (i.e., carbon fibre, resin...) and the need to offer the best conditions in the workshop;</para></listitem>
<listitem>
<para><emphasis><emphasis role="underline">Workshop electrical consumption:</emphasis></emphasis></para>
<para>Currently there is no information about machine downtimes produced by electrical failures. Electrical consumption monitoring will help to provide a reliable maintenance of the machine, engines or any other electrical installation in order to find the possible causes of downtimes or malfunctions.</para></listitem>
</itemizedlist>
<para>Data to be collected from the pultrusion machine:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis><emphasis role="underline">Pull-Clamp system</emphasis></emphasis>:</para>
<para>It is one of the main pultrusion line subsystems responsible for moving the profile along the machine and its subsystems while the different treatments to produce composite profiles are performed. The data that is missing regarding this subsystem is the one related to production speed, pull force and presses oil status:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis>Production speed</emphasis> in order to detect machine behavioural patterns and anomalies;</para></listitem>
<listitem>
<para><emphasis>Pull force</emphasis> in order to know the appropriate amount of oil for lubricating this subsystem;</para></listitem>
<listitem>
<para><emphasis>Oil tank</emphasis> presses system. Generally, oil contamination is one of the major causes of hydraulic system and lubricating system failures. The oil inside the hydraulic system is changed from time to time by the operators, based on the process engineer instructions through visual check. Therefore, the continuous supervision of oil state plays a vital role in predictive maintenance systems. The status of oil humidity and temperature are parameters that influence on the quality and profitability of process;</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para><emphasis><emphasis role="underline">Injection Chamber Resin system</emphasis></emphasis>:</para>
<para>It is the responsible for impregnating the fiber with resin. The profile quality depends on the homogeneity of the spray pattern within this process. There are several parameters that will help to analyze an adequate maintenance of this system, such as:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis>Injection System Temperature</emphasis>: To determine possible malfunctions that can impact on the proper resin status and manufacture profile quality;</para></listitem>
<listitem>
<para><emphasis>Thermocouples break detection,</emphasis> in order to avoid malfunctions and errors in the production line;</para></listitem>
<listitem>
<para><emphasis>Resin Header</emphasis> (resin pressure or stream flow rate): It was an unreliable system without any sensor installed. Measuring pressure or resin stream flow rate proved to help to identify when drain maneuvers are needed to avoid breakdown/malfunction of resin injection system;</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para><emphasis><emphasis role="underline">Compressed Air:</emphasis></emphasis></para>
<para>Monitoring pressure, humidity and temperature for ensure a proper maintenance of this system. Continuous monitoring reduces ongoing operation costs, cuts investment costs for new compressors and ensures availability around-the-clock. Compressed air subsystem is one of the most expensive systems in production plants. Many companies are not aware of the fact that the generation and treatment of compressed air accounts for up to 20% of their overall energy costs;</para>
<para>Measurement of compressed air maintenance activities is the first important step towards a cost-conscious and efficient approach to energy consumption and to increase the life-time of the system. Detailed knowledge of the actual compressed air is the basis for reducing energy costs and is an important indicator for investment decisions;</para>
<para>Dew point, pressure, temperature and flow monitoring makes a significant contribution to quality assurance in expensive systems and the products produced there. Only sufficiently dry compressed air can reduce the risk of corrosion, machine failures and low-quality end products;</para>
<para>The correct maintenance of this system will assure:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>To ensure efficiency</para>
<para>Permanently record, monitor and optimize the effectiveness and efficiency of compressed air generation and treatment processes;</para></listitem>
<listitem>
<para>To assure product quality</para>
<para>A change in consumption of compressed air in a production plant is a first indication of possible deviations in the production process. Sufficiently dry compressed air assures the quality of the system and the pultrusion pieces produced;</para></listitem>
<listitem>
<para>For accounting</para>
<para>Billing individual costs for compressed air according to actual consumption can contribute significantly to enhancing a cost-conscious system, and it can suggest whether the compressed air is dry enough and can thereby avoid unnecessary operating costs for compressed air treatment;</para></listitem>
<listitem>
<para>To detect leaks</para>
<para>25&#x02013;40% of the compressed air generated is lost through leaks. Consumption of compressed air in a system that is switched off is a clear indication that there is a leak.</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
<para>After some analysis and lab tests, the most suitable sensors were selected for gathering the data.</para>
<para>The main drivers and constrains that were considered are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>An efficient sensor installation process was needed, as the pultrusion line is continuously working with a few limited stops per months. Moreover, these stops usually are very short;</para></listitem>
<listitem>
<para>Wires installation has been limited, both for communication and for powering the devices;</para></listitem>
<listitem>
<para>The required maintenance for the monitoring system should be minimum;</para></listitem>
<listitem>
<para>The ambient conditions inside the workshop are far from a friendly environment, so all devices installed need to be protected.</para></listitem>
</itemizedlist>
<para>An overview of the architecture deployed in the use case demonstrator is given in <link linkend="F7-15">Figure <xref linkend="F7-15" remap="7.15"/></link>, and the list of installed sensors is given in <link linkend="T7-1">Table <xref linkend="T7-1" remap="7.1"/></link>. The figure represents all equipment, communications and software services needed to transfer the information provided by sensors to the specific point where the HMI is going to show all the information to the user.</para>
<para>The system implemented can be divided in different parts or subsystems:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Sensors;</para></listitem>
<listitem>
<para>Local Sensors Controller</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Zigbee Root Node</para></listitem>
<listitem>
<para>X86 Gateway</para></listitem>
<listitem>
<para>LTE Router</para></listitem>
<listitem>
<para>Wi-Fi Access Point</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>Local HMI</para></listitem>
<listitem>
<para>Cloud Servers</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>OpenMQ Server</para></listitem>
<listitem>
<para>Database Server</para></listitem>
<listitem>
<para>Web Server</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
<fig id="F7-15" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-15">Figure <xref linkend="F7-15" remap="7.15"/></link></label>
<caption><para>Monitoring system.</para></caption>
<graphic xlink:href="graphics/ch07_fig015.jpg"/>
</fig>
<table-wrap position="float" id="T7-1">
<label><link linkend="T7-1">Table <xref linkend="T7-1" remap="7.1"/></link></label>
<caption><para>List of installed sensors</para></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/ch07_tab1.jpg"/>
</table-wrap>
</section>
<section class="lev2" id="sec7-2-4">
<title>7.2.4 Human Machine Interfaces</title>
<para>Given the different roles of the professionals who use this system, there have been developed and deployed two different HMIs: Local HMI and Remote HMI.</para>
<para><emphasis role="strong">Local HMI:</emphasis></para>
<para>Our use case has particularities due to 24h/day production in the pultrusion line. In general, local HMI is focused on providing useful information for the machine operators and process engineers, displaying instant parameter values and alerts detected by the sensors (<link linkend="F7-16">Figure <xref linkend="F7-16" remap="7.16"/></link>). The display is located near the pultrusion line machine, in a very visible spot for the workers. Whenever an alert occurs in the local HMI (due to data out of range) the operators and process engineers who are near the local HMI are able to see the alert and they will act in order to face the problem when possible. The operators, following process engineer directions, get the job done and in case of any nonconformity in the process the operator must report the problem to the process engineer.</para>
<para>All the information and alarms should be shown without any interaction. There are several priorities of alert level composed by warnings and alarms.</para>
<fig id="F7-16" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-16">Figure <xref linkend="F7-16" remap="7.16"/></link></label>
<caption><para>Local HMI. PC, Tablets, mobile devices.</para></caption>
<graphic xlink:href="graphics/ch07_fig016.jpg"/>
</fig>
<para><emphasis role="strong">Remote HMI:</emphasis></para>
<para>This HMI, on the other hand, is focused on displaying historical information and high or low level alerts detected by the sensors devices (data out of range, the latter will be the same alert as in local HMI). Historical information and its representation using graphs can help the process engineer or production manager to anticipate possible failure and plan maintenance tasks.</para>
<fig id="F7-17" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-17">Figure <xref linkend="F7-17" remap="7.17"/></link></label>
<caption><para>Remote HMI. Home menu.</para></caption>
<graphic xlink:href="graphics/ch07_fig017.jpg"/>
</fig>
<para>The objective of this HMI, represented in Figures 7.17 and 7.18, is to provide a powerful tool not only for data visualisation but also for showing any analysis that the production manager (or the maintenance staff) would require.</para>
<para>Some of the features include:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>User management;</para></listitem>
<listitem>
<para>Notifications configuration;</para></listitem>
<listitem>
<para>Alarms configuration;</para></listitem>
<listitem>
<para>Report generation;</para></listitem>
<listitem>
<para>Data graphs.</para></listitem>
</itemizedlist>
<para>The HMI is able to display alerts if values are out of range (warning or alarms) or if possible anomalies or failures are detected from historical data. Historical information includes input of time-stamped data. The detection of anomalies could have future implications in the machine with the resulting risk of failures. The system involved in the maintenance task would be identified, together with what is wrong according to the sensing devices.</para>
<fig id="F7-18" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-18">Figure <xref linkend="F7-18" remap="7.18"/></link></label>
<caption><para>Remote HMI. Temperature, Humidity &#x00026; Light (THL) menu.</para></caption>
<graphic xlink:href="graphics/ch07_fig018.jpg"/>
</fig>
</section>
<section class="lev2" id="sec7-2-5">
<title>7.2.5 Maintenance Optimization and Validation Results</title>
<para>This section focuses on the results achieved on the pilots by the MANTIS techniques and monitoring process.</para>
<section class="lev3" id="sec7-2-5-1">
<title>7.2.5.1 Temperature control system located in the mixing area and in the storage area</title>
<para>It is required to keep the storage and mixing temperature of some products used during the pultrusion process lower than a specific value due to safety and quality issues. Consequently, the temperature of the storage and mixing area (<link linkend="F7-19">Figure <xref linkend="F7-19" remap="7.19"/></link>) must be controlled, in this case using an air cooling system.</para>
<fig id="F7-19" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-19">Figure <xref linkend="F7-19" remap="7.19"/></link></label>
<caption><para>Mixing area at workshop.</para></caption>
<graphic xlink:href="graphics/ch07_fig019.jpg"/>
</fig>
<para>The environmental parameters are gathered through the monitoring system. The operator checks out on the local display warnings and alarms in case that abnormal values of temperature, flow, or air pressure were detected. The Process Engineer checks out the system performance through the Remote HMI according to the Control Program.</para>
<para>In the case that a temperature alarm has been triggered and if the maximum temperature allowed is reached (safe temperature is recommended in the safety-sheet of the stored products), explosive products should be moved to an alternative secure area before carrying out the reparation of the cooling system and the safety manager should be reported. If failure was not critical, the problem could be solved during a scheduled stop of the line according to the maintenance program.</para>
<para>Figures 7.20 and 7.21 show some values of temperature and relative humidity of the mixing zone during a continuous parameter registration between February and March 2018. The plots illustrate periods where the machine is not working, achieving temperatures between 15 to 23 degrees Celsius, as well as temperature ranges where the pultrusion line is working during production, achieving temperatures between 22 to 24 degrees Celsius. Relative humidity is always below 50%.</para>
<para>No warning and alarms (yellow and red color alarms) were detected by the operator in the local and remote user interfaces according to the obtained data. The temperature and humidity were kept within the optimal parameters in the mixing area during the pultrusion process tests.</para>
<fig id="F7-20" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-20">Figure <xref linkend="F7-20" remap="7.20"/></link></label>
<caption><para>Temperature control in the Mixing Area. (Monitoring period 21/02/2018 -06/03/2018).</para></caption>
<graphic xlink:href="graphics/ch07_fig020.jpg"/>
</fig>
<fig id="F7-21" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-21">Figure <xref linkend="F7-21" remap="7.21"/></link></label>
<caption><para>Humidity control in the Mixing Area. (Monitoring period 21/02/2018 -06/03/2018).</para></caption>
<graphic xlink:href="graphics/ch07_fig021.jpg"/>
</fig>
</section>
<section class="lev3" id="sec7-2-5-2">
<title>7.2.5.2 Cooling system for the injection chamber</title>
<para>The cooling system (<link linkend="F7-22">Figure <xref linkend="F7-22" remap="7.22"/></link>) maintains the temperature of the injection chamber low enough to avoid premature gelification of the resin inside the injection chamber. The temperature of the chamber is controlled through a liquid chiller. The cooling fluid must be water, optionally mixed with a certain percentage of ethylene glycol (to prevent freezing), depending on the water outlet temperature. The water is cooled using a refrigeration circuit.</para>
<fig id="F7-22" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-22">Figure <xref linkend="F7-22" remap="7.22"/></link></label>
<caption><para>Cooling system from injection chamber.</para></caption>
<graphic xlink:href="graphics/ch07_fig022.jpg"/>
</fig>
<para><link linkend="F7-23">Figure <xref linkend="F7-23" remap="7.23"/></link> shows the elementary scheme of the circuit that allows a refrigerating cycle. <link linkend="F7-24">Figure <xref linkend="F7-24" remap="7.24"/></link> shows a schema of the pipeline circuit used to cool the injection chamber.</para>
<para>The cooling system is monitored detecting the liquid temperature and flow as well as its level in the chiller&#x02019;s deposit, using wireless sensors located in the liquid tank (level) and at the pipelines that connect the chiller with the refrigeration circuit of the injection chamber (temperature and flow).</para>
<fig id="F7-23" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-23">Figure <xref linkend="F7-23" remap="7.23"/></link></label>
<caption><para>Schema of the chiller cooling system from injection chamber.</para></caption>
<graphic xlink:href="graphics/ch07_fig023.jpg"/>
</fig>
<fig id="F7-24" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-24">Figure <xref linkend="F7-24" remap="7.24"/></link></label>
<caption><para>Schema of the injection chambers cooling circuit.</para></caption>
<graphic xlink:href="graphics/ch07_fig024.jpg"/>
</fig>
<para>The operator checks the possible warnings and alarms on the local display in the case that abnormal values of temperature, level or flow of refrigerant liquid were detected. The process engineer views the system performance through the Remote HMI according to the Control Program. The specialized operator, who is allowed to manage the machine in emergency cases, will be responsible for checking the local display.</para>
<para>If failures were not critical, the problems would be solved during a scheduled stop of the line according to the maintenance program or keeping the machine running (for example, when the chiller&#x02019;s deposit needs to be filled or if a leak occurs). Repair on the fly is allowed in this scenario because the chiller is not physically integrated in the pultrusion machine so it is easier and safer to solve small complications. If it was critical, a non-scheduled stop would be performed and the process manager would be reported. Then, both the process engineer and the process manager would analyse the data available and decide if it is necessary to modify the maintenance program or instead, if it is an unforeseen failure and it is preferable to contact the specialised technical service.</para>
<para>The limits of the temperatures were defined by means of previous manufacturing experience and taking into account that the set temperature of the chiller must be different than the ones we obtain inside the manufacturing site. During the winter the chiller set temperature is between 22&#x02013;24<superscript>&#x02218;</superscript>C and during summer it is adjusted to 16&#x02013;18<superscript>&#x02218;</superscript>C to obtain the desired process parameters. It is expected that after a complete year of monitoring a more accurate relationship between these two parameters can be obtained. The installed sensors can be seen in <link linkend="F7-25">Figure <xref linkend="F7-25" remap="7.25"/></link>.</para>
<fig id="F7-25" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-25">Figure <xref linkend="F7-25" remap="7.25"/></link></label>
<caption><para>Sensors installed in the refrigeration circuit.</para></caption>
<graphic xlink:href="graphics/ch07_fig025.jpg"/>
</fig>
<para>Time series data of the liquid temperature have been recorded for 3 months with the sensors installed as part of the Mantis project. Due to some problems between the sensors and the data acquisition program it has not been possible to see the flow measurements, but it can be predicted that some of the variations seen in the temperature could be directly related with it.</para>
<para>The graphical display for room temperatures and the inlet circuit temperature of the refrigeration liquid is shown in <link linkend="F7-26">Figure <xref linkend="F7-26" remap="7.26"/></link>. It can be clearly seen that although the exterior temperature suffers large variations than the temperature in the refrigeration circuit. An important point to check is that the refrigeration circuit operates correctly and that the external factors are not directly affecting the manufacturing process.</para>
<fig id="F7-26" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-26">Figure <xref linkend="F7-26" remap="7.26"/></link></label>
<caption><para>Room temperature vs. liquid inlet temperature control chart.</para></caption>
<graphic xlink:href="graphics/ch07_fig026.jpg"/>
</fig>
<para>Taking into account upper and lower limits for the alarms defined, it can be seen that some points must be checked although none of them are in the red alarm range. In the first part of the control chart the variation of the temperature is larger than in the last stage. The increase of the temperature (Early January 2018) is due to problems derived from non-constant flows, the water stagnates in the pipes and the temperature increases. On the other hand, low temperatures (middle January 2018) are due to some maintenance stops for the installation of new sensors and connections. As the room temperature is low the refrigeration system decreases its temperature.</para>
<para>In the chart in <link linkend="F7-27">Figure <xref linkend="F7-27" remap="7.27"/></link>, which corresponds to the production of the pultruded profiles, the temperature of the refrigeration circuit is more stable. It can be seen that some points are in the rage of the yellow alarm. As can be seen by comparing Figures 7.27 and 7.28, these increases in the temperature occur at the same time that the level of the liquid in the refrigeration circuit decrease.</para>
<fig id="F7-27" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-27">Figure <xref linkend="F7-27" remap="7.27"/></link></label>
<caption><para>Refrigeration circuit inlet temperature control chart.</para></caption>
<graphic xlink:href="graphics/ch07_fig027.jpg"/>
</fig>
<para><link linkend="F7-28">Figure <xref linkend="F7-28" remap="7.28"/></link> shows the evolution of the refrigeration liquid tank level. The consumption of the tank depends on different factors such as exterior temperature, working hours, number of equipment connected to the circuit, etc. so it is not easy to predict a constant behavior. But it has been seen that a reduction of more than a 35% of the level of the tank has a direct impact in the refrigeration circuit temperature, so it has been decided to refill the tank once a day above the 80%. The operator must be in charge of this action and also must be aware of any other alarm that can appear.</para>
<fig id="F7-28" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-28">Figure <xref linkend="F7-28" remap="7.28"/></link></label>
<caption><para>Refrigeration liquid tank level control chart.</para></caption>
<graphic xlink:href="graphics/ch07_fig028.jpg"/>
</fig>
<para>All the charts in Figures 7.26&#x02013;7.28 are displayed in the data recorded by the remote HMI system. Any data out of the target values will give an alarm that must be checked out by the process engineer.</para>
<para>It has been seen that the monitoring system of the refrigeration temperature and tank level exhibit useful information for the control of the manufacturing process. The system works correctly and in-situ checks with the control through the monitoring system could avoid production problems and advise of maintenance needs that otherwise are difficult to detect. A good control of the refrigeration circuit will allow reducing the purge needs of the system and the cleaning operations.</para>
</section>
<section class="lev3" id="sec7-2-5-3">
<title>7.2.5.3 Compressed air system from pulling system</title>
<para>For this application, the goal is to assure the correct workings of the compressor (<link linkend="F7-29">Figure <xref linkend="F7-29" remap="7.29"/></link>) and, on the other hand to detect any anomaly in the injection circuit (<link linkend="F7-30">Figure <xref linkend="F7-30" remap="7.30"/></link>) (pressurized tanks, pipelines, connections, etc.).</para>
<para>Using wireless sensors located at the compressor outlet and at the pipelines that connect to the injection chamber, air pressure, temperature and flow are monitored. The operator checks out on the local display warnings and alarms in case of abnormal values of temperature, flow or air pressure were detected. The process engineer checks out the system performance through the Remote HMI according to the Control Program.</para>
<fig id="F7-29" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-29">Figure <xref linkend="F7-29" remap="7.29"/></link></label>
<caption><para>Detail of the main compressor that supplies compressed air to the resin injection system.</para></caption>
<graphic xlink:href="graphics/ch07_fig029.jpg"/>
</fig>
<fig id="F7-30" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-30">Figure <xref linkend="F7-30" remap="7.30"/></link></label>
<caption><para>(a) Schema of the injection chamber resin injection circuit. (b) Detail of the resin injection circuit.</para></caption>
<graphic xlink:href="graphics/ch07_fig030.jpg"/>
</fig>
<fig id="F7-31" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-31">Figure <xref linkend="F7-31" remap="7.31"/></link></label>
<caption><para>Data record of the evolution of the air pressure obtained from the Remote HMI interface.</para></caption>
<graphic xlink:href="graphics/ch07_fig031.jpg"/>
</fig>
<para><link linkend="F7-31">Figure <xref linkend="F7-31" remap="7.31"/></link> shows the data record of the evolution of the air pressure obtained from the Remote HMI interface for a period of two and a half months of production (Monitoring period 01/01/2018&#x02013;14/03/2018).</para>
<para><link linkend="F7-31">Figure <xref linkend="F7-31" remap="7.31"/></link> shows that the air pressure remained above 8 bar and thus above the operational limits fixed during the period studied. In fact, the Control Program states that:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Value > 6 bar that corresponds to regular operational conditions no alarm;</para></listitem>
<listitem>
<para>L1: Value &#x0003C; 6 that corresponds to non-regular operational conditions, but not critical alarm turns to yellow;</para></listitem>
<listitem>
<para>L2: Value &#x0003C; 2 bar that corresponds to critical operational conditions alarm turns to red.</para></listitem>
</itemizedlist>
<table-wrap position="float">
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="center"></td>
<td valign="top" align="center"><emphasis role="strong">Green Alarm</emphasis></td>
<td valign="top" align="center"><emphasis role="strong">Yellow Alarm</emphasis></td>
<td valign="top" align="center"><emphasis role="strong">Red Alarm</emphasis></td></tr>
</thead>
<tbody>
<tr>
<td valign="top" align="center"><emphasis role="strong">Temperature</emphasis></td>
<td valign="top" align="center">10&#x02013;25<superscript>&#x02218;</superscript>C</td>
<td valign="top" align="center">5&#x02013;10<superscript>&#x02218;</superscript>C//25&#x02013;30<superscript>&#x02218;</superscript>C</td>
<td valign="top" align="center">> 30<superscript>&#x02218;</superscript>C and &#x0003C; 5<superscript>&#x02218;</superscript>C</td>
</tr>
<tr>
<td valign="top" align="center"><emphasis role="strong">Pressure</emphasis></td>
<td valign="top" align="center">>6%</td>
<td valign="top" align="center">2&#x02013;6%</td>
<td valign="top" align="center">&#x0003C; 2%</td></tr>
<tr>
<td valign="top" align="center"></td></tr>
</tbody>
</table>
</table-wrap>
<para>The air pressure remained stable oscillating between 8&#x02013;9.5 bars (<link linkend="F7-32">Figure <xref linkend="F7-32" remap="7.32"/></link>). This oscillation is normal within the regular operating regime of the compressor because the compressor&#x02019;s engine does not work constantly. It automatically turns on when the pressure inside the pressurized tank falls below a consigned value (about 8 bar).</para>
<fig id="F7-32" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-32">Figure <xref linkend="F7-32" remap="7.32"/></link></label>
<caption><para>Detail of the data record of the evolution of the air pressure obtained from the Remote HMI interface.</para></caption>
<graphic xlink:href="graphics/ch07_fig032.jpg"/>
</fig>
<fig id="F7-33" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-33">Figure <xref linkend="F7-33" remap="7.33"/></link></label>
<caption><para>Data record of the evolution of the temperature obtained from the Remote HMI interface.</para></caption>
<graphic xlink:href="graphics/ch07_fig033.jpg"/>
</fig>
<para>Operators did not detect any alarm trigger in relation to air pressure during this period except 19 January when the compressor shut down due to a power cut. Because of this event, the pressure decreased, triggering the yellow alarm. However, the red alarm was not activated since the pressure remained above two bars until the power was recovered. The compressor was able to keep the air moderately pressurized inside the tank for few hours (<link linkend="F7-31">Figure <xref linkend="F7-31" remap="7.31"/></link>).</para>
<para>Therefore, the in situ control carried out by the operators matched with the data recorded by the remote HMI system and checked out by the process engineer.</para>
<para>The temperature inside the room where the compressor is placed was also recorded. <link linkend="F7-33">Figure <xref linkend="F7-33" remap="7.33"/></link> shows the data record of the evolution of the temperature obtained from the Remote HMI interface for a period of two and a half months of production (Monitoring period 01/01/2018&#x02013;14/03/2018).</para>
<para>Even though the temperature oscillates between 10 and 25<superscript>&#x02218;</superscript>C inside the room, the effect on the air pressure control is negligible (<link linkend="F7-32">Figure <xref linkend="F7-32" remap="7.32"/></link>). The heat generated by the operation of the compressor causes this oscillation. The room is conditioned by means of an extraction system that brings out the room the hot air when the temperature inside the room reaches 25<superscript>&#x02218;</superscript>C. On weekends, highlighted in red, the compressor remains off and the temperature is closer to the exterior room temperature.</para>
<para>During the monitoring period, the temperature values stayed within the operational limits allowed, between 5&#x02013;30<superscript>&#x02218;</superscript>C. Therefore, the current conditioning system is enough to keep the temperature within the recommended operation values.</para>
<para>In addition to the sensors installed in the compressor room, another sensor has been installed at the furthest point from the compressed air circuit (<link linkend="F7-34">Figure <xref linkend="F7-34" remap="7.34"/></link>). This sensor will help to check if the pressure in the compressor maintains the same or if there is any loss along the circuit. The installation of the sensor had to be done while there was no production in progress and complies with different security.</para>
<fig id="F7-34" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-34">Figure <xref linkend="F7-34" remap="7.34"/></link></label>
<caption><para>Pressure and temperature sensor installed in the compressed air circuit.</para></caption>
<graphic xlink:href="graphics/ch07_fig034.jpg"/>
</fig>
</section>
</section>
</section>
<section class="lev1" id="sec7-3">
<title>7.3 Maintenance in Press Forming Machinery</title>
<blockquote role="flushleft">
<para><emphasis>Contributors: Urko Zurutuza, Javier Fernandez-Anakabe, Ekhi Zugasti, Petri Helo, Lauri V&#x000E4;limaa, Mathias Gr&#x000E4;dler, Mikel Mondragon, Andoitz Aranburu, David Chico, Oier Sarasua, Mar&#x000ED;a Aguirregabiria, Xabier Eguiluz, Iosu Gabilondo, Eduardo Saiz, Iban Barrutia Inza, Mikel Viguera, F&#x000E9;lix Larrinaga Barrenechea, Mikel Anasagasti, Jon Olaizola and Ricardo Romero.</emphasis></para></blockquote>
<para>This use cases focuses on stamping press machines, which are metal working machines used to shape or cut metal by deforming it with a die. See <link linkend="F7-35">Figure <xref linkend="F7-35" remap="7.35"/></link> for some examples of this kind of machinery.</para>
<para>This kind of press is built by FAGOR and, during its active lifetime, might be capable of giving more than 40 million strokes characterized by impressive force and precision, insofar as the press is used and maintained appropriately. This use case considers two scenarios.</para>
<para>The first scenario focuses on the press forming machinery itself. The customers expect both high quality of the pieces produced by the machine, and high availability, which led FAGOR to incorporate cutting-edge technologies in their products as a means of enhancing products robustness and functionality in order to facilitate proactive maintenance activities.</para>
<para>The second scenario considers the pneumatic Clutch Brake, which is a critical device within the mechanical press machine. The Clutch Brake is responsible of activating and stopping the tool of the press machine in order to perform different processes. In this scenario, the clutch brake became a CPS itself, able to provide data regarding its own health conditions.</para>
<fig id="F7-35" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-35">Figure <xref linkend="F7-35" remap="7.35"/></link></label>
<caption><para>Fagor Arrasate mechanical and servo driven presses.</para></caption>
<graphic xlink:href="graphics/ch07_fig035.jpg"/>
</fig>
<section class="lev2" id="sec7-3-1">
<title>7.3.1 Introduction</title>
<para><emphasis role="strong">FAGOR ARRASATE S.COOP</emphasis> is a company of 800 employees specialized in designing, manufacturing and supplying sheet metal forming machine tools. Fagor Arrasate was created in 1957 and, since then, has expanded its products and business in an significant manner, being now one of the world leaders in the field. It is one of the 5 biggest manufacturers in the world in terms of turnover and the first one considering the product&#x02019;s portfolio.</para>
<para>The Company is located in the Basque Country, in the north of Spain, very close to the French border in the most industrialized area of the country and surrounded by a traditionally metallurgical and exporting environment.</para>
<para>FAGOR is a world leader in the design and manufacture of mechanical and hydraulic presses, complete stamping systems, transfer presses, robotised press lines, press hardening, forging; Cut-to-Length, Slitting, Combi and multiblanking lines; Processing lines as pickling lines, skin passes, reversible mills, painting, galvanizing or levelling lines; special metal part forming systems, strip roll forming, flexible roll forming, rotor/stator cutting equipment, dies and many other types of equipment.</para>
<para>Fagor Arrasate serves to numerous sectors, with a particular focus on the car industry, the domestic appliances industry, the Steel Industry and Service Centres. For Fagor Arrasate a key goal is the constant collaboration with its customers, so there is a close and continuous presence in order to give solutions for any process with the most adequate technology.</para>
<para><emphasis role="strong">GOIZPER S.COOP</emphasis> is one of the leading technology suppliers in power transmission components, such us brakes, clutches, turning systems, gear boxes, cams or elevators. GOIZPER designs, manufactures and supplies customized power transmission components to meet market needs in sectors like metal forming, automotive, aeronautics, packaging, construction, marine, machine tools, etc.</para>
<para>As mentioned, Clutch Brakes and gearboxes are products consumed in automotive industry and Fagor Arrasate is currently using them within their mechanical press machines (see <link linkend="F7-36">Figure <xref linkend="F7-36" remap="7.36"/></link>). In other words, Goizper is one of Fagor Arrasate&#x02019;s current power transmission components supplier.</para>
<fig id="F7-36" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-36">Figure <xref linkend="F7-36" remap="7.36"/></link></label>
<caption><para>Clutch Brake (left) and gearboxes (center) by Goizper, used in Fagor&#x02019;s mechanical press machines (right).</para></caption>
<graphic xlink:href="graphics/ch07_unfig001.jpg"/>
</fig>
<para>GOIZPER&#x02019;s headquarters and productive plant are located in Antzuola, Spain, and almost 80% of the sales come from exports all over the world. The maintenance of sold parts has become an issue due to the different locations of the parts around the world.</para>
<para>GOIZPER also has another division, totally different, focused on the design, manufacture and marketing of manual sprayers and dusters for treatments in farming, gardening, industry, construction, cleaning, pest control and vector control.</para>
</section>
<section class="lev2" id="sec7-3-2">
<title>7.3.2 Scope and Logic</title>
<para>The final customers of FAGOR ARRASATE produce products with high levels of quality and availability seeking a drastic reduction of high cost caused by production downtimes with required maintenance-repair operations and a better delivery times&#x02019; compliance. This is why FAGOR ARRASATE needs to increase the reliability of machines and components. To meet this challenge FAGOR ARRASATE is continuously incorporating cutting-edge technologies in their products as a means of enhancing products robustness and functionality in order to facilitate proactive maintenance activities.</para>
<section class="lev3" id="sec7-3-2-1">
<title>7.3.2.1 Background information on the press machine</title>
<para>Mechanical and servo driven press machine elements have been analysed using sensor technologies in order to improve maintenance strategies for detecting early failures on the cranks and in forming elements of the stamping press.</para>
<para>A platform has been developed where the data from different components of a press machine could be captured, monitored, transmitted, stored and analyzed in order to come to reliable predictive and proactive maintenance. The data is analyzed and monitored via local or cloud level (see <link linkend="F7-37">Figure <xref linkend="F7-37" remap="7.37"/></link>).</para>
<fig id="F7-37" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-37">Figure <xref linkend="F7-37" remap="7.37"/></link></label>
<caption><para>Predictive maintenance HMI platform.</para></caption>
<graphic xlink:href="graphics/ch07_fig036.jpg"/>
</fig>
<para>The components of the press machine that require sensors with innovative CPSs are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Bushings (Temperature and oil condition status);</para></listitem>
<listitem>
<para>Bolster (Relative displacements);</para></listitem>
<listitem>
<para>Head (Structural health);</para></listitem>
<listitem>
<para>Gear axis (Torque);</para></listitem>
<listitem>
<para>Engine (Tension and current);</para></listitem>
<listitem>
<para>Connecting rod (Displacement, forces).</para></listitem>
</itemizedlist>
<para><link linkend="F7-38">Figure <xref linkend="F7-38" remap="7.38"/></link> illustrates the location of the source of each component data.</para>
<fig id="F7-38" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-38">Figure <xref linkend="F7-38" remap="7.38"/></link></label>
<caption><para>Location of data sources.</para></caption>
<graphic xlink:href="graphics/ch07_fig037.jpg"/>
</fig>
<para>The objectives of this use case are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Maintenance Cloud Platform development;</para></listitem>
<listitem>
<para>Torque measurement using wireless sensors;</para></listitem>
<listitem>
<para>Head structural health monitoring;</para></listitem>
<listitem>
<para>Torque measurement using wireless sensors;</para></listitem>
<listitem>
<para>Bushing status measurement;</para></listitem>
<listitem>
<para>Gears wear measurement;</para></listitem>
<listitem>
<para>Crank strain and force measurement;</para></listitem>
<listitem>
<para>Press unbalances forces measurement;</para></listitem>
<listitem>
<para>Press cutting shock measurement.</para></listitem>
</itemizedlist>
</section>
<section class="lev3" id="sec7-3-2-2">
<title>7.3.2.2 Background information on the clutch brake component</title>
<para>A critical device within the mechanical press machine is the pneumatic clutch brake. The Clutch Brake is the responsible component of activating and stopping the tool of the press machine in order to perform different processes. In this case, metal forming process is considered, where the clutch brake works as a mechanical commutator. The clutch brake components suffer from degradation during operation. Component degradation usually causes machine failures and downtime, generating unwanted and unexpected costs. <link linkend="F7-39">Figure <xref linkend="F7-39" remap="7.39"/></link> shows how a pneumatic clutch brake looks like.</para>
<fig id="F7-39" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-39">Figure <xref linkend="F7-39" remap="7.39"/></link></label>
<caption><para>GOIZPER pneumatic Clutch Brake outside and inside.</para></caption>
<graphic xlink:href="graphics/ch07_fig038.jpg"/>
</fig>
<para>The focus lies on identifying issues related to the maintenance of the clutch brake, adopting strategies to monitor and to make decisions against those issues.</para>
<para>Downtimes caused by the clutch brake have been listed taking into account the number of stops. From this list, the following topics were picked in order to analyse and solve.</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Friction material slippage detection at clutching;</para></listitem>
<listitem>
<para>Friction material slippage detection at braking;</para></listitem>
<listitem>
<para>Friction material wear and misalignment;</para></listitem>
<listitem>
<para>Piston chamber air leakage;</para></listitem>
<listitem>
<para>Brake springs degradation.</para></listitem>
</itemizedlist>
<para>For the Clutch Brake scenario, two demonstrators have been used. One of them is situated at MGEP facilities in Mondragon, Spain, and it consists of a Fagor mechanical press machine that contains a GOIZPER&#x02019;s Clutch Brake component. The other one is located at GOIZPER&#x02019;s facilities in Antzuola, Spain, and it consists of a Clutch Brake wear test bench.</para>
<section class="lev4">
<title>MGEP Press Machine demonstrator</title>
<para>In <link linkend="F7-40">Figure <xref linkend="F7-40" remap="7.40"/></link>, Fagor&#x02019;s press machine demonstrator&#x02019;s front and back sides are shown. This machine contains a GOIZPER Clutch Brake and it is located in MGEP shopfloor for small size metal parts forming.</para>
<fig id="F7-40" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-40">Figure <xref linkend="F7-40" remap="7.40"/></link></label>
<caption><para>MGEP press machine, demonstrator 1.</para></caption>
<graphic xlink:href="graphics/ch07_fig039.jpg"/>
</fig>
<para>This machine is a mechanical press machine used for low duty metal forming processes. It contains a pneumatic Clutch Brake at the back side (shown in <link linkend="F7-41">Figure <xref linkend="F7-41" remap="7.41"/></link>) in order to activate and deactivate the ram of the press. The ram is the orange part of the press which performs the action of metal forming.</para>
<fig id="F7-41" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-41">Figure <xref linkend="F7-41" remap="7.41"/></link></label>
<caption><para>GOIZPER clutch brake in MGEP demonstrator before MANTIS.</para></caption>
<graphic xlink:href="graphics/ch07_fig040.jpg"/>
</fig>
<fig id="F7-42" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-42">Figure <xref linkend="F7-42" remap="7.42"/></link></label>
<caption><para>Clutch Brake in MGEP demonstrator with the sensors installed.</para></caption>
<graphic xlink:href="graphics/ch07_fig041.jpg"/>
</fig>
<para><link linkend="F7-42">Figure <xref linkend="F7-42" remap="7.42"/></link> shows the back side of the demonstrator, and the GOIZPER clutch brake next to the flywheel (orange). Sensors have been installed within the Clutch Brake in order to capture data and execute the algorithms for the preventive maintenance of the component. The electric motor, power source of the application, is located at the top side (not visible) connected to the flywheel by means of a black belt.</para>
<para>Within the pneumatic circuit, the electro valve is located at the right side of the picture, opening and closing the air flow into the clutch brake. This air is introduced through the black tube connected to the application axis, in the middle of the picture. Sensors have been installed in order to receive data from the Clutch Brake. These installed sensors are visualized in <link linkend="F7-40">Figure <xref linkend="F7-40" remap="7.40"/></link>, which indicates each sensor&#x02019;s location.</para>
</section>
<section class="lev4">
<title>GOIZPER Test Bench</title>
<para>The second demonstrator is the test bench in GOIZPER&#x02019;s installations (<link linkend="F7-43">Figure <xref linkend="F7-43" remap="7.43"/></link>). Friction discs accelerated degradation has been forced in order to get the data from the beginning (%0 of wear) until the end of the friction discs life (%10 of wear).</para>
<para>The installed sensors are not giving direct information, all the captured data needs a processing stage (<link linkend="F7-44">Figure <xref linkend="F7-44" remap="7.44"/></link>) in order to know the actual health of the Clutch Brake. From these calculations, different problematic scenarios have been identified. Some of the algorithms are located in a local data logger and the rest are located within the cloud.</para>
<fig id="F7-43" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-43">Figure <xref linkend="F7-43" remap="7.43"/></link></label>
<caption><para>GOIZPER Clutch Brake test bench, demonstrator 2.</para></caption>
<graphic xlink:href="graphics/ch07_fig042.jpg"/>
</fig>
<fig id="F7-44" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-44">Figure <xref linkend="F7-44" remap="7.44"/></link></label>
<caption><para>Local data logger for data capturing and processing.</para></caption>
<graphic xlink:href="graphics/ch07_fig043.jpg"/>
</fig>
</section>
</section>
</section>
<section class="lev2" id="sec7-3-3">
<title>7.3.3 MANTIS Solutions for Press Machine</title>
<para>This section focuses on the first scenario of the use case, and thus on the solution implemented by FAGOR for the press machine itself.</para>
<section class="lev3" id="sec7-3-3-1">
<title>7.3.3.1 Maintenance cloud platform</title>
<para>A new demand of technical solutions and services aiming at improving the efficiency of maintenance and repair operations is arising. In line with this need, FAGOR ARRASATE wants to offer to its customers a broad range of maintenance services based on digital technologies that allow the company to collect real-time data from the press machines installed all over the world. As Fagor and Goizper are different firms, each company is developing its own Cloud solution. However, interoperability has been taken into account for the cases that both partners work together.</para>
<section class="lev4" id="sec7-3-3-1-1">
<title>7.3.3.1.1 Solution approach</title>
<para>The solution selected by FAGOR has been to develop a digital cloud platform where data from different components of a remote press machine can be collected, monitored, transmitted, stored and analysed providing services for reliable predictive and proactive maintenance. The cloud platform has an architecture divided into two different environments: <emphasis role="strong">On-cloud</emphasis> where data coming from the different sources are stored, processed and analysed and <emphasis role="strong">On-premise</emphasis> where is the data acquisition system to extract the data from the different sensors of the machines. Data format is based on the <emphasis role="strong">Event Information Model</emphasis> adopted for the present project and communications among both environments are secured by using VPN tunnels that guarantee the data integrity, confidentiality and availability.</para>
<para>In the Cloud, a Big Data architecture following the MANTIS reference architecture principles and based on different applications has been designed aiming at supporting fault-tolerance and high scalability. In this way, each part of the system is independent and loosely coupled. The general architecture of this approach is illustrated in <link linkend="F7-45">Figure <xref linkend="F7-45" remap="7.45"/></link>.</para>
<fig id="F7-45" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-45">Figure <xref linkend="F7-45" remap="7.45"/></link></label>
<caption><para>Architecture design.</para></caption>
<graphic xlink:href="graphics/ch07_fig044.jpg"/>
</fig>
<para>On-Cloud architecture consists of the following core components:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Elastic Search:</emphasis> The data coming from the different manufacturing facilities will be persisted in a NoSQL database. It allows the storage of huge volumes of data as well as an optimised search mechanism with a flexible approach to perform a number of aggregations;</para></listitem>
<listitem>
<para><emphasis role="strong">[Apache Kafka]:</emphasis> A distributed queue message system to decouple the different applications by following the publisher-subscriber communication pattern. This technology uses a topic approach to categorise the data. In this work the different data natures are published through different topics;</para></listitem>
<listitem>
<para><emphasis role="strong">Proxy:</emphasis> This application is an HTTP proxy that receives the data from on-premise sources and categorises them in different data natures by considering their origin. Afterwards, the data is published in different Kafka topics;</para></listitem>
<listitem>
<para><emphasis role="strong">Real Time Processing:</emphasis> this application has three objectives:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Extracting the raw data categorised by means of the Kafka topics;</para></listitem>
<listitem>
<para>Persisting the data in elastic search also categorised by data nature;</para></listitem>
<listitem>
<para>Executing data analytics to detect possible alarms (that are published to Kafka), and performing a predictive maintenance.</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para><emphasis role="strong">API:</emphasis> A REST API in charge of exposing the functionalities supported by the Cloud to allow the connection with a front-end (App Web). Among the functionalities, they are worth emphasising the possibility of querying an Elastic Search NoSQL database to obtain historic data by applying distinct filters, the execution of [CRUD] operations over the resources required and the users&#x02019; management system. This API is implemented as a [Spring Boot] application, a framework to simplify the creation and development of Java Web applications. In addition, OAuth 2.0 protocol [Aaron Parecki, 2018] is adopted in order to guarantee the security in the communications;</para></listitem>
<listitem>
<para><emphasis role="strong">Alarm:</emphasis> This application consumes from Kafka the alarms generated from real-time processing and push the corresponding values to a front end to trigger an alarm when necessary.</para></listitem>
</itemizedlist>
<para>On-premise is considered as the technological solution deployed on the manufacturing plants of the customers where their machines are located. This environment usually has limited hardware, software and network resources. This leads to frequently delegate the high-consume processes and the exhaustive exploitation of the data to the cloud environment. Therefore, in some cases data will be directly submitted to the Cloud, while in other cases data will be normalised, standardised and persisted in a local database to subsequently be submitted to the Cloud. In the On-premise environment there is an Industrial PLC that provides resources with which the information from the automation is obtained from the sensors of the machines. This computer executes the following modules:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Machine2Raw:</emphasis> This is the system in charge of extracting the data from the different PLCs through a PLC obtaining, in turn, the data from the different machines of the customers;</para></listitem>
<listitem>
<para><emphasis role="strong">Datalogger:</emphasis> This system is responsible for storing in a local database the data coming from the sensors and the systems that are being monitored;</para></listitem>
<listitem>
<para><emphasis role="strong">Local Database:</emphasis> A local database in SQL server to allocate the raw data structures provided by the sensors. In this database there are some triggers to centralize the information in a table that is the entry point for Apache NiFi;</para></listitem>
<listitem>
<para><emphasis role="strong">Apache NIFI</emphasis> [The Apache Software Foundation, 2018]<emphasis role="strong">:</emphasis> A technology that processes the raw data stored in the local database by the DataLogger. This system allows defining data-flows in a visual way. It is fault-tolerant, has a low-latency and is able to manage a high volume of data. After processing the data, Apache NiFi transfers the data to the cloud through a proxy.</para></listitem>
</itemizedlist>
</section>
<section class="lev4" id="sec7-3-3-1-2">
<title>7.3.3.1.2 Results</title>
<para>As a first result, FAGOR ARRASATE ended up having a cloud platform (<link linkend="F7-46">Figure <xref linkend="F7-46" remap="7.46"/></link>) to monitor the status of the press machine park running on their customers&#x02019; premises. This platform provides several functionalities to create the network of press machines, to collect and monitor data from selected components, to analyse them and triggering alarms.</para>
<fig id="F7-46" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-46">Figure <xref linkend="F7-46" remap="7.46"/></link></label>
<caption><para>FALINK Cloud Platform.</para></caption>
<graphic xlink:href="graphics/ch07_fig045.jpg"/>
</fig>
<para>For operating the cloud platform, a control room (<link linkend="F7-47">Figure <xref linkend="F7-47" remap="7.47"/></link>) is set up at FAGOR&#x02019;s headquarters in Arrasate. This will allow the company to offer new maintenance services to its customers looking for increasing their press machines performance and availability. This way, the company aims to strengthen their market position and to create new business opportunities.</para>
<fig id="F7-47" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-47">Figure <xref linkend="F7-47" remap="7.47"/></link></label>
<caption><para>FAGOR ARRASATE future control room.</para></caption>
<graphic xlink:href="graphics/ch07_fig046.jpg"/>
</fig>
</section>
</section>
<section class="lev3" id="sec7-3-3-2">
<title>7.3.3.2 Torque measurement using wireless sensors</title>
<para>Press machines manufacturers are confronted with increasing technological and cost pressure. Many customers demand faster and more precise presses. The more precisely force is applied in a press machine, the higher is the quality of the parts manufactured. Thus, increasing the press machines accuracy is one of the most important challenges for manufacturers of these assets. In addition, the market requires increasingly faster press machines that, at the same time, offer higher bandwidth to increase production output in existing systems.</para>
<para>Nowadays, the torque of the press gear shaft is measured indirectly from the force that is applied in the connecting rod. This measure is quite precise but it needs to be continuously recalibrated to keep it accurate. To solve this problem, the torque is measured directly by using wireless sensors placed in the press gear shaft.</para>
<section class="lev4" id="sec7-3-3-2-1">
<title>7.3.3.2.1 Solution approach</title>
<para>As a solution, IKERLAN has designed and manufactured a prototype of a shaft-adapted wireless sensor node that comprises a transducer based on torque oriented gauges, a signal conditioning circuit and a signal processing software, the latter allowing local preprocessing and treatment of the collected data by means of intelligent functions.</para>
<para>The design process has been made following two main phases:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Phase 1: Testbed validation</emphasis></para></listitem>
</itemizedlist>
<para>Before starting the development of the wireless torque sensor, a preliminary validation step was made in testbeds both in IKERLAN and in the Try-Out press machine of FAGOR ARRASATE. This was an initial requirement to ensure the proper functioning of gauges, generic electronics and wireless communication for working in press-based conditions.</para>
<para>Two types of strain gauges were used: FCT strain gauges designed for torque measurement and FCA gauges chosen as a pair of strain gages oriented at 90. Also, a generic signal conditioning circuit was used where the signal is then processed by a low power microcontroller, which transmits data wirelessly with a radio module. The data is received by a similar access point, which is controlled by a LabVIEW National Instruments Corporation interface (configuration and visualisation).</para>
<para>Initial tests were carried out on two testbeds scenarios. In the first testbed (<link linkend="F7-48">Figure <xref linkend="F7-48" remap="7.48"/></link>), different weights, which correspond to corresponding micro strains, were loaded on the bar and static measurements were performed in half bridge and full bridge configuration for gauge calibration. In the second one (<link linkend="F7-49">Figure <xref linkend="F7-49" remap="7.49"/></link>), the test was made using a motorized test bench. To check if measurements were suitable, the electric engine speed was slowly increased resulting in increases of the measured torque.</para>
<fig id="F7-48" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-48">Figure <xref linkend="F7-48" remap="7.48"/></link></label>
<caption><para>Static testbed scenario.</para></caption>
<graphic xlink:href="graphics/ch07_fig047.jpg"/>
</fig>
<fig id="F7-49" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-49">Figure <xref linkend="F7-49" remap="7.49"/></link></label>
<caption><para>Dynamic testbed scenario.</para></caption>
<graphic xlink:href="graphics/ch07_fig048.jpg"/>
</fig>
<para>With regards to wireless communications, two main challenges were tested: (i) signal attenuation due to the rotation of the emitter around the shaft and (ii) multipath fading due to RF signal reflections in the metallic (steal) elements of the head of the press in which the torque sensor will be installed. Tests were successful, taking into account that depending on the angular position of the shaft, and therefore, on the relative position of the transmission and reception antennas, more or less amount of power is received periodically.</para>
<para>A similar test has been performed in the Try-Out press machine from FAGOR ARRASATE. In this case, both the emitter and the reception antenna have been placed in a realistic place within the head of the press machine as in can be seen in <link linkend="F7-50">Figure <xref linkend="F7-50" remap="7.50"/></link>.</para>
<fig id="F7-50" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-50">Figure <xref linkend="F7-50" remap="7.50"/></link></label>
<caption><para>Measurement setup in the press machine.</para></caption>
<graphic xlink:href="graphics/ch07_fig049.jpg"/>
</fig>
<para>Once the top cover is closed, creating a complete metallic case, it was observed how the received signal was not as clean as the one in the previous measurements due to multipath reflections. The statistical features obtained from this signals were used in the selection of the most suitable wireless communication technology to be used for the torque sensor.</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Phase 2: Design and development</emphasis></para></listitem>
</itemizedlist>
<para>Once the concept and the elements of the device (gauges, conditioning and processing, radio) were validated in a rotational environment, the system design and development was started taking into account the following Try-Out press specifications:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Shaft material: F1140 (C45E) steel;</para></listitem>
<listitem>
<para>Shaft dimensions:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Diameter: &#x003A6; 310.07 mm</para></listitem>
<listitem>
<para>External diameter: &#x003A6; 360 mm</para></listitem>
<listitem>
<para>Width: 150 mm</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>System thickness: 25mm max</para></listitem>
<listitem>
<para>Electronics &#x00026; Cover:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>No screwing</para></listitem>
<listitem>
<para>Speed: 88 rpm</para></listitem>
<listitem>
<para>Expected torque:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>188762 Nm</para></listitem>
<listitem>
<para>&#x0223C;200 microstrain</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist></listitem>
<listitem>
<para>Environment:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Temperature: &#x0003C;45<superscript>&#x02218;</superscript>C</para></listitem>
<listitem>
<para>Subjected to oil: CLP-150ftesp @ 400cm3/mm</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para>Placement: Head of the press machine (see <link linkend="F7-50">Figure <xref linkend="F7-50" remap="7.50"/></link>)</para></listitem>
</itemizedlist>
<para>From these specifications, a prototype of the wireless sensor node was designed and developed. It consists of a single PCB with the necessary interfaces to attach torque gauges, besides the conditioning, processing and wireless communication electronics. The whole system is powered by a rechargeable lithium ion polymer battery and it is encapsulated and protected by a plastic cover in the shape of the press&#x02019; secondary driving shaft, which is prepared to avoid oil leakage (see <link linkend="F7-51">Figure <xref linkend="F7-51" remap="7.51"/></link>).</para>
<fig id="F7-51" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-51">Figure <xref linkend="F7-51" remap="7.51"/></link></label>
<caption><para>Wireless sensor node.</para></caption>
<graphic xlink:href="graphics/ch07_fig050.jpg"/>
</fig>
<para>In order to configure the system and show the measured data, an user interface was designed in LabVIEW. From this interface, the Gage Factor, Poisson Ratio, Young modulus and the bar diameter can be configured. Typically the amount of received data is huge, so data values are averaged and only the average value is visualised in the user interface.</para>
</section>
<section class="lev4" id="sec7-3-3-2-2">
<title>7.3.3.2.2 Results</title>
<para>Once the design and fabrication of the wireless torque sensor was finished, the sensor was installed in the Try-Out press machine from FAGOR ARRASATE. First tests regarding the overall performance of the sensor were successful providing signals with the torque measurements were sent to an external laptop were they could be visualized (<link linkend="F7-52">Figure <xref linkend="F7-52" remap="7.52"/></link>).</para>
<fig id="F7-52" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-52">Figure <xref linkend="F7-52" remap="7.52"/></link></label>
<caption><para>Torque measures.</para></caption>
<graphic xlink:href="graphics/ch07_fig051.jpg"/>
</fig>
<para>Later, the complete validation process was carried out. This process aimed to test de accuracy of the sensor&#x02019;s measurement against several torque and speeds and the robustness of the wireless communication protocol employed.</para>
<para>15 different tests were carried out combining 30%, 60% and 87% of the nominal torque of the press, 57%, 78% and 100% of the nominal speed and several configurations of the sensing electronics. These results were compared with an estimation of the torque at the drive shaft obtained from an overload pressure evolution analysis. Besides, some measurements regarding the performance of the wireless communication were also taken.</para>
<para><link linkend="F7-53">Figure <xref linkend="F7-53" remap="7.53"/></link> shows the results of the test in which the maximum torque (87%) and the maximum speed (100%) were configured at the press machine.</para>
<fig id="F7-53" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-53">Figure <xref linkend="F7-53" remap="7.53"/></link></label>
<caption><para>Comparison between estimated and measured torque values (87% of the nominal torque, 100% of the nominal speed and gain 1000).</para></caption>
<graphic xlink:href="graphics/ch07_fig052.jpg"/>
</fig>
<para>The measured torque values at almost each stroke are close to 60 kNm, which fit the estimated torque values. Moreover, the clutch brake engage and disengage events were captured.</para>
<para>In general terms, it is considered that the obtained results are valid, taking into account that they are compared with estimated values and not with another measurement obtained by a commercial system. However, regarding the amount of data shown at the measured torque values, some data can be missed either on the positive or the negative peaks, as the same amplitude should be acquired for each stroke. With regards to wireless communications, in general the expected performance in terms of data throughput and network availability has been achieved. However, the loss of some data packets has been detected, which should be corrected in future versions.</para>
<para>As future work, the use of antenna diversity inside the shell of the press machine will improve the communication between emitter and receiver hence decreasing the number of packets lost. Besides, being the energy management of the system a key feature if it is pretended to leave it permanently attached to the press machine&#x02019;s drive shaft, a more energy efficient redesign will be carried out together with the development of an energy harvesting system to power up the wireless sensor node.</para>
</section>
</section>
<section class="lev3" id="sec7-3-3-3">
<title>7.3.3.3 Head structural health monitoring</title>
<para>The press structural components are welded steel structures where, on rare occasions, cracks may appear. The crack initiation is usually associated to fatigue damage in the welds and weld transitions (maximum design load is not exceeded due to the overload security devices). Fatigue is a cumulative phenomenon due to fluctuating loads, when material is subjected to repeated loading and unloading. The nominal stress for such loads may be much less than the ultimate tensile stress limit of the yield stress. If the loads are above a certain threshold, microscopic cracks begin to form at the stress concentrators. Eventually, a crack could reach a critical size, suddenly spread and provoke the fracture of the structure.</para>
<section class="lev4" id="sec7-3-3-3-1">
<title>7.3.3.3.1 Solution approach</title>
<para>Physics based degradation models (see Section 3 of <link linkend="ch05">Chapter <xref linkend="ch05" remap="5"/></link> on RUL) are implemented to detect most damaged zones of the press structural components. Based on the classical high cycle fatigue damage model, a damage indicator map is implemented. A damage threshold is set (Damage threshold = 1.3, for example), which is associated with a minimum length crack (2 mm, for example) appearance probability. Based on the measurements of real forces in the press rods and the stresses calculated at every point with Finite Element Models, 3 methods recommended by the International Institute of Welding are used to calculate the damage indicators at the welded structural components (mean stress, hot spot and notch stress).</para>
<para>As the real forces are being measured, dynamic and asymmetric loading effects are taken into account and accumulated over time. At the same time the damage indicator is calculated, the remaining time to reach the predefined threshold (see Section 3 of <link linkend="ch05">Chapter <xref linkend="ch05" remap="5"/></link> on RUL) is also calculated. Identification of unexpected premature occurrences can easily be identified and analyse probable associated Root Cause (see Section 2 of <link linkend="ch05">Chapter <xref linkend="ch05" remap="5"/></link> on RCA).</para>
<para>Additionally, for certain cases, a minimum crack length is supposed and a second physics based degradation model is applied to study the fracture mechanics. This is the field of mechanics concerned with the study of the propagation of cracks in materials. During the second stage of crack propagation or stable propagation stage, Paris&#x02019; law is used to estimate the crack propagation under certain load. The time needed to reach a threshold critical length is calculated. This is interesting when a crack is detected and the evolution needs to be estimated in order to schedule the corrective maintenance action.</para>
<para>Two different crack sensors were tested to detect or measure crack propagation: a commercial local crack gauge and a conductive ink sensor at a stage of development. Once a crack has been localised, the RUL to a critical crack length is calculated. If crack length data is available, Particle Filter method is used. This method combines the physical model and the available measurements in order to improve the RUL estimation to the critical crack length.</para>
<para>Algorithms and sensors are applied in a testbed prototype before the final application in the press machine head, shown in <link linkend="F7-54">Figure <xref linkend="F7-54" remap="7.54"/></link>. As a result of this, the structural health of the head of the press machine is monitored by means of two developments: i) The setup of a testbed for structural failure prediction and simulation and, ii) the analysis of conductive inks for crack detection.</para>
<fig id="F7-54" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-54">Figure <xref linkend="F7-54" remap="7.54"/></link></label>
<caption><para>Selected structural detail from press head.</para></caption>
<graphic xlink:href="graphics/ch07_fig053.jpg"/>
</fig>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Testbed for structural failure prediction and simulation</emphasis></para></listitem>
</itemizedlist>
<para>Due to the difficulty to artificially create structural failures in a real press, the press head Structural Health Monitoring scenario has two demonstrators:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>A fatigue testbed, where welded structural details are tested until complete fracture. The algorithms and sensors are applied and checked in this demonstrator prior to applying them in a real press machine;</para></listitem>
<listitem>
<para>A real press, where a structural damage and associated RUL indicator is applied, taking into account the results obtained in the fatigue testbed demonstrator and the features of the real press.</para></listitem>
</itemizedlist>
<para>In the case of the testbed, a fully sensor welded specimen has been submitted to fatigue loading until its complete fracture. The welded specimen is selected according to a structural detail located in the press head. The material of the specimen is the same of the press head as well as the welding procedure. The thicknesses of the sheets have been reduced according to the testbed load capacity.</para>
<fig id="F7-55" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-55">Figure <xref linkend="F7-55" remap="7.55"/></link></label>
<caption><para>Fully sensorised welded specimen in the testbed.</para></caption>
<graphic xlink:href="graphics/ch07_fig054.jpg"/>
</fig>
<para>A Finite Element Model of the specimen (<link linkend="F7-55">Figure <xref linkend="F7-55" remap="7.55"/></link>) was built in order to estimate the stresses at any location of the welded specimen.</para>
<para>The damage and RUL are estimated every minute during the test. Complete sensor data for a period of 20 seconds were stored every hour. Some results to remark are that the complete fracture occurs at 3.1E6 cycles, 3.8 times compared to the design life for 95% (8.4E5). The macrocrack initiation (failure) is estimated to occur near 1.25E6 cycles, 50% above the estimated design life. Crack initiation occurs where predicted by the FE model and the weld damage methods (<link linkend="F7-56">Figure <xref linkend="F7-56" remap="7.56"/></link>).</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Conductive inks</emphasis></para></listitem>
</itemizedlist>
<para>The objective is to proof the concept of using conductive inks to detect cracks in a mechanical test specimen. Tests with several conductive inks have been carried out. Different circuits are painted by this conductive ink on top of a test specimen coated manually with an insulating layer (e.g., Magnesia 919). The specimen is heated at different temperatures to increase the ink conductivity. During this heating step, cracks appeared in the insulating layer, which is interesting for the application since the objective is to study the bottom crack effect in the conductive ink. It is observed that it is difficult to apply the conductive ink homogeneously along the insulating layer surface.</para>
<fig id="F7-56" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-56">Figure <xref linkend="F7-56" remap="7.56"/></link></label>
<caption><para>Stress plot in the testbed due to a unitary load.</para></caption>
<graphic xlink:href="graphics/ch07_fig055.jpg"/>
</fig>
<para>After these preliminary tests, a conductive ink is deposited on top of a mechanical test specimen prior to subjecting it to a mechanical fatigue test to create a crack, which should be detected by the conductive ink.</para>
<para>First, the &#x0201C;silver conductive adhesive pro&#x0201D; from RS is deposited on top of a mechanical test specimen shown in <link linkend="F7-57">Figure <xref linkend="F7-57" remap="7.57"/></link>, next to the pre crack, in perpendicular direction. First of all, an insulating layer is bonded directly on the test specimen surface. Secondly, different insulating adhesive layers are tested: a kapton adhesive film (A) and a Teflon adhesive film (B). Then, a rectangle of silver conductive ink (4 x 40 mm2) is painted on top of each insulating layer, using the paint brush provided by the ink. Finally, electric cables are bonded in the four ends using a silver conductive epoxy adhesive.</para>
<fig id="F7-57" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-57">Figure <xref linkend="F7-57" remap="7.57"/></link></label>
<caption><para>Mechanical test specimen with two sensors based on silver conductive adhesive deposited directly on top of the surface.</para></caption>
<graphic xlink:href="graphics/ch07_fig056.jpg"/>
</fig>
<para>Regarding the experimental conditions for fatigue test, a mechanical test specimen made of S235JR steel material and CT Compact Tension geometry is created. The stress is applied in a different way in two different phases: (i) 225000 cycles of 0.8&#x02013;8 kN with a frequency of 10 Hz and (ii) 109000 cycles of 1&#x02013;10 kN with a frequency of 10 Hz. In both phases, every 1000 cycles, a constant force is applied (F = (Fmax + Fmin)/2) during 10 seconds. The subsequent post processing is carried out with data acquired in these intervals of 10 seconds. The sampling frequency is 20Hz.</para>
<para>Apart from the signal of conductive inks, two other signals are acquired in the test, in order to be able to compare afterwards the results with a reference (a commercial crack detection gauge and a commercial extensometer, see <link linkend="F7-58">Figure <xref linkend="F7-58" remap="7.58"/></link> for all the elements used in the test, and <link linkend="F7-59">Figure <xref linkend="F7-59" remap="7.59"/></link> for the final set-up for the experiments).</para>
<fig id="F7-58" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-58">Figure <xref linkend="F7-58" remap="7.58"/></link></label>
<caption><para>(a) Two sensors based on conductive inks placed in one side of the specimen; (b) Commercial crack detection gauge placed on the opposite side of the specimen; (c) Commercial extensometer placed in the crack, during the fatigue test.</para></caption>
<graphic xlink:href="graphics/ch07_fig057.jpg"/>
</fig>
<fig id="F7-59" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-59">Figure <xref linkend="F7-59" remap="7.59"/></link></label>
<caption><para>Picture of experimental set-up.</para></caption>
<graphic xlink:href="graphics/ch07_fig058.jpg"/>
</fig>
</section>
<section class="lev4" id="sec7-3-3-3-2">
<title>7.3.3.3.2 Results</title>
<para>Structural damage and associated RUL indicator based on fatigue damage are ready to be integrated in a real press machine. Stress at critical positions is obtained from a finite element model and online experimentally measured forces.</para>
<para>With regards to conductive inks, the following results were achieved. The blue line in <link linkend="F7-60">Figure <xref linkend="F7-60" remap="7.60"/></link> corresponds to ink B, which is placed closer to the pre-crack and is deposited on top of the Teflon insulating layer. On the other hand, the red line corresponds to ink A, which is placed farther away from the pre-crack and is deposited on top of the Kapton insulating layer.</para>
<para>The signal increase of ink B indicats the presence of a crack. However, this increase started too late, when the tests specimen was completely broken. The behaviour of ink A is similar, where the signal began to increase later, again when the test specimen was completely broken.</para>
<fig id="F7-60" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-60">Figure <xref linkend="F7-60" remap="7.60"/></link></label>
<caption><para>Ink signal during the second phase of fatigue test and ink pictures corresponding to the two points indicated in the graph. Pictures of both sides of the test specimen, at the end of the fatigue test.</para></caption>
<graphic xlink:href="graphics/ch07_fig059.jpg"/>
</fig>
<para>It is concluded that although apparently the ink performance is correct, the bottom insulating layer behaviour is not as expected. It is too flexible and does not transfer in a correct way the crack from the test specimen to the ink. Both insulating layers are extended too much before cracking due to their high flexibility.</para>
<para>Direct crack detection methods are going to be tested (ink sensors and crack gauges, acoustic emission sensors) and an indirect model based damage detection method (Extended Constitutive Relation Error approach). This method compares the measured strain at different moments of the system with the ones expected by the model. It is a method to identify, localize and determine the severity of the damage.</para>
<para>Three research lines of interest are identified:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Probabilistic approach of the fatigue damage;</para></listitem>
<listitem>
<para>Stress estimation based on model and sensor data. Improvement of stress estimation and hence, structural damage and associated RUL indicators;</para></listitem>
<listitem>
<para>Crack length estimation based on fracture mechanics and RUL estimation using particle filter based prognostics algorithm.</para></listitem>
</itemizedlist>
<para>The last two will improve the estimation of the stress or the crack length by combining physical models and experimental data.</para>
</section>
</section>
<section class="lev3" id="sec7-3-3-4">
<title>7.3.3.4 Bushing status measurement</title>
<para>Bushings are critical parts in press machines to reduce friction between rotating shafts and stationary support members. Depending on the working conditions and oil status, the bushing can increase its temperature getting stuck with the connecting rod. This failure forces to stop the stamping process and the time to repair it is about one working week. Taking into account that the one just described is the best case scenario when there are spare parts available in stock, due to the magnitude of the problem, it was necessary to tackle it within the MANTIS project.</para>
<section class="lev4" id="sec7-3-3-4-1">
<title>7.3.3.4.1 Solution approach</title>
<para>It is considered that bushing failure due to seizure can be anticipated, and that it is possible to estimate its RUL by collecting and analysing defined measured data. Bushings were tested and run to failure at different ranges of working conditions in a test bench. A wear out model is created based on the time data obtained from the sensors and control installed in the test bench. For the analysis, regression based models are used as a first approach.</para>
<para>Bushing temperature and oil sensor signal based alarms are set in order to prevent seizure failure. This is done by limit and trending checking. A physical model represented in <link linkend="F7-61">Figure <xref linkend="F7-61" remap="7.61"/></link> and based on DIN 31652 has been programmed in a simulator to characterise the theoretical behaviour of the bushing when it is working in hydrodynamic ideal conditions. For some given working conditions, the model calculates the expected equivalent friction, temperature rising in the bushing as well as the lubrication through the oil film thickness (<link linkend="F7-62">Figure <xref linkend="F7-62" remap="7.62"/></link>). This is done in two steps:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Firstly, it is checked if the model describes the real behaviour of the bushings during the tests;</para></listitem>
<listitem>
<para>Secondly, the model is used to define a safety working condition where bushing seizure should not occur. The main idea lies in detecting bushing abnormal behaviour when temperature alarm triggers in the defined safety working conditions.</para></listitem>
</itemizedlist>
<fig id="F7-61" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-61">Figure <xref linkend="F7-61" remap="7.61"/></link></label>
<caption><para>Bushing seizure model.</para></caption>
<graphic xlink:href="graphics/ch07_fig060.jpg"/>
</fig>
</section>
<section class="lev4" id="sec7-3-3-4-2">
<title>7.3.3.4.2 Results</title>
<para>Some improvements were made to the test bench in order to improve the automation of the tests and lubrication conditions. Further tests are carried out in different working conditions and bushing types to complete the correlation between the measurements and the model, characterising the theoretical safe working conditions zone. Additionally, analysis is conducted to establish the safe working condition area based on real data measured from the tests.</para>
<para>Results of this analysis are expected to be applied in the future to press machines by installing temperature and oil sensors in critical bushings for collecting, monitoring and analysing real time data. This is in line with the company strategy of minimising these kinds of incidents.</para>
</section>
</section>
<section class="lev3" id="sec7-3-3-5">
<title>7.3.3.5 Gears wear measurement</title>
<para>Another failure that happens in press hardening machinery is that, depending on the force and working conditions of the press, the gear can be damaged. Sometimes this damage is caused by wear with the passing of the years or working hours, sometimes the problem can raise much earlier caused by bad working conditions.</para>
<section class="lev4" id="sec7-3-3-5-1">
<title>7.3.3.5.1 Solution approach</title>
<para>To predict that the gear is wearing out and must be replaced, it is possible to analyze the oil condition. Taking into account that the more hours the press works the more metal particles appears in the oil, it is possible to predict if the gear is wearing by analysing the oil itself.</para>
<para>The selected sensor for acquiring and monitoring the particles present in fluids is the OilWear S100. This sensor can classify particles larger than 20 &#x003BC;m according to their size and shape, to determine the root cause: fatigue, sliding or cutting.</para>
<fig id="F7-62" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-62">Figure <xref linkend="F7-62" remap="7.62"/></link></label>
<caption><para>OilWear sensor.</para></caption>
<graphic xlink:href="graphics/ch07_fig061.jpg"/>
</fig>
<para>The OilWear S100 is located in the hydraulic tube from where the oil returns to the tank.</para>
<para>Data is captured every scan cycle. The different data values are stored on a local server.</para>
<para>Once data is storage in the local server, data pre-processing is done in local mode, taking into account requited parameters such as the maximum and minimum values as KPIs. After data pre-processing, particles data that FAGOR ARRASATE considers are critical for their press machine working conditions will be analyse using specific algorithms for that.</para>
<para>Apart from acquiring oil particles characteristics data, some other data is acquired, in order to know exactly the main reason of the gears wear:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Die reference;</para></listitem>
<listitem>
<para>Press machine Total Strokes;</para></listitem>
<listitem>
<para>Die Total Strokes;</para></listitem>
<listitem>
<para>Press machine Speed: Stroke Per Minute;</para></listitem>
<listitem>
<para>Press machines maximum force;</para></listitem>
<listitem>
<para>Stamping force.</para></listitem>
</itemizedlist>
</section>
<section class="lev4" id="sec7-3-3-5-2">
<title>7.3.3.5.2 Results</title>
<para>The OilWear S100 sensor is now installed in FAGOR ARRASATEs press machine. During next months, data will be captured and preprocessed locally tacking into account defined KPIs.</para>
<para>The next step will be to integrate the data sources within the cloud platform.</para>
</section>
</section>
<section class="lev3" id="sec7-3-3-6">
<title>7.3.3.6 Press forces measurement</title>
<para>The ram force is a fundamental parameter that affects the final quality of the produced workpieces. Furthermore, its deviations could cause damage to the press machine&#x02019;s components. Besides, in presses with multiple cranks, an unbalanced forces could appear due to an imbalance of the cranks or other components, affecting both the quality of the produced workpieces and the integrity of the press. The cutting shock effect is another undesired effect that has to be taken into account during any process of metal forming. If a big enough cutting shock is exerted, many components of the press can suffer damages.</para>
<para>These force measurements usually are carried out by hardware sensors located throughout the press structure and tooling, whose calibration loss are caused by the strongest forces the press experiences during its life cycle.</para>
<section class="lev4" id="sec7-3-3-6-1">
<title>7.3.3.6.1 Solution approach</title>
<para>In order to overcome the limitations associated to the hardware sensors, indirect measurements are proposed by means of model based soft sensors that leverage the existing signals and the knowledge about the process that the system performs.</para>
<para>The servo driven press machine is modelled as slider - connecting rod - crank mechanism, considering also the gearbox between the connecting rod shaft and the servomotor shaft. Along with the servo driven press machine model signals are measured in order to use them as inputs for the model, such as the servomotor current signals, voltage signals and rotor position signal. A <emphasis>Prediction Error Method</emphasis> based soft sensor is used for estimating the coefficients of a friction model that completes the model of the system.</para>
<fig id="F7-63" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-63">Figure <xref linkend="F7-63" remap="7.63"/></link></label>
<caption><para>Servo driven press machine model diagram.</para></caption>
<graphic xlink:href="graphics/ch07_fig062.jpg"/>
</fig>
<para>As visible in <link linkend="F7-63">Figure <xref linkend="F7-63" remap="7.63"/></link>, the servo driven press machine model is formed by three sub-models, each one generating a torque that interacts with the rest of the system.</para>
<para>After estimating the coefficients of the friction model, the whole model is utilised for estimating important coefficients that are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Press ram/slider force;</para></listitem>
<listitem>
<para>Cutting shock effect;</para></listitem>
<listitem>
<para>Unbalanced forces.</para></listitem>
</itemizedlist>
<para><emphasis role="underline">Model</emphasis></para>
<para>Initially the system model mathematical representation is developed applying Euler-Lagrange function considering all the mechanical elements of the system. The shortened mathematical model is reported in Equation (7.5):</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq7-5.jpg"/></para>
<para>Where <emphasis>M(&#x03B8;)</emphasis>, <emphasis>N(&#x03B8;)</emphasis> and <emphasis>O(&#x03B8;)</emphasis> represent the inertia and mass of the gearbox, crank, connecting rod and the slider of the system. Respectively, &#x003C4;<subscript><emphasis>e</emphasis></subscript> is the electric torque, &#x003C4;<subscript><emphasis>lb</emphasis></subscript> is the load balancer torque and &#x003C4;<subscript><emphasis>fric</emphasis></subscript> is the torque exerted by the friction of the system. At this point, the <emphasis>Prediction Error Method</emphasis> based soft sensor is applied for estimating the aforementioned friction model coefficients, yielding a friction related torque that acts against the electric torque. For this purpose, unladen tests (without strokes) are carried out at different press speeds.</para>
<fig id="F7-64" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-64">Figure <xref linkend="F7-64" remap="7.64"/></link></label>
<caption><para>Estimated and actual angular positions.</para></caption>
<graphic xlink:href="graphics/ch07_fig063.jpg"/>
</fig>
<para>Once the model is fitted, the estimated angular position is plotted against the measured one, tracing a path similar to the measured one, as shown in <link linkend="F7-64">Figure <xref linkend="F7-64" remap="7.64"/></link>.</para>
</section>
<section class="lev4">
<title>Estimation of the objectives</title>
<para>The estimation of the previously mentioned objectives is performed adding another stroke related torque to the Equation (7.1), which yields the Equation (7.6):</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq7-6.jpg"/></para>
<para>where &#x003C4;<subscript><emphasis>S</emphasis></subscript> represents the torque generated by the slider stroke. This new term collects the applied force during a cycle, and thus, many metal forming processes are monitored through the measured signals, the model (<link linkend="F7-65">Figure <xref linkend="F7-65" remap="7.65"/></link>), and a soft sensor.</para>
<fig id="F7-65" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-65">Figure <xref linkend="F7-65" remap="7.65"/></link></label>
<caption><para>Step by step Bayesian soft sensor estimation performance diagram.</para></caption>
<graphic xlink:href="graphics/ch07_fig065.jpg"/>
</fig>
<para>In order to estimate the states (position and velocity of the slider, process force) of the model, a step by step Bayesian soft sensor is used, which can perform real time estimations of the system states. On a first stage, all the system inputs are recorded in advance and are used in a computer environment. The Bayesian soft sensor takes system inputs and measured outputs and is able to estimate unmeasured system states as shown in <link linkend="F7-65">Figure <xref linkend="F7-65" remap="7.65"/></link>.</para>
<fig id="F7-66" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-66">Figure <xref linkend="F7-66" remap="7.66"/></link></label>
<caption><para>Estimated electric torque for different tonnage tests.</para></caption>
<graphic xlink:href="graphics/ch07_fig066.jpg"/>
</fig>
<para>The three estimation objectives of the use case are related to the slider stroke force.</para>
</section>
<section class="lev4" id="sec7-3-3-6-2">
<title>7.3.3.6.2 Results</title>
<para>Preliminary results of measured electric magnitudes of the servomotor are discussed in this section. Some of those test results are displayed in Figures 7.66 and 7.67 where the relation between electric magnitudes of the servomotor and the applied ram force is shown.</para>
<fig id="F7-67" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-67">Figure <xref linkend="F7-67" remap="7.67"/></link></label>
<caption><para>Calculated active power for different tonnage tests.</para></caption>
<graphic xlink:href="graphics/ch07_fig067.jpg"/>
</fig>
<para>These tests reveal that the servomotor magnitudes (electric torque and active power) change accordingly with the applied process force, feature that is used for estimating the applied real force directly from measurement of the servomotor magnitudes.</para>
<para>Regarding the cutting shock effect, <link linkend="F7-68">Figure <xref linkend="F7-68" remap="7.68"/></link> shows the estimated electric torque and its corresponding tonnage for two cutting shock tests. In the test where the narrowest metal sheet is cut, the cutting process finishes earlier comparing to the widest metal sheet cutting test and besides, the generated cutting shock compensation electric torque is smaller than in the other test. The image on the right side of <link linkend="F7-68">Figure <xref linkend="F7-68" remap="7.68"/></link> depicts the applied force and the cutting shock effect quantification which are of 5 and 21 tons for the 300mm and 500mm wide metal sheets respectively.</para>
<fig id="F7-68" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-68">Figure <xref linkend="F7-68" remap="7.68"/></link></label>
<caption><para>Estimated electric torque for 300mm and 500mm wide tough metal sheet cutting process.</para></caption>
<graphic xlink:href="graphics/ch07_fig068.jpg"/>
</fig>
<para><link linkend="F7-69">Figure <xref linkend="F7-69" remap="7.69"/></link> displays the consumed active power by the PMSM. As in <link linkend="F7-68">Figure <xref linkend="F7-68" remap="7.68"/></link>, the cutting process and the cutting shock effect compensation shape the PMSM power consumption.</para>
<fig id="F7-69" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-69">Figure <xref linkend="F7-69" remap="7.69"/></link></label>
<caption><para>Calculated active power for 300mm and 500mm wide tough metal sheet cutting process.</para></caption>
<graphic xlink:href="graphics/ch07_fig069.jpg"/>
</fig>
<para>With respect to the unbalanced forces, tests have not revealed any difference looking at the analysed electric magnitudes between the balanced and unbalanced process. <link linkend="F7-70">Figure <xref linkend="F7-70" remap="7.70"/></link> shows a similar electric torque for the unbalanced and balanced processes. Estimated electric torques for unbalanced and balanced processes are also unable to determine differences between the two processes, as displayed in <link linkend="F7-71">Figure <xref linkend="F7-71" remap="7.71"/></link>.</para>
<fig id="F7-70" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-70">Figure <xref linkend="F7-70" remap="7.70"/></link></label>
<caption><para>Calculated active power of centred and off-centred tests.</para></caption>
<graphic xlink:href="graphics/ch07_fig070.jpg"/>
</fig>
<fig id="F7-71" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-71">Figure <xref linkend="F7-71" remap="7.71"/></link></label>
<caption><para>Estimated electric torque for a centred and off-centred tests.</para></caption>
<graphic xlink:href="graphics/ch07_fig071.jpg"/>
</fig>
<para>As future work, the soft sensor approach will be employed once the final test is done in order to measure indirectly (estimate) the applied ram force.</para>
</section>
</section>
</section>
<section class="lev2" id="sec7-3-4">
<title>7.3.4 MANTIS Solutions for Clutch Brake</title>
<para>GOIZPER considers critical to increase machines and components reliability. To meet this challenge GOIZPER decided to incorporate cutting-edge technologies in their products as a means of enhancing product robustness and functionality in order to facilitate proactive-predictive maintenance activities.</para>
<para>GOIZPER decided to investigate different ways of sending data of its components to the Cloud. The main objective is to find the most robust and reliable architecture, and for this reason, two different data fluxes were developed in order to send the Clutch Brakes information to the cloud. One of the architectures is coordinated and developed by MGEP and the second architecture is coordinated by TEKNIKER. These two different approaches (at edge/sensor, and at platform level) are explained in detail in the next sections.</para>
<section class="lev3" id="sec7-3-4-1">
<title>7.3.4.1 Maintenance cloud platform by MGEP</title>
<para>The platform presented in this section is concerned with analysing a clutch brake system and its components in press machines to detect the most important failure sources and be able to perform predictive maintenance in those press machines. Analysis techniques and algorithms, to be used on the assets data, were implemented in the platform with the aim to support predictive maintenance of clutch-brake. These technologies are (1) Root Cause Analysis powered by Attribute Oriented Induction Clustering and (2) Remaining Useful Life powered by Time Series Forecasting. The implementation of that platform was previously published in a conference paper [Larrinaga et al., 2018].</para>
<section class="lev4" id="sec7-3-4-1-1">
<title>7.3.4.1.1 Background</title>
<para>The overall objective sought by GOIZPER is to early detect internal wear of a clutch-brake. To do that, the moving parts of the clutch-brake were sensorized. By continuously monitoring the system conditions proper operation of the clutch-brake can be ensured. Moreover, the most critical operating variables are registered in the platform in order to analyze the working process and prevent misuses. The data is uploaded enabling the holistic analysis of the clutch-brake system, with the aim to determine/detect the main causes of failure and the components&#x02019; remaining useful life.</para>
<fig id="F7-72" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-72">Figure <xref linkend="F7-72" remap="7.72"/></link></label>
<caption><para>Edge Tier.</para></caption>
<graphic xlink:href="graphics/ch07_fig072.jpg"/>
</fig>
</section>
<section class="lev4" id="sec7-3-4-1-2">
<title>7.3.4.1.2 Solution approach</title>
<para>The architecture implementation agrees with the three-tier architecture presented in [Hegedus, 2018] and in <link linkend="ch03">Chapter <xref linkend="ch03" remap="3"/></link>. There are three levels: Edge tier, Platform tier and Enterprise tier. The <emphasis role="strong">Edge Tier</emphasis> is concerned with the technological solution deployed on the sites where the Press Machines (including the Clutch Brake component) are located. At this level, data acquisition systems to extract the data from the different sensors and SCADA systems connected to the machines are deployed. <link linkend="F7-72">Figure <xref linkend="F7-72" remap="7.72"/></link> depicts the elements of this tier. An Industrial PLC based on the B&#x00026;R X20CP1382 module was connected to the sensors attached to the Clutch-Brake (including the intelligent soft-sensors). The module collects all the measurements from the sensors and runs local code to pre-process the signals and produce a set of parameters that are able to characterize the overall status of the Clutch Brake. The module stores these parameters in a local file to act as a Datalogger for the cyber physical system. A second embedded computer (Edge Gateway) is attached to the Datalogger, and it retrieves the parameter files and creates IoT-A CEP Events [Internet of Things Architecture] that are sent to the cloud platform as messages.</para>
<fig id="F7-73" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-73">Figure <xref linkend="F7-73" remap="7.73"/></link></label>
<caption><para>Platform Tier.</para></caption>
<graphic xlink:href="graphics/ch07_fig073.jpg"/>
</fig>
<para>At platform level, data coming from the different sources is persisted and different applications that allow analyzing of this data are available. The specific modules for the <emphasis role="strong">Platform Tier</emphasis> are presented in <link linkend="F7-73">Figure <xref linkend="F7-73" remap="7.73"/></link>, and are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Edge Brokers:</emphasis> It maintains the connection between the edge devices and edge tier, and it includes a data distributor. The distributor is a message-oriented component to collect and redistribute the in- and outbound messages between components. In this use case, this module is a publish/subscribe system that receives the data from edge tier in different queues and publish the message received to the modules in the platform that have subscribed to each queue. The technological solution for the Edge Broker is RabbitMQ. RabbitMQ [RabbitMQ, 2018] is an open source message broker that supports multiple messaging protocols [Albano et al., 2015] such as Advanced Message Queuing Protocol (AMQP) [Vinoski, 2006];</para></listitem>
<listitem>
<para><emphasis role="strong">Converters:</emphasis> Software components or modules to translate edge cloud interface (IoTA CEP Events) into a database interface (MIMOSA API Rest [MIMOSA Consortium, 2003]) or files system interface (HDFS API). The converter is implemented using the translation capabilities of an Enterprise Service Bus (ESB) named WSO2 [WSO, 2018];</para></listitem>
<listitem>
<para><emphasis role="strong">Data Storage systems</emphasis> store the information coming from Cyber Physical Systems (CPS) and results obtained from data analysis maintenance actions and algorithms. Two storage systems are employed:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>MIMOSA DB: This is a database compliant with the ISO-13374 Standard (Condition Monitoring and Diagnostic of Machines) [ISO, 2018]. One of the main objectives of the MIMOSA CBM architecture is to standardize the information flow between the various blocks, so that equipment from different vendors could be interoperable. The MIMOSA database is deployed in SQL Server and API REST is used to access data from applications;</para></listitem>
<listitem>
<para>Hadoop Distributed File System (HDFS): This is a distributed file system designed to run on commodity hardware. Designed to be deployed on low-cost hardware, HDFS is highly fault-tolerant and provides high throughput access to application data, which makes it suitable for applications that have large data sets;</para></listitem>
</itemizedlist>
</listitem>
<listitem>
<para><emphasis role="strong">Batch Processing:</emphasis> data analysis and processor mechanisms to enable the management of large volumes of data, fetched from storage systems and process on demand. This is implemented over Apache Spark [Apache, 2018]. The batch processor units implement the o&#x0FB04;ine analytics capabilities of the platform. These technologies are (1) RCA powered by Attribute Oriented Induction Clustering and (2) RUL powered by Time Series Forecasting;</para></listitem>
<listitem>
<para><emphasis role="strong">API or WS:</emphasis> To interact with the Enterprise Tier an API offering services is provided. This component provides information and functionality (for example RUL) to components external to the platform such as HMI, applications (ERP) or even other platforms that lack certain maintenance algorithms.</para></listitem>
</itemizedlist>
<para>The <emphasis role="strong">enterprise level</emphasis> is concerned with the applications that integrate information from one/several sites to enhance the global decision-making process using monitoring through Human Machine Interfaces (HMI) and data aggregation and analysis.</para>
<para>In relation to CBM-based PM the following aspects have been addressed for this scenario:</para>
<para><emphasis role="strong">Equipment Failure Root Cause Analysis:</emphasis> The RCA is the first and necessary step to identify the main equipment failure causes. An AOI algorithm is used as the principal RCA algorithm. AOI is considered a hierarchical clustering algorithm, it is considered a rule-based concept hierarchy algorithm, and it was first proposed by [Han et al., 1992] Jiawei Han et al. as a method for knowledge discovery in databases. The representation of the knowledge is structured in different generalization-levels of the concept hierarchy with IF-THEN rules. The execution of the algorithm AOI follows an iterative process in which each variable (also referred as attribute) is generalized based on its own hierarchy-tree. This step is denoted as concept-tree ascension [Cheung et al., 1994]. To ensure the correct functioning of the algorithm, it is necessary to establish background knowledge, which specifies attribute generalization levels.</para>
<para><emphasis role="strong">Equipment Remaining Useful Life estimation:</emphasis> The main objective of the RUL estimation process is to estimate the useful life of an asset before a catastrophic failure occurs. The RUL estimation process is performed as a combination of AOI algorithm outcome and Auto Regressive Integrated Moving Average (ARIMA) statistical time series forecasting models. A common objective of Time Series Forecasting methods is to learn from previous data in order to be able to make predictions of future behaviours. In order to estimate the RUL, the first step is to evaluate a new variable to represent the machine behaviour correction factor, denoted as Normality Factor. The Normality Factor quantifies the extent of the damage of the machine. By applying ARIMA time series forecasting models, the Normality Factor evolution is modelled. As a final result, the Normality Factor model allows to predict the wear of the Normality Factor, providing the machine RUL in terms of clutch-brake cycles. Finally, clutch-brake cycles are translated into days, by combining the number of cycles the clutch-brake system does per day.</para>
</section>
<section class="lev4" id="sec7-3-4-1-3">
<title>7.3.4.1.3 Results</title>
<para>Regarding the implementation of the reference architecture, a platform that accommodates different industrial processes and assets data for CBM analysis was built. The platform integrates an interoperable data model for CBM. Additionally a data/protocol converter that enables translations between most common data formats and protocols was developed.</para>
<para>Regarding data analysis, preliminary results performed as a proof of concept show the capability of the proposal. For the experiment, several features of the clutch-break machine have been used (trigger, angular position, application pressure, line pressure and flywheel speed). Once the knowledge-base has been created applying AOI and the most significant cluster-appearance order for the working cycles was calculated, the anomaly detection step is processed using the Normality Factor as threshold (value of 0.70). The Normality Factor evolution signal shown in <link linkend="F7-74">Figure <xref linkend="F7-74" remap="7.74"/></link>, is the result of applying ARIMA model over the training data utilized to generate the knowledge-base. In this experiment, around two hundred and fifty &#x02018;break&#x02019; working cycles have been predicted. As it can be observed, there are five different work cycles cutting the established Normality Threshold; thus, it can be inferred that five different anomalies were detected. The next step is to analyze the characteristics of the anomalies, inspecting the reasons of their occurrence. For example, if there is any cluster in the abnormal work cycle that is not registered in the knowledge-base, it is recommendable to check the features or the grips of features in which the new values have occurred in order to establish the reason of the failure; if the order of the clusters inside the abnormal working cycle is significatively different respect to the ones registered in the knowledge-base, it can be reasonable to check the evolution of the values of the features in order to specify the reasons of the failure.</para>
<fig id="F7-74" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-74">Figure <xref linkend="F7-74" remap="7.74"/></link></label>
<caption><para>Evolution of the Normality Factor over work cycles.</para></caption>
<graphic xlink:href="graphics/ch07_fig074.jpg"/>
</fig>
</section>
</section>
<section class="lev3" id="sec7-3-4-2">
<title>7.3.4.2 Maintenance cloud platform by Tekniker</title>
<para>As it is stated at the beginning of Section 7.3.4, the main objective of Tekniker&#x02019;s platform is the analysis of clutch-brake systems in order to detect failures. The platform supports Smart-G, a cyber-physical system that compiles critical process values and condition-related parameters, performs pre-processing based on algorithms specifically designed for this purpose, and offers a first level of monitoring and decision support directly back at the edge tier.</para>
<para>In addition, this valuable information recorded locally can be sent to the cloud platform where the user can access the entire historical information related to the use of the component. Therefore, the objective of this platform is to support the knowledge of GOIZPER and give predictive maintenance capabilities using different algorithms integrated in the system.</para>
<section class="lev4" id="sec7-3-4-2-1">
<title>7.3.4.2.1 Background</title>
<para>The main objective is to understand clutch-brake wear in order to give services and advices to the customers. All the critical signals are acquired, stored and processed in a PLC-based device called &#x0201C;Smart-G&#x0201D;, and then sent to the cloud platform, where they are stored and processed to predict failures and give visual decision support capabilities. Therefore, information on the condition of machine components and operating processes is recorded locally on the Smart-G devices, and transmitted remotely to reduce unforeseen downtime and increase equipment availability.</para>
</section>
<section class="lev4" id="sec7-3-4-2-2">
<title>7.3.4.2.2 Solution approach</title>
<para>The platform is built using the Microsoft Azure cloud services, and is represented in <link linkend="F7-75">Figure <xref linkend="F7-75" remap="7.75"/></link>.</para>
<fig id="F7-75" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-75">Figure <xref linkend="F7-75" remap="7.75"/></link></label>
<caption><para>Azure-based Maintenance Cloud Platform.</para></caption>
<graphic xlink:href="graphics/ch07_fig075.jpg"/>
</fig>
<para>The system is designed using a typical pattern in big data scenarios known as &#x0201C;lambda architecture&#x0201D;, with uses three layers to solve the computing problem: speed layer, batch layer and serving layer. The batch layer holds an immutable, read-only master database, and it pre-computes a batch view with indicators and aggregated data. The speed layer deals with recent data only and executes quick algorithms (rules and machine learning algorithms) to produce a speed view with alarms and predictions. The serving layer is composed of the batch view and the speed view mentioned before.</para>
<para>Exploitation and visualization of data relies on the Microsoft Power Business Intelligence capabilities to show aggregated information, indicators and transient raw data of the monitored assets.</para>
<para>Application of big data techniques, combined with machine learning for pattern identification, and complex event processing for the detection in real time of the learned patterns, is the approach for reaching a high ratio of availability and operational performance.</para>
<para>The Azure services used to build the platform are described below:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Azure Event Hubs: It is a highly scalable publish-subscribe message broker for event ingestion, with a partition-based approach to support the ingestion of millions of events per second;</para></listitem>
<listitem>
<para>Azure Stream Analytics: This is an event data processing service for real-time analysis of streaming data. It uses a SQL-like language to create rules, and can sent requests to the Azure Machine Learning Service to execute algorithms in real time;</para></listitem>
<listitem>
<para>Azure Machine Learning: It is a cloud service for the implementation of predictive analytical solutions. It provides a big number of built-in packages, and allows the customization of new ones;</para></listitem>
<listitem>
<para>Azure HDInsight: This is a highly scalable solution used to prepare the data in the batch layer. It allows the combination of data and statistical equations providing many possibilities for data enrichment;</para></listitem>
<listitem>
<para>Azure Blob Storage: It is a cloud service to store unstructured data as blobs. A variety of data files can be stored, for example binary, text, documents, multimedia files, etc.;</para></listitem>
<listitem>
<para>Azure SQL Database: This is a relational database that is used to store alarms and aggregated values.</para></listitem>
</itemizedlist>
<para>Additionally, a business intelligence tool (Microsoft Power Business Intelligence) is used for reporting and visualization of data. This tool facilitates the representation of information in attractive panels and reports that can be customized in a very flexible manner.</para>
</section>
</section>
<section class="lev3" id="sec7-3-4-3">
<title>7.3.4.3 Friction material slippage</title>
<para>Two kind of slippages can arise during operation, clutch side slippages and brake side slippages. Each of them are caused by different reasons. Clutch side slippages cause a transmitted torque loss to the output shaft, which is dissipated as heat. Brake side slippages also cause a transmitted torque loss and in turn, a delay of the shaft&#x02019;s braking time.</para>
<section class="lev4" id="sec7-3-4-3-1">
<title>7.3.4.3.1 Solution approach</title>
<para>In the proposed solution, many factors have been considered in order to identify slippages causes for each case. For the slippages that come up during clutching, air leakages and clutch side friction material degradation have been analysed. On the other hand, brake side slippages are produced due either to brake springs degradation or brake side friction material degradation.</para>
<para>Slippages can be detected directly from encoder velocity and acceleration signals as shown in Figures 7.76 and 7.77.</para>
<fig id="F7-76" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-76">Figure <xref linkend="F7-76" remap="7.76"/></link></label>
<caption><para>Clutching and braking velocity and acceleration without slippage.</para></caption>
<graphic xlink:href="graphics/ch07_fig076.jpg"/>
</fig>
</section>
<section class="lev4" id="sec7-3-4-3-2">
<title>7.3.4.3.2 Results</title>
<para><link linkend="F7-76">Figure <xref linkend="F7-76" remap="7.76"/></link> shows velocity and acceleration profiles when no slippage is generated. As it is noticed, clutching and braking velocity and acceleration curves are continuous.</para>
<fig id="F7-77" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-77">Figure <xref linkend="F7-77" remap="7.77"/></link></label>
<caption><para>Clutching and braking velocity and acceleration with slippage at clutching.</para></caption>
<graphic xlink:href="graphics/ch07_fig077.jpg"/>
</fig>
<para>In <link linkend="F7-77">Figure <xref linkend="F7-77" remap="7.77"/></link>, one can notice an interruption in the velocity and acceleration rising curves generated during clutching. This interruption is provoked by slippages that have emerged due to the different reasons mentioned above.</para>
</section>
</section>
<section class="lev3" id="sec7-3-4-4">
<title>7.3.4.4 Brake spring degradation</title>
<para>Brake springs degradation reduces braking torque, increasing braking time. This effect may put the integrity of the produced work pieces at risk, as well as operators&#x02019; integrity.</para>
<section class="lev4" id="sec7-3-4-4-1">
<title>7.3.4.4.1 Solution approach</title>
<para>A soft sensor approach is developed for estimating the brake springs stiffness. Several test were done with many brake spring combinations for simulating degradation. The set-up is shown in <link linkend="F7-78">Figure <xref linkend="F7-78" remap="7.78"/></link>. The pressure sensor P1 measured the line pressure, the pressure sensor P2 measured clutch brake input port pressure and the pressure sensor P3 measured the chamber pressure.</para>
<para>The developed soft sensor is able to estimate several clutch brake states and a parameter by measuring only the line pressure P1 and the clutch brake input port pressure P2. The estimated states were piston displacement, velocity and acceleration, and the inner chamber pressure evolution over time. The estimated parameter was the brake springs stiffness.</para>
<fig id="F7-78" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-78">Figure <xref linkend="F7-78" remap="7.78"/></link></label>
<caption><para>Prototyping clutch brake set-up.</para></caption>
<graphic xlink:href="graphics/ch07_fig078.jpg"/>
</fig>
</section>
<section class="lev4" id="sec7-3-4-4-2">
<title>7.3.4.4.2 Results</title>
<para>For each test the inner brake springs were changed and the soft sensor is able to estimate the brake springs stiffness with an error less than 5%. The estimation results are depicted in <link linkend="F7-79">Figure <xref linkend="F7-79" remap="7.79"/></link>.</para>
<fig id="F7-79" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-79">Figure <xref linkend="F7-79" remap="7.79"/></link></label>
<caption><para>Estimated vs actual inner chamber pressure and brake springs stiffness.</para></caption>
<graphic xlink:href="graphics/ch07_fig079.jpg"/>
</fig>
<para>In the case of the estimated stiffness, the estimation line converges quite well with the actual value of the brake springs. The estimation of the chamber pressure does not converge so well due to the leakages that are not yet taken into account in the model.</para>
</section>
</section>
<section class="lev3" id="sec7-3-4-5">
<title>7.3.4.5 Friction material wear</title>
<para>During operation, clutch and brake sides&#x02019; friction material suffer from wear. These issues could cause some problems such as delays at braking and clutching and, at the same time, they could also cause the degradation of other components.</para>
<section class="lev4" id="sec7-3-4-5-1">
<title>7.3.4.5.1 Solution approach</title>
<para>The wear of the clutch side and brake side friction material is monitored by means of a soft sensor that takes advantage of the already installed pressure sensors signals. Some metrics are defined in order to relate the air mass flow and the instantaneous pressure level with the wear of both friction materials. As in all cases where soft sensors have been applied, this solution needs a model of the analysed system.</para>
<para>Many tests have been carried out combining different wear levels for friction material for both sides. Analysed magnitudes or metrics have revealed a similar behaviour for an identical friction material wear in the same side, either in clutch side or brake side. <link linkend="F7-80">Figure <xref linkend="F7-80" remap="7.80"/></link> shows air pressure vs air mass curves shapes for different friction material wear combinations. The percentages that appear in the figure legend represent the wear level of both sides, being left side percentage brake side wear and right side percentage clutch side&#x02019;s.</para>
<fig id="F7-80" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-80">Figure <xref linkend="F7-80" remap="7.80"/></link></label>
<caption><para>Air pressure vs air mass representation.</para></caption>
<graphic xlink:href="graphics/ch07_fig080.jpg"/>
</fig>
</section>
<section class="lev4" id="sec7-3-4-5-2">
<title>7.3.4.5.2 Results</title>
<para><link linkend="F7-81">Figure <xref linkend="F7-81" remap="7.81"/></link> shows a zoomed-in view of the brake and clutch related curve sections and shows how the curves track the same path for an identical wear level for both sides.</para>
<fig id="F7-81" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-81">Figure <xref linkend="F7-81" remap="7.81"/></link></label>
<caption><para>Left side, brake friction material related curve section. Right side, clutch friction material related curve section.</para></caption>
<graphic xlink:href="graphics/ch07_fig081.jpg"/>
</fig>
<para>From those results, the soft sensor was able to estimate the wear level of the friction material attached to the friction discs, establishing some thresholds for different friction wear levels.</para>
</section>
</section>
<section class="lev3" id="sec7-3-4-6">
<title>7.3.4.6 Piston chamber air leakage</title>
<para>Air leakages during a clutching operation imply an engaging force loss, which in turn is associated to economic losses since more compressed air must be provided to the clutch brake in order to compensate those leakages.</para>
<section class="lev4" id="sec7-3-4-6-1">
<title>7.3.4.6.1 Solution approach</title>
<para>The air leakages are detected and measured by the air mass flowmeter. The mass flow of air reveals directly the air leakages during the press machine operation while the clutch brake is clutched.</para>
</section>
<section class="lev4" id="sec7-3-4-6-2">
<title>7.3.4.6.2 Results</title>
<para><link linkend="F7-82">Figure <xref linkend="F7-82" remap="7.82"/></link> depicts the evolution of the air mass flow during a single stroke operation of the press machine. The portion surrounded in purple quantifies the air leakage that the whole system has experimented.</para>
<fig id="F7-82" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-82">Figure <xref linkend="F7-82" remap="7.82"/></link></label>
<caption><para>Air mass flow profile during a single stroke cycle.</para></caption>
<graphic xlink:href="graphics/ch07_fig082.jpg"/>
</fig>
</section>
</section>
</section>
</section>
<section class="lev1" id="sec7-4">
<title>7.4 Fault Detection for Metal Benders</title>
<blockquote role="flushleft">
<para><emphasis>Contributors: Rafael Rocha, Michele Albano, Luis Lino Ferreira, Hugo Ferreira, Catarina F&#x000E9;lix, Carlos Soares, Goreti Marreiros, Diogo Martinho, Isabel Pra&#x000E7;a, Giovanni Di Orio, Pedro Mal&#x000F3;, Asif Mohammed, Rui Casais</emphasis></para></blockquote>
<para>The objective of this use case is to apply anomaly detection algorithms to the data from the the CNC of metal benders and additional sensors in order to detect failures. The idea is that the machine tool will have an expected behaviour given a set of environmental and production parameters. If at any point in time this behaviour deviates from the expected, then we can assume that something is wrong. In such cases, it is necessary to flag these states and warn both machine operator and maintenance personnel of this event. More concretely an analytics module that processes the data generates alerts or alarms according to the deviation detected from normal operation. These alerts are sent to a monitoring system were they can be viewed and analysed by the machine operators. This can save on downtime by both supporting the diagnosis of the potential malfunction, collecting spare parts in advance, and repairing the machine before the malfunction occurs [Ferreira et al., 2017].</para>
<section class="lev2" id="sec7-4-1">
<title>7.4.1 Introduction to Press Braking</title>
<para>Press braking (brake forming) is the process of deforming a sheet of metal along an axis by pressing it between a clamp (tool) (<link linkend="F7-83">Figure <xref linkend="F7-83" remap="7.83"/></link>), performed by <emphasis>metal sheet bender machines</emphasis>, such as the one in <link linkend="F7-84">Figure <xref linkend="F7-84" remap="7.84"/></link>. A single sheet metal may be subject to a sequence of bends resulting in complex metal parts. Such operations can be used to produce a wide variety of products ranging from electrical lighting posts to metal cabinets.</para>
<fig id="F7-83" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-83">Figure <xref linkend="F7-83" remap="7.83"/></link></label>
<caption><para>Bending process.</para></caption>
<graphic xlink:href="graphics/ch07_fig083.jpg"/>
</fig>
<para>In the case at hand, brake forming can bend sheet-metal (thus the common name of metal sheet benders) from 0.6 to 50 mm thick and lengths from 150 mm to 8 m long. The sheet metal bender machine considered in this section is a top of the line model and pertains to the Greenbender family [Ferreira et al., 2017], manufactured and commercialized by ADIRA. The machine (<link linkend="F7-84">Figure <xref linkend="F7-84" remap="7.84"/></link>) is able to exert a force up to 2200 kN using 2 electric motors of 7.5 kW each, and it is able to bend metal with high precision while saving a considerable amount of energy in the process, as per the EcoDesign (2005/32/CE) European directives.</para>
<para>The angle and type of the bend are determined by the shape of the punch and die and the depth with which a punch penetrates a die. The dies can have &#x0201C;U&#x0201D;, &#x0201C;V&#x0201D; or channel shapes. A movable ram is attached to the beam and is covered by a shroud. The punch is attached to the bottom of the ram and the die to the top of bed (covered by the lower shroud). When the ram descends on the table, a bending force is exerted on the sheet-metal between the punch and the die. The bending force and bending speed must be carefully controlled in order for the material being used to maintain their physical characteristics and insure the required bending precision. Additionally, the machine structure deforms due to the forces involved and those deformations have to be compensated in order to guarantee the machine precision.</para>
<para>Although this may seem simple enough, these machines require very accurate control to ensure the required bending precision (in the order of tens of microns). This accuracy is critical when the bending axis is long. The success of the operations depends on many variables including for example the tensile strength and thickness of the work piece, the type of tools (punch and die) used and the type of bend required.</para>
<fig id="F7-84" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-84">Figure <xref linkend="F7-84" remap="7.84"/></link></label>
<caption><para>Sheet metal machine.</para></caption>
<graphic xlink:href="graphics/ch07_fig084.jpg"/>
</fig>
<para>To ensure the quality of the final product, the bending process comprises of several sophisticated control methods that include:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Calculating the deformation of the workpiece based on the metals characteristics, tool geometry and the desired bend;</para></listitem>
<listitem>
<para>Compensating for spring back by measuring the deviations and repeating the bending process;</para></listitem>
<listitem>
<para>Compensating for frame deformation by measuring changes in the machine tools structure and adapting the pressing accordingly;</para></listitem>
<listitem>
<para>Compensating for deflections in the bed by changing the shape of the bed during the pressing of the workpiece.</para></listitem>
</itemizedlist>
<para>The machines used for this process consist of a hydraulic system that applies pressure to sheet metal thereby deforming the workpiece to the required specifications. The hydraulic and mechanical systems (pistons, valves, tubes, pumps) are subject to high pressures and may fail. Mechanical wear and tear also occur, which may result in damaged axis, shafts, bearings and tools (punch and dies). The hydraulic systems depend on the correct function of the electromechanical valves and motors. These elements are also subject to electrical failure. Many supplies are consumed periodically, such as hydraulic fluid, air filters and oil filters. These elements may also be subject to failure.</para>
<para>These machines have stringent safety requirements that also impose certain restriction on its operation. In addition to this, the production efficiency is also a very important factor in its operation. Moreover, since these machines are used for very costly manufacturing processes, downtime is extremely hurtful to the company that bought the machine. All these requirements mean that the various components such as the hydraulic system, tools (punch, die and bed) and back gauges, which are subject to extreme pressure, must operate in the best of conditions. It is therefore important to predict, detect and correct any failures that will either generate scrap, put an operator&#x02019;s life at risk, or cause downtime.</para>
<para>Next section describes the PM platform implemented to support this use case, while Section 7.4.3 provides insights regarding the employed data analysis techniques and their results.</para>
</section>
<section class="lev2" id="sec7-4-2">
<title>7.4.2 Design &#x00026; Implementation</title>
<para>Proactive maintenance strategies are implemented on the sheet metal bender machine by means of a distributed platform compliant with the MANTIS architecture described in chapter 3, and tailored to the work at hand. In particular, the focus of the platform is on sensor data acquisition, data transmission and storage, and forecasting and machine learning techniques.</para>
<para>The deployed platform is represented in <link linkend="F7-85">Figure <xref linkend="F7-85" remap="7.85"/></link> and is described in the rest of this subsection. In particular, the components are put into relation with the three tiers the architecture is divided into (edge, platform, and enterprise).</para>
<fig id="F7-85" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-85">Figure <xref linkend="F7-85" remap="7.85"/></link></label>
<caption><para>Design of the Proactive Maintenance solution.</para></caption>
<graphic xlink:href="graphics/ch07_fig085.jpg"/>
</fig>
<section class="lev3" id="sec7-4-2-1">
<title>7.4.2.1 Data collected by the machine&#x02019;s sensors</title>
<para>Data on the machine are collected by means of sensors that are part the machine&#x02019;s control systems or from sensors which were added specifically for maintenance purposes. In fact, the machines under study are advanced and heavily sensorized. The existing sensors can be grouped into three different data sources: the <emphasis>Programmable Logic Controller</emphasis> (PLC), the machine&#x02019;s <emphasis>Computerized Numerical Controller</emphasis> (CNC) that controls the machine, merging data from the PLC-connected sensors and actuators, and the <emphasis>Safety PLC</emphasis>.</para>
<para>The sensors of the PLC are used internally to control its operation. These range from buttons and pedals to advanced electric motor drives, with positioning information. Although used primarily for control functions, these sensors can also be used to determine anomalous events or states, to diagnose problems and even to infer the root cause of problems.</para>
<para>The PLC works in close cooperation with the CNC controlling all automation functionalities and, at the same time, it can send information from its sensors to the CNC. The Safety PLC handles only safety-related functions for the machine, such as preventing humans from being too close while the machine is working, detecting critical conditions, etc. Data from these sensors is mainly used to distinguish between component failures and safety-related events.</para>
<para>Data are collected indirectly from the CNC of the Green Bender Press Brake machine, which in turn collects information from the PLC control system of the machine and from its Safety PLC. An application on the CNC stores data regarding raised alarms, machine configuration and ERP-related information (e.g.,: production related data such as type of metal and bend) on a Microsoft Access database. Note that the CNC is based on a Windows machine. The same application stores data collected from existing machine sensors (e.g.,: extensometers, pressure sensors, oil temperature and oil quality), which is collected from shared memory and to a file in the CNC filesystem. The information stored is then sent to the Edge Local node using the OPC-UA protocol. The ideal solution would be to access the shared memory directly by a single software module, but this solution was a compromise in order to ensure the safety and certification of the machine control system.</para>
<para>The application can be tailored and configured to different machines and applications, but the current pilot collects data from 50 machine sensors, with a periodicity of 20 ms, and from the MS Access database, which is scanned every second. The amount of data generated and transmitted to the cloud depends on the machine operation cycles. However, according to the data collected so far, we can extrapolate that it averages 300 MB per working day.</para>
</section>
<section class="lev3" id="sec7-4-2-2">
<title>7.4.2.2 Wired nodes: The oil sensor</title>
<para>The application installed on the CNC also receives data directly from some of the sensors that were installed for maintenance-specific purposes and integrated with the PLC module. In particular, an oil sensor was the only wired sensors installed explicitly for Proactive Maintenance operations.</para>
<para>Oil condition sensors have the capability to detect ferrous particles, water, viscosity changes, etc., to detect lubricant related engine wear and lubricant quality degradation, among other problems. The installed sensor monitors the oil that lubricates the machine&#x02019;s hydraulic circuits, both in terms of its temperature and its quality, the latter being related to presence of contaminations like water, particles, glycol and other impurities in the oil.</para>
<para>The system that analyses the oil consists of two parts, the sensor unit (Hydac Sensor AS1008), and the data acquisition and computation board. The sensor reads temperature from &#x02013;25 to 100<superscript>&#x02218;</superscript>C, and saturation from 0% to 100%. Both signals are reported using a 4&#x02013;20 mA interface. The data acquisition/computation module receives the signals, convert them, and exports the data through an analogic voltage signal with a range from 0V to 10V to the machine&#x02019;s CNC. The CNC digitalizes and sends the data through a communication middleware to the cloud for storage and processing, the latter being the comparison with custom thresholds.</para>
</section>
<section class="lev3" id="sec7-4-2-3">
<title>7.4.2.3 Wireless nodes: The accelerometer</title>
<fig id="F7-86" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-86">Figure <xref linkend="F7-86" remap="7.86"/></link></label>
<caption><para>Sensor component hardware</para></caption>
<graphic xlink:href="graphics/ch07_fig086.jpg"/>
</fig>
<para>Each machine is equipped with two wireless accelerometer sensors, represented in <link linkend="F7-86">Figure <xref linkend="F7-86" remap="7.86"/></link>. The sensors monitor the blade that actually performs the bending of the metal sheet. The sensors collect the data from the own movement of the blade in the press, especially from the vibration patterns that are caused by the hydraulics. In fact, given the fact that the vibratory pattern can be associated to the condition of the machine&#x02019;s bending motors, the collected data can be used to perform PM of the machine.</para>
<para>The wireless protocol for the communication with the accelerometers is Bluetooth Low Energy (BLE), which enables data collection while maintaining energy consumption low. BLE is optimized for low power use at low data rates, and was designed to operate from simple lithium coin cell batteries.</para>
<para>The sensors are based on the Arduino 101 platform, which provide a 3-axis accelerometer with a maximum amplitude range of 8g. They are powered by two 9V batteries, and the sensors are configured for a lower measurement range (between 0 and 2g), aiming to attain a better accuracy. The sensors are able to perform self-calibration, synchronization and security, and the CurieBLE library is used to support communication between the sensors and the Edge Gateway by using of the Generic Attribute Profile (GATT). According to some preliminary experiments, the maximum distance for this technology is 30 meters, which corresponds with the BLE specifications.</para>
</section>
<section class="lev3" id="sec7-4-2-4">
<title>7.4.2.4 Edge gateway</title>
<para>The Edge Gateway used for this deployment is located in the factory and it isolates the latter from the outside world, at the same time providing some functionalities at local level. From the security point of view, the Edge gateway creates a DeMilitarized Zone (DMZ) in the sense that it is the only module in the factory premises that has network access, and thus concentrates all the security requirements on itself. Communication with the cloud is mediated by the AMQP protocol, and in particular through a RabbitMQ bus.</para>
<para>On the other hand, communication with the wireless sensors in a factory is done through BLE protocol, and the rest of the systems (CNCs of the machines and the wired sensors) is done by means of the OPC-UA protocols, which allows for a number of capabilities, including node discovery, data caching, and some degree of security.</para>
<para>Node discovery is used to enable the fast configuration of new machines in a factory. The CNC and the wireless sensors act as OPC-UA server and are discoverable by the Edge Gateway, which then provides the servers with mechanisms to support communication and management of the data acquired across multiple heterogeneous and distributed data sources. This is accomplished by providing an abstraction layer that detaches the application development from the intricacies of the lower level details. It acts as a virtualization platform and as data broker that connects the Machine logical block to the Cloud Middleware, capable of extracting, collecting, distributing/sharing, pre-processing, compressing, and semantically enhancing the data produced in an efficient manner. Therefore, the one of the fundamental goals of the Edge Gateway is &#x02013; from one side is to support the data integration of multiple data sources and &#x02013; from the other side is the provisioning of data to the cloud where more complex and resource consuming data processing takes place.</para>
<para>Finally, the Edge Gateway of this pilot comprises a database to cache collected data, and a local HMI service responsible for visualizing all the necessary information generated within the factory, such as list of machines available and their conditions, and data readouts.</para>
</section>
<section class="lev3" id="sec7-4-2-5">
<title>7.4.2.5 Communication in the cloud</title>
<para>A few components on the cloud (Messaging Bus, Management Panel, Edge Broker and Database) manage the data, by storing and transporting them between the Edge Gateways of the factories and the Data Analysis and HMI modules. The communication mechanisms are implemented on top of a message-oriented bus and allow the interaction of the factories, mediated by their Edge Gateway, with the rest of the Platform tier, and the Enterprise tier. Communication is performed on top of a RabbitMQ bus, which is the most popular implementation for the AMQP protocol.</para>
<para>The basic elements of the message distribution system are the exchange and the queue. The exchange is the recipient of a message from a message producer, and its duty is to deliver the message to one or more queues, the latter being buffers from which the message consumers will pull the messages. An exchange can be connected to multiple queues, and the exchange can be configured to treat messages in different ways, such as relaying the messages to the queues in a round-robin fashion or broadcasting the messages to all the queues. Finally, the decision on which queue(s) receives each message from the exchange, is done by means of a routing key, which is a meta-datum assigned to each message. The messaging system can also implement Remote Procedure Calls (RPC) mechanisms, see for example <link linkend="F7-87">Figure <xref linkend="F7-87" remap="7.87"/></link>.</para>
<fig id="F7-87" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-87">Figure <xref linkend="F7-87" remap="7.87"/></link></label>
<caption><para>Consumer Remote procedure call (RPC).</para></caption>
<graphic xlink:href="graphics/ch07_fig087.jpg"/>
</fig>
<para>In the pilot at hand, Edge Gateways send data to the RabbitMQ server, where the routing process is used to deliver specific messages to the other components, and in particular the Edge Broker and the Data Analysis components. The RabbitMQ Bus is configured using a RabbitMQ Management panel (which is part of the Enterprise tier as far as the MANTIS reference architecture is concerned) and that obeys the REST architectural pattern. The component respects the reactive programming properties, namely, Responsive, Resilient, Elastic, and Message Driven. For example, the RabbitMQ platform is fault tolerant since, if a message delivery fails, the queue buffers the messages and retransmits them when the message consumer is back online. Moreover, if the broker malfunctions, messages in the queues are not lost since they are saved in the persistent memory of the broker.</para>
<para>The Edge Broker is the main peer that receives messages from several Edge Gateway devices through the queues and saves the data to a database module, which is structured according to the MIMOSA standard, which is described in <link linkend="ch03">Chapter <xref linkend="ch03" remap="3"/></link>. Current implementation of the DB is based on Microsoft IIS, even though an alternative design based on No-SQL Mongo is available. The RabbitMQ Bus provides queues for generic RPC connectors for handling database queries, to expose stored data to other components. The HMI is allowed to make queries trough RPC patterns to the database. Anyway, when data from a device is received in the middleware, the Data System Module component and HMI component will receive asynchronous messages.</para>
</section>
<section class="lev3" id="sec7-4-2-6">
<title>7.4.2.6 Components for data analysis</title>
<para>The Data Analysis techniques and results for this pilot are described in next subsection, while this subsection is focused on the implemented components that support the techniques.</para>
<para>The main component for Data Analysis is the <emphasis>Intelligent Maintenance Decision Support System</emphasis>, which is used to manage the models (model generation, selection, training and testing), for example on reception of training data. The <emphasis>Intelligent Maintenance Decision Support System</emphasis> is composed of a Knowledge Base that uses diagnosis and prediction models and the data sent by sensors. On top of this Knowledge Base there will be a Rule based Reasoning Engine which includes all the rules that are necessary to deduce new knowledge that helps the maintenance crew to diagnose failures.</para>
<para>In addition to the data and algorithms, expert knowledge has been encoded as a set of rules that are used to detect and flag possible failures. Each rule indicates what sensor and CNC signals need to be acquired, how they are segmented, the type of analysis to be executed and what failure is associated with these signals.</para>
<para>As an example, let us consider when the brake press is working in automatic mode, terminates its bend cycle and has parked the ram on the top position waiting for the next task. If no failure exists then the ram must remain still in the same position where it stopped. Because the hydraulic system is constantly losing pressure, the CNC compensates for any deviation. Normally, such deviations are minor (imperceptible to the naked eye) and occur at very low rates. However, if a hydraulic pump fails or a hydraulics tube ruptures, leaks will cause large deviations as the CNC compensates for this.</para>
<para>In order to detect such problem, the positions of the pistons are recorded when the control signal indicates that the ram is at top dead center (segmentation). Statistical tests are used to check that the deviation is within a specific tolerance threshold. This threshold is determined via the machine learning algorithm (stream based) and is tweaked in order to reduce the false positive and negative detection rates.</para>
</section>
<section class="lev3" id="sec7-4-2-7">
<title>7.4.2.7 Human machine interface</title>
<para>This module is part of the Enterprise tier, and it provides a Human interface for the proactive maintenance system. The HMI follows a web-oriented design and therefore can be accessed from anywhere, at any time and through all sort of electronic devices with the only requirement being the use of the Internet to do so. This allows both remote (administrative) and on-site operations such as analyzing the machine&#x02019;s state or view its past performance. It has two main modes, one for data visualization and another for data management.</para>
<para>In the visualization mode, it is possible to view historical and live data, which is collected from specific machine sensors (e.g., machine status, speed, positioning and pedals state). It is also possible to show the results generated by the data analysis module, more specifically the alarms for unusual sensor data and the warnings regarding impending failures. It is possible to match the warnings from the Data Analysis block with historical data collected from the sensors. All data retrieved from the machine and the Data Analysis logical block can be inspected through graphs and tables, provided by the HMI module (see <link linkend="F7-88">Figure <xref linkend="F7-88" remap="7.88"/></link>). Also, each user can be promptly notified of any event on the machine through a text notification.</para>
<para>The HMI also displays the results of data aggregation and calculation of statistics. Several descriptive statistics provide useful (albeit simple) indicators that support the decisions making by those responsible for maintenance and design. These indicators are therefore available to the Data Analysts. The results of the machine learning algorithms are displayed when they generate alerts and alarms, but they can also be visualized as historical data.</para>
<fig id="F7-88" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-88">Figure <xref linkend="F7-88" remap="7.88"/></link></label>
<caption><para>Checking the &#x0201C;Bend&#x0201D; variable on a GreenBender machine.</para></caption>
<graphic xlink:href="graphics/ch07_fig088.jpg"/>
</fig>
<para>The Management mode allows for all the administrative operations, like users and roles management, as well as factories and machines setup. Role management allows to dynamically assign specific permissions to each type of user, which can be Operator, Data Analyst, and Maintenance Manager.</para>
<para>Operator can view historical and live data only. The Data Analyst role allows inspecting live streamed data collected from the machine, such as oil-flow and temperature sensors. The data is displayed in near real-time (<link linkend="F7-89">Figure <xref linkend="F7-89" remap="7.89"/></link>). The Data Analyst role also allows the visualization of historical data by selecting the data variables to be shown as well as the desired time-frame. The Maintenance Manager role allows to view some statistics, e.g., the type of components substituted and the frequency rate of the replacements. This should be specified for each monitored parameter according to the current number of cycles performed and to the maintenance actions of the machine tools. The user can also choose to display the results of the Data Analysis logical block, in the form of alarms, alerts and reports, which are displayed highlighting the relevant information for the maintenance manager and allowing the consultation of details on their provenance. These notifications include alarms that indicate unusual sensor data (for example based on simple statistics) and, unexpected behavior (for example, using outlier detection algorithms). These notifications only allow the detection of failures (corrective maintenance), but in the future may also be used to plan preventive maintenance tasks.</para>
<para>Security is implemented by means of SSL/TLS, for both the communication with the Cloud Middleware and the access to the HMI webpage (HTTPS). It is also important to note that the web-based HMI is running in the same node as Cloud Middleware, in order to reduce system complexity and to be able to access the same Database.</para>
</section>
</section>
<section class="lev2" id="sec7-4-3">
<title>7.4.3 Data Analysis</title>
<para>The analysis on data collected in the pilot is limited to a set of signals that can be divided into 3 groups: a) movements of the mechanical parts via hydraulic or electric power, b) temperature of the hydraulic fluid and c) extensometers that measure the deformation of the press brake machine structure. Additional signals are also available such as error codes and from the machine&#x02019;s soft numerical controller and alarms generated by the safety system. However these are not processed statistically and are only forwarded as alerts to a monitoring application.</para>
<para>The machine&#x02019;s designers and maintenance engineers determine the signals that need to be analysed. They are compiled as a set of rules that indicate not only the signals that must be monitored, but also when these signals should be collected and analysed and how the testing should be performed.</para>
<fig id="F7-89" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-89">Figure <xref linkend="F7-89" remap="7.89"/></link></label>
<caption><para>Example data from the controller.</para></caption>
<graphic xlink:href="graphics/ch07_fig089.jpg"/>
</fig>
<para>The pilot applies anomaly detection algorithms to the data from the machine tools controllers and sensors in order to detect failures. The idea is that the machine tool will have an expected behaviour given a set of environmental and production parameters. If at any point in time this behaviour deviates from the expected, then we can assume that something is wrong. In such cases, it is necessary to flag these states and warn both machine operator and maintenance personnel of this event. More concretely an analytics module that processes the data generates alerts or alarms according to the deviation detected from normal operation. These alerts are sent to a monitoring system were they can be viewed and analysed by the machine operators. Anomaly detection done in this work uses simple statistical testing (See for example <link linkend="F7-89">Figure <xref linkend="F7-89" remap="7.89"/></link>). The control signals are those signals generated by the machine controller that are used to activate components and/or inform of the various process phases. Examples include: the <emphasis>Top Dead Centre</emphasis> signal that is active when the press ram is parked at the top position, the pedal signals (indicating if the operator pressed the up or down pedals), the axis in position signal (indicates when the backgauge is in position), the bending signal (that indicates if the ram is deforming the workpiece in manual or automatic mode), the speed change points (when the ram slows down for a press), the pinch points (indicating the start and end of the press phase).</para>
<para>Domain knowledge was elicited from brake press machine experts, which was then encoded as a set of rules that allow for the detection of possible failures. These rules indicate how to use the control signals to sample the process signals, thereby sampling data only during valid periods of time. It also substantially reduces the amount of data that is processed and analysed by the analytics module (for example, certain signals need not be collected when the machine is in standby or parked).</para>
<para>The machine tool manufacturer has provided a list of components that may be the cause of the failure for a given rule. For example, any deviations in the ram&#x02019;s position when parked may be indicative of a failure in hydraulic system. This list includes leaking tubes or oil sump (due to ruptures), malfunctioning valves, broken oil pumps and clogged tubes or pumps. Note that this information does not indicate the root cause of the failure. For example, a ruptured tube may be due to the hydraulic fluid being below the required temperature or a broken pump may have either electrical or mechanical malfunctions in other subcomponents. Currently this information is not used but could conceivably be combined with the alerts to help in diagnosing problems.</para>
<section class="lev3" id="sec7-4-3-1">
<title>7.4.3.1 Data pre-processing</title>
<para>Data related to the machine&#x02019;s behaviour are received from the machine in chunks and are sent to the Data Analytics module, which has two functions. The first is the off-line learning and tweaking of the statistical anomaly detection models and the second is the on-line use of the models to detect failures and generate alerts. The combination of the alerts generated by the analytics module and the signal data collected from the machine tool facilitate the diagnosis of failures by machine operators and maintenance personnel (data selection and visualization).</para>
<para>An initial pre-processing phase of data processing will first segment and collect only those signals that, according to the rules we have previously described, are required. For example, only the movement data is collected for the &#x0201C;parking rule&#x0201D; when the <emphasis>Top Dead Centre</emphasis> signal is on or we only analyse the ram speeds and synchronization when the pedal down control signal is on.</para>
<para>A second pre-processing phase will transform data in order to be able to either learn the statistics or use the statistical tests to check for anomalies. These transformations depend on the type of test to perform, which is determined by the rule. For example, this can involve calculating the difference between the two signals of piston distances when testing for synchronization between those pistons.</para>
<para>A third transformation of the signals is the calculation of the temporal difference of the piston displacement signals. This is used in the rules that use speed as a basis for comparison and checking. Note that usually such difference introduces significant noise into the signal and this may require additional filtering (smoothing).</para>
<para>From the experiments, it appears that the signals from the numeric controller and the oil quality sensor are clean. Even though the accelerometer and extensometers data are relatively noisy, no additional sophisticated pre-processing was used because the related failures could not be evaluated.</para>
</section>
<section class="lev3" id="sec7-4-3-2">
<title>7.4.3.2 Failure detection</title>
<para>Statistical hypothesis testing allows one to compare two processes by comparing the distributions generated by the random variables that describe those processes [Stuart et al., 1999]. If the distributions are not equivalent, then we assume a failure occurred. A p-value is used as a threshold in order to detect any deviations from the expected process with the goal of reducing the number of false positives and negatives. In the case of the formal statistical tests, if the p-value does not allow us to reject the null hypothesis we cannot infer that the machine has no failure (type 2 error). However, here we assume that this is true, and alerts are only sent when the null hypothesis is successfully rejected. A distribution of a given process may be described by one or more random variables. Here we limited our analysis to the use of univariate statistics only using both parametric and non-parametric models.</para>
<section class="lev4" id="sec7-4-3-2-1">
<title>7.4.3.2.1 Parametric models</title>
<para>In the case of the parametric models two basic tests were performed: univariate signal that should be close to a constant (within an unknown threshold), or two signals must not diverge from each other (no more than an unknown threshold). In both cases we can use the signals that indicate velocity, distance, heat and acceleration. In either case, if we can perform a parametric Gaussian test on the mean (using a t-Test) or deviation (directly compare deviations, F-test, Bonett&#x02019;s test and Levene&#x02019;s test).</para>
<para>In addition to the tests described above, additional na&#x000EF;ve statistical tests were used. In both cases an online Gaussian model is obtained via the calculation of a mean and variance. These means and variances are then directly compared to the continually sampled signals from a working machine. If a significant divergence is found, alerts are generated. Due to the high false positive and false negative rates (type 1 and type 2 respectively), an additional multiplicative threshold (in respect to one standard deviation) is used when comparing deviations. The initial values of this threshold are set automatically by selecting the lowest possible threshold that reduces type 1 errors.</para>
</section>
<section class="lev4" id="sec7-4-3-2-2">
<title>7.4.3.2.2 Non-parametric models</title>
<para>As with the case of parametric statistic, the process of defining the hypothesis, sampling data and establishing a significance level as a threshold were the same. However, not all of the non-parametric methods provide a p-value for a significance level comparison. Three types of statistical tests were used in our work: the Kolmogorov-Smirnov test (K-S test), the Mann-Whitney U test and the use of a kernel density estimator (KDE). In the case of the U statistic, which is approximated by a normal distribution, a p-value is available to establish a threshold for accepting or rejecting the null hypothesis.</para>
<para>In the case of the KDE, an online algorithm was used that generates a dynamic number of (Guassian) kernels. The kernels and respective parameters of two different distributions cannot be directly compared. Experimentation shows that the estimated densities may be visually very similar, but the kernels themselves differ significantly. However, because we can use the kernel to sample the underlying estimated distribution we used the parametric statistical tests to compare the samples (both for the parametric cases and non-parametric cases using the Kolmogorov-Smirnov test and Mann-Whitney U test respectively). Here we could have also opted for the use of alternative algorithms such as the earth movers distance but did not do so because the na&#x000EF;ve parametric tests seemed to be working well [Levina and Bickel, 2001] (see Section 7.4.3.2.1).</para>
</section>
<section class="lev4" id="sec7-4-3-2-3">
<title>7.4.3.2.3 Evaluation and interpretation</title>
<para>For anomaly detection the positive labels are indicative of failures. Expected failure rates (as reported by the machine tool manufacturer) are very low. We therefore need to deal with data-sets that will be highly skewed (very few positive labels). This brings with it two challenges.</para>
<para>The first is that false negatives are more important than the false positives. This is due to the fact that if we incorrectly predict that the machine is failing (false positive), the alert/alarm message will be sent and the operator, which will verify that the machine is not in fact failing. On the other hand, if we incorrectly predict the machine is not failing (false negative) it could have serious consequences, since the operator will not receive any warning message.</para>
<para>The second challenge is related to the selection of the appropriate metrics used in the evaluation of the model&#x02019;s performance. Accuracy is not a viable metric because a biased prediction of no failure will always result in high accuracy values. We considered the following metrics: the AUC-ROC (Area under the Curve for the ROC Curve) and a set of relations involving a combination of true positive (TP), true negative (TN) and false positive (FP) and false negative (FN) counts:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq7-7.jpg"/></para>
<para>F0.5, F1 and F2 measures (<emphasis>F</emphasis><subscript>&#x003B2;</subscript> for &#x003B2; &#x02208;&#x0007B;0.5, 1, 2&#x0007D;), <inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg412-1.jpg"/></para>
<para>Mathews correlation coefficient,</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq7-10.jpg"/></para>
<para>Exploratory work was done using precision, accuracy and the MCC metrics. Final comparisons were done using MMC. Future work will consider the metric <emphasis>F</emphasis><subscript>&#x003B2;</subscript> and how to establish an appropriate value of &#x003B2;.</para>
<para>Due to lack of data, initial exploratory work used artificial (synthetic) data. This data was generated using normal (Gaussian), Bimodal (composed of two joined normal distributions), Pareto and Weibull distributions because we had no way of checking the signals distributions. We assumed a failure rate of 5%. Each of these distributions was tested for differences in the mean and/or standard deviation using both the parametric and non-parametric statistics referred to above (T-test, Kolmogorov-Smirnov Test, Mann-Whitney U test). We also generated non-parametric statistical models of the synthetic data using the Kernel Density Estimation (KDE). We then used this model to generate a reference distribution that was used in the statistical test. In addition to this we also modelled the data as a Gaussian distribution and naively compared the means and deviations to determine if the processes are different.</para>
<para>We found that statistical methods tend to be very brittle and usually result in many false positives. Increasing the threshold would result in an unreasonably high rate of false positives. The reason is that the formal statistical tests make important assumptions about the distributions. For example, the t-Test is only valid if the distributions have the same standard deviation. The best results were obtained using the Mann-Whitney U test and the na&#x000EF;ve Gaussian tests.</para>
<para>In the next phase of the work we opted to use the na&#x000EF;ve Gaussian tests because it allowed for the easy generation and update of the model. The data we got from the machine seems to be Gaussian (we say seems to be because at the time of writing, data with failures had not been obtained and cannot therefore confirm this). During this period we collected data from a single machine tool executing a preprogramed sequence of operations under optimal conditions for several days.</para>
<para>The pre-processing steps described above are applied to this data (each segment collected for a given rule) and the mean and standard deviation are calculated using a robust algorithm. The Gaussian model is initially generated using the first 10% of the data stream. During the calibration we use the same 10% of the data set to establish the threshold so that no false positives are detected. The threshold indicates by how much a mean or standard deviation must differ from the base models&#x02019; mean or standard in order to flag a failure. This is not an appropriate way of setting this threshold, but because no failure data exists, using a ROC curve to establish a good compromise between false positives and negatives is not possible. We then tested the failure detection models of each rule calculating the Gaussian parameters of the segmented data and comparing those parameters to the base model. We expected a false positive rate to be close to 0 but got the results in <link linkend="T7-2">Table <xref linkend="T7-2" remap="7.2"/></link>.</para>
<table-wrap position="float" id="T7-2">
<label><link linkend="T7-2">Table <xref linkend="T7-2" remap="7.2"/></link></label>
<caption><para>True negatives (OK) and false positives (ALERT, ALARM) detected</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="center">rule</td>
<td valign="top" align="center">trained</td>
<td valign="top" align="center">OK</td>
<td valign="top" align="center">ALERT</td>
<td valign="top" align="center">ALARM</td></tr>
</thead>
<tbody>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="center">1</td>
<td valign="top" align="center">112</td>
<td valign="top" align="center">1004</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">7</td>
</tr>
<tr>
<td valign="top" align="center">2</td>
<td valign="top" align="center">83</td>
<td valign="top" align="center">749</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">7</td></tr>
<tr>
<td valign="top" align="center">3</td>
<td valign="top" align="center">112</td>
<td valign="top" align="center">1011</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">0</td>
</tr>
<tr>
<td valign="top" align="center">4</td>
<td valign="top" align="center">146</td>
<td valign="top" align="center">1310</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">5</td></tr>
<tr>
<td valign="top" align="center">5</td>
<td valign="top" align="center">187</td>
<td valign="top" align="center">1631</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">152</td>
</tr>
<tr>
<td valign="top" align="center">6</td>
<td valign="top" align="center">187</td>
<td valign="top" align="center">1631</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">52</td></tr>
<tr>
<td valign="top" align="center">7</td>
<td valign="top" align="center">112</td>
<td valign="top" align="center">1010</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">0</td>
</tr>
<tr>
<td valign="top" align="center">8</td>
<td valign="top" align="center">90</td>
<td valign="top" align="center">813</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">0</td>
</tr>
<tr>
<td valign="top" align="center">9</td>
<td valign="top" align="center">120</td>
<td valign="top" align="center">1027</td>
<td valign="top" align="center">53</td>
<td valign="top" align="center">0</td>
</tr>
<tr>
<td valign="top" align="center">101</td>
<td valign="top" align="center">152</td>
<td valign="top" align="center">1368</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">1</td>
</tr>
<tr>
<td valign="top" align="center">102</td>
<td valign="top" align="center">95</td>
<td valign="top" align="center">855</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">3</td>
</tr>
<tr>
<td valign="top" align="center">11</td>
<td valign="top" align="center">117</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">1058</td>
</tr>
<tr>
<td valign="top" align="center">12</td>
<td valign="top" align="center">8109</td>
<td valign="top" align="center">8109</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">0</td>
</tr>
<tr>
<td valign="top" align="center">13</td>
<td valign="top" align="center">14938</td>
<td valign="top" align="center">74690</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">0</td>
</tr>
<tr>
<td valign="top" align="center">14</td>
<td valign="top" align="center">95</td>
<td valign="top" align="center">734</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">124</td>
</tr>
<tr>
<td valign="top" align="center">15</td>
<td valign="top" align="center">184</td>
<td valign="top" align="center">1665</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">0</td></tr>
</tbody>
</table>
</table-wrap>
<para>As referred above, a multiplicative factor can be tweaked to increase or decrease the false positives rates, but there is no way to measure the false negatives and hence use the MCC metric (Equation (7.3)). However, these test serve as an important sanity check and allow us to detect and correct several issues.</para>
<para>The first issue is that there is a lag between the control and measured signals (due to sampling delays and (mostly) due to mechanical inertia). This means that, for example, when a control signal indicated the start or stop of the ram, the corresponding sensor reading of the ram displacement does not show up immediately. This also means that the control signals are not perfectly aligned with the ram movement (for example ram movement occurs after a stop signal is sent). To solve this, in certain rules, only the final signal samples are used to generate and compare the models.</para>
<para>Another issue is that oil temperature varies widely during the operation of the machine. This depends not only on the load, but also on the rate of the ram movement and the environmental temperature. In addition to this, several brake press machines have a heating element to warm up the oil to acceptable operational levels. The only way to truly solve this is to increase the sampling population to several machine tools operating in very diverse conditions. More important however is the fact that the machines&#x02019; oil temperature will initially rise significantly compared to its initial operation. This means that we cannot limit our sampling of the oil temperature (or any other signal) to the start of the data stream when generating the models.</para>
</section>
</section>
</section>
<section class="lev2" id="sec7-4-4">
<title>7.4.4 Conclusions</title>
<para>The pilot described in this section is used to experiment with building a PM platform able to collect data in an effective and efficient manner from a metal sheet bender machine, and with machine learning techniques applied to collected data to find misbehaviors.</para>
<para>The implemented platform is used to collect data from a single machine in a factory, and transport data to a cloud for processing. It is straightforward to align all of the pre-existing and new components with concepts from the MANTIS reference architecture, and the design of the platform is used to validate the solution described in <link linkend="ch03">Chapter <xref linkend="ch03" remap="3"/></link>. The platform uses different protocols in the different tiers of the platform, and in particular OPC-UA in the Edge tier, AMQP between the Edge Gateway and the Edge Broker and between the components in the Platform tier, and HTTPS for the components in the Enterprise tier since these latter components are web-based.</para>
<para>A na&#x000EF;ve Gaussian model is used to identify failures in a brake press machine using signal readings of position, speed, temperature, and acceleration. The type of failures and the respective signal analysis is selected and encoded by domain experts. Due to a lack of failure data (positive labels) it is not possible to evaluate the effectiveness of the models using standard metrics such as the F score and MCC (Equation (7.3)). On the other hand, it is possible to partially validate the solution based on the measure of false positives. The results seems to indicate that for this specific case, the na&#x000EF;ve Gaussian model may be a viable solution. We are able to determine how to implement and test the experts&#x02019; checks by applying a simple set of pre-processing steps and using Gaussian means and standard deviations. More importantly, it enabled us to identify and resolve some issues regarding the sampling and use of the signals (delayed signals due to inertia, time series with very high variability of the oil temperature).</para>
</section>
</section>
<section class="lev1" id="sec7-5">
<title>7.5 Off-road and Special Purpose Vehicles</title>
<blockquote role="flushleft">
<para><emphasis>Contributors: Ansgar Bergmann</emphasis></para></blockquote>
<para>Off-road and special purpose vehicles include lorries (trucks), buses, agricultural machinery, construction machinery, and forklift trucks. These types of vehicles share many common characteristics, and offer the possibility of the development of related technical solutions and technologies under technically similar challenges. In addition to the property as an investment and a working machine in production and value-added processes, there are facts like high complexity, low-volume, high variety and high quality and reliability requirements over a long lifetime.</para>
<section class="lev2" id="sec7-5-1">
<title>7.5.1 Introduction to the Use Case on Vehicles</title>
<para>STILL supplies customized internal logistics solutions and implements the intelligent management of material handling equipment, software and services worldwide. With over 7000 employees, four production facilities, 14 branches in Germany and 20 international subsidiaries as well as a global dealer network, STILL is a successful international player. Today and in the future, STILL fulfils the requirements of small, medium-sized and large companies with highest quality, reliability and innovative technology. STILL&#x02019;s forklift trucks are operating in a variety of areas and conditions with often totally different application profiles, which differ not only in the temporal use (one to multi shift) as also in the environmental conditions (from easy hall operation up to use in the heavy industry or fishing industry). This results in high demands concerning ensuring the availability of vehicles, recognition of special types of damage and an optimized maintenance scenario, which fits to the special needs of this types of usage and thereby minimizes potential and also safety-critical damages resulting from this.</para>
<para>STILL currently collects its data only for internal processes and customer applications. A key problem lies in the mobility of the machines, which does not allow large amounts of data to be transferred on a wireless way without high costs. In addition, the systems are operated in a wide variety of environmental conditions. Within this project, the aim is to determine whether existing data collection mechanisms are already sufficient for the desired objectives or whether new solutions have to be chosen, to be an enabler for new service solutions and other maintenance based products. Basing on these fundamental analyses, options for business actions can be derived.</para>
<fig id="F7-90" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-90">Figure <xref linkend="F7-90" remap="7.90"/></link></label>
<caption><para>Smart service landscape as a possible result of the Mantis basis work.</para></caption>
<graphic xlink:href="graphics/ch07_fig090.jpg"/>
</fig>
<para>Smart services (see <link linkend="F7-90">Figure <xref linkend="F7-90" remap="7.90"/></link>) are interesting for a company as they:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Enable higher added value (optimization of service in combination with intelligent products) and better service quality, e.g., through shorter reaction and repair times;</para></listitem>
<listitem>
<para>Increase user-friendliness;</para></listitem>
<listitem>
<para>Open up new markets for services and data-driven business models;</para></listitem>
<listitem>
<para>Enable a drastically increased efficiency of service-based operating models;</para></listitem>
<listitem>
<para>Result in higher machine availability;</para></listitem>
<listitem>
<para>Guarantee a more detailed planning of machine operation and downtime times;</para></listitem>
<listitem>
<para>Improve component design through monitoring.</para></listitem>
</itemizedlist>
</section>
<section class="lev2" id="sec7-5-2">
<title>7.5.2 Scope and Logic</title>
<para>Currently, STILL has established systems to support all standard issues of maintenance, where actions and processes are mainly based on existing technical and historical know-how and supplier specifications. There is no off-the-shelf solution for such brand specific processes on the market, all the tools which are used to day have been custom-made for the company. This is also due to the fact that this is property and essential technical expertise of the company, so there is a need to protect this intellectual property. But because of the increasing demands of the market and the increasing price pressure, processes have to be shortened and optimized. In addition, the complexity and variability of forklift trucks is increasing, so solutions have to be implemented to support the service technician in his work. In order to ensure the next steps in the service evolution, clear statements in the process chain must make fault detection clearer. First time fix is one of the most important goals of the future. By analyzing the internal system values of machine components and all other existing databases related to the service process (master data, repair databases, customer information) a fundamental base for this steps will be generated. For this reason, both the technical know-how from the examination of defective parts and the big data analysis will be used to identify specific patterns in the application and the environmental conditions that lead to breakdowns or high service costs. STILL GmbH is focusing on two main topics - Wear and Root cause analysis. Both topics are too extensive, that a global solution found in one research project, so the expectation lies in creating first demonstration cases on this topics, which are to be refined over a longer observation period.</para>
<para>So concerning wear, STILL is finally focusing now on the subject of tires, because there are measurable conditions. All other relevant topics of wear of filters, mechanical and electronic parts have been discussed in the first phase of this project, but the necessary measures go beyond the possibilities of MANTIS. The study focuses in particular on the relationship between usages of the forklift by the driver and wear, since this knowledge can also be used in other business models (e.g., pay per use). Due to the lack of other environmental information like temperature, humidity and quality of the ground (this information is not recorded by the truck automatically) which are also important on wear, their influence has to be estimated.</para>
<para>In the topic of RCA, STILL uses the existing error messages of the forklift trucks with regard to the cumulated service reports and internal knowledge. At present, most error messages have no clear reference to the existing error screens in the field, since they are created as developer knowledge under laboratory conditions. Influencing conditions such as the environment or faulty interaction with other damaged components can be difficult to simulate in the laboratory. Under real conditions, however, the causal chains can differ considerably, so that errors can have other causes or effects. A broken wire can cause the display of a device defect, although the control unit is fully functional. The aim is therefore to achieve useful results through pattern finding and cooperative decision-making. The project will initially focus on some electronic errors to validate these results. However, these analyses are made more difficult by the fact that the error reports are freely formulated and do not show any clarity either.</para>
<fig id="F7-91" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-91">Figure <xref linkend="F7-91" remap="7.91"/></link></label>
<caption><para>Overall scenario of a resulting field configuration and interaction.</para></caption>
<graphic xlink:href="graphics/ch07_fig091.jpg"/>
</fig>
<para>The illustrated example (<link linkend="F7-91">Figure <xref linkend="F7-91" remap="7.91"/></link>) shows the complete range of components required for a powerful future concept in the field. Most of these modules are only listed for the purpose of being complete, but will not be considered in the following context.</para>
</section>
<section class="lev2" id="sec7-5-3">
<title>7.5.3 Data Platform and Sensors</title>
<para>As mentioned above, most industrial companies, as well as STILL, tend to have a grown data infrastructure. For these reasons, data are neither harmonized nor centralized. The data used can therefore be based on platforms whose technologies are up to 10 years apart. The demands on the merge are enormous, especially since many data have not been checked for their electronic processing capability in the past. It must therefore be checked whether it is meaningful and effective to integrate these data via interface, or whether one prefers to collect the data in a new and future-proof way. This is also due to the fact that 80% of the time spent on data analysis flows into data preparation. The following graphic (<link linkend="F7-92">Figure <xref linkend="F7-92" remap="7.92"/></link>) shows the necessary data sources for the following analytic steps.</para>
<fig id="F7-92" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-92">Figure <xref linkend="F7-92" remap="7.92"/></link></label>
<caption><para>Data sources for downstream analytics.</para></caption>
<graphic xlink:href="graphics/ch07_fig092.jpg"/>
</fig>
<para>Since the data landscape of STILL GmbH cannot be completely replicated for MANTIS, a demonstration solution on MS Azure was chosen (see <link linkend="F7-93">Figure <xref linkend="F7-93" remap="7.93"/></link>). The main reasons for this decision is the great flexibility of the step and the simple scalability. All Elements can be selected without having a negative impact on normal business processes. In the initial phase, a small number of internal forklifts (up to 5) are connected to this system. The advantage of this approach lies in the traceability of the use of these forklifts. The results can thus be validated directly. Once usable results are achieved, the integration of forklifts can also be extended to rental forklifts, for example.</para>
<fig id="F7-93" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-93">Figure <xref linkend="F7-93" remap="7.93"/></link></label>
<caption><para>MS Azure demonstrator architecture approach.</para></caption>
<graphic xlink:href="graphics/ch07_fig093.jpg"/>
</fig>
<para>The presented architecture serves as a basis for the validation and further development of the resulting knowledge from the project. As data sources, the architecture includes vehicle data as well as data from company databases and from results of the interaction with the service technician. It is deliberately designed in such a way that individual blocks can be extended or replaced as required. Communication with the forklift trucks is via Microsoft&#x02019;s IoT Hub<footnote id="fn7_1" label="1"><para>An IoT Hub is a site focused on the connectivity between software, the cloud and the devices used in everyday business operations.</para></footnote>. The industrial protocol MQTT is used for data transmission. The incoming data can be pre-processed via various stream analytics blocks and thus either reformatted or already evaluated in parts for further steps. The incoming data is stored in the data lake for the next analysis steps. This enables the possibility of o&#x0FB04;ine processing. Modified analysis methods can then be applied to vehicle data several times and the results can be compared directly with each other. This part is mainly used for RUL. For RCA additional the event hub is used, which can convert online analysis results directly into an action. This will be used for pattern recognition. To make the results available for the service technician in a special smartphone app, STILL uses the API app from Azure. Company data from the SAP system is coupled in via the Azure data factory element.</para>
<para>As mentioned in the previous paragraphs, forklift truck data in particular are used for the analysis. The forklift truck itself has a large number of sensors that are used to control its movements and processes to the operator. These sensors are primarily intended for internal control operations, so their measurement data are not transferred to the outside world for further processing as a standard. Therefore, there is a need for an additional external sensor that records on one hand all relevant external measurement data autonomously. So this sensor should provide the following external information to draw a precise picture of the environment conditions:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Acceleration in x, y and z direction;</para></listitem>
<listitem>
<para>Ambient temperature;</para></listitem>
<listitem>
<para>Humidity;</para></listitem>
<listitem>
<para>Pollution degree.</para></listitem>
</itemizedlist>
<para>On the other hand he should also have access to internal control values, with the aim to draw a precise picture of the detailed usage profile and transfer this data to the Azure cloud.</para>
<fig id="F7-94" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-94">Figure <xref linkend="F7-94" remap="7.94"/></link></label>
<caption><para>Raspberry Pi based sensor based.</para></caption>
<graphic xlink:href="graphics/ch07_fig094.jpg"/>
</fig>
<para>This sensor is shown in <link linkend="F7-94">Figure <xref linkend="F7-94" remap="7.94"/></link> and it was built prototypically, but due to the complex mechanical interfaces, e.g., for dirt detection, the sensor could not be implemented in this project for a real validation on the forklift truck. Nevertheless, the functional principle is promising and will be used in further developments. Without this sensor, the environmental data are initially determined via questionnaires and weather information from the internet. The vehicle information is provided by so-called soft sensors. Soft sensors are software solutions that convert existing system variables into algorithmic data.</para>
<fig id="F7-95" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-95">Figure <xref linkend="F7-95" remap="7.95"/></link></label>
<caption><para>Data Logging on the internal CAN bus.</para></caption>
<graphic xlink:href="graphics/ch07_fig095.jpg"/>
</fig>
<para>First, all available and relevant data for the creation of possible soft sensors were recorded via data logging and visualized for the forklift. For logging, powerful multi-channel industrial data loggers are used, which record the data traffic on the can bus in real time (see <link linkend="F7-95">Figure <xref linkend="F7-95" remap="7.95"/></link>). It is important to ensure that the process does not interfere, as the vehicles have a high risk potential during operation and are subject to the Machinery Directive. Especially since most measurements have to be carried out on real customers in order to obtain data sets that are as variable as possible. These several gigabytes of data were then processed via Matlab<footnote id="fn7_2" label="2"><para>Matlab (spelling: MATLAB) is a commercial software from the US company MathWorks for solving mathematical problems and displaying the results graphically.</para></footnote> to determine the most promising constellation.</para>
</section>
<section class="lev2" id="sec7-5-4">
<title>7.5.4 Data Analytics and Maintenance Optimization</title>
<para>The core problem for the subsequent analytical processes lies in both data preparation and the transfer of expert knowledge as described in the chapters before. The data is therefore first processed after the ETL<footnote id="fn7_3" label="3"><para>ETL is a process in which data from several possibly differently structured data sources is combined in a target database.</para></footnote> process to extract the core information. The data situation in mobile systems is fundamentally problematic for statistical processes, since the high transfer costs and the low storage depth of these devices mean that less data is available about these systems than needed for statistical analytics. The high influence of the environment also reduces the possibility to extract algorithms from the data, because each additional element need more validation data. The last burdening influence on analytics is the lack of precision in describing the problem. Many data sources (e.g., service reports) are not designed for later analysis processes. Several preliminary studies are conducted to find out which data sources and constellations are suitable for the expansion of its existing processes and business models. Among other things, tools such as Spotfire<footnote id="fn7_4" label="4"><para>TIBCO Spotfire Analytics is a commercial software platform for business intelligence solutions for the systematic analysis of internal and external data.</para></footnote> and programming languages such as R<footnote id="fn7_5" label="5"><para>R is a free programming language for statistical calculations and graphics.</para></footnote> are used for faster analysis visualization.</para>
<fig id="F7-96" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-96">Figure <xref linkend="F7-96" remap="7.96"/></link></label>
<caption><para>Example for graphical pattern recognition by Spotfire.</para></caption>
<graphic xlink:href="graphics/ch07_fig096.jpg"/>
</fig>
<para>The example in <link linkend="F7-96">Figure <xref linkend="F7-96" remap="7.96"/></link> illustrates the analysis of vehicle usage data with regard to fault events for RCA. In particular, pattern analysis plays an important role. Are there recurring patterns to which special errors can be assigned? Once such patterns are identified, appropriate processes can be stored to perform automatic service actions. For this reason, STILL initially selected errors for its basic analysis that have little scope for interpretation. For meaningful results, the ambient boundary conditions are also of great importance (see also <link linkend="ch05">Chapter <xref linkend="ch05" remap="5"/></link> Section 2). For the final results in real-time operating systems, it is necessary to choose learning systems that allow an adequate response also to each new situation.</para>
<para>In the area of wear, as mentioned above, STILL has concentrated on tires, assuming that the tires in particular give a good picture of how the truck is used by the driver and of the interaction with the environmental situation. STILL has analyzed the data sets of approximately 70 forklift trucks in the topic RUL in terms of use and tire wear.</para>
<para>However, the first approaches basing on the service reports were problematic, since the reason for replacing a tire can be quite different:</para>
<fig id="F7-97" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-97">Figure <xref linkend="F7-97" remap="7.97"/></link></label>
<caption><para>Tire changed due to wear and damage.</para></caption>
<graphic xlink:href="graphics/ch07_fig097.jpg"/>
</fig>
<para>Not every tire is changed because it is worn. A lot of tires are also changed, due to damages (see <link linkend="F7-97">Figure <xref linkend="F7-97" remap="7.97"/></link>), so the existing database is faulty with respect to precise information about the real wear. Due to this fact we start with a basic analytic, which will be optimized during operation.</para>
<para>In order to find a pragmatic solution, STILL approached the problem by analyzing the behavior of the different forklift trucks. This means that the forklift trucks are examined over a long period of time and their behavior is classified. The basis for this assumption is the fact that there must be a statistical correlation between physical usage and the resulting wear. Therefore, various operating conditions of the forklift truck are considered with regard to their possible influence. It turned out that agility seems to be the one of the most promising values, since its influence on wear is disproportionately high. For the first analytics the agility was divided into 100 elements, where 100 is representing an extreme dynamic driving (<link linkend="F7-98">Figure <xref linkend="F7-98" remap="7.98"/></link>). As a rule, the forklift trucks are observed over a period of one year, so that it can be assumed that we have an average of the typical usage of the trucks. The individual dynamic elements are summed up per vehicle and per class. Due to a simple weighting of the dynamic classes, a ranking of the dynamic use could be created, which corrodes in parts with the wear. Since the ground condition, load and environmental conditions were not considered, certain deviations can be explained. For the complete model further analyses are necessary, but these are not carried out in the course of this project.</para>
<fig id="F7-98" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-98">Figure <xref linkend="F7-98" remap="7.98"/></link></label>
<caption><para>Agility heat map of analysed forklift trucks.</para></caption>
<graphic xlink:href="graphics/ch07_fig098.jpg"/>
</fig>
<para>The graphic in <link linkend="F7-98">Figure <xref linkend="F7-98" remap="7.98"/></link> also shows that the weighting does not always allow an explicit conclusion. Vehicles with low dynamics can be clearly identified and the result fits to their wear. But there are also vehicles in the midfield in particular that have entries in the very dynamic classes. However, due to the weighting selected so far, they are classified as medium in terms of wear. Since the associated tire wear does not match the placement, it can be assumed that a linear weighting between the classes does not seem to finally apply. There are currently too few data sets available for a clear statement, so that the results still need to be sharpened with appropriate self-learning mechanisms.</para>
<fig id="F7-99" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-99">Figure <xref linkend="F7-99" remap="7.99"/></link></label>
<caption><para>Second step in architecture to improve process results.</para></caption>
<graphic xlink:href="graphics/ch07_fig099.jpg"/>
</fig>
<para>The extent to which the knowledge gained can be used to improve maintenance processes must be demonstrated by subsequent field studies. An installed example architecture (shown in green in <link linkend="F7-99">Figure <xref linkend="F7-99" remap="7.99"/></link>) will produce over a period of time to be defined, results of the algorithm. This results are then rated and used to refine the analytics.</para>
<para>The high level of service product requirements is due to the fact that a large part of these products are subject to a fee for the customer. For this reason, the highest diligence is required when developing solutions so that these do not have a negative cost effect on the customer and burden the business relationship. However, there is always a need to improve processes, since all services provided are subject to considerable cost pressure from the market. For this reason, the provision of individualized solutions and the increase in effectiveness is of great importance to the company.</para>
<para>The developed solution modules from the wear area are used in particular for the optimization of demand-based billing of services like rental or full service, while the RCA solutions are to be used for increasing the effectiveness in the processing of defects in the field.</para>
</section>
<section class="lev2" id="sec7-5-5">
<title>7.5.5 Conclusions</title>
<para>Considerations in the project have shown that mobile systems with small data volumes are particularly difficult to handle. For this reason, technical aspects are required in such systems in a much higher than expected dimension. Purely statistical observation cannot lead to success with such conditions.</para>
<para>It became clear that especially the knowledge of the required data is of great importance and an important key to success. It also became clear that the quality of the data has to be checked at an early stage and ensured by appropriate measures.</para>
<para>For STILL, these considerations provide clear starting points that can be used in future applications. Due to the very complex relationships between physical vehicle use and the interaction with environmental influences, an expansion of the data structures will be necessary in the future. The results achieved so far are refined in further validation loops and enriched by artificial intelligence. This is necessary in order to realize the high potential in the general area of optimization of maintenance processes. However, in addition to the classical support in problem solving tips and analyses results, there are also elements that will result from visual and/or HMI technology. These will even open up opportunities for further business areas.</para>
<para>A chosen system architecture must therefore always offer the possibility of expansion. The platform of this system architecture plays a rather subordinate role, as the solutions present themselves as a kind of modular system rather than an integrated solution.</para>
</section>
</section>
<section class="lev1" id="sec7-6">
<title>7.6 Proactive Maintenance of Railway Switches</title>
<blockquote role="flushleft">
<para><emphasis>Contributors: Csaba Heged&#x00171;s, Paolo Ciancarini, Attila Frank&#x000F3;, Ale&#x00161; Kancilija, Istv&#x000E1;n Moldov&#x000E1;n, Gregor Papa, &#x00160;pela Poklukar, Mario Riccardi, Alberto Sillitti, Pal Varga, Paolo Sannino, Salvatore Esposito, and Antonio Ruggieri</emphasis></para></blockquote>
<para>A larger pressure on the railway infrastructure has been created by a strong necessity for faster mass transport with high capacities and frequent runs. This makes it fundamental to continuously monitor the technical equipment of the railroad tracks in the most efficient way possible. By detecting fatigue wear of the track system in an early stage &#x02013; due to issues such as broken rails or increased rail wear, caused by natural hazards or by excess loading on the track system &#x02013; it is possible to avoid serious damage, also by means of correct interpretation of collected data, which allows for rapid intervention on the track system.</para>
<para>The railway use-case within the MANTIS project [The MANTIS consortium] is dealing with these issues through a proactive maintenance approach of the railway system [Hegedus et al., 2018]. This concerns the interlocking system and the study of possible complications that affect railway signalling &#x02013; i.e., non-functional and out of control situations &#x02013; with its main focus on describing the development of a set of approaches and support tools which allow to continuously analyse the status of specific components within the infrastructure. The use case aims at determining whether and within which limits it is possible to make reliable predictions for improving the maintenance process. In particular, it targets identifying anomalies and reducing emergency maintenance, since it is very costly and cases major train delays.</para>
<section class="lev2" id="sec7-6-1">
<title>7.6.1 Introduction to Railway Monitoring</title>
<para>In the railway infrastructure, the prevailing maintenance approach is still following the preventive model where most of the maintenance operations are based on periodical check-ups and substitutions of parts when a failure is detected. These tasks are carried out at given periodical intervals designed to mitigate risk with a considerable safety margin involving having to send maintenance staff to the asset site on a regular basis, exposing them to the usual safety risks of a running railway [Cocciaglia, 2012].</para>
<para>Modern railways have very low level of signalling installed. For switches, this comprises only of the detection if a switch is in the correct end position and locked. The development of new maintenance systems, including the integration of heterogeneous monitoring and diagnostic technologies, plays a key role in the improvement of railway safety operations. Existing monitoring solutions show some limitations due to their non-standardized, proprietary nature and very low integration level. Consequently, they are not able to monitor properly the degradation of complex asset, and to detect correlations between the condition of assets [Cocciaglia, 2012].</para>
</section>
<section class="lev2" id="sec7-6-2">
<title>7.6.2 Scope and Logic</title>
<para>A railroad switch allows trains to change tracks (<link linkend="F7-100">Figure <xref linkend="F7-100" remap="7.100"/></link>). When a train is destined to run on another track, the switch-man on the train or another employee in the railroad yard will turn the switch to direct the train toward the chosen direction. The railroad switch is activated by moving a long arm from side to side and moving the train tracks to the desired position. While many railroad switch activations are accomplished by hand, nowadays some are electronic and can be changed by an employee in an elevated office at the railroad yard.</para>
<fig id="F7-100" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-100">Figure <xref linkend="F7-100" remap="7.100"/></link></label>
<caption><para>Simple turnout schematic. Moving switch blades are shown in red.</para></caption>
<graphic xlink:href="graphics/ch07_fig100.jpg"/>
</fig>
<para>A realistic proactive maintenance solution for railway switches is based on the concept of Cyber-Physical Systems (CPSs), where a cyber-twin of the physical system is modelled, and its status is kept up-to-date through data collection from physical sensors deployed on-site.</para>
<para>The rest of this section provides some details on data processing, presents the proactive measurement system, and describes the developed data visualization subsystem.</para>
</section>
<section class="lev2" id="sec7-6-3">
<title>7.6.3 Data Processing</title>
<para>The data processing requires the consideration of all the available datasets connected to each switch. In particular, the data available are in the form of time series and can be grouped as follows:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Control:</emphasis> data generated by the switch control unit. They include commands sent to the switch (i.e., start, stop, etc.), and some feedback data provided by already existing sensors in the switch. The collected information is coarse grained, with a log sequence structure;</para></listitem>
<listitem>
<para><emphasis role="strong">Physical:</emphasis> data generated by sensors temporary added to the switches to measure some specific parameters. The most interesting data used in this analysis are the electric current consumed during the movement [Ampere], the duration of the movement [second], and the environmental temperature [<superscript>&#x02218;</superscript>C].</para></listitem>
</itemizedlist>
<para>The goal is to identify anomalies that could require a maintenance activity. We have focused on the following behaviours:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Drifts of the profiles:</emphasis> could be caused by accumulation of dust on the switch resulting in an increased amount of current leading also to a failure of the switch;</para></listitem>
<listitem>
<para><emphasis role="strong">Unexpected behaviour:</emphasis> could be caused by physical obstacles in the switch that may cause damages to the device.</para></listitem>
</itemizedlist>
<para>During data exploration, we have identified the following behaviours:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Behaviour 1</emphasis> (<link linkend="F7-101">Figure <xref linkend="F7-101" remap="7.101"/></link>)<emphasis role="strong">:</emphasis> This is a very noisy profile that makes the identification of the behaviour difficult and may highlight problems in the data collection. In particular, in the correct positioning of the sensor and/or the presence of sources of noise that may alter the collection;</para>
<fig id="F7-101" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-101">Figure <xref linkend="F7-101" remap="7.101"/></link></label>
<caption><para>Switch Data &#x02013; Behaviour 1.</para></caption>
<graphic xlink:href="graphics/ch07_fig101.jpg"/>
</fig>
</listitem>
<listitem>
<para><emphasis role="strong">Behaviour 2</emphasis> (<link linkend="F7-102">Figure <xref linkend="F7-102" remap="7.102"/></link>)<emphasis role="strong">:</emphasis> Similar to Profile 1 but with a limited amount of noise;</para>
<fig id="F7-102" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-102">Figure <xref linkend="F7-102" remap="7.102"/></link></label>
<caption><para>Switch Data &#x02013; Behaviour 2.</para></caption>
<graphic xlink:href="graphics/ch07_fig102.jpg"/>
</fig>
</listitem>
<listitem>
<para><emphasis role="strong">Behaviour 3</emphasis> (<link linkend="F7-103">Figure <xref linkend="F7-103" remap="7.103"/></link>)<emphasis role="strong">:</emphasis> Expected profile of a double switch;</para>
<fig id="F7-103" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-103">Figure <xref linkend="F7-103" remap="7.103"/></link></label>
<caption><para>Switch Data &#x02013; Behaviour 3.</para></caption>
<graphic xlink:href="graphics/ch07_fig103.jpg"/>
</fig>
</listitem>
<listitem>
<para><emphasis role="strong">Behaviour 4</emphasis> (<link linkend="F7-104">Figure <xref linkend="F7-104" remap="7.104"/></link>)<emphasis role="strong">:</emphasis> Expected profile of a switch;</para>
<fig id="F7-104" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-104">Figure <xref linkend="F7-104" remap="7.104"/></link></label>
<caption><para>Switch Data &#x02013; Behaviour 4.</para></caption>
<graphic xlink:href="graphics/ch07_fig104.jpg"/>
</fig>
</listitem>
<listitem>
<para><emphasis role="strong">Behaviour 5</emphasis> (<link linkend="F7-105">Figure <xref linkend="F7-105" remap="7.105"/></link>)<emphasis role="strong">:</emphasis> Profile of a switch with an abnormal behaviour.</para>
<fig id="F7-105" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-105">Figure <xref linkend="F7-105" remap="7.105"/></link></label>
<caption><para>Switch Data &#x02013; Behaviour 5.</para></caption>
<graphic xlink:href="graphics/ch07_fig105.jpg"/>
</fig>
</listitem>
</itemizedlist>
<para>The profiles of the current depend on several physical variables linked to the mechanical and electrical components that compose switches. These profiles are linked to the specific model of the switch from which the data are collected. The current is influenced also by environmental factors: temperature, humidity, and dust.</para>
<fig id="F7-106" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-106">Figure <xref linkend="F7-106" remap="7.106"/></link></label>
<caption><para>Identification of bands for quartiles and outliers.</para></caption>
<graphic xlink:href="graphics/ch07_fig107.jpg"/>
</fig>
<para>Due to the large variability of the profiles, our main problem is the identification of an approach to define the default correct behaviour. This can be achieved in different ways:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Physics:</emphasis> this approach is able to define the physical model of each switch, and it is able to predict the correct behaviour in many different environmental conditions. However, it requires building a model for each kind of switch and tuning the parameters for each installation;</para></listitem>
<listitem>
<para><emphasis role="strong">Statistics:</emphasis> this approach requires the collection of data from a wide set of devices in different operating conditions to define the default behaviours that are known with some level of uncertainty, but it does not require the manual development of a physical model for each switch. The model can be derived from the data and can be adapted to different switches collecting additional data.</para></listitem>
</itemizedlist>
<para><link linkend="F7-106">Figure <xref linkend="F7-106" remap="7.106"/></link> shows the statistical features of a specific switch. The black line is generated calculating the median of the different time series representing the current profiles of hundreds of movements that happened correctly in the past.</para>
<para>Threshold detection was used. If the current is outside the bounds, a warning is risen. The definition of proper bounds is of a great importance for the detection of an abnormal behaviour. As the distribution of the samples at each time instance is not normal, the outliers&#x02019; bounds are defined using the 1st and 3rd quartiles, according to the Tukey&#x02019;s range test for outliers [q1-1.5&#x0002A;IRQ; q3+1.5&#x0002A;IRQ].</para>
<para>There is quite a large range, especially at the end of the movement. After a deeper investigation, it was found that the behaviour was caused by the fact that the analysed single set of data was hiding two different data sets. Actually, the behaviour of the switch in the summer and in the winter is different due to the temperature sensitiveness. There are several aspects that depend on the temperature, such as the duration of the movements (longer in winter) and the current peaks (higher in winter). For these reasons, the statistical model had to consider the current season (the temperature of the environment). Therefore, the same analysis was repeated, but the dataset was divided into two sets based on the time of the year of the data.</para>
<para>The model validation was done by a bootstrap approach building the model using a random subset of correct movements in the same season and verifying it with the rest of the data. This analysis was helpful for defining the statistically correct behaviour of a switch using data coming from the field and tuning a model without any specific knowledge of the internal structure of the switch. Such a model can be easily adapted to different switches working in different conditions, and it was tuned using data in the different seasons.</para>
<para>The data can be analysed by log analysis approaches. However, the coarse grain and the lack of a sufficient information from the field (including tagged data describing anomalies) resulted in an analysis that was not able to build a relevant model that can actually be used.</para>
<para>So, the aim for analysing such data by applying those different statistical approaches is to determine a model of the default behaviour of the switch, and to identify anomalies in the behaviour. Among various purposes of diagnosis and prognosis [Jantunen et al., 2016], this can be used for failure prediction [Fronza et al., 2013], and other proactive maintenance purposes [Lenarduzzi et al., 2017], including root cause analysis and the calculation of remaining useful life.</para>
<fig id="F7-107" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-107">Figure <xref linkend="F7-107" remap="7.107"/></link></label>
<caption><para>Measurement system setup.</para></caption>
<graphic xlink:href="graphics/ch07_fig106.jpg"/>
</fig>
</section>
<section class="lev2" id="sec7-6-4">
<title>7.6.4 Measurement System for Proactive Maintenance of Railway Switches</title>
<para>For failure prediction and diagnostics, a new maintenance system was needed. We built a new, low cost non-invasive measurement system that can be attached in retrofit to operational switches. The measurement system measures the factors that affect the life expectancy of the railway infrastructure. The choice of the appropriate attributes is based on expert knowledge since the different types of switches are not equally affected by these impacts.</para>
<para>The data acquisition system is part of an architecture which is based on the MANTIS platform [Hegedus et al., 2018] and complies with the architecture of the platform in full extent. The architecture of the system is shown on <link linkend="F7-107">Figure <xref linkend="F7-107" remap="7.107"/></link> and consists on the following modules: (1) Standalone data gathering edge device; (2) Edge broker implementing MQTT; (3) MIMOSA database on a Microsoft SQL Server; (4) Data analytic modules; and (5) MANTIS Human-Machine Interface (HMI).</para>
<para>The edge device &#x02013; which is the embedded subsystem deployed with the railway switch &#x02013; is not only responsible for gathering new data but pre-processing and forwarding it to the cloud in an appropriate, MANTIS-enabled message format. The heart of this device is an STM32F4 series MCU (Microcontroller Unit) which employs a single ARM-Cortex-M4 core is capable of collecting, storing and pre-processing the information, while also handing the communication tasks as well. It offers numerous interfaces &#x02013; including UART, SPI, I2C &#x02013;, and 12-bit analogue-to-digital converters; thus both analogue and digital sensors can be used.</para>
<para>In this use case, the edge device contains one digital integrated humidity and ambient temperature sensor, a digital temperature sensor and four analogue displacement sensors.</para>
<section class="lev3" id="sec7-6-4-1">
<title>7.6.4.1 New factors collected</title>
<para>The system measures several factors that can affect the wear of the railway switch over time. These expert-identified factors can be divided into two groups:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Operational factors:</emphasis> These parameters are directly related to the operation of switches &#x02013; they have direct impact on condition deterioration. In our implementation, we measure lateral and longitudinal displacement of point blades. These point blades direct trains to one of the possible paths, i.e., they are the moving parts of a switch. Here the excepted resolution is high, and we are interested in gathering data only during switching sequences;</para></listitem>
<listitem>
<para><emphasis role="strong">Environmental factors:</emphasis> These parameters are well-known to affect almost every cyber-physical system. The most significant one is temperature. Both the ambient temperature and the temperature of the rails are measured. The rail temperature can cause dilation of rails thus it affects the operation of switches indirectly. Another environmental factor is humidity, which plays a lead role in corrosion. Since the ambient parameters are changing slowly, reading the values periodically, every half an hour provides appropriate accuracy and resolution for this use-case.</para></listitem>
</itemizedlist>
<para>The blade movement measurement must be event driven, data is collected when a switching of blades occurs. Therefore, the switching itself must be detected. Detecting the start of a switchover is tricky, since measurement noise and passing trains may interfere. Therefore, a threshold based triggering is used together with a pre-fetch measurement phase, as <link linkend="F7-108">Figure <xref linkend="F7-108" remap="7.108"/></link> shows. The thresholds are set high enough to avoid false positive triggers, and the pre-fetch phase ensures that the acquired data contains the full movement of the blades. Moreover, if the device starts a measurement and the actual position does not reach the end position &#x02013; just nearly approaches it &#x02013;, the measurement cycle will not stop. In this case all information about the movement between the real end positions and the threshold levels would be lost.</para>
<para>The state-transitions of the measurement are presented on <link linkend="F7-109">Figure <xref linkend="F7-109" remap="7.109"/></link>. In the case when the measurement takes longer than a predefined (expected) interval, the measurement stops and triggers the device to send a warning message to the central cloud. This function indicates an error, which means that the point blades cannot reach their end position &#x02013; so the switching operation failed.</para>
<fig id="F7-108" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-108">Figure <xref linkend="F7-108" remap="7.108"/></link></label>
<caption><para>Switch&#x02019;s point blades displacements.</para></caption>
<graphic xlink:href="graphics/ch07_fig108.jpg"/>
</fig>
<fig id="F7-109" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-109">Figure <xref linkend="F7-109" remap="7.109"/></link></label>
<caption><para>Measurement state diagram.</para></caption>
<graphic xlink:href="graphics/ch07_fig109.jpg"/>
</fig>
<section class="lev4" id="sec7-6-4-1-1">
<title>7.6.4.1.1 Platform level</title>
<para>The gathered information is encoded in an interoperable JSON-based message format developed within the MANTIS project, based on the MIMOSA [MIMOSA Consortium] domain ontology. The messages contain not only the results of measurements, but additional information: (i) exact timestamp, (ii) duration of the measurement, (iii) identifier of the edge device instance and (iv) additional values that help the re-assembly of the message at broker side.</para>
<para>The messages are transmitted via the MQTT protocol over TCP/IP. The wireless connection between the edge device and the central cloud is provided by a SIMcom SIM800 based GPRS modem which is attached to the MCU via serial line. The central cloud contains an MQTT Edge Broker, which handles the messaging, while both the Low-level Device and the cloud have an MQTT implementation each.</para>
<para>In the central cloud, the message is received by a Mosquitto MQTT broker [Mosquitto<superscript>TM</superscript>, 2010] with a parser client. The information is then stored into a MIMOSA OSA-CBM database, which is a standard architecture for condition-based maintenance systems. The parsed datasets will be processed o&#x0FB04;ine by data mining and analysing tools. Future work includes that the incoming message can be analysed online, automatically by a stream processor. This will enable an automated alerting and forecasting system.</para>
<para>The processed and analysed information is stored in the database, thus the central cloud can provide relevant information to different parts of the MANTIS architecture, for example for the Human-Machine Interfaces.</para>
</section>
</section>
</section>
<section class="lev2" id="sec7-6-5">
<title>7.6.5 Data Visualization</title>
<para>To increase the efficiency of the maintenance personnel and to evaluate the results of the data analysis [Koro&#x00161;ec et al., 2013], an intelligent HMI had to be developed. Following the scenario-based design approach, user needs and the context of use were described in the human-machine interaction scenarios. In the iterative process of scenarios refinement, five main human roles have been identified, ranging from the maintenance technician to the business manager. Further refinement of the scenarios led to identification of three main functionalities of the user interface:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Monitoring the parameters given by the measurement box;</para></listitem>
<listitem>
<para>Displaying the alarms that indicate the abnormal movement of the railway switch;</para></listitem>
<listitem>
<para>Displaying the task schedule for the maintenance service.</para></listitem>
</itemizedlist>
<para>The interface was developed on top of the generic MANTIS HMI, described in <link linkend="ch06">Chapter <xref linkend="ch06" remap="6"/></link>. It supports multiple users with different roles, where each user or role can be presented with one or more dashboards covering their intended interaction. Dashboard is customizable and does not require any web development skills.</para>
<fig id="F7-110" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-110">Figure <xref linkend="F7-110" remap="7.110"/></link></label>
<caption><para>Graphics, displaying the state and the position of the switch (left) and the instant values of the environmental parameters (right).</para></caption>
<graphic xlink:href="graphics/ch07_fig110.jpg"/>
</fig>
<para>As it can be seen in <link linkend="F7-110">Figure <xref linkend="F7-110" remap="7.110"/></link>, the HMI allows the user to quickly see the position of the railway switch through the corresponding graphics. An additional graphics with the IoT image indicates the connection to the measurement box to ensure the reliability of the data. When the connection is established, the image turns green. Otherwise, the user should not assume that the data presented on the interface is accurate or up to date. In addition to the switch position and state, instant parameter values related to the railway switch, such as the rail temperature, switch status, ambient temperature and humidity, are displayed. If the alert thresholds for a measurement are set, values out of range will also be shown as alarms. Historic values of the raw sensor measurements and environmental parameters are displayed as a graph (<link linkend="F7-111">Figure <xref linkend="F7-111" remap="7.111"/></link>). Visualization of the data analysis and prediction results is done through the same monitoring widgets, which can also show predictions, remaining useful life estimations. For more in-depth analysis, Kibana visualizations are integrated.</para>
<fig id="F7-111" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-111">Figure <xref linkend="F7-111" remap="7.111"/></link></label>
<caption><para>Graph displaying blade displacement sensor measurements.</para></caption>
<graphic xlink:href="graphics/ch07_fig111.jpg"/>
</fig>
<para>Scheduled maintenance tasks are currently displayed in the alarms table, but they can also be displayed separately. The table is editable, filterable and sortable and allows the user to acknowledge the task/alarm as well as to enter textual feedback. To assist the maintenance personnel working on the field, a map with the location of the railway switch is displayed on a separate widget.</para>
<para>Several context-awareness features, mainly based on location and the user role, have been proposed to assist the maintenance personnel in performing their tasks. Such features proved to be most useful in performing the maintenance actions on the field, where the visualization of the information varies depending on the location of maintenance team. Another such example is a personalised suggestion of the user&#x02019;s next step according to their past interaction with the interface. In this way, the users are provided with the right information in the right moment and context.</para>
</section>
<section class="lev2" id="sec7-6-6">
<title>7.6.6 Conclusion</title>
<para>The Industrial Internet of Things, the concept of CPS and the industrial initiatives force the evolution of proactive maintenance solutions for industrial systems. This section presented some results, where the MANTIS concepts were applied to the use-case of railway switches. These include the measurement method, the analysis of the measured data, and the result visualization HMI tailored for various types of contexts and users &#x02013; ranging from the maintenance technician to the business manager.</para>
</section>
</section>
<section class="lev1" id="sec7-7">
<title>7.7 Fault Detection for Photovoltaic Plants</title>
<blockquote role="flushleft">
<para><emphasis>Contributors: Achim Woyte, Babacar Sarr, Karel De Brabandere, Tom Tourw&#x000E9;</emphasis></para></blockquote>
<para>The high-level objectives of the Photovoltaic plants use case are mainly concerning the reduction of efforts for operation and maintenance (O&#x00026;M) by more cost-efficient monitoring. This is pursued through smart sensing and data acquisition as well as through analysis and decision-making functions applied in real time to the operational data. These developed analytical methods are meant to reduce downtime and subsequent losses in electricity production due to component failures. Also one of the main objective is to improve the energetic performance of the plants by detecting design flaws, bad installation and maintenance practices, performance degradation of components over time, and sudden changes in the performance of components. Such possibility of early detections requires improvements of O&#x00026;M scheduling departing from root cause analysis, alerting and prediction functions, and maintenance optimisation.</para>
<section class="lev2" id="sec7-7-1">
<title>7.7.1 Introduction to PV Plants</title>
<para>Established in 1999, 3E is an independent technology and consultancy company. 3E provides solutions as well as guidance to improve renewable energy system performance, to optimise energy consumption and facilitate grid and power market interaction. 3E pursues innovation to provide leading energy intelligence and practical solutions to its customers and it disposes of long-term monitoring data sets recorded with high time resolution for more than 3000 PV installations distributed over the world with a total installed capacity of more than 2GW via its monitoring service SynaptiQ. 3E has worked on projects in more than 40 countries and operates with an international team of around 100 experts from its headquarters in Brussels and offices in Toulouse, Paris, Beijing, Istanbul, Cape Town, and London. 3E is certified ISO 9001:2008 since early 2010.</para>
<para>A huge potential for more effective pro-active maintenance actions passing through automated fault detection and identification lies in the large amounts of PV monitoring data that are recorded but currently used in a very limited way. It is in this perspective of exploiting this potential that 3E has been developing analysis and decision-making functions for proactive maintenance of Photovoltaic (PV) plants.</para>
<para>In view of the objectives of this use case, mentioned at the beginning of this section, the exploration and validation of analysis and decision-making functions for proactive maintenance of PV plants has therefore been the principal focus. 3E and its partners have been developing intelligent functions for pyranometers sensors, and overall automatic PV plant analysis to assess the &#x0201C;health&#x0201D; of the plant.</para>
<para>The following section provides an overview of a practical application of RCA developed for the Photovoltaic Plants use case lead by 3E. It focuses on the illustration of one of the techniques used for fault detection: Limit checking applied on the PV use case.</para>
</section>
<section class="lev2" id="sec7-7-2">
<title>7.7.2 Practical Application of Root Cause Analysis in Photovoltaic Plants</title>
<para>Photovoltaic plants are energy conversion systems. They convert the power of light, i.e., photon beams or electromagnetic waves, into electricity that can be used in an off-grid system and/or fed into the public utility grid in terms of frequency and voltage. The efficiency of this energy conversion step is influenced primarily by ambient temperature and secondarily by wind speed (Sw). Wind speed is often neglected.</para>
<para>Consequently, the primary input variables for this energy conversion process are the solar irradiance in the plane of the PV array (GPOA) and the ambient temperature (Tamb). The output variable is the electric AC power to the grid (PAC) as indicated in <link linkend="F7-112">Figure <xref linkend="F7-112" remap="7.112"/></link>. The PV module temperature (Tmod), the DC voltage, current and power at the output of the PV array (VDC, IDC, PDC, respectively), and the AC voltage and current to the grid (VAC, IAC, respectively) may be considered measurable state variables of this conversion process. The so-called yields (Y) and losses (L) describe the energy balance throughout the system in operation and are represented in <link linkend="F7-112">Figure <xref linkend="F7-112" remap="7.112"/></link> at the different stages of the plant.</para>
<fig id="F7-112" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-112">Figure <xref linkend="F7-112" remap="7.112"/></link></label>
<caption><para>Energy flow in a grid-connected photovoltaic system.</para></caption>
<graphic xlink:href="graphics/ch07_fig112.jpg"/>
</fig>
<para>Data received from the PV array comes in the form of time-series data. After standard data cleaning procedures, based on pre-mentioned variables listed above, normalized performance parameters are derived. They allow to quantify the energy flow and losses through the PV array per loss type.</para>
<para>They are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Availability loss: due to unavailability of grid, inverter, or DC input. This is encountered in the situation when power monitored from the plants is equal to zero while there is still light coming from the sun;</para></listitem>
<listitem>
<para>Array-current loss: due to deviations of the measured DC current from proportionality with irradiance through STC (Standard Test Conditions), for times when the plant is available;</para></listitem>
<listitem>
<para>Array-voltage loss: due to deviations of measured DC voltage from &#x02018;STC voltage&#x02019;;</para></listitem>
<listitem>
<para>Inverter loss: due to deviations between measured AC and DC power.</para></listitem>
</itemizedlist>
<para>The yields and losses are typically hourly average values but can be integrated over time.</para>
<para>The main variables used for limit checking are solar irradiance in the plane of the PV array (GPOA), ambient temperature (Tamb), PV module temperature (Tmod), DC voltage and current at the output of the PV array (VDC, IDC) and electric AC power injected to the grid (PAC). The AC voltage (VAC) and power factor (PF) are not used for limit checking.</para>
<para>For checking the operational performance over different energy conversion steps, a performance loss ratio per step is defined. This performance loss ratio is computed for a given time span, e.g., a day up to several months. It is the useful energy lost over the energy conversion step divided by the energy available, i.e., the incoming solar energy on the PV array as represented by the solar irradiance in the plane of the PV array (GPOA); all normalized to standard rating conditions of the PV array. Accordingly, the overall performance of a PV plant is described by the performance ratio (PR), i.e., 100% minus the sum of all performance losses.</para>
<para>In practice, we compare the performance loss ratios from measurements to model-based performance loss ratios and thresholds. The model is fed with measured values of GPOA and Tamb. The model parameters can be set from data sheet parameters of the devices in the PV plant or identified from measurements from the plant in a healthy state. Accordingly, adequate limits can be derived either from tolerances on the data sheet parameters or from choosing percentiles from the healthy plant. Both the model-based performance loss ratios and their limit values vary depending on the PV plant and the weather during the evaluation period.</para>
<fig id="F7-113" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-113">Figure <xref linkend="F7-113" remap="7.113"/></link></label>
<caption><para>Example of limit checking results for the energy conversion process in a PV plant; performance loss ratios per conversion step are compared to the model for each conversion step.</para></caption>
<graphic xlink:href="graphics/ch07_fig113a.jpg"/>
</fig>
<para><link linkend="F7-113">Figure <xref linkend="F7-113" remap="7.113"/></link> illustrates this application of limit checking for a PV plant located in Belgium. The current-related array losses (&#x02018;Array (current)&#x02019;) in the upper half of <link linkend="F7-113">Figure <xref linkend="F7-113" remap="7.113"/></link> by far exceed the threshold. During a thorough maintenance action after this problem was detected, several smaller PV module failures were fixed. After maintenance action, all performance loss ratios were back within their expected ranges, yielding a much higher PR of 82.9% (lower half of <link linkend="F7-113">Figure <xref linkend="F7-113" remap="7.113"/></link>).</para>
</section>
</section>
<section class="lev1" id="sec7-8">
<title>7.8 Conventional Energy Production</title>
<blockquote role="flushleft">
<para><emphasis>Contributors: Matti Kaija, Antti Niemel&#x000E4;, Juha Valtonen, Ville Rauhala, Veli-Pekka Salo, Erkki Jantunen</emphasis></para></blockquote>
<para>The most important goal of a power plant operator is to maximize the availability and performance of their power plant and to minimize the generating costs of energy. To reach this goal, the maintenance of the power plant is crucial and maintenance can account up to 30 % of the cost of the produced energy.</para>
<para>Most modern power plants use a combination of scheduled maintenance and corrective maintenance. In this strategy, critical components, such as turbines and pumps, are serviced based on statistical trend data and less critical components are left to a run-to-failure strategy, where maintenance is not done until the machinery fails. These maintenance strategies are not cost optimized since scheduled maintenance usually results in either unnecessary repairs if servicing is done too frequently or potentially catastrophic failures if service is neglected. Ineffective maintenance not only wastes materials and resources, but lost operating time resulting from equipment failure is a significant expenditure to a power plant operator.</para>
<para>Effective maintenance planning and scheduling is essential for shortening revision and downtime periods. Most of the revision work is done by external contractors and work delays can be expensive. By maintaining components based on their condition, revision work and cost can be optimized.</para>
<para>Current condition based maintenance strategies usually rely on periodical measurements done by experts, because the power plant staff do not have the necessary skills to analyse the measurement data. Most power plants collect information that could be utilized in condition based maintenance, but the data is not used because it is difficult to interpret or because it lacks parameters for fault prediction. Implementing new sensors in power plant equipment can be difficult.</para>
<section class="lev2" id="sec7-8-1">
<title>7.8.1 Introduction to the Plant Under Study</title>
<para>The J&#x000E4;rvenp&#x000E4;&#x000E4; power plant is combined heat and power (CHP) power plant owned by Fortum Power and Heat Oy and operated by Maintpartner Oy. The power plant is located in the city of J&#x000E4;rvenp&#x000E4;&#x000E4; in the Tuusula municipality in southern Finland. The power plant main boiler K1 is a fluidized bed boiler (BFB) with a fuel capacity of 76 MW. The K1 boiler was commissioned in 2013 and has a wide range of utilizable fuels, mainly wood biomass, peat and waste based fuels. The plant also has three natural gas boilers for peak- and reserve situations. Yearly the power plant generates 250-330 GWh heat for the district heating network and 100 GWh electricity.</para>
<para>The flue gas recirculation blower was selected as the monitored machine due to its easy accessibility and constant run status. The blower is used to recirculate scrubbed flue gas back to the boiler to lift the fluidized bed in the boiler in order to cool the bottom of the boiler furnace. The recirculation fan is not an immediately process critical component, but a failure can affect the combustion process and prolonged operation can reduce the boiler lifecycle by overheating the boiler grate.</para>
<para>The blower consists of three parts, the engine, bearing and impeller inside the impeller housing. Most common faults for any rotating machines are bearing failures, imbalance and misalignment. All of these causes have distinct vibration patterns that can be identified with proper vibration instrumentation. The pilot instrumentation consisted of vibration sensors installed to the blower bearing and a tachometer for measuring the rotation frequency of the blower. The monitored blower is shown in <link linkend="F7-114">Figure <xref linkend="F7-114" remap="7.114"/></link>.</para>
<fig id="F7-114" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-114">Figure <xref linkend="F7-114" remap="7.114"/></link></label>
<caption><para>Flue gas recirculation blower monitored in the pilot project.</para></caption>
<graphic xlink:href="graphics/ch07_fig114.jpg"/>
</fig>
</section>
<section class="lev2" id="sec7-8-2">
<title>7.8.2 Scope and Logic</title>
<para>The pilot project consists of instrumenting a flue gas recirculation blower at J&#x000E4;rvenp&#x000E4;&#x000E4; power plant with several different types of vibration sensors and creating a data collection and data storage system for the data. Also implementing data collection from the plant process data to the same data storage as the vibration data is studied. The goal of the project is generate a pilot condition monitoring system including installed sensors, data connection and collection and data analysis and display possibilities. The current planned pilot structure is presented in <link linkend="F7-115">Figure <xref linkend="F7-115" remap="7.115"/></link>.</para>
<fig id="F7-115" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-115">Figure <xref linkend="F7-115" remap="7.115"/></link></label>
<caption><para>Conventional energy production use case platform.</para></caption>
<graphic xlink:href="graphics/ch07_fig115.jpg"/>
</fig>
<para>The pilot has been divided into three phases for management purposes. The first phase consists of installation of the sensors and local data collectors at the site. The installation takes into account the current Finnish standards regarding vibration measurements and monitoring [PSK Standards Association, 2006, 2007]. In the second phase, the local data collectors are connected to the MANTIS data storage that is created following the MIMOSA data structure presented in the OSA-CBM and OSA-EAI standards [MIMOSA, 2010, 2014]. The data storage also utilizes the reference architecture [Mantis Consortium, 2016]. From the MANTIS data storage, the data is distributed to the individual systems. Some of the sensors also provide direct access via internet that is used in the first phases of the pilot to collect preliminary measurement data. The third phase focuses on the analysis of the data and failure prediction.</para>
<para>The pilot structure allows for comparison of several different sensor types to find the most applicable sensor for this type of condition monitoring if such a preference can be made. The standardized data structures for data communication and storage provide a basis for collaborative use and development of the MANTIS platform.</para>
<para>The analysis and failure prediction study of the pilot is focused on rotating machines. The applied techniques will be specified later in the project after a preliminary technical study is complete and the data collection system provides preliminary results for analysis.</para>
<para>As several partners already have commercial sensor and data collection solutions available, these are used for maximum benefit in order to focus the research work to areas that have not yet been studied. The main focus of the pilot is the collaboration of different standardized solutions and protocols and the utilization of the collected data to create new value. The data analysis and failure prediction is a relatively new area that is not implemented in the commercial systems and will require research on the analysis methods, not only implementation and collaboration research.</para>
</section>
<section class="lev2" id="sec7-8-3">
<title>7.8.3 Monitoring Rolling Element Bearings</title>
<para>Rolling element bearings are vital components of rotating machinery and they can be found in almost all rotating machines. In theory they are usually designed so that they should last the whole life time of the machine assuming that proper lubrication is provided and that no over-loading takes place. Unfortunately in real life it is rather typical that something goes wrong either with the lubrication or the loading and that initiates the wear of the bearing. Assuming that the outer or inner rolling surface has suffered, the wear takes place with increasing speed because the wear particles tend to cause further wear and worn surfaces cannot withstand the loading as well as new intact surfaces. Consequently bearing wear develops in exponential way. From financial point of view, it&#x02019;s important to know if the worn bearing will last until the next planned stoppage since in many cases it is very costly to stop the production in order to change one bearing. Naturally, one option could be to have redundancy i.e., a spare machine that could be used while the one suffering from bearing wear is repaired. However, it is easy to understand how costly this kind redundancy would be as in such a case two factories would be needed to do the work of one.</para>
<para>Traditionally the condition monitoring of bearings is carried out manually so that a trained professional is manually doing measurements once a month or every two weeks and possibly so that if something strange has been noticed in the measured signals the time period between consecutive measurements has been reduced to much shorter level e.g., once a day measurements. Clearly, this kind of manual monitoring is rather costly and takes a lot of effort in industry because of the high number of potential bearing failure objects. The aim is to automate the above described measuring process and also to be able to carry out the diagnosis automatically i.e., define whether a bearing fault is initiated with signal analysis and diagnosis based on artificial intelligence. In addition, the capability of predicting the remaining useful life is one of the objectives. In practice this means that we can predict when the latest date would be when the bearing has to be changed.</para>
<para>During recent years, the price of sensors and processors has reduced dramatically and this is the reason why three different type of vibration monitoring solutions for the detection of bearing wear have been tested. One of the tested solutions is the Nome nmas system, which is developed for this type of purpose i.e., condition monitoring of machinery. The second tested solution WRM can be seen as a more generic platform that supports not only vibration measurement but also a wide range other kind of techniques. The third option is based on low cost components Mems sensor and RaspberryPi [Junnola, 2017] for processing the data, (see <link linkend="F7-116">Figure <xref linkend="F7-116" remap="7.116"/></link>). The idea is to find out how well this kind of solution performs when compared to more sophisticated equipment. The results of that comparison are discussed in chapter 4 of this book.</para>
<fig id="F7-116" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-116">Figure <xref linkend="F7-116" remap="7.116"/></link></label>
<caption><para>RaspberryPi based measuring system with a mems accelerometer.</para></caption>
<graphic xlink:href="graphics/ch07_fig116.jpg"/>
</fig>
<para>The most common technique today to diagnose whether a bearing fault is present or not is so called envelope detection, which is based on the detection of the vibration impulses that are caused when a rolling element hits the worn surface. The impulses vibrate at the first natural frequency of the structure in question i.e., at relatively high frequency 0.5&#x02013;5 kHz. The frequency at which the impulses take place reveal which kind of fault is present. Is it outer or inner race or possible cage fault? Naturally the reliable detection of bearing faults at an early stage is the key action in condition monitoring. Quite a lot of effort is dedicated in being able to predict the development of the wear process. This is a very challenging task as there are so many factors that influence this phenomenon. For example, the loading and lubrication conditions together with the bearing geometry and material have an influence. The fact that the signal analysis indicators that reveal the existence of a bearing fault do not increase linearly but instead both increase and decrease at certain phases. The development in the end of life is so rapid that it makes the prognosis process challenging [38]. These techniques have been more thoroughly discussed in chapter 5 of this book.</para>
<para>With the introduction of new sensor types and processing power of current CPSs, the amount of data that needs to be managed increases dramatically and this in turn emphasises the role of the platform that is for this purpose. In this use case the idea is to use hierarchical data structure so that most of the data is processed locally and only meaningful information of exceptions is passed to higher levels. For example, the RaspberryPi processor is capable of carrying out the necessary signal analysis tasks together with the diagnosis of possible faults. The maths are programmed with Python which is an open programming language dedicated to mathematical programming. It should be noted that due the openness of the programming language a lot of useful material is available free of charge. The data is at all levels (locally e.g., RaspberryPI, plant, cloud -Azure in this use case-, service centre) managed with MIMOSA [Gorostegui, 2017]. This is again an open solution for maintenance related data. MIMOSA can hold data of the bearing type and geometry, its maintenance history (who, when, what &#x02026;), measurement data, data to support the making of diagnosis and prognosis, results of diagnosis and prognosis. Basically, this is all data that is needed for a CMMS that is handled with MIMOSA. MIMOSA has also served the purpose of integration (installation in Kemi) between the various systems that the individual partners have been using. Mimosa is the platform for supporting the development of OSA-CBM Web Services that are developed for diagnostic and prognostic purposes.</para>
</section>
<section class="lev2" id="sec7-8-4">
<title>7.8.4 IoT-Ticket Platform</title>
<para>One of the measurement platforms used in the pilot was industrial IoT platform IoT-Ticket and reference edge computing device WRM247+ both developed and owned by Wapice. For data connectivity IoT-Ticket offers several possibilities to connect into data sources. These are e.g., OPC, OPC UA, MQTT and other industrial standard communication methods. Custom connectivity is possible through ready-made developer libraries and REST API. Using the WRM247+ multi-purpose data collection and edge computing device it is also possible to execute vibration measurements. As an off-the-shelf solution for the vibration measurements it is possible to connect the device into IFM VSE sensor gateway solutions. Connectivity to any industrial standard IEPE 4-20mA vibration sensor is built during the pilot phase in addition to existing methods. In pilot setup the WRM 247+ device was directly connected to vibration sensors using a signal condition amplifier. The benefit of this approach is that it gives a full control of all measurements that are done in the device. IMI 603C01 sensors are selected and connected to the system under test using magnetic connectors.</para>
<para>A support for computing RMS (Root Mean Square), Peak (Maximum Peak) and Crest (Peak/RMS) time domain analyses (KPIs) and FFT analysis for sampled signal is implemented and configured appropriately. The pre-processed data is then uploaded to IoT-Ticket server in regular intervals for further analysis and condition dashboards and reports (see <link linkend="F7-117">Figure <xref linkend="F7-117" remap="7.117"/></link>).</para>
<fig id="F7-117" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-117">Figure <xref linkend="F7-117" remap="7.117"/></link></label>
<caption><para>IoT-Ticket dashboard utilizing the 3D model built by LapinAMK. Live values are fused into 3D model.</para></caption>
<graphic xlink:href="graphics/ch07_fig117.jpg"/>
</fig>
<para>Using the Interface Designer and graphical flow programming tools the connection to MIMOSA database is implemented by creating the necessary flows to connect, read and analyze required data. From IoT-Ticket connectivity to MIMOSA was done using a standard flow-component that allows connectivity to external REST sources. As a parameter this component takes a combination of username and password, source URL and REST method (contains the XML/JSON payload). Virtual data tags allow forwarding REST response into IoT-Ticket&#x02019;s system. In order to post-process the data further several diagnostics flows are created to automatically monitor the vibration levels. Data driven events are enabled by utilizing the IoT-Ticket alarms and reports feature. Automatic reporting was setup to trigger if something exceptional would happen in the diagnosed data. This could be for example an exceptional signal level in vibration measurements or a decision based on statistical information computed inside IoT-Ticket or one of the external prognostics service providers available through MIMOSA.</para>
</section>
<section class="lev2" id="sec7-8-5">
<title>7.8.5 nmas Measuring System</title>
<para>Nome used and further developed the nmas monitoring system in the use case. The nmas monitoring system is developed and owned by Nome and is made for condition monitoring of rotating machinery. Connectivity to MIMOSA database is made through a REST API. The pilot case monitoring system included a local measurement unit capable of measuring, calculating, and storing data. A remote connection is build using 4G connection and locally the measurement unit is connected through WLAN.</para>
<para>Industrial standard 100mV/g IEPE acceleration sensors and optical tachometer are used for measurements. All measurement channels are measured simultaneously. Local device calculated velocity RMS (Root Mean Square) values and acceleration peak value continuously and stored time signals in defined intervals. Pre-Alarm and alarm limits are set according to ISO10816 standards.</para>
<para>Analysing of signals is done using nmas Analysator or View Java based analysing tools. A browser based viewing interface is made to view results with cellular phone or tablet. Viewing software and browser interface displays measurement trends, raw time signal, velocity spectrums, and alarm statuses. More sophisticated analysis including different mathematical functions, band-pass filters, window functions, different spectrums are available with analysis software. Basic view of nmas View and Analysator software is presented in <link linkend="F7-118">Figure <xref linkend="F7-118" remap="7.118"/></link>.</para>
<para>During the project Nome developed nmas Simple measuring device that is highly adaptive and easy to install condition monitoring system for local and remote monitoring of critical machinery. Nmas Simple measurement device is presented in <link linkend="F7-119">Figure <xref linkend="F7-119" remap="7.119"/></link>.</para>
<fig id="F7-118" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-118">Figure <xref linkend="F7-118" remap="7.118"/></link></label>
<caption><para>nmas Analysator UI basic view.</para></caption>
<graphic xlink:href="graphics/ch07_fig118.jpg"/>
</fig>
<fig id="F7-119" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-119">Figure <xref linkend="F7-119" remap="7.119"/></link></label>
<caption><para>nmas Easy condition monitoring device.</para></caption>
<graphic xlink:href="graphics/ch07_fig119.jpg"/>
</fig>
</section>
<section class="lev2" id="sec7-8-6">
<title>7.8.6 Mantis Cloud Platform</title>
<para>A Microsoft Azure based MIMOSA deployment is used as an information exchange platform. A REST interface into MIMOSA was developed by Lapland University of Applied Sciences (LUAS) in order to integrate various different measurement platforms together. It&#x02019;s been developed in JAVA using existing libraries such as Jersey2 and Jackson to provide basic REST and JSON functionalities. The REST interface is named MIREI short for MIMOSA REST Interface. RESTful approach is chosen due to its simplicity, ease of use and bi-directionality. Any possible performance related downsides in the way REST operates is outweighed by the overall ease of integration and flexibility of the approach.</para>
<para>MIREI was initially released as two separate variants with slightly different HTTP command structures; the standard MIREI and the experimental MIREI. The standard MIREI provides helpers to make CRUD operations more streamlined and concise. It also can contain vendor specific data mappers that allow systems to store and retrieve native data from MIMOSA while still retaining MIMOSA compliancy. Such a mapper is developed for Nome&#x02019;s measurement system. The standard MIREI release however is only available for specific tables and does not allow access to other MIMOSA tables without further expansion. The MIREI experimental enabled access to all MIMOSA tables. This, however, requires better understanding of the MIMOSA data model to be of use. It does not contain any helper functions and will require users to fill in all the fields marked NOT NULL and adhere to the constraints existing within the data model. Though the new commands automatically fill the information related to row updates.</para>
<para>Later the two were merged into a single release enabling both functionalities, however the REST URLs are still kept separate. This is done in order to make it easier to access the REST commands in scenarios where it is necessary to use both the helpers and have a more complete access to the underlying MIMOSA database. <link linkend="F7-120">Figure <xref linkend="F7-120" remap="7.120"/></link> shows the REST interface and its role in this use case.</para>
<fig id="F7-120" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-120">Figure <xref linkend="F7-120" remap="7.120"/></link></label>
<caption><para>REST interface for this use case.</para></caption>
<graphic xlink:href="graphics/ch07_fig120.jpg"/>
</fig>
<para>The type of data inserted into the MIMOSA database is mostly focused on measurement data and generated data provided by the analytics, prognostics and simulation tools developed by VTT. Envisioned CMMS and ERP data integration is not completed, however the REST interface would make this possible. The MIMOSA database is used to store different types of location information used in augmented reality and virtual reality applications. This is accomplished by creating new data types to the MIMOSA and using the existing tables related to assets, segments and measurement locations.</para>
<para>HTTP Basic authentication is used to restrict access to MIREI. Partners received their own username and password for the system. There is also possibility to enable more secure token-based authentication for the MIREI. In the token-based authentication method the client needs to request authentication token from the server using the login information provided and then include this token in the REST headers.</para>
</section>
<section class="lev2" id="sec7-8-7">
<title>7.8.7 Data Analytics and Maintenance Optimization</title>
<para>The main optimization problem in power plant maintenance is to minimize the total lifecycle maintenance costs and to maximize the total lifecycle availability. This means that it could sometimes be beneficial to select &#x0201C;sub-optimal&#x0201D; operational modes to reduce maintenance costs or vice-versa. However these decions need to be made based on actual data that is carefully analyzed.</para>
<para>The success/failure of lifecycle optimization is measured by total plant availability and maintenance costs. However power plants are designed to operate for 25&#x02013;40 years making lifecycle optimization difficult. In order to collect better data and thus make more informed decisions, better data collection and analysis tools need to be developed. Modern data collection, remote monitoring and analysis tools (neural networks, statistical analytics, physical models) also allow cost effective implementation of more advanced maintenance methods on less critical components.</para>
</section>
<section class="lev2" id="sec7-8-8">
<title>7.8.8 Conclusions</title>
<para>This use case represents a quite normal situation that occurs in power plant environments. Power plants are long-term investments and house IT systems from several different vendors and technologies that need to be integrated with each other and need to be able to communicate. Each of the different technologies operate in their individual fields of expertise, but can have common elements such as data collection and databases.</para>
<para>This use case represents such an integrated system and the research done in the use case provides a reference architecture of how such a system can be built as well as benchmarking some of the open source technologies such as MIMOSA database and REST API needed for the integration.</para>
<para>Each of the partners also continued developing the individual components of the integrated system to provide improved analysis, connectivity and prognosis capabilities for the users. HMI and AR/VR is developed by the partners and is presented and discussed in <link linkend="ch06">Chapter <xref linkend="ch06" remap="6"/></link>.</para>
</section>
</section>
<section class="lev1" id="sec7-9">
<title>7.9 Health Equipment Maintenance</title>
<blockquote role="flushleft">
<para> <emphasis>Contributors: Jeroen Gijsbers, Mauro Barbieri, Verus Pronk, Hans Sprong, Jaap van der Voet, Godfried Webers, Karel Eerland, Marcel Boosten, Kees Wouters, Mike Holenderski, and Alp Ak&#x000E7;ay</emphasis></para></blockquote>
<para>This chapter provides an overview of practical application of several elements developed for the Health equipment use-case lead by Philips Healthcare. The chapter focusses on the most essential activities carried out by Philips Healthcare and its research partners, as the whole use-case description would go into too much detail for a chapter.</para>
<section class="lev2" id="sec7-9-1">
<title>7.9.1 Introduction to Health Imaging Systems</title>
<para>Healthcare Imaging Systems are essential for the diagnosis and treatment of patients in hospital and private clinics (<link linkend="F7-121">Figure <xref linkend="F7-121" remap="7.121"/></link>). Due to the complexity of these systems and the large costs involved, it is not economically feasible to implement backup systems. Therefore, system uptime has to be maximized, planned downtime has to be minimized and unplanned shutdown has to be prevented. To cope with the exploding cost of healthcare, the cost of ownership has to be reduced, which also implies that maintenance budgets are under pressure. In response, Philips Healthcare has developed maintenance services for hospitals based on remote monitoring of their systems.</para>
<fig id="F7-121" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-121">Figure <xref linkend="F7-121" remap="7.121"/></link></label>
<caption><para>Unplanned system shutdown has a large impact on patients and hospital staff.</para></caption>
<graphic xlink:href="graphics/ch07_fig121.jpg"/>
</fig>
<para>The biggest challenge there is to retrieve, store and analyze large amounts of data from globally distributed systems such that predictive maintenance can be offered instead of maintenance at fixed time intervals. Furthermore, an alerting system is necessary when the online data analysis detects a threat of shutdown.</para>
<para>Due to the large purchase cost and the cost of housing, unplanned shutdown has a large impact on the hospitals and on the patients who may not get the care they need. Philips Healthcare made use of MANTIS Reference Architecture for equipment asset optimization, thereby aiming to move from a reactive to proactive and predictive maintenance.</para>
<para>The objective is to accurately predict upcoming failures by mining large amounts of data from heterogeneous systems distributed globally, such that maintenance can be timely scheduled or in urgent cases, the responsible person can be alerted. The main challenge is understanding how to get from large amount of data to accurate and precise failure detection and prediction.</para>
<para>Next to that, the availability of data and the analytical outcomes can give additional opportunities to exploit this information.</para>
<fig id="F7-122" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-122">Figure <xref linkend="F7-122" remap="7.122"/></link></label>
<caption><para>Graphic depiction of the healthcare use-case.</para></caption>
<graphic xlink:href="graphics/ch07_fig122.jpg"/>
</fig>
<para>Every Healthcare Imaging Systems contains many sensors and generates large log files daily. Since these systems are heterogeneous by nature, the first challenge is to optimize logging such that data mining success can be optimized (anamnesis in <link linkend="F7-122">Figure <xref linkend="F7-122" remap="7.122"/></link>). The next challenge is to make all data available worldwide in the cloud (transport in <link linkend="F7-122">Figure <xref linkend="F7-122" remap="7.122"/></link>). Once the data is centrally available, it has to be translated to behavioral models and consolidated in a limited set of relevant parameters (translation in <link linkend="F7-122">Figure <xref linkend="F7-122" remap="7.122"/></link>). This translation requires significant computing power and storage space (infrastructure in <link linkend="F7-122">Figure <xref linkend="F7-122" remap="7.122"/></link>). Next, the obtained parameters have to be analyzed with respect to the maintenance challenges (analytics in <link linkend="F7-122">Figure <xref linkend="F7-122" remap="7.122"/></link>) and the results have to be visualized by end-users (visualization in <link linkend="F7-122">Figure <xref linkend="F7-122" remap="7.122"/></link>).</para>
<para>From the Healthcare Imaging Systems division of Philips two modalities participate in the MANTIS project: the <emphasis>Magnetic Resonance</emphasis> modality and the Interventional X-ray modality recently renamed to Image Guided Therapy Systems. They are introduced separately in the following sections.</para>
<section class="lev3" id="sec7-9-1-1">
<title>7.9.1.1 Introduction to magnetic resonance</title>
<para>In the Business Unit Magnetic Resonance, medical <emphasis>Magnetic Resonance</emphasis> systems are developed (see <link linkend="F7-123">Figure <xref linkend="F7-123" remap="7.123"/></link>). These systems are mainly used to diagnose diseases. There is a variety of Magnetic Resonance system configurations to cover different magnetic field strengths, different gradient power strengths and different clinical application areas.</para>
<fig id="F7-123" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-123">Figure <xref linkend="F7-123" remap="7.123"/></link></label>
<caption><para>Magnetic Resonance Unit.</para></caption>
<graphic xlink:href="graphics/ch07_fig123.jpg"/>
</fig>
<para>A Magnetic Resonance system contains a cryogenically cooled superconductive coil that generates a static magnetic field. When the cryogenic cooling shuts down, the liquid helium level reduces and, within a few days, the superconducting coil loses its superconducting properties and results in a quench. The net effect is that the static magnetic field is lost and operation of the Magnetic Resonance system is no longer possible. It requires days to refill and ramp up the magnetic field back. Early detection of the loss or reduction of the cryogenic cooling function may stop the cascade of events that lead to a quench.</para>
</section>
<section class="lev3" id="sec7-9-1-2">
<title>7.9.1.2 Introduction to IGT systems</title>
<para>In the Business Unit Image Guided Therapy Systems, medical X-ray systems in a C-arm configuration are developed (see <link linkend="F7-124">Figure <xref linkend="F7-124" remap="7.124"/></link>). The focus will be on the larger motorized &#x02018;fixed&#x02019; systems (also called Cath labs), though some of the ideas may also be viable for the smaller &#x0201C;mobile&#x0201D; C-arms.</para>
<fig id="F7-124" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-124">Figure <xref linkend="F7-124" remap="7.124"/></link></label>
<caption><para>Image Guided Therapy Unit.</para></caption>
<graphic xlink:href="graphics/ch07_fig124.jpg"/>
</fig>
<para>These larger X-ray systems can be used for diagnostics, but are most useful in minimal invasive interventional X-ray procedures like the treatment of coronary disease (Dotter treatment or stenting), structural heart disease (valve placement or repair), stroke treatment (aneurisms or stenosis), vascular disease (aortic aneurism or revascularization of limbs) and many other less known treatments. Because vital organs are often targeted, it is essential that the doctor can follow accurately what he is doing inside the patient&#x02019;s body. A wrong movement with a lead wire or catheter may cause serious harm to the patient or even death. Hence, IGT systems philosophy of avoiding interruptions of the image chain while there is a patient being treated. In the equipment of interventional X-ray systems, there obviously is a serious need for reliability.</para>
</section>
</section>
<section class="lev2" id="sec7-9-2">
<title>7.9.2 Data Platform</title>
<para>The existing data storage platform is outdated, not scalable, unstable, and requires too much maintenance. The focus therefore is to transform the existing platform towards a platform that adheres to the MANTIS Reference Architecture. This is the foundation for all other activities in the project as well for future activities. <link linkend="F7-125">Figure <xref linkend="F7-125" remap="7.125"/></link> shows a high-level overview of the implemented data architecture.</para>
<fig id="F7-125" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-125">Figure <xref linkend="F7-125" remap="7.125"/></link></label>
<caption><para>High-level overview of the data architecture.</para></caption>
<graphic xlink:href="graphics/ch07_fig125.jpg"/>
</fig>
<para>Healthcare Imaging Systems can upload their data via the Philips Remote Services VPN. A Data Lake provides the high-volume storage required to store large amounts of historical device files. The device files in the Data Lake are processed by parallel ETL scripts to extract more useful information that is then stored in a Data Warehouse. The Data Warehouse is also the destination for structured data coming from administrative databases such as MS SQL Server, SAP and Teradata. Data analytics can be performed on the data stored in the Data Warehouse by using RapidMiner or programming languages such as R, Python or Java.</para>
<para>The next sections provide more details on the individual elements of the architecture.</para>
</section>
<section class="lev2" id="sec7-9-3">
<title>7.9.3 Data Lake</title>
<para>Data from Healthcare Imaging Systems (<emphasis>Installed base</emphasis>) has been centrally collected by Philips via the Philips Remote Services VPN and stored in a Data Lake. The Data Lake is a component allowing low-cost secure storage of large amounts of data. The current capacity is of a few hundred terabytes but the system can be scaled out to multiple petabytes. It is realized using a Spectrum Scale storage cluster with GPFS file system and disaster recovery that is mounted as network shares on Linux and Windows machines.</para>
<para>Approximately 3 years of data have been stored in the Data Lake. Logically, the data lake space is divided in an archive and a live data-landing zone. The archive contains all the historical data available while the live data-landing zone contains the data received from medical devices that has still to be processed.</para>
<para>Both the archive and the live data-landing zone are further divided per imaging modality (e.g., <emphasis>Magnetic Resonance</emphasis>, IGT) and, within each imaging modality, each medical device has its own space. Within the reserved space for a medical device, the archive and the live data-landing zone are further divided per year, month and day.</para>
</section>
<section class="lev2" id="sec7-9-4">
<title>7.9.4 ETL Scripts</title>
<para>Although it is possible to run analysis scripts on the Data Lake, this is not done very often due to the costs of developing parallel scripts and the lack of interactivity. The preferred way to analyze the data is by interactively querying a data warehouse that contains a pre-processed version of the information contained in the log files.</para>
<para>Log files come in many different file formats depending on the type of equipment, type of log file, type of device and release. For example, IGT equipment stores log events in textual format, parameters in a Microsoft Windows registry format, and configuration information in XML. Magnetic Resonance equipment produces log event files and parameters in textual format, XML, and proprietary binary formats.</para>
<para>For each log file format, a parser has been developed or adapted from existing tools to process the files and extract relevant data that can support root cause analysis and predictive maintenance. Given the diverse nature, formats and the large size of the log files from Healthcare Imaging Systems, ETL scripts have been developed to extract known, potentially useful, information from the log files. This information is stored in a Data Warehouse along with structured data to enable the scenarios. The ETL scripts are written in Java, Ruby, and Python and run on a cluster of computers that uses TORQUE, a distributed resource manager for cluster management, to parallelize the import operations. Each script is therefore written to support parallel processing and be resilient to failures.</para>
<para>To this aim, two aspects are very important: <emphasis>idempotence</emphasis> and ensuring correct <emphasis>data provenance</emphasis>. With <emphasis>idempotence</emphasis> we mean that an ETL script can be applied multiple times to the same input file without resulting in the data being imported more than once. This facilitates re-running ETL jobs that have failed or partially failed. With <emphasis>data provenance,</emphasis> we mean that every record in the Data Warehouse should have associated at least three key pieces of information:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>A reference to the Healthcare Imaging System, preferably a direct reference (e.g., serial number) that does not require joins with other tables;</para></listitem>
<listitem>
<para>The file it originates from with full path and last modification timestamp;</para></listitem>
<listitem>
<para>The version of the ETL script that created the record;</para></listitem>
<listitem>
<para>The date and time at which the record was created.</para></listitem>
</itemizedlist>
</section>
<section class="lev2" id="sec7-9-5">
<title>7.9.5 Data Warehouse</title>
<para>For the architecture, we choose a distributed column-based storage solution (Vertica) that allows to store large amounts of data and to perform SQL queries as if it were a &#x0201C;standard&#x0201D; relational database. The main characteristics of the chosen data warehouse are the following:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Distributed: this allows scaling-out when more storage or speed is required, increases robustness by replicating data on multiple nodes, increases speed of access by storing multiple copies of the data and by distributing load across nodes;</para></listitem>
<listitem>
<para>Column-oriented: optimized for data access. Data is logically organized in tables as in traditional relational databases, though on disk, the data is stored &#x0201C;per column&#x0201D; instead of &#x0201C;per row&#x0201D;. This allows speeding up queries by only reading the files of the columns involved in the queries;</para></listitem>
<listitem>
<para>Advanced compression: aggressive compression of the data is used to replace slow disk I/O for fast CPU cycles. Because the data is stored in columns, different compression schemes can be used depending on the property of a column (e.g., type, cardinality, order) achieving extremely high compression rates;</para></listitem>
<listitem>
<para>SQL-compatible: data can be retrieved using standard SQL queries and via ODBC/JDBC connections. This makes it easy for the applications and ETL scripts to upload data and for the users to retrieve it and to perform data analytics. Additionally, structured data from existing relational databases can be imported directly 1-1 without having to change the data models.</para></listitem>
</itemizedlist>
<para>Before designing and implementing ETL scripts, domain knowledge experts and data scientists decide which information from device log files they consider useful for further analysis. This is typically done in an interactive way that involves interviews with R&#x00026;D experts, gathering of documentation and specification documents of the Healthcare Imaging System, gathering and manual exploration of sample data as well as automatic analysis to determine: structure, data types, data value boundaries, etc.</para>
<para>A &#x0201C;data model document&#x0201D; provides the specification of the ETL task as well as the final format in which the data will be stored in the Data Warehouse. Identifying the dataset to import is also part of the specification (e.g., which files should be skipped or declared invalid).</para>
<para>After the data model document is approved by domain knowledge experts and data scientists, the ETL is designed, implemented (this may include porting or adapting an existing parser), tested and applied to the set of historical device files. The resulting data is then verified using basic analytics (e.g., checking data boundaries, number of records, etc.) and validated by data scientists and domain knowledge experts. The process is repeated until the desired quality level is achieved.</para>
<para>Note that the data in the Data Warehouse is typically written once and read multiple times. Unless the data needs to be corrected, updates and deletes will be done rarely. Furthermore, due to the column-oriented nature of the chosen Data Warehouse technology, the data models are de-normalized to achieve fast access and simplify data analysis.</para>
<para>The data model document has also the function of <emphasis>data dictionary</emphasis> documenting the definitions and the logic behind the records stored in the Data Warehouse.</para>
<para>The ingestion in the Data Warehouse of structured data from existing databases follows a similar process with the difference that instead of ETL scripts, data loading scripts are written that use JDBC/ODBC connections or simple CSV files to move sets of relational data from the sources to the Data Warehouse.</para>
</section>
<section class="lev2" id="sec7-9-6">
<title>7.9.6 Sensors</title>
<para>To monitor the performance of critical components in a Healthcare Imaging System, intelligent components are developed. These intelligent components sense and record their state in real-time. For components that have a wear-out mechanism that would make them fail within the estimated lifetime of a system, the known wear can be removed from computation by means of calibration during planned maintenance activities. For the IGT equipment one of these calibrations is automated in real-time with the use of an intelligent function. There are more opportunities to automate calibrations in the future, but therefore the required feedback loops need to be in place, which is not always the case.</para>
<fig id="F7-126" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-126">Figure <xref linkend="F7-126" remap="7.126"/></link></label>
<caption><para>e-Alert sensor context diagram.</para></caption>
<graphic xlink:href="graphics/ch07_fig126.jpg"/>
</fig>
<para>In the installed base, there are still many Healthcare Imaging Systems that lack the required sensing for critical components. For those devices an intelligent sensor, e-Alert sensor, has been developed. This stand-alone sensor has embedded sensors to measure physical properties of the equipment or environment, to process the signals and to send alerts. Such a sensor is not available in the Healthcare Imaging System itself. The sensor is connected to the healthcare facility network and can communicate via E-Mail and/or SMS. <link linkend="F7-126">Figure <xref linkend="F7-126" remap="7.126"/></link> shows the context diagram of the e-Alert sensor. In case of unfavorable conditions that require a corrective action to resolve the issue, messages can be sent directly to customers as well as to service engineers. Next to that, the e-Alert sensor can be connected to the Philips Remote Service Network. The e-Alert sensor uploads sensor logs and alert logs to Philips Remote Service Network, where the data is stored and pre-processed. This data is accessible via a Philips Remote Service portal to enable Philips to determine operational profiles (aggregated or on e-Alert sensor level). This information is used to define control limits to keep the Healthcare Imaging System in optimal operational condition.</para>
</section>
<section class="lev2" id="sec7-9-7">
<title>7.9.7 Analysis and Decision Making Functionalities</title>
<para>The next sections explain on high level some topics with respect to analysis and decision making functionalities. First, the scoring of predictive models using live data, data sources and data preparation steps for the Log Pattern Finder is explained. Then physical modeling of a unit that dominantly fails due to wear-out, and finally a mathematical model that optimizes the decision making process of remote service engineers in the presence of imperfect failure alerts.</para>
<section class="lev3" id="sec7-9-7-1">
<title>7.9.7.1 Predictive model deployment and live scoring</title>
<para>The data ingested in the data warehouse is used to develop predictive models for particular failure modes. Predictive models are software programs that take as input &#x0201C;live&#x0201D; and &#x0201C;historical&#x0201D; data from a Healthcare Imaging System and calculate a probability of failure of a certain component or group of components within a given period. Predictive models can be based on simple and static rules and thresholds when the failure modes are well understood and can be easily modelled with the data provided by the Healthcare Imaging System. In the case of static rules and thresholds, historical data is typically used to choose the best threshold values that minimize the false positive rate while providing a high number of true positives.</para>
<para>Very often, simple and static rules and thresholds are not sufficient to predict failures with the desired level of accuracy. In these cases, statistical learning is applied. A machine-learning algorithm, such as a neural network or a support vector machine, is trained on a historical dataset until it reaches sufficient predictive performances on historical data.</para>
<para>Once a predictive model has been developed (and trained or tuned with historical data), the model is deployed in a <emphasis>Quality Assurance</emphasis> environment where it is scored daily with new data coming from the Healthcare Imaging Systems in the field. The results of the models, called &#x0201C;alerts&#x0201D;, are stored in the data warehouse for being consumed by a web application called the &#x0201C;remote monitoring dashboard&#x0201D; where a team of remote monitoring engineers evaluates them for their accuracy and predictive power. During this evaluation, the remote monitoring engineers check whether an alert actually corresponds to a situation of imminent failure using all the knowledge at their disposal. Furthermore, the remote monitoring engineers provide feedback to the model development team on the text and plots to be used in the alerts in order to make them more actionable and easier to understand. The model development team uses the feedback of the remote monitoring team to improve the predictive models until the alerts they generate in <emphasis>Quality Assurance</emphasis> are deemed good enough for being promoted to production. At this stage, a predictive model is deployed into production and its alerts are displayed in the production remote monitoring dashboard. These alerts are used to create the actual proactive cases for Healthcare Imaging Systems in the field.</para>
</section>
<section class="lev3" id="sec7-9-7-2">
<title>7.9.7.2 Log pattern finder data</title>
<para>In the previous section, we have seen how predictive models can be developed and used within the data processing architecture. A particular set of models that is useful in this context are the so-called log patterns. A log pattern is a logical sequence of log events that correlates with a particular failure mode. Ideally, we would like to be able to discover log patterns automatically using data. In this section, an approach for automatically finding log patterns is described.</para>
</section>
<section class="lev3" id="sec7-9-7-3">
<title>7.9.7.3 Data sources</title>
<para>For the log pattern finder for IGT equipment, we make use of various data sources. The primary data source is the <emphasis>calls</emphasis> data source. As the objective is to find reactive patterns, we make use of a calls table to identify (<emphasis>i</emphasis>) the time of the call, (<emphasis>ii</emphasis>) the system to which the call applies, (<emphasis>iii</emphasis>) the parts that were replaced, if any, and the log events that were generated by the identified system during a time window prior to the call. These four data sources are linked as shown in <link linkend="F7-127">Figure <xref linkend="F7-127" remap="7.127"/></link>, where an arrow from data source A to B indicates that one or more fields from an entry in A are used to identify the proper entry or entries in B.</para>
<fig id="F7-127" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-127">Figure <xref linkend="F7-127" remap="7.127"/></link></label>
<caption><para>Overview of the main data sources used.</para></caption>
<graphic xlink:href="graphics/ch07_fig127.jpg"/>
</fig>
<para>For any call, there is always a single associated Healthcare Imaging System, but there may have been various parts replaced, depending on the outcome of a root-cause analysis by the service engineers. There are also calls where no part has been replaced, but other actions have been undertaken to resolve the issue. We do not consider these and concentrate on those calls where at least one part has been replaced.</para>
<para>For collecting the appropriate set of log events, we use various methods, ranging from taking a fixed <emphasis>observation</emphasis> interval of <emphasis>n</emphasis> days prior to the call in order to retrieve a set of log events to more carefully selecting an observation interval containing a sufficient number of log events.</para>
</section>
<section class="lev3" id="sec7-9-7-4">
<title>7.9.7.4 Inspect and normalize the data</title>
<para>It was necessary to invest quite some time into getting to &#x0201C;know&#x0201D; the data. Blindly applying an algorithm to the data is usually not a good idea. For example, for log events only a portion of the available fields is used to obtain a concise representation of an event. This approach, however, results in millions of different log events, a situation that is not appropriate in the current context. We discovered that event ids with multiple descriptions are present. The main reason for this is that these descriptions contain numbers, dates, times, ip-addresses and such. As these numbers are most likely not interesting, a <emphasis>normalization step</emphasis> is applied to reduce the number of different descriptions. A specific filter has to be created for each specific issue. These filters are handcrafted, based on the inspection of numerous descriptions.</para>
<para>As another example, there is the issue of dates and times. As the Healthcare Imaging Systems are located in many different countries, one inevitably has to investigate how time is represented in the data coming from the various countries. This encompasses time zones, daylight savings time, and the use of local time without any time zone. Especially for identifying the proper log events, the call-open date is used, so the time information for log events and calls should be encoded in the same way.</para>
</section>
<section class="lev3" id="sec7-9-7-5">
<title>7.9.7.5 Data pre-processing</title>
<para>As retrieving a complex set of data from various large databases can be quite time consuming, we decided to implement a pre-processing step to generate intermediate data, stored in easily accessible files. In particular, for a certain configuration (e.g., a time window, a selected set of system codes and system releases and the (maximal) length of the observation interval) we determine for each part replaced, during a call that opened in the given time window, details of all part replacements during this time window. These details include the system for which the replacement was done, the exact system release, various parameters and, finally, all the log events that occurred.</para>
<para>Although we do not consider calls wherein no part has been replaced, we do generate data for these calls, as their contribution in all calls is significant. We do this by creating a virtual part &#x0201C;NO_PART&#x0201D; and treating all calls without any part replacements as if this &#x0201C;NO_PART&#x0201D; has been replaced. By consistently doing this during the pre-processing step, we already prepare ourselves to look in more detail into these calls.</para>
<para>As this pre-processing is only done once for each configuration, much time is saved and the time required to run the experiments is greatly reduced.</para>
</section>
<section class="lev3" id="sec7-9-7-6">
<title>7.9.7.6 Data representation</title>
<para>Finding log patterns entails combining the occurrence of combinations of log events during an observation interval and the replacement of a specific part during the associated call. To enable an efficient implementation, we decided to encode calls, together with their observation intervals, as integers, meaning that one integer is used to encode both a call and the associated observation interval. In this way, a part <emphasis>p</emphasis> can be represented by a set of integers, i.e., those that encode calls in which <emphasis>p</emphasis> has been replaced. A log event <emphasis>e</emphasis> can analogously be represented as a set of integers, i.e., those that encode observation intervals in which <emphasis>e</emphasis> has occurred. Note that log patterns in general can be represented in exactly the same way.</para>
<para>This representation of parts and log patterns allows efficient computation of all kinds of logical operations. For example, all calls wherein part <emphasis>p</emphasis> has been replaced and before which log event <emphasis>e</emphasis> has occured is represented by the intersection of their respective sets. A log pattern <emphasis>AND</emphasis>(<emphasis>E</emphasis><subscript>1</subscript>,<emphasis>E</emphasis><subscript>2</subscript>) is also represented as the intersection of the representation of its two arguments.</para>
<para>Another advantage of this representation is that, by precomputing a <emphasis>fingerprint</emphasis> for each part and log pattern, checking equality can be done very efficiently. Concatenating the integer elements of a set in a string in increasing order, separated by an appropriate character, allows string comparison to be used. This will be further elaborated upon in the next subsection.</para>
<para>In addition, log events are encoded to allow efficient computation. The encoding is by using integers, prepended with the letter E. This allows quick comparison of log events, as well as a concise notation of log patterns. It is noted that especially the additional info field can be as large as a few KBs, so that a concise representation not only aids in readability, it can potentially save a considerable amount of memory during computations.</para>
</section>
<section class="lev3" id="sec7-9-7-7">
<title>7.9.7.7 Equivalent log patterns</title>
<para>When two log patterns have the same representation in terms of their sets or fingerprints, they are called equivalent. This is handy when looking for log patterns: of the equivalence classes that can be created by this equivalence relation, only one representative needs to be used, as the others will give the same results. This can significantly reduce the number of log patterns to be searched through when looking for good log patterns, and even more when, testing for a limited number of hypotheses, the total number of log patterns to be taken into account gets small.</para>
<para>Although two equivalent log patterns give the same results, this does not mean that they are equivalent in every sense. A service engineer may make a distinction between two log patterns based on the available information. It is also important to mention that this equivalence relation depends on the dataset at hand. On a different dataset, e.g., the test set, they may not be equivalent. Therefore, they may have differing performance. A service engineer may be able to identify the most suitable candidate from an equivalence class to act as representative. Of course, multiple log patterns from an equivalence class can be selected.</para>
</section>
<section class="lev3" id="sec7-9-7-8">
<title>7.9.7.8 Log pattern selection problem</title>
<para>Once candidate log patterns have been identified, they will have a certain performance in terms of the number of true positives (TP) and false positives (FP) found in the training data. During the search for log patterns, these numbers have been subject to a number of constraints in order to generate log patterns of sufficient quality.</para>
<para>Part of the functionality of the log pattern finder is the <emphasis>false-positive analysis</emphasis>. Once, for a given part <emphasis>p</emphasis>, a log pattern with sufficient quality in terms of the number of true positives (TPs) and a sufficiently low false discovery rate (FDR) has been identified, an important and useful exercise is to investigate why a false positive (FP) ended up as such. This gives us a better insight into the issues that caused these false positives and allows us to improve the log pattern finder.</para>
</section>
<section class="lev3" id="sec7-9-7-9">
<title>7.9.7.9 Design decisions</title>
<para>In order to limit the wide spectrum of possibilities, we have made a number of simplifying assumptions and design decisions. The most important one is choosing a fixed observation interval length of one day. This has disadvantages, but we have experimented with longer intervals, and, although the results are different, they were not significantly better. Currently, investigations into choosing a proper observation interval length are ongoing. Note that the observation interval length could be chosen differently for different parts.</para>
<para>Another important decision is that we look at binary occurrences rather than frequencies of occurrences. In other words, either a log pattern occurs or it does not occur during an observation interval. We also do not consider the ordering of individual log events in a log pattern containing more than one log event. Yet another decision is not to apply processing on the description and additional info other than normalization.</para>
<para>Further, we restrict ourselves to log events that only occur a limited number of times, i.e., at most 400 times per year. This is a heuristic that we introduce to deal with the issue of significance of individual log events in a log pattern. We could have been stricter by also here considering <emphasis>p</emphasis>-values, but this is for further investigation.</para>
<para>Finally, we adopt a file naming approach in order to facilitate the management of these files. Although seemingly unimportant, we prepend all generated files with a timestamp, so that the order in which they are generated can be reflected in the directory and files are never overwritten. We use one timestamp for each individual run of the software, so that multiple runs can be performed in parallel.</para>
</section>
<section class="lev3" id="sec7-9-7-10">
<title>7.9.7.10 Output</title>
<para>As output, we create files for individual parts and list all patterns found, their performance, their equivalents, as well as the results of the FP-analysis. We also report on the number of possible, allowed and actually generated hypotheses. In the end, the individual log patterns are combined into an overall log pattern, consisting of ORs of ANDs of individual log events. Several dozen of these log patterns are now actively monitoring thousands of Healthcare Imaging Systems daily and a few dozen are still under development.</para>
</section>
<section class="lev3" id="sec7-9-7-11">
<title>7.9.7.11 Failure prediction</title>
<para>To come to a prediction, research is done on neural networks and their capability to predict failure of a specific component, a high power amplifier in equipment of the Business Unit Magnetic Resonance.</para>
<para>A neural network algorithm is used to predict failures in the high power amplifiers, based on system utilization data and power amplifier demands. For this purpose, utilization and demand data prior to an amplifier replacement was compared to utilization and demand data after an amplifier replacement. Data prior to amplifier replacement is considered as Failure data whereas data after an amplifier replacement is considered as Good data. The Failure data is gathered from a period of 17 days prior to device failure to 4 days prior to device failure. The Good data for period of 13 days is collected after 10 days of part replacement. Please refer to <link linkend="F7-128">Figure <xref linkend="F7-128" remap="7.128"/></link> for a summary of the timeline for data collection. The cooling period masks the uncertainty in dates where the amplifier was actually installed and masks potential &#x0201C;burn-in&#x0201D; related failures.</para>
<fig id="F7-128" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-128">Figure <xref linkend="F7-128" remap="7.128"/></link></label>
<caption><para>Timeline of gathered categories of data from a single system.</para></caption>
<graphic xlink:href="graphics/ch07_fig128.jpg"/>
</fig>
<para>The problem of fault prediction is here onwards addressed as a classification problem where, the model reads 13 days of historical data and predicts if it is Failure data or Good data. If it is predicted as Failure data, then it is more likely that the device is going to become faulty at least after 4 days and if it is predicted as Good data, the device is less likely to become faulty in the next few days. The dataset covers 219 amplifier replacements in a period of 1.5 years (July 2015 until the end of 2016). However, due to connectivity issues, we have got data for 134 systems only prior to amplifier replacement, and data of 154 machines only after amplifier replacement. Summarizing, we have 134 failure data points and 154 good data points. For each replacement, 16 features (F16) have been defined. Hence, the total data set consists of (134 [replacements] + 154 [Good]) &#x0002A; 16 [features] &#x0002A; 13 [days data/features] = 59.9k feature points. The data points are ordered as a single dimensional array (1D) (see <link linkend="F7-129">Figure <xref linkend="F7-129" remap="7.129"/></link>). The 1D arrays are formed by lexicographical ordering of the above features where F1 D1 denotes first feature recorded at Day 1.</para>
<fig id="F7-129" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-129">Figure <xref linkend="F7-129" remap="7.129"/></link></label>
<caption><para>Conversion of features into arrays.</para></caption>
<graphic xlink:href="graphics/ch07_fig129.jpg"/>
</fig>
<para>The neural network used in this research is an Artificial Neural Networks (ANN), where the input is fed as a single dimensional array. The ANN used consists of 7 fully connected layers. The initial 6 layers consist of 50 neurons and a relu activation function and the last layer consists of two neurons and a softmax activation function. The network consists of two dropouts. The initial dropout in forward direction drops 25% of the features and the next dropout eliminates 50% of the features. This helps the network to develop a generalization that prevents overfitting. Once trained, the output layer of the model returns a probability of how it is likely that the device corresponding to the input data fails.</para>
<fig id="F7-130" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-130">Figure <xref linkend="F7-130" remap="7.130"/></link></label>
<caption><para>ANN Architecture.</para></caption>
<graphic xlink:href="graphics/ch07_fig130.jpg"/>
</fig>
<para>The data is split into training and testing sets at a proportion of 70% of samples and 30% of samples respectively. The ANN architecture, represented in <link linkend="F7-130">Figure <xref linkend="F7-130" remap="7.130"/></link> was built and the training data was used to train it for 20,000 epochs. Figures 7.131 shows the learning curves (accuracy, loss, validation accuracy, validation loss) for increasing epochs of the ANN architecture.</para>
<fig id="F7-131" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-131">Figure <xref linkend="F7-131" remap="7.131"/></link></label>
<caption><para>Learning curves.</para></caption>
<graphic xlink:href="graphics/ch07_fig131.jpg"/>
</fig>
<para>It was necessary to train for 1500 epochs to attain the target accuracy of 95% over 1000 epochs. It can also be observed that the learning late curves are oscillatory in nature. This is because of the dropout which helps to reach a higher level of generalizability for the model. The resulting validation accuracy is 67%. Subsequently, the algorithm was fed with data from the years 2016+2017, covering more power amplifier replacements. The resulting validation accuracy is 54%; we could not achieve better validation accuracy. This is understandable since the number of amplifier replacements is still limited and deep learning architectures generally require thousands to obtain a generalised model. However, the ability of the deep learning architectures to perform satisfactorily on the limited data is promising. Additional work will be carried out to collect more data in order to achieve a validation accuracy of more than 90%.</para>
</section>
<section class="lev3" id="sec7-9-7-12">
<title>7.9.7.12 Physical modeling</title>
<para>This section will reflect on the 10-step method used to come to a RUL prediction for the X-ray tubes. X-ray tubes are the most expensive parts to replace of IGT equipment and they are as such a major concern for the service organization. It is known that X-ray tubes are subject to wear-out, so it can be expected that they fail after some usage. Because of the major impact (in terms of downtime and cost) of an X-ray tube replacement, it is important to understand the failure of X-ray tubes better in order to improve the service to customers.</para>
<section class="lev4">
<title>Step 1: Weibull analysis</title>
<para>As a starting point, do a Weibull analysis of the failures of the unit you want to make a RUL prediction for. The presence of a significant wear-out means that a RUL prediction is feasible. Next to working on the prediction of the wear-out, a second activity should be started to eliminate other failure modes as much as possible.</para>
</section>
<section class="lev4">
<title>Step 2: Identify the responsible failure mode</title>
<para>The Weibull plot for the unit only tells whether there are failures due to wear-out, but not which is the related failure mode. After a candidate failure mode is found, a Weibull plot for just that failure mode should confirm that this is indeed the failure mode responsible for wear-out. As a rule of thumb, at least 80-85% of the failures should be related to wear-out in order for the prediction to be useful.</para>
</section>
<section class="lev4">
<title>Step 3: Gather knowledge about the physics of the failure mode</title>
<para>X-ray tube cathode filaments wear, because they are heated to a high temperature during an X-ray run to emit sufficient electrons to produce the desired X-ray dose, but the heat causes the Tungsten, the material they are made of, to evaporate. Over time, a hot spot forms where heat and evaporation are exponentially increased until the material melts at the hot spot resulting in the opening of the filament. Understanding the stress and having a damage indicator at hand are two prerequisites for finalizing this step and continue with the next step in the method.</para>
</section>
<section class="lev4">
<title>Step 4: Establish the relationship between stress and damage indicators and an end criterion in controlled experiments</title>
<para>Controlled experiments can be performed in the lab or in the field, but they usually require extra instrumentation. The controlled experiments will allow modeling a first order relationship between damage and stress for the unit, without the interference of secondary influences. These will appear later in the field, but can be recognized by comparing them with the first order model.</para>
</section>
<section class="lev4">
<title>Step 5: Measure variables with a strong relationship to the stress and damage</title>
<para>Sometimes stress and damage can be measured directly, but often only indirect methods are available in the field. Make sure that the relation between the field data and the variables in the wear model are well established and understood.</para>
</section>
<section class="lev4">
<title>Step 6: Collect, plot, and monitor the data</title>
<para>Once the field data becomes available, they can be collected and plotted. To be able to predict the amount of time that is left before a failure, you need to know how much wear is accumulated per time unit. A convenient way to represent the data is a (Damage/Load/Time) DLT plot. An example of a DLT-plot is given in <link linkend="F7-132">Figure <xref linkend="F7-132" remap="7.132"/></link>.</para>
<fig id="F7-132" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-132">Figure <xref linkend="F7-132" remap="7.132"/></link></label>
<caption><para>Example DLT plot.</para></caption>
<graphic xlink:href="graphics/ch07_fig132.jpg"/>
</fig>
<para>In <link linkend="F7-132">Figure <xref linkend="F7-132" remap="7.132"/></link>, the blue curve represents the ln(c-factor) against the linear wear (damage vs. load). This curve is called the <emphasis>wear curve</emphasis>. The orange curve represents the linear wear against calendar dates (load vs. time). It is remarkably straight for this particular DLT plot, indicating this user on average keeps on using the system in the same way every day over a long period of time. The blue curve is called the <emphasis>usage curve</emphasis>. At this stage, it is important to look for anomalies in the data, be able to explain these anomalies, and see if the failure is actually close when the wear curve starts to bend.</para>
</section>
<section class="lev4">
<title>Step 7: Segment the usage curve</title>
<para>The usage curve in <link linkend="F7-132">Figure <xref linkend="F7-132" remap="7.132"/></link> shows a relatively straight line for the linear wear against time plot. There are examples, where the usage curve changes significantly as can be seen in <link linkend="F7-133">Figure <xref linkend="F7-133" remap="7.133"/></link>.</para>
<fig id="F7-133" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-133">Figure <xref linkend="F7-133" remap="7.133"/></link></label>
<caption><para>Segmenting the usage curve.</para></caption>
<graphic xlink:href="graphics/ch07_fig133.jpg"/>
</fig>
<para>When the usage changes, the best predictor for the future will also change. A change in usage depends on human decisions. In our case, it is related to the number of patients being treated on the system per unit time, the type of treatments, the availability of staff and all kinds of causes that would make the use of the system fluctuate. Therefore it is necessary to look for significant changes in the usage and adjust the prediction when such a significant change is detected. This can be done and tuned just based on the shape of the usage curves themselves.</para>
<para>In <link linkend="F7-133">Figure <xref linkend="F7-133" remap="7.133"/></link>, the dashed line represents the prediction of load at a particular time considering the usage change. In this particular case, the usage curve was segmented in two segments and the dashed line is based only on the second segment and obtained by means of linear regression.</para>
</section>
<section class="lev4">
<title>Step 8: Collect sufficient wear curves to failure and establish an end criterion</title>
<para>When sufficient units have actually failed with reasonable certainty that they failed with the failure mode related to wear-out, the ends of their wear curves can be used to establish values for end criteria. In step 4, the variable for the end criterion is selected. In our case, this is the slope of the wear curve. When sufficient units have worn out, an average value at which failure happens can be established. This may not be necessary in all cases, for instance there may be a limit value already defined (like a minimum profile depth of a car tire).</para>
</section>
<section class="lev4">
<title>Step 9: Predict the wear curve progression for curves approaching failure</title>
<para>Once the end criteria are found, predictions can be made for units approaching failure.</para>
</section>
<section class="lev4">
<title>Step 10: Assess the prediction stability</title>
<para>As time goes by, more data will be available and this will have influence on the prediction. At a certain moment, the prediction should however become more or less stable.</para>
</section>
</section>
<section class="lev3" id="sec7-9-7-13">
<title>7.9.7.13 Maintenance and inventory optimization</title>
<para>Philips Healthcare is interested in developing a cost effective proactive maintenance strategy, relying on mathematical tools for statistical life-cycle and reliability analysis. The goal is to come to the optimal planning of maintenance and related resource planning. This consists of two different elements:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Develop maintenance optimization models that can be used to make a balanced tradeoff between the cost of a failure and the cost of proactive maintenance, taking the uncertainty in the prediction models into account;</para></listitem>
<listitem>
<para>Create a decision support system for remote monitoring engineers to come to the business optimal decision, and to guide them on what to do in case of imperfect predictions.</para></listitem>
</itemizedlist>
<para>In addition to providing a recipe for the remote monitoring engineers to follow, we expect that the maintenance optimization model will also shed light on how much can be invested in improving the predictive/proactive models, i.e., the value of improved predictive/proactive models will be revealed by comparing the optimal expected costs under different levels of imperfectness in alert predictions.</para>
</section>
<section class="lev3" id="sec7-9-7-14">
<title>7.9.7.14 Model and analysis</title>
<para>This section describes the creation of a mathematical model that supports remote monitoring engineers in their proactive maintenance decision making. The model should support in the decisions that follow on an alert raised by predictive models. Whenever an alert is seen, a remote monitoring engineer needs to decide whether to create a case or to reject the alert.</para>
<section class="lev4">
<title>Decision variables</title>
<para>As reported in <link linkend="T7-3">Table <xref linkend="T7-3" remap="7.3"/></link>, there are two decision variables in the model: <emphasis>a</emphasis> represents the decision to create a maintenance case and <emphasis>y</emphasis> represents the decision to combine the case with an already scheduled maintenance activity.</para>
<table-wrap position="float" id="T7-3">
<label><link linkend="T7-3">Table <xref linkend="T7-3" remap="7.3"/></link></label>
<caption><para>Decision variables of the model</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left"><emphasis role="strong">Decision variable</emphasis></td>
<td valign="top" align="center"><emphasis role="strong">Notation</emphasis></td></tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Initiate maintenance actions</td>
<td valign="top" align="center"><emphasis>a</emphasis> &#x02208; {0, 1}</td>
</tr>
<tr>
<td valign="top" align="left">Combine with next maintenance activity</td>
<td valign="top" align="center"><emphasis>y</emphasis> &#x02208; {0, 1}</td></tr>
</tbody>
</table>
</table-wrap>
</section>
<section class="lev4">
<title>Model parameters</title>
<para><link linkend="T7-4">Table <xref linkend="T7-4" remap="7.4"/></link> summarizes all the parameters used in the model to support the remote monitoring engineers.</para>
<table-wrap position="float" id="T7-4">
<label><link linkend="T7-4">Table <xref linkend="T7-4" remap="7.4"/></link></label>
<caption><para>Parameters used in the model</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left"><emphasis role="strong">Parameter</emphasis></td>
<td valign="top" align="center"><emphasis role="strong">Notation</emphasis></td></tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">No Facility Systems Engineer on site</td>
<td valign="top" align="center"><emphasis>m</emphasis> &#x02208; {0, 1}</td>
</tr>
<tr>
<td valign="top" align="left">No open case</td>
<td valign="top" align="center"><emphasis>o</emphasis> &#x02208;&#x0007B;0, 1&#x0007D;</td>
</tr>
<tr>
<td valign="top" align="left">Customer contract</td>
<td valign="top" align="center"><emphasis>v</emphasis> &#x02208; {0, 1, ...,	6}</td>
</tr>
<tr>
<td valign="top" align="left">Outside working hours coverage</td>
<td valign="top" align="center"><emphasis>owh</emphasis></td>
</tr>
<tr>
<td valign="top" align="left">Costs of downtime per unit time</td>
<td valign="top" align="center"><emphasis>c<subscript>d</subscript></emphasis></td>
</tr>
<tr>
<td valign="top" align="left">Downtime compensation</td>
<td valign="top" align="center"><emphasis>cc<subscript>dt</subscript></emphasis> &#x02208; {0, 1}</td>
</tr>
<tr>
<td valign="top" align="left">Customer&#x02019;s region monitored by RME</td>
<td valign="top" align="center"><emphasis>r</emphasis> &#x02208; {0, 1}</td>
</tr>
<tr>
<td valign="top" align="left">Expected costs of PdM</td>
<td valign="top" align="center"><emphasis>c<subscript>PdM</subscript></emphasis></td>
</tr>
<tr>
<td valign="top" align="left">Expected costs of PdM combined with PM</td>
<td valign="top" align="center"><emphasis>c&#x02032;<subscript>PdM</subscript></emphasis></td>
</tr>
<tr>
<td valign="top" align="left">Expected costs of CM</td>
<td valign="top" align="center"><emphasis>c<subscript>cm</subscript></emphasis></td>
</tr>
<tr>
<td valign="top" align="left">Expected costs of diagnostics</td>
<td valign="top" align="center"><emphasis>c<subscript>diagnostics</subscript></emphasis></td>
</tr>
<tr>
<td valign="top" align="left">SLA Response time for contract <emphasis role="strong"><emphasis role="bf">V</emphasis></emphasis></td>
<td valign="top" align="center"><emphasis>Tr<subscript>v</subscript></emphasis></td>
</tr>
<tr>
<td valign="top" align="left">Estimated time for diagnostics</td>
<td valign="top" align="center"><emphasis>t<subscript>diagnostics</subscript></emphasis></td>
</tr>
<tr>
<td valign="top" align="left">Estimated repair time</td>
<td valign="top" align="center"><emphasis>t<subscript>r</subscript></emphasis></td>
</tr>
<tr>
<td valign="top" align="left">Time to next Planned Maintenance</td>
<td valign="top" align="center"><emphasis>t<subscript>sm</subscript></emphasis></td>
</tr>
<tr>
<td valign="top" align="left">RUL</td>
<td valign="top" align="center"><emphasis>X</emphasis></td>
</tr>
<tr>
<td valign="top" align="left">Time to on-site maintenance</td>
<td valign="top" align="center"><emphasis>T<subscript>os</subscript></emphasis></td>
</tr>
<tr>
<td valign="top" align="left">Probability that the alert is true</td>
<td valign="top" align="center"><emphasis>P</emphasis></td></tr>
</tbody>
</table>
</table-wrap>
</section>
<section class="lev4">
<title>Timeline of events and probabilistic scenarios in the model</title>
<para>The model includes two random variables, <emphasis>X</emphasis> and <emphasis>T<subscript>os</subscript></emphasis>. These variables measure the time from the alert arrival to the failure and on-site maintenance, respectively (see <link linkend="F7-134">Figure <xref linkend="F7-134" remap="7.134"/></link>). When the remote monitoring engineer decides to create a case, the local service organization schedules the maintenance activities to resolve the case. Since <emphasis>X</emphasis> is random, we have to distinguish several scenarios when a case is created (see <link linkend="T7-5">Table <xref linkend="T7-5" remap="7.5"/></link>).</para>
<para>A proactive case is scheduled on <emphasis>T<subscript>os</subscript></emphasis> when <emphasis>a</emphasis> = 1, <emphasis>y</emphasis> = 0, and <emphasis>T<subscript>os</subscript> &#x0003C; t<subscript>sm</subscript></emphasis>. We assume that if the realization of <emphasis>T<subscript>os</subscript></emphasis> is greater than <emphasis>t<subscript>sm</subscript></emphasis>, the LSO makes the decision to combine the case with the already scheduled maintenance case on t<subscript>sm</subscript>. It makes no sense to execute the case later than this moment because it will lead to an additional visit and costs. Therefore, a proactive case is scheduled on <emphasis>t<subscript>sm</subscript></emphasis> when <emphasis>a</emphasis> = 1 and <emphasis>y</emphasis> = 1 or when a = 1, <emphasis>y</emphasis> = 0, and <emphasis>T<subscript>os</subscript></emphasis> > <emphasis>t<subscript>sm</subscript></emphasis>.</para>
<para>The equipment can fail before the case is solved. This happens if <emphasis>X &#x0003C; T<subscript>os</subscript></emphasis> when the case is scheduled on <emphasis>T<subscript>os</subscript></emphasis> or if <emphasis>X &#x0003C; t<subscript>sm</subscript></emphasis> when the case is scheduled on <emphasis>t<subscript>sm</subscript></emphasis>. The downtime is equal to the response time plus the repair time when the equipment fails before the case is solved; corrective maintenance (CM) costs are incurred.</para>
<para>The proactive maintenance case can also prevent a failure. This happens when <emphasis>X > T<subscript>os</subscript></emphasis> or <emphasis>X > t<subscript>sm</subscript></emphasis>. The downtime is equal to the repair time and proactive maintenance costs are incurred. Costs of <emphasis>c&#x02032;<subscript>PdM</subscript></emphasis> are incurred if the proactive case is combined with another case, and costs of <emphasis>c<subscript>PdM</subscript></emphasis> are incurred when the proactive case is not combined with another case.</para>
<fig id="F7-134" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-134">Figure <xref linkend="F7-134" remap="7.134"/></link></label>
<caption><para>Timeline of Events (1: failure occurs before maintenance, 2: maintenance occurs before failure).</para></caption>
<graphic xlink:href="graphics/ch07_fig134.jpg"/>
</fig>
<para>When no case is created and the equipment fails, CM and diagnostic costs are incurred because the local service organization did not receive a case. The problem needs to be diagnosed first because the problem is unknown, when the customer calls. In this situation, the downtime consists of response time, time for diagnostics and repair time.</para>
<para>When the remote monitoring engineer decides to create a case, it is always possible that the alert was false. The local service organization discovers that the alert was false and costs of <emphasis>c<subscript>FP</subscript></emphasis> are incurred in such a scenario.</para>
<table-wrap position="float" id="T7-5">
<label><link linkend="T7-5">Table <xref linkend="T7-5" remap="7.5"/></link></label>
<caption><para>Possible scenarios for the maintenance cases that are created</para></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/ch07_tab5.jpg"/>
</table-wrap>
</section>
<section class="lev4">
<title>Mathematical formulation of the model</title>
<para>After an alert arrival, the remote monitoring engineer has to make the decision such that the expected costs and downtime are minimized. We define two objective functions. The first one aims to minimize the expected costs, and the second one is used to minimize the expected downtime:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq7-11.jpg"/></para>
<para>s.t.</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq7-13.jpg"/></para>
<para>Equation (7.4) represents the minimization objective of the expected costs because of the decisions made by the remote monitoring engineer. <inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/pg487-1.jpg"/> represents the expected costs of creating and sending a case to the local service organization (<emphasis>a</emphasis> = 1, <emphasis>y</emphasis> = 0). <emphasis>E[C<subscript>combine</subscript>]</emphasis> represents the expected costs of creating a case and suggesting to combine it with an already scheduled case (<emphasis>a</emphasis> = 1, <emphasis>y</emphasis> = 1). <emphasis>E[C<subscript>SNAR</subscript>]</emphasis> represents the expected costs of SNAR (<emphasis>a</emphasis> = 0). The expected downtime costs are represented by cc<subscript>dt</subscript> &#x022C5; <emphasis>E[D]</emphasis> &#x022C5; <emphasis>c<subscript>d</subscript></emphasis> and are only incurred when the customer is entitled to downtime compensation (<emphasis>cc<subscript>dt</subscript></emphasis> = 1).</para>
<para>We can calculate the expected costs of each action by summing the multiplications of scenario probabilities for that action with the associated costs. These expected costs expressions for the different actions are given below:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq7-16.jpg"/></para>
<para>Equation (7.5) represents the minimization objective of the expected downtime. <emphasis>E[D<subscript>case</subscript>]</emphasis> represents the expected downtime of creating and sending a case to the local service organization (<emphasis>a</emphasis> = 1, <emphasis>y</emphasis> = 0). <emphasis>E[D<subscript>combine</subscript>]</emphasis> represents the expected downtime of creating a case and suggesting to combine it with an already scheduled case (<emphasis>a</emphasis> = 1, <emphasis>y</emphasis> = 1). <emphasis>E[D<subscript>SNAR</subscript>]</emphasis> represents the expected downtime of SNAR (<emphasis>a</emphasis> = 0). We can calculate the expected downtime of each action in the same way as the expected costs. The expected downtime expression for each action are given below:</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/eq7-18.jpg"/></para>
<para>Equation (7.6) makes sure that an alert is SNARed if, the customer&#x02019;s region is not monitored (<emphasis>r</emphasis> = 0), the customer has no contract (<emphasis>v</emphasis> = 0), the Facility Systems Engineer is already on-site (<emphasis>m</emphasis> = 0), or there is already a maintenance case opened for the system (<emphasis>o</emphasis> = 0). Alerts with such characteristics should be SNARed automatically.</para>
<para>Equation (7.7) enforces that a case can only be combined with an existing case when the remote monitoring engineer decides to create a case for the alert.</para>
<para>Equation (7.8) ensures that <emphasis>a</emphasis> and <emphasis>y</emphasis> can only take binary values.</para>
</section>
</section>
<section class="lev3" id="sec7-9-7-15">
<title>7.9.7.15 Results and insights</title>
<para>The model is implemented in a case study for flat detectors, which converts X-ray into electrical signals. Since we have two objective functions, there is not always a single solution for the optimization problem. Lower costs can result in higher downtimes. The remote monitoring engineer can take three different decisions: (<emphasis>i</emphasis>) SNAR the alert (<emphasis>a</emphasis> = 0), (<emphasis>ii</emphasis>) create a case (<emphasis>a</emphasis> = 1, <emphasis>y</emphasis> = 0), or (<emphasis>iii</emphasis>) create a case and combine it with an already scheduled case (<emphasis>a</emphasis> = 1, <emphasis>y</emphasis> = 1). The model aims to evaluate all three options in terms of expected costs and expected downtimes. The output of Model 1 consists of a summary of each option with the expected costs and downtime of each option. This gives the remote monitoring engineer support in their decision making because they can account for the possible consequences of their decisions.</para>
<para>With the Flat-Detector-specific default values, creating a case is the optimal decision to make by the remote monitoring engineers. In <link linkend="F7-135">Figure <xref linkend="F7-135" remap="7.135"/></link>, we see that the optimal action outperforms the other actions in both expected costs and expected downtime at a specific value of the model parameter <emphasis>P</emphasis>.</para>
<fig id="F7-135" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-135">Figure <xref linkend="F7-135" remap="7.135"/></link></label>
<caption><para>Model output default values for flat detector with P=0.80.</para></caption>
<graphic xlink:href="graphics/ch07_fig135.jpg"/>
</fig>
<para>Notice that <link linkend="F7-135">Figure <xref linkend="F7-135" remap="7.135"/></link> is made for a credible alert with <emphasis>P</emphasis> = 0.8. If we use the same input values but set <emphasis>P</emphasis> to 0.15, we receive the plot in <link linkend="F7-136">Figure <xref linkend="F7-136" remap="7.136"/></link>. It can be seen that there is no optimal decision to make. No action outperforms all others in terms of both expected downtime and expected costs. Create a case is the best option in terms of costs while combining the case is the best option in terms of downtime.</para>
<fig id="F7-136" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-136">Figure <xref linkend="F7-136" remap="7.136"/></link></label>
<caption><para>Model output default values for flat detector with P=0.15.</para></caption>
<graphic xlink:href="graphics/ch07_fig136.jpg"/>
</fig>
<para>If we vary <emphasis>P</emphasis> from 0 to 1 with steps of 0.01, we receive the plots in <link linkend="F7-137">Figure <xref linkend="F7-137" remap="7.137"/></link> and <link linkend="F7-138">Figure <xref linkend="F7-138" remap="7.138"/></link>. We observe the existence of a probability threshold for creating a case.</para>
<fig id="F7-137" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-137">Figure <xref linkend="F7-137" remap="7.137"/></link></label>
<caption><para>Influence of P on expected costs in flat detector case.</para></caption>
<graphic xlink:href="graphics/ch07_fig137.jpg"/>
</fig>
<fig id="F7-138" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-138">Figure <xref linkend="F7-138" remap="7.138"/></link></label>
<caption><para>Influence of P on expected downtime in flat detector case.</para></caption>
<graphic xlink:href="graphics/ch07_fig138.jpg"/>
</fig>
<para>We next evaluate the influence of the service contracts on the optimal decisions according to the model. We create three fictitious customers, which we refer to as Customer A, Customer B and Customer C. Customer A has a contract with the most extensive entitlements. Customer B has a contract with no coverage options. Customer C has the most basic service contract.</para>
<para>We vary <emphasis>P</emphasis> to find out if different probability thresholds exist for different types for customer. <link linkend="F7-139">Figure <xref linkend="F7-139" remap="7.139"/></link> shows the influence of <emphasis>P</emphasis> on the expected costs and downtime for the different types of customers.</para>
<fig id="F7-139" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-139">Figure <xref linkend="F7-139" remap="7.139"/></link></label>
<caption><para>Influence of P on expected costs and downtime for different customers.</para></caption>
<graphic xlink:href="graphics/ch07_fig139.jpg"/>
</fig>
<para>We observe that the expected costs incurred for Customer A are the highest. This is due to the compensation of downtime received by this type of customer. The expected downtime of all actions is the lowest for Customer A. The reason behind this is that shorter on-site response times are offered to customers with higher contracts. After a customer call, the field service engineer is faster on-site to conduct maintenance on the failed equipment. In addition, the customer is entitled for maintenance outside operating hours. For Customer B and C, the expected costs of each action are equal. The value <emphasis>cc<subscript>dt</subscript></emphasis> is the only customer-specific parameter that influences the expected costs.</para>
</section>
<section class="lev3" id="sec7-9-7-16">
<title>7.9.7.16 Visualization and HMI</title>
<para>Two HMIs were created related to the Philips Healthcare use case. One related to the development of the e-Alert sensor and one to assist the remote monitoring team. Both are explained in the sections below.</para>
</section>
<section class="lev3" id="sec7-9-7-17">
<title>7.9.7.17 E-Alert portal</title>
<para>The e-Alert sensor provides a web-based user interface to configure sensors and to configure the control limits. When the healthcare facility allows it, the staff of the healthcare facility can access the user interface of the e-Alert sensor. This user interface provides capabilities to view the history of sensor values to support root cause analysis. Service engineers can also access the user interface of the e-Alert sensor. The user interface provides capabilities to view the history of sensor values, to configure the control limits and to update the embedded software.</para>
<para>Each e-Alert sensor is able to upload its sensor logs and alerts to the Philips Remote Service Network. A portal (see <link linkend="F7-140">Figure <xref linkend="F7-140" remap="7.140"/></link>) has been developed to gain access to these logs and alerts for o&#x0FB04;ine data analysis. This enables Philips to determine operational profiles, specific to the Healthcare Imaging System where the e-Alert sensor is connected. This information can be used to fine-tune the configured control limits for that specific Healthcare Imaging System to keep it in optimal operational conditions.</para>
<fig id="F7-140" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-140">Figure <xref linkend="F7-140" remap="7.140"/></link></label>
<caption><para>Portal to gain access to e-Alert sensor data.</para></caption>
<graphic xlink:href="graphics/ch07_fig140.jpg"/>
</fig>
</section>
<section class="lev3" id="sec7-9-7-18">
<title>7.9.7.18 Remote monitoring dashboard</title>
<para>A dashboard to created, to provide an overview of all Health Imaging Systems that generated an alert for the remote monitoring engineers. On the highest level of the dashboard an overview of all alerts is presented, as shown in <link linkend="F7-141">Figure <xref linkend="F7-141" remap="7.141"/></link>.</para>
<fig id="F7-141" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-141">Figure <xref linkend="F7-141" remap="7.141"/></link></label>
<caption><para>Remote monitoring dashboard high level overview.</para></caption>
<graphic xlink:href="graphics/ch07_fig141.jpg"/>
</fig>
<para>When one of the alerts is selected, a detailed report is presented (see <link linkend="F7-142">Figure <xref linkend="F7-142" remap="7.142"/></link>). This report contains all details for the related Health Imaging System like system type, model version, software release, alert history, parameter trends, and maintenance history. This will support the remote monitoring engineer to be able to take further action. When further action is required, the remote monitoring engineer can directly initiate a service work order. Depending on the urgency of the service action, the engineer can combine it with an already scheduled appointment.</para>
<fig id="F7-142" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F7-142">Figure <xref linkend="F7-142" remap="7.142"/></link></label>
<caption><para>Remote monitoring dashboard detailed report.</para></caption>
<graphic xlink:href="graphics/ch07_fig142.jpg"/>
</fig>
</section>
<section class="lev3" id="sec7-9-7-19">
<title>7.9.7.19 Conclusions</title>
<para>For the Philips Healthcare use-case, quite a lot of effort has been spent on the data storage platform and the analysis and decision making functionalities of the Mantis project. Having a large set of historical data already available, made it possible to make a lot of progress from the start. It also made the cooperation much smoother with both our main research partners, Philips Research and the Technical University of Eindhoven, because from an early stage we were able to share data. In our case, the misconception that only providing data eventually will lead to results is once again proven true. Without decent domain knowledge, it is virtually impossible to get usable results. Besides explaining the research partners what all the available data means, domain knowledge is also needed to review and filter possible outputs. This means that in the beginning intensive collaboration is required to get the research partners up-to-speed, followed by frequent updates. This is something that needs to be taken into consideration.</para>
<para>The Philips Healthcare use case featured a few challenging topics and goals set at the beginning of the project. Most of them were realized. The data storage platform is mature and ready to be extended, sensors and intelligent functions have been designed, and a considerable amount of predictive and proactive models has been created. All the different aspects combined, resulted in a remote monitoring capability. We envision that, from 2018 onwards, one in every five system service events worldwide will be triggered by careful analysis of system data &#x02013; and will therefor take place before any major issues arise. This maintenance can also be planned so there is no disruption to the workflow.</para>
</section>
</section>
</section>
<section class="lev1" id="sec7-10">
<title>References</title>
<orderedlist numeration="arabic" continuation="restarts" spacing="normal">
<listitem>
<para>Aaron Parecki, C. M. &#x02018;OAuth 2.0,&#x02019; [Online]. Available: https://oauth.net/2/. [Accessed 01/04/2018].</para></listitem>
<listitem>
<para>Albano, M., Ferreira, L. L., Pinho, L. M., and Alkhawaja, A. R. (2015) Message-oriented middleware for smart grids, Computer Standards &#x00026; Interfaces. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Albano%2C+M%2E%2C+Ferreira%2C+L%2E+L%2E%2C+Pinho%2C+L%2E+M%2E%2C+and+Alkhawaja%2C+A%2E+R%2E+%282015%29+Message-oriented+middleware+for+smart+grids%2C+Computer+Standards+%26+Interfaces%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Apache, &#x02018;Spark Apache,&#x02019; [Online]. Available: https://spark.apache.org/. [Accessed April 2018].</para></listitem>
<listitem>
<para>The Apache Software Foundation. (2018) &#x02018;Apache NiFi,&#x02019; [Online]. Available: https://nifi.apache.org/. [Accessed 01/04/2018].</para></listitem>
<listitem>
<para>Carroll, J. M. (2002) Making Use: Scenario-Based Design of Human-Computer Interactions, The MIT Press. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Carroll%2C+J%2E+M%2E+%282002%29+Making+Use%3A+Scenario-Based+Design+of+Human-Computer+Interactions%2C+The+MIT+Press%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Cheung, D. W., Fu, A. W.-C., and Han, J. (1994) Knowledge Discovery in Databases: A Rule-based Attribute-oriented Approach, Berlin: Springer. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Cheung%2C+D%2E+W%2E%2C+Fu%2C+A%2E+W%2E-C%2E%2C+and+Han%2C+J%2E+%281994%29+Knowledge+Discovery+in+Databases%3A+A+Rule-based+Attribute-oriented+Approach%2C+Berlin%3A+Springer%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Cocciaglia, D. (2012) &#x02018;Case Study: Switch &#x00026; Crossing Diagnostics,&#x02019; in Track Maintenance &#x00026; Renewals Congress, London, United Kingdom. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Cocciaglia%2C+D%2E+%282012%29+%27Case+Study%3A+Switch+%26+Crossing+Diagnostics%2C%27+in+Track+Maintenance+%26+Renewals+Congress%2C+London%2C+United+Kingdom%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>CRUD, Managing the Data-base Environment, p. 381, at Google Books.</para></listitem>
<listitem>
<para>De Barr, A. E. and Oliver, D. A. (1975) Electrochemical Machining, Surrey: Unwin Brothers Ltd., 1975. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=De+Barr%2C+A%2E+E%2E+and+Oliver%2C+D%2E+A%2E+%281975%29+Electrochemical+Machining%2C+Surrey%3A+Unwin+Brothers+Ltd%2E%2C+1975%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Ferreira L. L.<etal/> (2017) &#x02018;A Pilot for Proactive Maintenance in Industry 4.0,&#x02019; in 13th IEEE International Workshop on Factory Communication Systems (WFCS 2017), Trondheim, Norway. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Ferreira+L%2E+L%2E+%282017%29+%27A+Pilot+for+Proactive+Maintenance+in+Industry+4%2E0%2C%27+in+13th+IEEE+International+Workshop+on+Factory+Communication+Systems+%28WFCS+2017%29%2C+Trondheim%2C+Norway%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Fronza, I.<etal/> (2013) &#x02018;Failure prediction based on log files using random indexing and support vector machines,&#x02019; Journal of Systems and Software, 86(1), pp. 2&#x02013;11. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Fronza%2C+I%2E+%282013%29+%27Failure+prediction+based+on+log+files+using+random+indexing+and+support+vector+machines%2C%27+Journal+of+Systems+and+Software%2C+86%281%29%2C+pp%2E+2-11%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Gorostegui, U. (2017) Automatic Integration of Cyber Physical Condition Monitoring Systems as Part of Computerized Maintenance Management Systems, Mondragon University. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Gorostegui%2C+U%2E+%282017%29+Automatic+Integration+of+Cyber+Physical+Condition+Monitoring+Systems+as+Part+of+Computerized+Maintenance+Management+Systems%2C+Mondragon+University%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Han, J., Cai, Y. and Cercone, N. (1992) &#x02018;Knowledge discovery in databases: An attribute-oriented approach,&#x02019; in 18th Int. Conf. Very Large Data Bases, Vancuver, Canada. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Han%2C+J%2E%2C+Cai%2C+Y%2E+and+Cercone%2C+N%2E+%281992%29+%27Knowledge+discovery+in+databases%3A+An+attribute-oriented+approach%2C%27+in+18th+Int%2E+Conf%2E+Very+Large+Data+Bases%2C+Vancuver%2C+Canada%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Hegedus, C., Varga, P. and Moldov&#x000E1;n, I. (2018) &#x02018;The MANTIS architecture for proactive maintenance,&#x02019; in <emphasis>5th International Conference on Control, Decision and Information Technologies CODIT</emphasis>, Thessaloniki.</para></listitem>
<listitem>
<para>Hegedus, C., Ciancarini, P., Frank&#x000F3;, A., Kancilija, A., Moldov&#x000E1;n, I., Papa, G., Poklukar, &#x00160;., Riccardi, M., Sillitti, A. and Varga, P. (2018) &#x02018;Proactive Maintenance of Railway Switches,&#x02019; in 5th International Conference on Control, Decision and Information Technologies CODIT, Thessaloniki, Greece. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Hegedus%2C+C%2E%2C+Ciancarini%2C+P%2E%2C+Frank%F3%2C+A%2E%2C+Kancilija%2C+A%2E%2C+Moldov%E1n%2C+I%2E%2C+Papa%2C+G%2E%2C+Poklukar%2C+&#x0160;%2E%2C+Riccardi%2C+M%2E%2C+Sillitti%2C+A%2E+and+Varga%2C+P%2E+%282018%29+%27Proactive+Maintenance+of+Railway+Switches%2C%27+in+5th+International+Conference+on+Control%2C+Decision+and+Information+Technologies+CODIT%2C+Thessaloniki%2C+Greece%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>https://kafka.apache.org/</para></listitem>
<listitem>
<para>http://projects.spring.io/spring-boot</para></listitem>
<listitem>
<para>IEEE Standard. (1998) &#x02018;IEEE 830-1998 &#x02013; Recommended Practice for Software Requirements Specifications,&#x02019;. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=IEEE+Standard%2E+%281998%29+%27IEEE+830-1998+-+Recommended+Practice+for+Software+Requirements+Specifications%2C%27%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>&#x02018;Internet of Things Architecture,&#x02019; [Online]. Available: http://www.iot-a.eu.</para></listitem>
<listitem>
<para>ISO, &#x02018;ISO 13374-1: Condition monitoring and diagnostics of machines &#x02013; Part 1: General Guidelines.&#x02019;</para></listitem>
<listitem>
<para>Jantunen, E.<etal/> (2016) &#x02018;Optimising maintenance: What are the expectations for cyber physical systems,&#x02019; in Emerging Ideas and Trends in Engineering of Cyber-Physical Systems (EITEC), Vienna, Austria. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Jantunen%2C+E%2E+%282016%29+%27Optimising+maintenance%3A+What+are+the+expectations+for+cyber+physical+systems%2C%27+in+Emerging+Ideas+and+Trends+in+Engineering+of+Cyber-Physical+Systems+%28EITEC%29%2C+Vienna%2C+Austria%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Junnola, J. (2017) The Suitability of Low-Cost Measurement Systems For Rolling Element Bearing Vibration Monitoring, University of Oulu.</para></listitem>
<listitem>
<para>Koro&#x00161;ec, P., Bole, U., and Papa, G. (2013) &#x02018;A multi-objective approach to the application of real-world production scheduling,&#x02019; Expert systems with applications, 40(15), pp. 5839&#x02013;5853.</para></listitem>
<listitem>
<para>Larrinaga, F.<etal/> (2018) &#x02018;Implementation of a reference architecture for cyber physical systems to support condition based maintenance,&#x02019; in 5th International Conference on Control, Decision and Information Technologies CODIT, Thessaloniki. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Larrinaga%2C+F%2E+%282018%29+%27Implementation+of+a+reference+architecture+for+cyber+physical+systems+to+support+condition+based+maintenance%2C%27+in+5th+International+Conference+on+Control%2C+Decision+and+Information+Technologies+CODIT%2C+Thessaloniki%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Lenarduzzi, V.<etal/> (2017) &#x02018;Analyzing forty years of software maintenance models,&#x02019; in 39th International Conference on Software Engineering (ICSE), Buenos Aires, Argentina. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Lenarduzzi%2C+V%2E+%282017%29+%27Analyzing+forty+years+of+software+maintenance+models%2C%27+in+39th+International+Conference+on+Software+Engineering+%28ICSE%29%2C+Buenos+Aires%2C+Argentina%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Levina, E. and Bickel, P. (2001) &#x02018;The EarthMover&#x02019;s Distance is the Mallows Distance: Some insights from statistics,&#x02019; in Proceedings of ICCV, Vancuver, Canada. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Levina%2C+E%2E+and+Bickel%2C+P%2E+%282001%29+%27The+EarthMover%27s+Distance+is+the+Mallows+Distance%3A+Some+insights+from+statistics%2C%27+in+Proceedings+of+ICCV%2C+Vancuver%2C+Canada%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>The MANTIS consortium, &#x02018;The MANTIS project website,&#x02019; [Online]. Available: http://www.mantis-project.eu/.</para></listitem>
<listitem>
<para>Mantis Consortium. (2016) &#x02018;D2.2 Reference architecture and design specification,&#x02019;.</para></listitem>
<listitem>
<para>McGeough, J. A. (1988) Advanced Methods of Machining, London &#x00026; New York: Chapman and Hall Ltd., pp. 55&#x02013;87.</para></listitem>
<listitem>
<para>McKay, M. (2018 to be published) Bearing Failure Prediction Based on a Wear Model, Aalto University.</para></listitem>
<listitem>
<para>MIMOSA. &#x02018;OSA-CBM 3.3.1 standard,&#x02019; 29/6/2010. [Online]. Available: http://www.mimosa.org/mimosa-osa-cbm. [Accessed 22/10/2015].</para></listitem>
<listitem>
<para>MIMOSA. &#x02018;OSA-EAI 3.2.3a standard,&#x02019; 12/5/2014. [Online]. Available: http://www.mimosa.org/mimosa-osa-eai. [Accessed 22/10/2015].</para></listitem>
<listitem>
<para>MIMOSA Consortium. &#x02018;An operations and maintenance information open system alliance,&#x02019; [Online]. Available: http://www.mimosa.org/mimosa/.</para></listitem>
<listitem>
<para>MIMOSA Consortium. &#x02018;The MIMOSA project site, [Online]. Available: http://www.mimosa.org/.</para></listitem>
<listitem>
<para>Mosquitto<superscript>TM</superscript>, (2010) &#x02018;An Open Source MQTT v3.1/v3.1.1 Broker,&#x02019; [Online].</para></listitem>
<listitem>
<para>National Instruments Corporation, &#x02018;LabVIEW,&#x02019; 2018. [Online]. Available: http://www.ni.com/en-gb/shop/select/labview/. [Accessed 01/04/2018].</para></listitem>
<listitem>
<para>PSK Standards Association. (2007) Vibration Measurement in Condition Monitoring. Choice and Identification of Measurement Point, 3nd edition.</para></listitem>
<listitem>
<para>PSK Standards Association. (2006) Vibration Measurement in Condition Monitoring. Selection and Mounting of Sensor, Connector and Cable, 4th edition.</para></listitem>
<listitem>
<para>&#x02018;RabbitMQ,&#x02019; [Online]. Available: http://www.rabbitmq.com. [Accessed April 2018].</para></listitem>
<listitem>
<para>Stuart, A., Ord, K., and Arnold, S. (1999) Kendall&#x02019;s Advanced Theory of Statistics: Volume 2A&#x02014;Classical Inference &#x00026; the Linear Mode, Wiley. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Stuart%2C+A%2E%2C+Ord%2C+K%2E%2C+and+Arnold%2C+S%2E+%281999%29+Kendall%27s+Advanced+Theory+of+Statistics%3A+Volume+2A-Classical+Inference+%26+the+Linear+Mode%2C+Wiley%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Vinoski, S. (2006) &#x02018;Advanced message queuing protocol,&#x02019; IEEE Internet Computing, 10, pp. 87&#x02013;89. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Vinoski%2C+S%2E+%282006%29+%27Advanced+message+queuing+protocol%2C%27+IEEE+Internet+Computing%2C+10%2C+pp%2E+87-89%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Wold, H. (1975) Path Models with Latent Variables: The NIPALS Approach, vol. Quantitative Sociology: International Perspectives on Mathematical and Statistical Modeling, A. A. F. M. B. &#x00026;. V. C. H. M. Blalock, Ed., New York: New York Academic, pp. 307&#x02013;357. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Wold%2C+H%2E+%281975%29+Path+Models+with+Latent+Variables%3A+The+NIPALS+Approach%2C+vol%2E+Quantitative+Sociology%3A+International+Perspectives+on+Mathematical+and+Statistical+Modeling%2C+A%2E+A%2E+F%2E+M%2E+B%2E+%26%2E+V%2E+C%2E+H%2E+M%2E+Blalock%2C+Ed%2E%2C+New+York%3A+New+York+Academic%2C+pp%2E+307-357%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>&#x02018;WSO,&#x02019; [Online]. Available: https://wso2.com/. [Accessed April 2018]. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=%27WSO%2C%27+%5BOnline%5D%2E+Available%3A+https%3A%2F%2Fwso2%2Ecom%2F%2E+%5BAccessed+April+2018%5D%2E" target="_blank">Google Scholar</ulink></para></listitem>
</orderedlist>
</section>
</chapter>
<chapter class="chapter" id="ch08" label="8" xreflabel="8">
<title>Business Models: Proactive Monitoring and Maintenance</title>
<para><emphasis role="strong">Michel I&#x000F1;igo Ulloa<superscript><emphasis role="strong">1</emphasis></superscript>, Peter Craamer<superscript><emphasis role="strong">2</emphasis></superscript>, Salvatore Esposito<superscript><emphasis role="strong">3</emphasis></superscript>, Carolina Mej&#x000ED;a Ni&#x000F1;o<superscript><emphasis role="strong">1</emphasis></superscript>, Mario Riccardi<superscript><emphasis role="strong">3</emphasis></superscript>, Antonio Ruggieri<superscript><emphasis role="strong">3</emphasis></superscript>, and Paolo Sannino<superscript><emphasis role="strong">3</emphasis></superscript></emphasis></para>
<para><superscript>1</superscript>MONDRAGON Corporation, Spain<?lb?><superscript>2</superscript>Mondragon Sistemas De Informacion, Spain<?lb?><superscript>3</superscript>Ansaldo STS, Italy</para>
<para>As stated by Brisk Insights market analysis [Brisk Insights, 2016] the global operational predictive maintenance market will grow at a CAGR of 26.6% within 2016&#x02013;2022, foreseeing a total market value of 2.900 million by the end of such period. This will be certainly boosted by the IIoT market rise, which is growing at a CAGR of 42%, and will act as an enabler for its rapid industrial penetration. One of the key sectors (among all industries), in which predictive maintenance will make a huge difference will be manufacturing. The European manufacturing sector accounts for 2 million companies and 33 million jobs, representing the 15% of the total European GDP. With the aim of increasing this contribution to 20% by 2020, European manufacturing industry faces a huge but promising challenge, given industry&#x02019;s potential in jobs and growth creation. However, industry&#x02019;s share in the European GDP has declined during the last years, mainly due to a deceleration of global investments, market uncertainty and production offshoring to low-cost countries. This applies to all actors of the manufacturing value chain, involving production asset end users, asset manufacturers and asset service providers.</para>
<para>In order to cope with that, the full digitization of European industrial ecosystems has been stated as the foundation upon which competitiveness goals will be achieved. Within this framework, predictive maintenance accounts for a huge improvement potential to all actors mentioned: relevant productivity increase (asset end users), new revenue streams with higher profit margins (asset manufacturers) and new business opportunities based on analytics (asset service providers). According to McKinsey, predictive maintenance in factories could cut maintenance costs down by 10 to 40 percent, leading to manufacturers savings of 215 to 580 Billion&#x20AC; in 2025 [McKinsey, 2015], resulting from reduced downtimes and minimized manufacturing defects among others. Despite this clear potential, maintenance strategies in place still rely on ineffective corrective and preventive maintenance actions, which have a high impact on productivity (higher production costs, delays on delivery, customer dissatisfaction, etc.). Not only available shop floor data and production assets behavior knowledge is underutilized, but also new businesses generation along the value chain is heavily hampered.</para>
<para>Regarding technology, there are several reasons behind the lack of adoption of predictive maintenance across European industries:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Production systems complexity:</emphasis> the majority of European industrial facilities is shaped by very heterogeneous assets, being the asset end user unable to gather deep knowledge about the behavior of each asset (expertise often retained by the asset manufacturer);</para></listitem>
<listitem>
<para><emphasis role="strong">Lack of interoperability among different assets:</emphasis> afraid of the possibility of having a 3<emphasis><superscript>rd</superscript></emphasis> party providing services on their production assets, asset manufacturers often apply vendor lock-in solutions to their products. This results in a huge IT integration work required to connect them, usually preventing end users from implementing predictive maintenance solutions;</para></listitem>
<listitem>
<para><emphasis role="strong">Non-reliable prognostics estimates at a system level:</emphasis> even though successful prognostics applications have been deployed at component and sub-system level, asset end users interest focuses on increasing the availability of the whole system, which has a direct impact on competitiveness. Thus, the lack of real prognostics and health management systems demonstrated at industrial level derives from a reluctance in early adopters.</para></listitem>
</itemizedlist>
<para>In order to overcome those limiting factors, there is a clear need of bringing together all value chain actors (gathering real-time data, asset behavior knowledge and analytics expertise); as well as taking advantage of advanced analytics technologies already applied in a wide range of sectors. This will enable to match predictive management system capabilities with real industrial needs, achieving downtime minimization and OEE maximization at a system level. Besides all above, several non-technological challenges (such as corporate culture) prevent the penetration of predictive maintenance technologies across industries. This applies especially to SMEs, being the most relevant the following challenges:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Uncertain RoI:</emphasis> industrial CAPEX plans are fully subject to their expected profitability, usually in a short-term (depending on the company&#x02019;s balance sheet, often 2&#x02013;3 years). Since the implementation of such predictive maintenance systems may imply investing in data acquisition, industrial communications and advanced analysis technologies (mainly regarding old production assets), companies often opt for more profitable investments (e-g purchasing new machinery, which leads to a direct productivity improvement);</para></listitem>
<listitem>
<para><emphasis role="strong">Required skills:</emphasis> despite the high level of automation in place in most of the European industrial companies, the implementation of Industry 4.0 (within which predictive maintenance is located) is currently requiring a shift from classical operators to highly analytical profiles. Industrial HMIs usually do not take advantage of available technologies such as adaptability, self-learning features, etc., resulting in workers frustration by not showing the right information to the right people.</para></listitem>
</itemizedlist>
<section class="lev1" id="sec8-1">
<title>8.1 Maintenance Present and Future Trends</title>
<para>Ever since asset failures have caused downtimes and extra costs, accidents or inefficiencies, businesses have supplied material and human resources to minimize their impact and avoid their re-occurrence. These resources have been different depending on a) the harm to be avoided, b) impact on the balance sheet and Profit &#x00026; Loss Statement, or c) competitive threats that hinder business survival causing very different grades of implementation depending on the sector and the type of asset.</para>
<para>Current approaches try to preserve function and operability, optimize performance and increase asset lifespan with optimal investments. This approach is the result of a significant evolution through time. According to some authors, four maintenance generations can be distinguished and each represents the best practices used in particular periods of history, as depicted in <link linkend="F8-1">Figure <xref linkend="F8-1" remap="8.1"/></link> or <link linkend="F8-2">Figure <xref linkend="F8-2" remap="8.2"/></link>, both from [Cristi&#x000E1;n M. Lincovil B and G. Ivonne Guti&#x000E9;rrez M, 2006]</para>
<fig id="F8-1" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-1">Figure <xref linkend="F8-1" remap="8.1"/></link></label>
<caption><para>Main maintenance objectives evolution [Cristi&#x000E1;n M. Lincovil B and G. Ivonne Guti&#x000E9;rrez M, 2006].</para></caption>
<graphic xlink:href="graphics/ch08_fig001.jpg"/>
</fig>
<fig id="F8-2" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-2">Figure <xref linkend="F8-2" remap="8.2"/></link></label>
<caption><para>Evolution of maintenance techniques [Cristi&#x000E1;n M. Lincovil B and G. Ivonne Guti&#x000E9;rrez M, 2006].</para></caption>
<graphic xlink:href="graphics/ch08_fig002.jpg"/>
</fig>
<para>Towards the end of the 90s, the development of 3rd generation maintenance included:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Decision-making tools such as risk management and error analysis;</para></listitem>
<listitem>
<para>New maintenance techniques like condition monitoring;</para></listitem>
<listitem>
<para>Design, with special relevance on reliability and ease of maintenance;</para></listitem>
<listitem>
<para>Wide-reaching organizational changes looking for employee input, teamwork and flexibility.</para></listitem>
</itemizedlist>
<para>The new approach of the fourth generation is centered on failure elimination using proactive techniques. It is no longer enough to eliminate failure effects but to pinpoint the root causes of malfunctions and avoid their re-occurrence.</para>
<para>Additionally, there are growing concerns about equipment reliability and thus maintenance is gaining more relevance starting from the design phase of the project. Also, it is very common to implement continuous improvement systems regarding preventive and predictive maintenance plans, applied to the planning and execution of maintenance.</para>
<para>Apart from the mentioned characteristics, there are other aspects whose importance had gotten considerable greater:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Risk management.</emphasis> The identification and control of possible incidents that have low probability but high impact (especially in high-risk industries) was gaining more relevance. The role of maintenance is key in this process and there are concerns that methodologies in use for &#x0201C;low probability / high impact&#x0201D; incidents are not effective, so new methodologies have to be developed;</para></listitem>
<listitem>
<para><emphasis role="strong">Failure patterns.</emphasis> Traditional thinking about the link between machine aging and malfunction is being shifted. In fact, there is evidence for some equipment that there is a low correlation between operation time and failure probability. The rate of machine failures can be represented in a bathtub curve (<link linkend="F8-3">Figure <xref linkend="F8-3" remap="8.3"/></link>) that shows that there are more probabilities of early and wear-out failures. [Arnold Vogt, 2016].</para></listitem>
</itemizedlist>
<fig id="F8-3" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-3">Figure <xref linkend="F8-3" remap="8.3"/></link></label>
<caption><para>Bathtub Curve used in reliability engineering [Arnold Vogt, 2016].</para></caption>
<graphic xlink:href="graphics/ch08_fig003.jpg"/>
</fig>
<para>Proactive maintenance consists of a step beyond when aiming to reduce failure probabilities. The main focus of proactive maintenance relies on eliminating failures, not their impact. For this purpose, root causes have to be removed, which requires deep knowledge of the system. Some tools like RCA are helpful although they are often used as reactive tools rather than proactive ones.</para>
<para>Methods for improving reliability can be divided into two families [Jantunen, 2016]:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Proactive methods:</emphasis> they seek to improve reliability using techniques that allow gradual removal of both persistent and potential failures;</para></listitem>
<listitem>
<para><emphasis role="strong">Reactive methods:</emphasis> they seek quick and effective solutions to daily problems and to avoid repetition of major failures. Basically, it consists of &#x0201C;post-mortem&#x0201D; methods and its main exponent is RCA.</para></listitem>
</itemizedlist>
<para>At this time, there are multiple maintenance techniques, methodologies and philosophies. The use of these methodologies is sector dependent and also differs between companies within the same sector. The main factors that drive the selection of a methodology are the risks and impact of the failures, the cost and implementation difficulties, and the competitive intensity within the sector. The most common tools, trends and methodologies are detailed below.</para>
<section class="lev2" id="sec8-1-1">
<title>8.1.1 Tools</title>
<para>This section gets through the most used strategies and techniques applied for advanced maintenance operations.</para>
<section class="lev3" id="sec8-1-1-1">
<title>8.1.1.1 Total productive maintenance</title>
<para>Known for the great benefits obtained in manufacturing companies and bold success stories in Japan, TPM emphasizes teamwork and it leans on correct cleaning and lubrication for chronic failure removal. It involves a strong team culture and sense of belonging to employees. When this is not applicable, a cultural shift is required as it is strongly related to continuous quality improvement and zero-defects philosophies. Anyway, this is difficult to implement in process-focused companies due to the ambiguity between the concepts of quality and defects [M&#x000E1;rquez et al., 2004].</para>
<para>When the roll-out of the methodology has been successful, there have been vast improvements in safety, reliability, availability and maintenance costs.</para>
</section>
<section class="lev3" id="sec8-1-1-2">
<title>8.1.1.2 Root-cause analysis</title>
<para>RCA is a very powerful technique that allows problem resolution with a short and mid-term view. It uses exhaustive research techniques with an intent to remove problem and failure root-causes. Its value is not only to avoid critical events but to eliminate chronic events that tend to consume maintenance resources. Gradual chronic and small problem removal require deeper analysis [M&#x000E1;rquez et al., 2004].</para>
</section>
<section class="lev3" id="sec8-1-1-3">
<title>8.1.1.3 Reliability centered maintenance</title>
<para>RCM is a technique that appeared in the late 90s to respond to high maintenance (preventive) costs in aircrafts. It proved its validity within the aerospace&#x02019;s sector, not only lowering costs and maintenance activities but also improving reliability, availability and security. This makes it appealing to other industries such as military, oil &#x00026; gas and utilities [Brisk Insights, 2016].</para>
<para>It is based on selecting maintenance only in the case that failure consequences so dictate. To do so, exhaustive studies have to be carried out for every function. RCM establishes priorities: safety and environment, production, repair costs. This has made this technique a very valuable tool in industries with high-security demands, generating excellent results.</para>
</section>
<section class="lev3" id="sec8-1-1-4">
<title>8.1.1.4 Improving operational reliability</title>
<fig id="F8-4" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-4">Figure <xref linkend="F8-4" remap="8.4"/></link></label>
<caption><para>Aspects of operational reliability.</para></caption>
<graphic xlink:href="graphics/ch08_fig004.jpg"/>
</fig>
<para><emphasis>Improving Operational Reliability</emphasis> gathers the best maintenance practices and operations with a business focus recognizing maintenance limitations to reach appropriate reliability levels [M&#x000E1;rquez et al., 2004]. This technique focuses on different aspects of operational reliability (see <link linkend="F8-4">Figure <xref linkend="F8-4" remap="8.4"/></link>).</para>
<para>It divides the techniques in:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Diagnosis: using short/mid-term reactive and long-term proactive opportunities;</para></listitem>
<listitem>
<para>Control: RCM as proactive technique and RCA as reactive. Also, Improving Operational Reliability used for static equipments;</para></listitem>
<listitem>
<para>Optimization: using advanced statistical risk management tools for optimal maintenance intervals, downtimes, inspections, etc.</para></listitem>
</itemizedlist>
</section>
<section class="lev3" id="sec8-1-1-5">
<title>8.1.1.5 Criticality analysis</title>
<para>This methodology allows establishing hierarchies or priorities of processes, systems, and equipment, creating a structure that makes decision making easier and effective. Efforts and resources are analyzed in key areas based on real data [4].</para>
<para>Once areas are identified it is much easier to design a strategy to make studies or projects that improve operational reliability, initiating applications to a group of processes or elements that form part of high criticality areas.</para>
</section>
<section class="lev3" id="sec8-1-1-6">
<title>8.1.1.6 Risk-based maintenance</title>
<para>This technique establishes inspection patterns for static equipment based on associated risks. It is a methodology for determining the most economical use of maintenance resources. This is done so that the maintenance effort across a facility is optimized to minimize any risk of a failure [M&#x000E1;rquez et al., 2004].</para>
<para>A risk-based maintenance strategy is based on two main phases:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Risk assessment;</para></listitem>
<listitem>
<para>Maintenance planning based on the risk.</para></listitem>
</itemizedlist>
</section>
<section class="lev3" id="sec8-1-1-7">
<title>8.1.1.7 Maintenance optimization models</title>
<para>These are mathematical models whose purpose is to discover a balance between costs and benefits, taking into account all type of restrictions.</para>
<para>The maintenance optimization models provide various results. First, they can be evaluated and used to compare different strategies regarding the characteristics of reliability and profitability. Second, the models can be monitored. Third, the models can determine how often to inspect or maintain the assets. Fourth, results of the evaluations of the models can be used for maintenance planning.</para>
</section>
<section class="lev3" id="sec8-1-1-8">
<title>8.1.1.8 Model-based condition monitoring</title>
<para>This tool involves the monitoring of one or more condition parameters in machinery (vibration, temperature etc.), in order to identify a significant change that is indicative of a developing fault.</para>
<para>The following list includes the main condition monitoring techniques applied in the industrial and transportation sectors:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Vibration Analysis and diagnostics;</para></listitem>
<listitem>
<para>Lubricant analysis;</para></listitem>
<listitem>
<para>Acoustic emission (Airborne Ultrasound);</para></listitem>
<listitem>
<para>Infrared thermography;</para></listitem>
<listitem>
<para>Ultrasound testing (Material Thickness/Flaw Testing);</para></listitem>
<listitem>
<para>Motor Condition Monitoring and Motor Current Signature Analysis.</para></listitem>
</itemizedlist>
<para>The use of Condition Monitoring is becoming very common. It was originally used by NASA to monitor and detect errors in the development of spacecraft engines.</para>
</section>
</section>
<section class="lev2" id="sec8-1-2">
<title>8.1.2 Trends</title>
<para>This section provides some background on a number of different strategies that are currently gaining interest in the field of maintenance.</para>
<section class="lev3" id="sec8-1-2-1">
<title>8.1.2.1 Servitization</title>
<para>Currently, traditional product-centric models are being transformed into customer-centric models where the focus shifts from creating the best product to the best solution for the customer. Product-based organizations focus their efforts on the development of products that they put on the market rather than the development of client-driven solutions. In the former model, companies sell and customers buy, whereas, in the latter, the company is focused on meeting customers&#x02019; needs. There is also a symbiotic model where both, the vendor and the client collaborate in a solid relationship in order to succeed (<link linkend="F8-5">Figure <xref linkend="F8-5" remap="8.5"/></link>).</para>
<fig id="F8-5" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-5">Figure <xref linkend="F8-5" remap="8.5"/></link></label>
<caption><para>Different models of collaboration between vender and client.</para></caption>
<graphic xlink:href="graphics/ch08_fig005.jpg"/>
</fig>
<para>Variations on the product-driven model have been predominant in the manufacturing industry so far and services have historically been a small fraction of revenue (around 10&#x02013;20%) for traditional manufacturing companies, although having greater gross margin compared to traditional goods&#x02019; sales.</para>
<para>In order to survive and thrive in a globalized market, companies are being forced to develop intelligent maintenance solutions and move towards a balanced mix of product- and client-focused approaches. Hence, companies will benefit from high service gross margins, much less adjusted than the traditional product-focused ones (<link linkend="F8-6">Figure <xref linkend="F8-6" remap="8.6"/></link>).</para>
<fig id="F8-6" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-6">Figure <xref linkend="F8-6" remap="8.6"/></link></label>
<caption><para>Margin in product-driven model vs margin in service models.</para></caption>
<graphic xlink:href="graphics/ch08_fig006.jpg"/>
</fig>
<para>Most companies recognize the urgency for moving their approach towards this new model, also known as Servitization, and thus are creating sophisticated platforms that leverage IT innovations such as cloud computing, machine learning, distributed databases, sensors, embedded solutions and so on, to create services that offer customized added value for the customer.</para>
</section>
<section class="lev3" id="sec8-1-2-2">
<title>8.1.2.2 Degree of automation</title>
<para>The level of maturity of the maintenance platforms and the degree of automation (<link linkend="F8-7">Figure <xref linkend="F8-7" remap="8.7"/></link>) differ greatly among the studied organizations.</para>
<para>For example, global 500 company Siemens has launched in Q1 2016 its platform &#x02013; MindSphere &#x02013; for optimizing asset performance, energy and resource consumption, maintenance and related services. Siemens will integrate it with existing platforms of the entire value-chain. SAP launched &#x0201C;SAP Predictive Maintenance and Service, cloud edition&#x0201D; on November 2014; IBM is able to predict and prevent asset failures, detect quality issues and improve operational processes since 2012 with its platform. Large Enterprises such as Rolls Royce have been performing predictive maintenance related activities for a long time being for them IoT and cloud-based platforms an evolution, instead of revolution. Trumpf founded Axoom (2015, within the framework of a consortium), a digital platform that covers the complete machine tool solution value-chain with the goal of shifting from being a machine vendor to a software vendor.</para>
<para>Current market solutions can be divided into <emphasis role="strong">a) customized solutions</emphasis> that cover the all maintenance services and <emphasis role="strong">b) standardized mainstream platforms</emphasis> [Arnold Vogt, 2016]. The current trend among large companies is to acquire the former due to the high level of customization and complexity that companies demand. Among SMEs, the standardized platforms are popular due to agile implementation efforts.</para>
<para>Vendors are trying to merge both solutions with the aim of providing customized vertical solutions (extensions) based on standardized platforms. New functional modules -verticals- are getting installed on top of horizontal core platforms and provide customized experiences to the user (<link linkend="F8-8">Figure <xref linkend="F8-8" remap="8.8"/></link>).</para>
<fig id="F8-7" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-7">Figure <xref linkend="F8-7" remap="8.7"/></link></label>
<caption><para>Degree of platforms automation among organizations.</para></caption>
<graphic xlink:href="graphics/ch08_fig007.jpg"/>
</fig>
<fig id="F8-8" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-8">Figure <xref linkend="F8-8" remap="8.8"/></link></label>
<caption><para>Customized vertical where solutions based on standard platform.</para></caption>
<graphic xlink:href="graphics/ch08_fig008.jpg"/>
</fig>
</section>
<section class="lev3" id="sec8-1-2-3">
<title>8.1.2.3 Top-down vs. bottom-up</title>
<para>Market leaders see Industry 4.0 as a Roadmap to build the ultimate CPS where smart products have all the information to be manufactured in every step of the value chain and the flexibility to adapt to changing conditions in an integrated production lifecycle, from product design to services (<link linkend="F8-9">Figure <xref linkend="F8-9" remap="8.9"/></link>). In short, market leaders are pursuing a strategic top-down approach.</para>
<para>Meanwhile, some Large Enterprises and SMEs are building platforms bottom-up in a Lean strategy, with a more limited functional scope to digitalize a particular area. In maintenance, this scope usually covers:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Monitorization, and smart data analytics;</para></listitem>
<listitem>
<para>Customized reports, alerts and warnings;</para></listitem>
<listitem>
<para>Consulting services.</para></listitem>
</itemizedlist>
<fig id="F8-9" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-9">Figure <xref linkend="F8-9" remap="8.9"/></link></label>
<caption><para>Top-down approach, where the product has all the information in every step of the value chain.</para></caption>
<graphic xlink:href="graphics/ch08_fig009.jpg"/>
</fig>
</section>
<section class="lev3" id="sec8-1-2-4">
<title>8.1.2.4 Smart products</title>
<para>Smart products are CPSs providing new features and functions based on connectivity. In industry 4.0 the smartification of the products and machinery is a key factor for the predictive maintenance. The information gathered from products and machines are translated into maintenance plans and products improvements (<link linkend="F8-10">Figure <xref linkend="F8-10" remap="8.10"/></link>).</para>
<fig id="F8-10" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-10">Figure <xref linkend="F8-10" remap="8.10"/></link></label>
<caption><para>Smart product as CPS providing new features and functions based on connectivity.</para></caption>
<graphic xlink:href="graphics/ch08_fig010.jpg"/>
</fig>
</section>
<section class="lev3" id="sec8-1-2-5">
<title>8.1.2.5 Machine learning</title>
<para>There are several software techniques and methods that uncover hidden patterns in large data-sets. ML in particular deals with algorithms that give computers the ability to learn based on empirical data.</para>
<para>ML makes software applications act without being explicitly programmed. ML has made possible the emergence of self-driving cars, effective medical diagnosis, chess champions (Deep Blue) and a deep inflection point due to AlphaGos win (Googles DeepMind). AlphaGo relies on several ML components, like deep neural networks and tree search.</para>
<para>Many companies are looking at using ML in their products. The manufacturing industry is not an exception, and it is currently developing algorithms to make data-driven predictions or decisions.</para>
</section>
</section>
</section>
<section class="lev1" id="sec8-2">
<title>8.2 Shift to a Proactive Maintenance Business Landscape</title>
<para>The current technical development in industry regarding information handling and digitalization leads to new ways of producing goods. The industry demands flexible, safe, environmental friendly and available production processes. At the same time, production processes become more automated, complex and dynamic. This approach, called Industry 4.0, focuses on delivering advanced technical solutions to manufacturing problems and supporting new industrial philosophies such as lean production, but in order to become successful these technical innovations in manufacturing must be connected with strategic business models.</para>
<para>In order to get an effective and successful business model, the characterization must be oriented in a dynamic market that demands flexible product-service model. The PMM business models analysed during the execution of the MANTIS project have this orientation.</para>
<para>Predictive/proactive maintenance solutions are common in processing industries like Oil &#x00026; Gas, Wind, Utilities, and aerospace. These solutions allow efficient critical asset management based on condition monitoring using predetermined models.</para>
<para>Monitored data correspond to parameters such as vibration, lubrication, temperature, and strength, among others. For example, vibration analysis can be used to detect regular mechanical defects like lack of alignment, component erosion, and union weakening.</para>
<para>A normality model is established as an ideal standard, consisting of a specification of how an asset behaves under certain conditions, and therefore, a deviation of the &#x0201C;normality model&#x0201D; indicates asset deterioration. The model has to be able to offer early predictions and exact diagnosis. The model of normality is built based on several approaches:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>An empirical approach based on historical asset data;</para></listitem>
<listitem>
<para>Engineering approach based on engineering principles that help describing how an asset should function under specific conditions;</para></listitem>
<listitem>
<para>A combination of the empirical and the engineering approaches.</para></listitem>
</itemizedlist>
<para>The design and implementation of a good model of normality require a precise asset&#x02019;s know-how in order to define failure patterns. A pivoting process readjusting parameters and KPIs helps to establish a robust normality model.</para>
<para>Manufacturing facilities also use PMM solutions to verify conditions of their most critical assets, although their degree of implementation is lower than in the processing industries. There are barriers that have to be overcome if PMM solutions are expected to be successful. They can be summarized as follows:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Complex and extended asset mix. Depending of the type of industry, the critical chain focuses on few or huge amount of equipment;</para></listitem>
<listitem>
<para>Capacity to function with changing conditions. The working conditions of assets are usually variable due to a high level of human interaction, a wide range of materials, multiple asset operations and changing processed end-products;</para></listitem>
<listitem>
<para>Asset data availability. Many critical assets are not monitored and not even have data capture options so they cannot be substituted by smart assets in a short period of time;</para></listitem>
<listitem>
<para>Confidence in PMM solutions. The high degree of skepticism regarding technologies capacity with respect to traditional maintenance methods.</para></listitem>
</itemizedlist>
<para>Advanced predictive/proactive maintenance solutions are supported by technological innovations. The rise of Cloud Computing, distributed databases and the establishment of machine learning algorithms, among other technologies, are helping companies create new sophisticated maintenance solutions and allowing new business models to emerge.</para>
<para>Proactive and Monitoring Maintenance solutions analyse data from multiple sources (Physical assets, SCADA, PDM, PLM, ERP, etc.) and provide recommended actions based on smart analytics, using advanced statistical models such as classification, regression, associations, clustering and many more.</para>
<section class="lev2" id="sec8-2-1">
<title>8.2.1 Key Success Factors</title>
<para>According to IBM [IBM, 2013], 83% of CIOs cited advanced analytics as the primary path to competitiveness enabling Asset Performance by a) improving quality and reducing failures and outages, b) optimizing services and support and c) optimizing operations, maintenance and product quality. As a title of example, <link linkend="T8-1">Table <xref linkend="T8-1" remap="8.1"/></link> maps this path to competitiveness onto the predictive maintenance use case.</para>
<table-wrap position="float" id="T8-1">
<label><link linkend="T8-1">Table <xref linkend="T8-1" remap="8.1"/></link></label>
<caption><para>Predictive maintenance solutions to client and vendor [IBM, 2013]</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="rows">
<thead>
<tr>
<td valign="top" align="left">Key Metric</td>
<td valign="top" align="left">Business Benefit</td>
<td valign="top" align="left">How Advanced Analytics enables Value</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left"><emphasis role="strong">Maximize Revenue</emphasis><?lb?><emphasis>Competitive Advantage</emphasis></td>
<td valign="top"><itemizedlist mark="bullet" spacing="normal"><listitem><para>Products and services</para></listitem>
<listitem><para>High availability</para></listitem>
<listitem><para>Lower Start-up costs</para></listitem></itemizedlist></td>
<td valign="top" align="left">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>New products and services, Up Sell opportunities, Higher product quality</para></listitem>
<listitem><para>Better asset utilization, more production cycles</para></listitem>
<listitem><para>Fewer reworks, fewer installation repairs</para></listitem>
</itemizedlist></td>
</tr>
<tr>
<td valign="top" align="left"><emphasis role="strong">Cost Savings</emphasis><?lb?><emphasis>Increased Reliability</emphasis></td>
<td valign="top"><itemizedlist mark="bullet" spacing="normal"><listitem><para>Less Unplanned downtime</para></listitem>
<listitem><para>Better productivity</para></listitem>
<listitem><para>Better quality</para></listitem></itemizedlist></td>
<td valign="top" align="left">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Fewer failures, faster problem identification, better process throughput</para></listitem>
<listitem><para>Issues cost avoidance, faster root cause, higher equipment utilization</para></listitem>
<listitem><para>Proactive monitoring, predictive performance, identification of factors likely to result in diminished quality</para></listitem>
</itemizedlist></td>
</tr>
<tr>
<td valign="top" align="left"><emphasis role="strong">O&#x00026;M Costs</emphasis><?lb?><emphasis>Increased Efficiency</emphasis></td>
<td valign="top"><itemizedlist mark="bullet" spacing="normal"><listitem><para>Non-production costs</para></listitem>
<listitem><para>Shorter maintenance</para></listitem>
<listitem><para>Lower warranty costs</para></listitem></itemizedlist></td>
<td valign="top" align="left">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Fewer failures, fewer emergencies, less need for excess MRO inventory</para></listitem>
<listitem><para>Predictive maintenance, better planning</para></listitem>
<listitem><para>Fewer Part failures, shorten issue resolution</para></listitem>
</itemizedlist></td>
</tr>
<tr>
<td valign="top" align="left"><emphasis role="strong">Customer Experience</emphasis><?lb?><emphasis>Increased Satisfaction</emphasis></td>
<td valign="top"><itemizedlist mark="bullet" spacing="normal"><listitem><para>Proactive management</para></listitem>
<listitem><para>Individual experience</para></listitem>
<listitem><para>Better collaboration</para></listitem></itemizedlist></td>
<td valign="top" align="left">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Fewer surprises, proactive communication</para></listitem>
<listitem><para>Focused communication, holistic view</para></listitem>
<listitem><para>Information integrated across industries, better insight across silos</para></listitem>
</itemizedlist></td>
</tr>
</tbody>
</table>
</table-wrap>
<para>Companies need transformations to be competitive as they face unprecedented global challenges. Thus, new type of organizations are emerging and highly skilled workforces are joining multidisciplinary teams that work with tools that improve communication and remote work. Offices are also transforming into open spaces that encourage creativity, ideas exchange and places where different teams work together in order to create innovative products and services. Organizations create internal structures to satisfy demands of new generation of users who make heavy use of smart devices in their everyday lives.</para>
<para>To elaborate the mentioned products and services, and to attract talent from other sectors, manufacturing industries need to take into account a number of points that are developed in following subsections, namely: User Experience, Privacy, Scalability, Technical Debt, Skills, Organizational structure, Technology.</para>
<section class="lev3" id="sec8-2-1-1">
<title>8.2.1.1 User experience</title>
<para>UX encompasses all aspects of the end-user&#x02019;s interaction with the company&#x02019;s services and products. UX is a broad term and should not be confused with User Interface (UI).</para>
<para>UX is how a person feels when interacting with a product. These feelings include usability, accessibility, performance, design/aesthetics, utility, ergonomics, overall human interaction and marketing.</para>
<para>Companies with highly effective UX have increased their revenue and the business benefits can be summed up into three categories:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Increased productivity: helping users to solve tasks faster and easier;</para></listitem>
<listitem>
<para>Reduced costs: less training and support for the end-users;</para></listitem>
<listitem>
<para>Increased sales: providing superior experience, and thus, market differentiation.</para></listitem>
</itemizedlist>
</section>
<section class="lev3" id="sec8-2-1-2">
<title>8.2.1.2 Privacy</title>
<para>With the emergence of Cloud-based software solutions that leverage underlying Big Data and Analytics power, concerns have risen regarding data privacy.</para>
</section>
<section class="lev3" id="sec8-2-1-3">
<title>8.2.1.3 Scalability</title>
<para>Successful product-market fit requires building platforms with strong scalability capacities. In addition, a correct balance between the different business models, full-service provider vs IoT platform vendor, as being successful in both is extremely difficult.</para>
<para>Platform based scalability can be achieved by strong IT infrastructures. Efforts that promote simplicity over complexity are preferable, as well as good usability design techniques.</para>
</section>
<section class="lev3" id="sec8-2-1-4">
<title>8.2.1.4 Technical debt</title>
<para>The surrounding infrastructure and platforms needed to support the system are vast and complex. A summary of the most relevant debts inherited from smart solutions follows:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">CPS Complexity.</emphasis> The implementation of intelligent CPS systems implies massive ongoing maintenance costs due to inherited-incidental and accidental-complexity;</para></listitem>
<listitem>
<para><emphasis role="strong">Data Dependencies.</emphasis> Data dependencies are not necessary causalities. Some data are unstable due to their changing behaviour over time and/or are underutilized/unneeded. This creates noise in the system so actions should be regularly put in place to identify and remove unnecessary data dependencies;</para></listitem>
<listitem>
<para><emphasis role="strong">Glue Code.</emphasis> Using general purpose libraries, third party open-source or proprietary packages, or cloud based solutions in the system results in glue code pattern;</para></listitem>
<listitem>
<para><emphasis role="strong">Configuration Debt.</emphasis> Systems allow to configure different options: features, data-mining techniques, data verification methods, etc. All this messiness makes configuration hard to modify correctly, hard to reason about and can lead to production issues;</para></listitem>
<listitem>
<para><emphasis role="strong">Data Testing Debt.</emphasis> Production issues are often caused by code bugs but also by data related issues, so some amount of data testing is critical to a robust and stable system;</para></listitem>
<listitem>
<para><emphasis role="strong">Reproducibility Debt.</emphasis> Re-running a process should return the same result but this is often not the case due to randomized algorithms, interactions with a changing physical world or similar causes. This adds complexity to the work of engineering teams;</para></listitem>
<listitem>
<para><emphasis role="strong">Process Management Debt.</emphasis> Real systems have various models running simultaneously. This raises many type of management problems that can cause production incidents;</para></listitem>
<listitem>
<para><emphasis role="strong">Cultural Debt.</emphasis> There is always a tension within a group of heterogeneous people (R&#x00026;D, engineering, support,..) but it is mandatory to create a working culture that promotes reduction of complexity, improvements in reproducibility, stability, and experience shows that strong multidisciplinary teams achieve better results.</para></listitem>
</itemizedlist>
</section>
<section class="lev3" id="sec8-2-1-5">
<title>8.2.1.5 Skills</title>
<para>Smart Factories in general and PM in particular demand new labour skills. The identification of these future skill-sets and competencies is a key factor and can determine the success or failure of the project.</para>
<para>Strong emphasis on IT related skills is clear: UI/UX design, mathematics, data-science, Front and Back-End development, AI, Machine Learning, etc. In summary, a hybrid approach where engineers and researchers are working together on the same teams.</para>
<para>In these knowledge-based factories work-related training becomes critical for both, employees and customers.</para>
</section>
<section class="lev3" id="sec8-2-1-6">
<title>8.2.1.6 Organizational structure</title>
<para>A major factor of successful organizations will be the ability to create a <emphasis>&#x0201C;</emphasis>learning organization<emphasis>&#x0201D;</emphasis>. The new digital paradigm demands continuous attitude to learn and share knowledge (<link linkend="F8-11">Figure <xref linkend="F8-11" remap="8.11"/></link>).</para>
<fig id="F8-11" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-11">Figure <xref linkend="F8-11" remap="8.11"/></link></label>
<caption><para>Dimension of managing change in digital transformation.</para></caption>
<graphic xlink:href="graphics/ch08_fig011.jpg"/>
</fig>
<fig id="F8-12" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-12">Figure <xref linkend="F8-12" remap="8.12"/></link></label>
<caption><para>Essential characteristics of an Agile operating model.</para></caption>
<graphic xlink:href="graphics/ch08_fig012.jpg"/>
</fig>
<para>According to Capgemini consulting (Capgemini Consulting, 2014), &#x0201C;Agility&#x0201D; (<link linkend="F8-12">Figure <xref linkend="F8-12" remap="8.12"/></link>) in manufacturing can be defined as a company<emphasis>&#x02019;</emphasis>s ability to thrive in a competitive environment of continuous and unanticipated change. Customers will demand constant changes and technological innovations that have arisen need agile organizations in order to gain competitive advantage. In short, the only constant is change.</para>
</section>
<section class="lev3" id="sec8-2-1-7">
<title>8.2.1.7 Technology</title>
<para>Using software terminology, it is being said that anyone slower than you is over-engineering, while anyone faster is introducing the above mentioned technical debt. Balance is not easy and building robust, scalable, user-friendly and smart prediction systems is not a trivial task.</para>
<para>Pivotal digital technologies are greatly affecting manufacturers, so organizations must master these technology ecosystems. Below, some technologies that are transforming the industry.</para>
</section>
<section class="lev3" id="sec8-2-1-8">
<title>Cloud Computing</title>
<para>Cloud-based platforms are on-demand infrastructures that provide shared processing resources and data to computers and other devices. Business leaders like Amazon and Microsoft offer -compelling- public clouds that guarantee scalability and fast-and-easy deployments of applications, with less maintenance in a &#x0201C;pay-as-you-go&#x0201D; model.</para>
<para>There are also many software platforms (Open or Proprietary) for creating public or private clouds that can be run on a set of privately owned servers. As a title of example, one of the most growing Open Source solutions is Openstack.</para>
</section>
<section class="lev3" id="sec8-2-1-9">
<title>Advanced analytics</title>
<para>These are techniques and tools that simulate the future, predict possible outcomes and understand data with the help of powerful dashboards, charts and graphs.</para>
<para>This is a hot topic due to the emergence of data-mining and AI techniques that have got huge popularity in the software industry in recent years.</para>
<para>The basic idea is to transform raw data structured (easily entered, stored, queried and analyzed) or unstructured (refers to content without underlying predefined model that are difficult to understand, whose knowledge to represent, using traditional approaches) into meaningful and useful information, and this information into knowledge that helps in decision making. Currently, sophisticated software systems that are being developed have human-like &#x0201C;memory&#x0201D; so they can time travel to the past and simulate (and predict) the future.</para>
<para>User-friendly dashboards that display the mentioned analytics in any device (smartphone, tablet, desktop) and smart graphs and charts are also very popular for monitoring.</para>
<para>In summary, advanced analytics provide insights and actionable events to improve operational efficiencies, extend asset life and reduce costs.</para>
</section>
<section class="lev3" id="sec8-2-1-10">
<title>Big data</title>
<para>Massive data sets that traditional databases and software applications are unable to store, process and analyze. Gartner&#x02019;s Magic Quadrant for Operational Database Management Systems used to be dominated by SAP, Oracle, IBM and Microsoft with their traditional databases. In contrast, we now see Amazon&#x02019;s DynamoDB, MongoDB and many more NoSQL companies in that report.</para>
<para>Apache Spark is the Big Data platform of choice for many enterprises as it provides dramatically increased data processing speed compared to competitors, and currently is the largest Big Data open source project.</para>
</section>
<section class="lev3" id="sec8-2-1-11">
<title>IoT</title>
<para>IoT is basically connecting devices with the Internet and/or with each other.</para>
<para>Gartner predicted that by 2020 there were going to be over 26 billion connected devices [GARTNER, 2013]. The IoT is a giant network of connected &#x0201C;things&#x0201D;: people-people, people-things, and things-things.</para>
</section>
<section class="lev3" id="sec8-2-1-12">
<title>Sensors</title>
<para>Devices that respond to physical stimuli (such as light, heat, pressure, magnetism, sound, etc.) and transmit a resulting impulse as a measurement or control. The corresponding output detect conditions that may affect the functioning of the machine and send the data to applications or clouds.</para>
</section>
<section class="lev3" id="sec8-2-1-13">
<title>Message broker</title>
<para>Responsible for taking input data or &#x0201C;messages&#x0201D; from physical or application layers, it performs some actions (transformations, aggregations, etc.) before sending the messages to destination (i.e.,: a database).</para>
<para>Thanks to message brokers, it is possible to integrate applications without enforcing a common interface and each application can also initiate interactions with other systems.</para>
<para>One very popular and heavily used open source library is Apache Kafka (<link linkend="F8-13">Figure <xref linkend="F8-13" remap="8.13"/></link>).</para>
<fig id="F8-13" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-13">Figure <xref linkend="F8-13" remap="8.13"/></link></label>
<caption><para>Message broker Apache Kakfka, by LinkedIn.</para></caption>
<graphic xlink:href="graphics/ch08_fig013.jpg"/>
</fig>
</section>
<section class="lev3" id="sec8-2-1-14">
<title>Legacy systems</title>
<para>Existing technologies, applications and databases that companies have been using (and want to keep) need connecting (as data source, reporting, etc.) to new advanced solutions.</para>
<para>The integrations with these legacy systems are important so great efforts have to be done to achieve smooth interoperability. Ideally integrations are done using web services (of REST type, for example) but due to old or proprietary technologies this is not always the case.</para>
</section>
</section>
</section>
<section class="lev1" id="sec8-3">
<title>8.3 Proactive Maintenance Business Model</title>
<para>For PM to bring to the next level its contribution to profits, productivity, and quality, it must be recognized as an integral part of the business strategy, taking into account the clients necessities as the production targets where the whole business model is affected.</para>
<para>The main advantages of the PM business model are explained in the present chapter, based on the structure provided by the nine dimensions of the canvas (see <link linkend="T8-2">Tables <xref linkend="T8-2" remap="8.2"/></link>&#x02013;<link linkend="T8-4"><xref linkend="T8-4" remap="8.4"/></link>). As a general overview, four of the nine dimensions are analyzed: value proposition, key activities, customer relationship and revenue streams. The five missing dimensions depend on the internal specifications of the companies related to the use case. For example, in some of the analyzed use cases, they were Key Partners, Key Resources, Customer Segments, Marketing Channels, and Cost Structure.</para>
<table-wrap position="float" id="T8-2">
<label><link linkend="T8-2">Table <xref linkend="T8-2" remap="8.2"/></link></label>
<caption><para>Asset Manufacturer Canvas&#x02019; key dimensions</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="cols">
<thead>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="center" colspan="4">ASSET MANUFACTURER</td>
</tr>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">Value Proposition</td>
<td valign="top" align="left">Key Activities</td>
<td valign="top" align="left">Customer Relationships</td>
<td valign="top" align="left">Revenue Streams</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left"><para><emphasis><emphasis role="underline">Asset design &#x00026; production</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Increase equipment lifespan</para></listitem> <listitem><para>Increase operational efficiency</para></listitem> <listitem><para>Improve quality</para></listitem> <listitem><para>Improve product lifecycle Improve and optimization of product design</para></listitem>
</itemizedlist><para><emphasis><emphasis role="underline">Innovation</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>New product introduction</para></listitem> <listitem><para>R&#x00026;D process optimization</para></listitem>
</itemizedlist><para><emphasis><emphasis role="underline">Supply chain &#x00026; logistics</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Delivery time optimization</para></listitem>
</itemizedlist><para><emphasis><emphasis role="underline">Customer experience</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Improvement of accuracy of warranty modelling</para></listitem> <listitem><para>Production output customization</para></listitem>
</itemizedlist></td>
<td valign="top" align="left"><para><emphasis><emphasis role="underline">Asset design &#x00026; production</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Asset reliability analysis</para></listitem> <listitem><para>Advanced HMI</para></listitem>
</itemizedlist><para><emphasis><emphasis role="underline">Innovation</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Analyse remote product performance</para></listitem>
</itemizedlist><para><emphasis><emphasis role="underline">Supply chain &#x00026; logistics</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Stock management</para></listitem>
</itemizedlist><para><emphasis><emphasis role="underline">Accounting</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Maintenance &#x00026; warranty limits</para></listitem>
</itemizedlist></td>
<td valign="top" align="left"><para><emphasis><emphasis role="underline">Asset design &#x00026; production</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Tap into customer base</para></listitem> <listitem><para>Machine virtualisation</para></listitem>
</itemizedlist><para><emphasis><emphasis role="underline">Innovation</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Customer connected products</para></listitem> <listitem><para>Integrate real-time customer feedback</para></listitem>
</itemizedlist><para><emphasis><emphasis role="underline">Supply chain &#x00026; logistics</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Increased customer satisfaction</para></listitem>
</itemizedlist><para><emphasis><emphasis role="underline">Customer experience</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Production/CRM integration</para></listitem> <listitem><para>Downtimes reduction</para></listitem> <listitem><para>Improved customization</para></listitem>
</itemizedlist><para><emphasis><emphasis role="underline">Accounting</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Tap into customer base</para></listitem> <listitem><para>New payment models which transform capex into opex for asset end-users</para></listitem> <listitem><para>Financial services</para></listitem> <listitem><para>Retain customer</para></listitem> <listitem><para>Gain new customers</para></listitem>
</itemizedlist></td>
<td valign="top" align="left"><para><emphasis><emphasis role="underline">Asset design &#x00026; production</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Idle time reduction</para></listitem> <listitem><para>Start-up time reduction</para></listitem>
</itemizedlist><para><emphasis><emphasis role="underline">Innovation</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Accelerate time to market</para></listitem>
</itemizedlist><para><emphasis><emphasis role="underline">Supply chain &#x00026; logistics</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Workpiece traceability</para></listitem>
</itemizedlist><para><emphasis><emphasis role="underline">Accounting</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Service business models</para></listitem> <listitem><para>Perpetuation of revenue streams instead of one-off asset sale for suppliers</para></listitem> <listitem><para>Customer&#x02019;s financial challenges overcome</para></listitem>
</itemizedlist></td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T8-3">
<label><link linkend="T8-3">Table <xref linkend="T8-3" remap="8.3"/></link></label>
<caption><para>Asset Service Provider Canvas&#x02019; key dimensions</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="cols">
<thead>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="center" colspan="4">ASSET SERVICE PROVIDER</td>
</tr>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">Value Proposition</td>
<td valign="top" align="left">Key Activities</td>
<td valign="top" align="left">Customer Relationships</td>
<td valign="top" align="left">Revenue Streams</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left"><para><emphasis><emphasis role="underline">Condition Monitoring</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Downtime reduction</para></listitem>
<listitem><para>Increase equipment lifespan</para></listitem>
<listitem><para>Operational Efficiency Increase</para></listitem>
<listitem><para>Quality Improvement</para></listitem>
<listitem><para>Maintenance cost reduction (for end-user)</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Asset retrofitting</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Increase equipment lifespan</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Other data-based services</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Add-on services for primary products (e.g., consulting on best usage of products)</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Supply chain &#x00026; logistics</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Delivery time optimization Spare part traceability</para></listitem>
</itemizedlist></td>
<td valign="top" align="left"><para><emphasis><emphasis role="underline">Condition Monitoring</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Asset failure demand prediction</para></listitem>
<listitem><para>Alarm triggering</para></listitem>
<listitem><para>Advanced reports</para></listitem>
<listitem><para>Remote monitoring</para></listitem>
<listitem><para>Assets usage analysis</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Other data-based services</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Advance training solutions</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Supply chain &#x00026; logistics</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Spare parts supply management</para></listitem>
<listitem><para>Connected supply chain network</para></listitem>
<listitem><para>Stock management</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Service operations</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Service (&#x00026; labour) planning &#x00026; scheduling</para></listitem>
<listitem><para>Business KPI definition and monitoring</para></listitem>
</itemizedlist></td>
<td valign="top" align="left"><para><emphasis><emphasis role="underline">Asset retrofitting</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Customised and systematic sales actions Customer inks consolidation</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Supply chain &#x00026; logistics</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Customer satisfaction degree increase</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Service operations</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Tap into customer base</para></listitem>
</itemizedlist></td>
<td valign="top" align="left"><para><emphasis><emphasis role="underline">Other data-based services</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Additional revenue streams</para></listitem>
<listitem><para>Potential sales argument for future offers</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Service operations</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Maintenance operation cost reduction</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Accounting</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Service business models</para></listitem>
<listitem><para>Customer&#x02019;s financial challenges overcome</para></listitem>
</itemizedlist></td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T8-4">
<label><link linkend="T8-4">Table <xref linkend="T8-4" remap="8.4"/></link></label>
<caption><para>Asset Service Provider Canvas&#x02019; key dimensions</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="cols">
<thead>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="center" colspan="4">ASSET SERVICE PROVIDER</td>
</tr>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">Value Proposition</td>
<td valign="top" align="left">Key Activities</td>
<td valign="top" align="left">Customer Relationships</td>
<td valign="top" align="left">Revenue Streams</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left"><para><emphasis><emphasis role="underline">Asset utilization</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Equipment lifespan increase</para></listitem>
<listitem><para>Maintenance cost reduction</para></listitem>
<listitem><para>Time to process orders reduction</para></listitem>
<listitem><para>Production costs reduction</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Employee Productivity</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Efficient training through collaborative solutions</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Sustainability</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Energy consumption reduction</para></listitem>
<listitem><para>Quality improvement</para></listitem>
<listitem><para>Operational efficiency increase</para></listitem>
<listitem><para>Rework &#x00026; scrap reduction</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Supply Chain &#x00026; Logistics</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Accident reduction Stock reduction</para></listitem>
</itemizedlist></td>
<td valign="top" align="left"><para><emphasis><emphasis role="underline">Asset utilization</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Remote monitoring</para></listitem>
<listitem><para>Connected factory</para></listitem>
<listitem><para>Downtime management</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Employee Productivity</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Training management</para></listitem>
<listitem><para>Advanced HMI</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Sustainability</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Energy management</para></listitem>
<listitem><para>Quality control</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Supply Chain &#x00026; Logistics</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Factory safety management</para></listitem>
<listitem><para>Connected supply chain</para></listitem>
</itemizedlist></td>
<td valign="top" align="left"><para><emphasis><emphasis role="underline">Employee Productivity</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Assistance systems</para></listitem>
<listitem><para>Staffing readiness</para></listitem>
<listitem><para>Worker mobility increase</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Supply Chain &#x00026; Logistics</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Product &#x00026; process traceability</para></listitem>
</itemizedlist></td>
<td valign="top" align="left"><para><emphasis><emphasis role="underline">Supply Chain &#x00026; Logistics</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Factory safety management</para></listitem>
<listitem><para>Connected supply chain</para></listitem>
<listitem><para>Product &#x00026; process traceability</para></listitem>
</itemizedlist>
<para><emphasis><emphasis role="underline">Supply Chain &#x00026; Logistics</emphasis></emphasis></para>
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Indirect monetization of insights from collected data</para></listitem>
</itemizedlist></td>
</tr>
</tbody>
</table>
</table-wrap>
<para>The targeted proactive maintenance in the MANTIS project takes into account assets such as manufacturing, wind energy production, special purpose vehicles, health imaging systems, etc. The eleven use-cases considered in the MANTIS projects, which comprise the 9 use cases described in <link linkend="ch07">Chapter <xref linkend="ch07" remap="7"/></link>, can be divided and classified into three main categories:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Asset manufacturers:</emphasis> machinery/equipment manufacturer, in many sectors being the same entities providing maintenance services and supplying spare parts;</para></listitem>
<listitem>
<para><emphasis role="strong">Asset service providers:</emphasis> entity in charge of building services upon production assets. Due to the transversality of data-based condition monitoring, often these entities cope with assets from very diverse vendors;</para></listitem>
<listitem>
<para><emphasis role="strong">Asset end users:</emphasis> exploiting assets capabilities. They have been traditionally the direct customers of the after-sales services. However, many of end users have adopted almost all preventive maintenance activities (even implementing some predictive features) of the equipment.</para></listitem>
</itemizedlist>
<para>The competitive advantage of applying the PM process to each of the value chain segments are analyzed below.</para>
<section class="lev2" id="sec8-3-1">
<title>8.3.1 Competitive Advantage for Asset Manufacturers</title>
<para>Providing advanced services upon asset maintenance has been identified by the VDMA (Mechanical Engineering Industry Association of Germany) as one of the most relevant strategic themes for European asset manufacturers growth, impacting on their digitization process and organizational structure. According to their report, companies expect to grow sales from digital business models from about 3% (2015) to 10% in 2020. Moreover, the development of advanced after sales services (such as predictive maintenance) has a huge growth potential: the previously named report states that it may impact on business profitability in a range from 1 to 4% (the average business profitability EBITDA margin is 6% within the European asset manufacturing industry). This is especially relevant for asset manufacturing SMEs, whose profitability barely reaches the average of 6%. Considering all factors, the implementation of PMM technologies will generate additional revenue sources (20% of revenue coming from services) as well as increase asset manufacturers business profitability (1 to 4% of EBITDA margin).</para>
</section>
<section class="lev2" id="sec8-3-2">
<title>8.3.2 Competitive Advantage for Asset Service Providers</title>
<para>Even though this role has been traditionally played by asset manufacturers, within the transition towards a digitized European industry, ICT oriented companies have entered the market to answer the data management and analytics technological challenge. This kind of entities operates either in collaboration with the asset manufacturer or directly with the production asset end-user. According to the report Manufacturing Analytics Market, this market is estimated to grow from 3 billion (2016) to 8,1 billion by 2021, at a growth rate of 21,9%. Overcoming one of the biggest challenges, MANTIS will link the OT &#x00026; IT worlds, by putting together production asset behaviour (mainly mechanical engineering knowledge) with failure data analysis knowledge, hence enabling data-driven maintenance business models. Tackling this problematic from a holistic point of view, MANTIS technologies will increase agility to respond to heterogeneous industrial shopfloors, enable cooperation with asset manufacturers to complement their value proposition, and provide better service to the production asset owners through adjacent predictive maintenance service.</para>
</section>
<section class="lev2" id="sec8-3-3">
<title>8.3.3 Competitive Advantage for Asset End Users</title>
<para>As stated by the Manufacturer IT Applications Study conducted by Industry Week, manufacturers in average have to deal with up to 800 hours of downtime annually. Besides, 30% of the facilities experienced incidents within the first four months of 2013. The end users are willing to offer their facilities to test the innovative framework, which could give response to their specific needs. Besides, they will be able to replicate the experience in other production facilities. This will enhance MANTIS replication in a wide range of industries, coping with different production assets, MES, ERPs, CMMS, etc.</para>
</section>
<section class="lev2" id="sec8-3-4">
<title>8.3.4 Value Chains</title>
<para>Depending on the business model archetype, it can be identified a series of value chains that can access or implement a PM business model on three stages of a company path: growth, digitization and organizational change, with reference to the framework defined by the business model archetypes designed by the VDMA in their [McKinsey&#x00026;Company, and VDMA, 2016], and explained in <link linkend="T8-5">Table <xref linkend="T8-5" remap="8.5"/></link>.</para>
<table-wrap position="float" id="T8-5">
<label><link linkend="T8-5">Table <xref linkend="T8-5" remap="8.5"/></link></label>
<caption><para>Strategic Business Models Archetypes [McKinsey&#x00026;Company, and VDMA, 2016]</para></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/ch08tab5.jpg"/>
</table-wrap>
</section>
<section class="lev2" id="sec8-3-5">
<title>8.3.5 Main Technological and Non-technological Barriers/Obstacles for the Implementation</title>
<para>Proactive maintenance implements safety in processes as the key element, needing a safety environment for every associated object and procedure involved. A function that is designed and developed to maintain safety is dependent on frequent maintenance in order to preserve its functionality and capability.</para>
<para>Barriers are often defined as an obstacle or function to prevent any form of risk to penetrate at an unwanted situation or process. The present situation in the industry indicates deviations in the common understanding and usage of barriers, and there are several areas of potential improvement [Moen, 2014].</para>
<para>For the Mantis project, several technological and non-technological barriers were identified, and are described below.</para>
<section class="lev3" id="sec8-3-5-1">
<title>8.3.5.1 Technological barriers</title>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Trust in prognostics</emphasis> <emphasis role="strong">results.</emphasis> Predicting failure of production assets (except isolated sectors, such as some defence applications) still remains a research field. This is mainly due to the fact that asset failure modes in industrial applications depend on a wide range of factors (often not monitored), and those assets (for obvious cost reasons) in most cases cannot be run to failure. That results in a lack of trust not only by asset end-users, but also by equipment and asset manufacturers;</para></listitem>
<listitem>
<para><emphasis role="strong">Ability to support a mix of large-scale, heterogeneous assets.</emphasis> A typical production plant employs multiple types of equipment identified as critical for production. Thus, predictive maintenance systems must be able to cope with a broad, complex and heterogeneous set of assets;</para></listitem>
<listitem>
<para><emphasis role="strong">Interoperability with a wide range of production management systems (mes, erp, scada, etc.).</emphasis> Each company has attained a different status regarding industrial digitization, which has a strong impact within predictive maintenance implementation (e.g., existence of a CMMS or not);</para></listitem>
<listitem>
<para><emphasis role="strong">Ability to function under dynamic operating conditions.</emphasis> Most of currently deployed manufacturing assets tend to operate with many product references, which imply highly dynamic operating conditions;</para></listitem>
<listitem>
<para><emphasis role="strong">Relyability.</emphasis> System robustness needs to be ensured with trial and error function. Model validation requires machines running to failure, which leads to high costs;</para></listitem>
<listitem>
<para><emphasis role="strong">Connectivity.</emphasis> The distributed sensors on machinery or other production assets need good wireless communication capabilities, which usually lack in industrial environments.</para></listitem>
</itemizedlist>
</section>
<section class="lev3" id="sec8-3-5-2">
<title>8.3.5.2 Non-technological barriers</title>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Conservative maintenance management culture.</emphasis> As clear example of Industrial Internet of Things (IIoT) application, maintenance activities of manufacturing companies (identified as Operational Technology - OT organizations) are usually change and risk averse;</para></listitem>
<listitem>
<para><emphasis role="strong">Need of training.</emphasis> The implementation of predictive maintenance technologies, besides from the obvious technological challenges, directly impacts on workers (maintenance and other departments) daily tasks. Operators from different ages, abilities, experience levels, will provoke resistances to PM implementation, which can prove difficult to overcome;</para></listitem>
<listitem>
<para><emphasis role="strong">Lack of resources.</emphasis> Specialized resources, such as data scientists, are needed for the evaluation and implementation process. These resources are on high demand and in some cases cannot be found or are too expensive;</para></listitem>
<listitem>
<para><emphasis role="strong">Legal.</emphasis> In the case of providing a PM service, it is necessary to have access to the machine&#x02019;s data, which is owned by the customer.</para></listitem>
</itemizedlist>
</section>
</section>
</section>
<section class="lev1" id="sec8-4">
<title>8.4 From Business Model to Financial Projections</title>
<para>Many services have been proposed to transform product-oriented into service-oriented businesses in industrial sector as can be seen in [Neely, 2007] and [Baines et al., 2009]. Besides, to systematically design and communicate ideas for new service business models, the SBMC has been proposed by [Zolnowski, 2015]. Another approach is the framework for service and maintenance business model development in 4 levels and 6 dimensions by [Kans and Ingwald, 2016] and [Kans 2016]. One of the key characteristics of component-based business model representations, such as BMC or the SBMC, is their qualitative nature, which is very suitable for developing and characterizing the business proposal [Zott et al., 2010]. However, to justify internal funding of service business developed, an assessment of the financial projection is required.</para>
<para>The use of &#x0201C;services&#x0201D; implies a connotation inside of BMC that in future PMM Business Model is implicit. The most relevant change in industrial sector is to incorporate and use ICT in all of life-cycle of product in industrial sector. Thanks to ICT the term &#x0201C;service&#x0201D; can be used to address more concrete opportunities and options inside of BMC. In case of industry and manufacturing sector, it is possible to establish processes and options, which were surveyed by [Zolnowski et al., 2011]:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Data collected in machines and products is analyzed for machine/product-related after-sales services, to increase internal knowledge, and improve performance/design or maintenance process;</para></listitem>
<listitem>
<para>Remote services give manufacturer direct access to its machines. Furthermore, it permits repairs and adjustments without geographical constraints;</para></listitem>
<listitem>
<para>Remote services help to improve internal processes via automation and parallelization in the whole life-cycle, to improve efficiency (cost reduction), and quality;</para></listitem>
<listitem>
<para>Collection and analysis of customer data are automated and help to receive customer knowledge and perform new products and services taking into account better understanding of specific customers needs;</para></listitem>
<listitem>
<para>Remote services and personalized offers facilitate relationship and send competitors to a situation of disadvantage;</para></listitem>
<listitem>
<para>Remote services improve performance maintenance. Therefore, maintenance costs and time requirements decrease;</para></listitem>
<listitem>
<para>Portfolio comprises entire chain, including plant layout, construction as general contractor, integration of inventory software, and after-sales services;</para></listitem>
<listitem>
<para>After-sales services can be improved by leveraging remote services e.g., plant monitoring, operator support, and software management;</para></listitem>
<listitem>
<para>Remote services allow for the development and provisioning of innovative services.</para></listitem>
</itemizedlist>
<para>All of the options listed above, to a greater or lesser extent, are analyzed in the MANTIS PMM Business Model and the consequence in the maintenance process can be summarized as per BMC [Zolnowski et al., 2011].</para>
<table-wrap position="float" id="T8-6">
<label><link linkend="T8-6">Table <xref linkend="T8-6" remap="8.6"/></link></label>
<caption><para>Impact of remote service technology on service business models in manufacturing [Zolnowski et al., 2011]</para></caption>
<table cellspacing="5" cellpadding="5" frame="box" rules="all">
<thead>
<tr>
<td align="left" valign="middle">Key Partners</td>
<td align="left" valign="middle">Key Activities<?lb?>Key Resources</td>
<td align="left" valign="middle" colspan="2">Value Proposition</td>
<td align="left" valign="middle">Customer Relationships Channels</td>
<td align="left" valign="middle">Customer Segments</td></tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Helps reducing the number of service orders placed with the partners of a company</para></listitem> <listitem><para>Requires the integration of new partners</para></listitem> <listitem><para>Collaboration with key partners allow the configuration of new advanced services</para></listitem>
</itemizedlist></td>
<td valign="top" align="left">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Increases process automation of a company</para></listitem> <listitem><para>Improve the product development process</para></listitem> <listitem><para>New research activities</para></listitem> <listitem><para>Increase the quality of the products and services</para></listitem> <listitem><para>Requires IT integration between shop floor, office and life cycle</para></listitem> <listitem><para>Increase efficacy and performance indicators</para></listitem> <listitem><para>Increase the level of skills to advanced tasks</para></listitem>
</itemizedlist></td>
<td valign="top" align="left" colspan="2">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Facilitates the introduction to the market</para></listitem> <listitem><para>Accumulates valuable information to offer better products and services</para></listitem> <listitem><para>Strengthen existing customer collaboration to have more knowledge and enable proactive service</para></listitem>
</itemizedlist></td>
<td valign="top" align="left">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Helps to intensity the direct contact to the customer</para></listitem> <listitem><para>New relationships through software or IT</para></listitem>
</itemizedlist></td>
<td valign="top" align="left">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Helps to increase satisfaction</para></listitem> <listitem><para>Increases the use of advanced technology in specific area</para></listitem> <listitem><para>Advanced Services can help to initiate new relationships to concrete possible customers</para></listitem>
</itemizedlist></td>
</tr>
<tr>
<td valign="top" align="left" colspan="3">Cost Structure</td>
<td valign="top" align="left" colspan="3">Revenue Streams</td>
</tr>
<tr>
<td valign="top" align="left" colspan="3">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Reducemaintenance and services costs</para></listitem> <listitem><para>New costs related to advanced services</para></listitem>
</itemizedlist></td>
<td valign="top" align="left" colspan="3">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Increase the revenues for new products and services</para></listitem> <listitem><para>New prices strategic for services</para></listitem> <listitem><para>New revenue streams due to new services make the company less dependent on the unpredictability of product sale and even out the cashflow.</para></listitem>
</itemizedlist></td>
</tr>
</tbody>
</table>
</table-wrap>
<para>Introducing a service business model directly affect economic evaluation and projection of a company, as seen in the <link linkend="T8-6">Table <xref linkend="T8-6" remap="8.6"/></link>. Moreover, in the characterization of SBMC it is necessary to not only take into account traditional cost structure in manufacturing but also consider the &#x0201C;service costs&#x0201D; to apply to the service model. As a consequence, investment on ICT such as specialized software and hardware are mandatory to allow for both savings through process improvements and revenues by additional services offered to customers. This leads to the first step to be taken in order to develop and quantify the potential impact on a business:</para>
<blockquote>
<para>To make the decision if to invest in a project, a Cost-Benefit-Analysis (CBA) has to be performed [Zolnowski et al., 2017; Boardman et al., 2017]</para>
</blockquote>
<para>A CBA is an established tool for evaluating the economic benefit of an investment. As such, it can point out whether a SBMC could be implemented or by the contrary a different BMC definition is required. Although a service business model can be complex because of affected factors, systematic capture and analysis of CBA-related nuisances is a desirable objective. The next step for companies is to make the effort to evolve from a qualitative perspective of component-based business model representations to a quantitative information enriching and completing the business plan cycle.</para>
<para>There are different models to implement a CBA; [Anke and Krenge, 2016] describes a method called meta-model for &#x0201C;Smart Services&#x0201D;, which was proposed for assessing data-driven services for connected products. The business case for smart services is associated with previous modelling processes. In their work, &#x0201C;smart services&#x0201D; are related to the process of digitalization and monitoring, leading to the result of a product/service with &#x0201C;connected smart connectivity&#x0201D;.</para>
<para>While the meta-model of [Anke and Krenge, 2016] is not directly related to a business model, it provides an interrelation between the service proposal and its financial evaluation. <link linkend="F8-14">Figure <xref linkend="F8-14" remap="8.14"/></link> shows the business model diagram and financial diagram.</para>
<fig id="F8-14" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-14">Figure <xref linkend="F8-14" remap="8.14"/></link></label>
<caption><para>Service and cost class diagram for the business case model [Anke and Krenge, 2016].</para></caption>
<graphic xlink:href="graphics/ch08_fig015.jpg"/>
</fig>
<para>[de Jesus, 2012] explaines how to proceed from BMCs to financial projection. Since in a company the earnings and revenues projections play a crucial role, the most relevant analysis aims to understand:</para>
<para><emphasis>&#x0201C;if the future company will be successful or not&#x0201D;</emphasis>.</para>
<para>Hence, it proposes to merge the BMCs framework with P&#x00026;L concepts. The P&#x00026;L establishes in a specified period of time the results of the business performance and is able to specify the Cost Structure and Revenues Streams building blocks of the canvas in a more tangible way.</para>
<para>P&#x00026;L is composed by the following items [de Jesus, 2012]:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Cost of sales:</emphasis> expected direct cost of developing and performing the product and/or service that a company carries out;</para></listitem>
<listitem>
<para><emphasis role="strong">Gross margin:</emphasis> revenues less cost of sales. This is a key indicator to know the earnings of the company;</para></listitem>
<listitem>
<para><emphasis role="strong">Operating expenses:</emphasis> expenses the company needs to operate and manufacture;</para></listitem>
<listitem>
<para><emphasis role="strong">Operating earnings/loss:</emphasis> gross margin minus operating expenses. This indicator is closer to reality;</para></listitem>
<listitem>
<para><emphasis role="strong">Income taxes:</emphasis> taxes that a company needs to pay. This item depends on which country the company establishes their activity in.</para></listitem>
</itemizedlist>
<para>The principal objective seeks that the revenue streams and the cost structure information are directly identified from the others seven blocks of the BMCs apart from specified cost and revenues blocks. In order to achieve it, [de Jesus, 2012] estimates the economic quantification of each element added in BMCs in terms of costs or revenues.</para>
<para>Finally, to complete the financial projections to the P&#x00026;L analysis, the calculations of financial indicators could be provided as an overview of the business. The calculated indicators are the following ones [de Jesus, 2012]:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">EBITDA:</emphasis> gives an overall idea about the potential capacity of the business to generate cash. It is a measure of a company&#x02019;s operating performance;</para></listitem>
<listitem>
<para><emphasis role="strong">Break even point:</emphasis> it is the point from which the company begins to have positive operating results. Typical question to address this indicator is, how much needs to be sold in order to cover all fixed costs?;</para></listitem>
<listitem>
<para><emphasis role="strong">Quantity critical point:</emphasis> represents the minimum amount that the firm should sell in order to have positive results;</para></listitem>
<listitem>
<para><emphasis role="strong">Security margin:</emphasis> expresses the distance on the level of activity achieved by the company for the critical point;</para></listitem>
<listitem>
<para><emphasis role="strong">Sales cost coefficient:</emphasis> gives part of sales after that the variable cost is paid, which will be left to pay for fixed costs.</para></listitem>
</itemizedlist>
<para>Aiming at supporting a better understanding by means of examples, in [de Jesus, 2012] there are 6 demonstrations of 6 different BMCs. For example, Portal dos Servi&#x000E7;os is a startup that aims to be a trusty intermediary between health care service providers and private clients. <link linkend="T8-7">Table <xref linkend="T8-7" remap="8.7"/></link> shows P&#x00026;L in BMCs.</para>
<table-wrap position="float" id="T8-7">
<label><link linkend="T8-7">Table <xref linkend="T8-7" remap="8.7"/></link></label>
<caption><para>Portal de Servi&#x000E7;os cost Business Model [de Jesus, 2012]</para></caption>
<table cellspacing="5" cellpadding="5" frame="box" rules="all">
<thead>
<tr>
<td align="left" valign="middle">Key Partners</td>
<td align="left" valign="middle">Key Activities<?lb?>Key Resources</td>
<td align="left" valign="middle" colspan="2">Value Proposition</td>
<td align="left" valign="middle">Customer Relationships Channels</td>
<td align="left" valign="middle">Customer Segments</td></tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Web providers</para></listitem> <listitem><para>Advertising companies 5.000</para></listitem>
</itemizedlist></td>
<td valign="top" align="left">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Platform Analysis and Design 40.000</para></listitem> <listitem><para>Specific and advanced profiles 20.000</para></listitem> <listitem><para>Logistics 30.000</para></listitem>
</itemizedlist></td>
<td valign="top" align="left" colspan="2">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Platform Development 70.000</para></listitem>
</itemizedlist></td>
<td valign="top" align="left">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Events 4.000</para></listitem> <listitem><para>Social Media 10.000</para></listitem> <listitem><para>Service center</para></listitem> <listitem><para>Press 2.000</para></listitem> <listitem><para>CRM 10.000</para></listitem>
</itemizedlist></td>
<td valign="top" align="left">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Private and liberal professional 3000</para></listitem>
</itemizedlist></td>
</tr>
<tr>
<td valign="top" align="left" colspan="3">Cost Structure</td>
<td valign="top" align="left" colspan="3">Revenue Streams</td>
</tr>
<tr>
<td valign="top" align="left" colspan="3">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Platform maintenance and development</para></listitem>
</itemizedlist></td>
<td valign="top" align="left" colspan="3">
<itemizedlist mark="bullet" spacing="normal"><listitem><para>Product sells 300.000</para></listitem> <listitem><para>Services sels 80.000</para></listitem>
</itemizedlist></td>
</tr>
</tbody>
</table>
</table-wrap>
<para>All of the parameters discussed above, namely EBITDA, break event point, quantity critical point, security margin and sales, cost coefficient, evolution and critical point, are shown in <link linkend="F8-15">Figure <xref linkend="F8-15" remap="8.15"/></link>:</para>
<fig id="F8-15" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-15">Figure <xref linkend="F8-15" remap="8.15"/></link></label>
<caption><para>Portal de Servi&#x000E7;os financial analysis [de Jesus, 2012].</para></caption>
<graphic xlink:href="graphics/ch08_fig017.jpg"/>
</fig>
<para>Wurzbach (2000) performs a web based cost benefit analysis method for predictive maintenance through two principal points:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Direct cost savings;</para></listitem>
<listitem>
<para>Indirect cost savings.</para></listitem>
</itemizedlist>
<para><emphasis role="strong">Direct cost savings</emphasis> are the principal group of costs to get benefit within a predictive maintenance program. These costs are recurring and usually established as annual direct cost reduction. Sometimes it can be new budgeted design and changes depending on information generated by the predictive maintenance program. Overall, these are one-time reductions but the savings can be relevant.</para>
<para><emphasis role="strong">Indirect costs savings</emphasis> affect manufacturing and maintenance process but with ups and downs. Since productive lines and products can have different values of OEMs, indirect costs savings could be increased if failures are detected before development ends. Even in some cases with less differences between predicted and an unanticipated failure there could be significant savings in maintenance within the scheduling environment. Indeed, advanced maintenance program get the most benefit due to associated costs of unanticipated or &#x0201C;emergent&#x0201D; maintenance repair activities. Finally, advanced maintenance program provides overtime and parts procurement reductions as well as impact to facility operations as indirect cost savings.</para>
<para>Zolnowski et al. (2017) explains the principal items to take into account inside of:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Cost;</para></listitem>
<listitem>
<para>Revenues;</para></listitem>
<listitem>
<para>Cost Savings from BMCs to financial projections.</para></listitem>
</itemizedlist>
<para>In case of Costs, [Zolnowski et al., 2017] references a variety of costs that are directly related to the development and management of a Data-Driven Business Model (DDBM). The costs are associated with the implementation of data-as-a-service or analytics-as-a-service technology and as a consequence close to the development of future PMM Business Model, some of the possible costs are listed below:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Monitoring and analysis of data system;</para></listitem>
<listitem>
<para>Extension of sensors in the key activities and use and allocation of data and systems;</para></listitem>
<listitem>
<para>Sensors, gadgets, data, and systems;</para></listitem>
<listitem>
<para>Use of data and systems in the key resources.</para></listitem>
</itemizedlist>
<para>A development phase is referenced, where the improvement or implementation of hardware infrastructure is an important cost factor as well as an industrial use case in MANTIS project. In manufacturing, customers have to implement:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Hardware such as embedded system and sensors in their machines to collect data;</para></listitem>
<listitem>
<para>Connectivity between customers machines and cloud servers.</para></listitem>
</itemizedlist>
<para>Another factor to reference cost is Software. Software can be developed, purchased or leased in ad-hoc, on-demand or Software as a service way. Any case individualized and specified algorithms are needed for companies. Moreover, these algorithms need to be maintained as well as the infrastructure to give continuity of Business Model.</para>
<blockquote><para>Hence, [Zolnowski et al., 2017] highlights the items of infrastructure (sensor and systems), software (algorithms and processing, monitoring viewing data) and connectivity input as Costs.</para></blockquote>
<para>In case of revenues, these can be enabled for any actor. From a company perspective, revenues can be generated as [Zolnowski et al., 2017] highlights:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Sales of new services;</para></listitem>
<listitem>
<para>Sales of data to third parties;</para></listitem>
<listitem>
<para>Sales of machine because of more competitive product.</para></listitem>
</itemizedlist>
<para>Some of the types of revenues analysed in the MANTIS use-cases highlight the sales of a new services and sales of the machines as the most representative returns.</para>
<para>While [Wurzbach, 2000] describes indirect costs, [Zolnowski et al., 2017] describes the importance of savings, especially for optimization of processes that lead to lower costs by the reduction of inventory (key resources) failures, personnel, and goods. These costs of operational processes are optimized for analysing data remotely, or to be replaced by automatized processes.</para>
</section>
<section class="lev1" id="sec8-5">
<title>8.5 Economic Tool to Evaluate Current and Future PMM Business Model</title>
<para>Once it has been identified the most important characteristics from &#x0201C;Business Model&#x0201D; to &#x0201C;Financial Projections&#x0201D;, the next step is to know the earnings, incomes or revenue streams and costs.</para>
<para>The current Business Model shows the actual business and results of a company performances during the last year with traditional activities in maintenance. On the other hand, future PMM Business Model shows the number of projections over the following five years with advanced services in maintenance.</para>
<para>At this point, the key question to ask is:</para>
<para><emphasis role="strong">Which are the items that you have to take into account in earnings, incomes or revenue streams and costs?</emphasis></para>
<para>[Burke, 2017; Pauceanu, 2016] or [Tanev et al., 2016] describes the principal elements to take into account related with earning, incomes or revenues streams and costs. In order to have a specific answer to the question above, the future business model must be considered as composed by two principal sections: <emphasis role="strong">(1)</emphasis> <emphasis role="strong">The Business and (2) Financial Data.</emphasis></para>
<para><emphasis role="strong"><emphasis>In the Business</emphasis></emphasis> section, it is necessary to develop a detailed plan that clarifies why the business is existing; the aims for the existence, client groups, products and services, and how it intends to develop and deliver those services and/or products. It acts as a road map for the organization and shows clearly the destination it seeks and the path to follow to get there.</para>
<para>This should also be accompanied by a description of the resources required to complete the Business journey.</para>
<para>Hence, you would describe and take into account the following items among others (<link linkend="F8-16">Figure <xref linkend="F8-16" remap="8.16"/></link>):</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Description of business;</para></listitem>
<listitem>
<para>Product and services;</para></listitem>
<listitem>
<para>Market analysis;</para></listitem>
<listitem>
<para>Marketing plan;</para></listitem>
<listitem>
<para>Operational plan:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Production;</para></listitem>
<listitem>
<para>Location and infrastructure;</para></listitem>
<listitem>
<para>Technology;</para></listitem>
<listitem>
<para>Human resources plan;</para></listitem>
<listitem>
<para>Spare parts/inventory/suppliers;</para></listitem>
<listitem>
<para>Competition.</para></listitem>
</itemizedlist>
</listitem>
</itemizedlist>
<para>In <emphasis role="strong"><emphasis>Financial Data</emphasis></emphasis> section, projected financial statements with revenues and costs would be indicated. Moreover, cash flow, balance sheets etc. should be taken into account. The financial data section is built over the Business section, and it is a critical section to verify that the evolution of the business is viable, useful and maintainable over time.</para>
<para>Companies should present their own values in such a way that it shows that the business will grow during time and there is a successful strategy at the horizon where they can make profits. The rule of thumb is to be realistic and credible, represent the growth trajectory in an understandable and detailed manner, and to figures into components by target market segment or by sales channels and provide realistic estimates for revenue and sales.</para>
<fig id="F8-16" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-16">Figure <xref linkend="F8-16" remap="8.16"/></link></label>
<caption><para>General components of business plan [Pauceanu, 2016].</para></caption>
<graphic xlink:href="graphics/ch08_fig018.jpg"/>
</fig>
<section class="lev2" id="sec8-5-1">
<title>8.5.1 Incomes Items</title>
<para>This section refers to elements pertaining to revenue streams, earnings and benefits. More concrete, companies should analyse market, product, defined services either current and future PMM Business Model or detailed model. Moreover, it considers how to reach the market through the marketing plan. All of these elements concur to determine the incomes. So, companies should know:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Unit sales:</emphasis> this is done by determining the number of products sold yearly;</para></listitem>
<listitem>
<para><emphasis role="strong">Price:</emphasis> this is the price of products and services that are sold during a year. The first consideration in pricing a product is the value that it represents to the customer. If, on the checklist of features, the product is truly ahead of the competition, it is possible to command a premium price, but in that case, it is better to offer some extra functionalities as companies performed in future PMM Business Model.</para></listitem>
</itemizedlist>
<para>In case of doubt, the market can set the price. Anyway, there are cases when the selling price does not exceed costs and expenses by the margin necessary to maintain your business robust. This leads to the utmost necessity to know competitors pricing policies.</para>
<para>To define the revenues, the type of company is a key aspect. <emphasis role="strong">In case of industrial use-case</emphasis> companies should indicate the incomes from product or service and also maintenance. At this point, companies should differentiate between traditional maintenance defined in current Business Models and advanced maintenance defined in future PMM Business Model (PMM BM).</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>In the first case (current Business Models), companies will provide information of the last year and probably have some maintenance contract and call service to find out solutions from issues during the life of the product;</para></listitem>
<listitem>
<para>In the second case (PMM BM), companies could offer some new services along with product. These services are probably related to data maintenance. As consequence, companies could have benefits from some applications that use this data such as desktop or mobile APP, dashboard HMI, Virtual or Augmented reality application, remote analysis etc. Companies can specify others if they consider. The values for the next five years are projections that could be accomplished or not at the end of the fifth year.</para></listitem>
</itemizedlist>
<para>Thereby, companies have used a template in excel to provide economic information. <link linkend="T8-8">Table <xref linkend="T8-8" remap="8.8"/></link> summarizes how to provide the information.</para>
<table-wrap position="float" id="T8-8">
<label><link linkend="T8-8">Table <xref linkend="T8-8" remap="8.8"/></link></label>
<caption><para>Revenue streams template for current business model and industrial use case</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<tbody>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">REVENUE STREAMS</td>
<td valign="top" align="left">LAST YEAR</td>
</tr>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">PRODUCT</td>
<td valign="top" align="left"></td></tr>
<tr>
<td valign="top" align="left">MAINTENANCE</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Contract</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Call service</td>
<td valign="top" align="left"></td>
</tr>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Others (specify)</td>
<td valign="top" align="left"></td>
</tr>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">TOTAL</td>
<td valign="top" align="left"></td></tr>
</tbody>
</table>
</table-wrap>
<para>One more time, to provide this information industrial companies can rely on previous year profit and loss statement. Product represents the industrial asset that companies sell. In case of maintenance, table represents the principal items as contract or call service but they can specify others.</para>
<para>In an <emphasis role="strong">industrial use case and future PMM Business Model</emphasis> <link linkend="T8-9">Table <xref linkend="T8-9" remap="8.9"/></link> indicates 5 year projections and new services maintenance.</para>
<table-wrap position="float" id="T8-9">
<label><link linkend="T8-9">Table <xref linkend="T8-9" remap="8.9"/></link></label>
<caption><para>Revenue stream projections template for future PMM Business Model and industrial use case</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<tbody>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">REVENUE STREAMS PROJECTION</td>
<td valign="top" align="left">YEAR 1</td>
<td valign="top" align="left">YEAR 2</td>
<td valign="top" align="left">YEAR 3</td>
<td valign="top" align="left">YEAR 4</td>
<td valign="top" align="left">YEAR 5</td>
</tr>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">PRODUCT</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td></tr>
<tr>
<td valign="top" align="left">MAINTENANCE SERVICE</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Tele-analysis</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Mobile APP</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;VR/AR application</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Dashboard HMI</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Others (specify)</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">TOTAL</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td></tr>
</tbody>
</table>
</table-wrap>
<para>Maintenance service shows the new services that companies offer along with product. <link linkend="T8-9">Table <xref linkend="T8-9" remap="8.9"/></link> enumerates some of possible items that companies could offer. Businesses need to plan the price and the number of clients to implant. Additionally, companies describe the principal keys of the current Business Model and future PMM Business Model, such as type or product, market, maintenance, KPIs, etc., in order to explains the future projections.</para>
<para>Overall, each year sales will be higher and incomes increase that would depend on an specific percentage to be decided depending on the market, commercial interests, commercial workers skills and added value of the product.</para>
<para>Regarding <emphasis role="strong">technological providers</emphasis>, a new Business Model will offer new software either ad-hoc or SaaS model. As a consequence, companies would estimate the revenues for 5 years in terms of consultancy, product license, customized etc. A unit price for each item should be set in order to concrete a technological revenue.</para>
</section>
<section class="lev2" id="sec8-5-2">
<title>8.5.2 Cost Items</title>
<para>According to [Pauceanu, 2016] and [Featherstone, 2015] it is better to calculate costs before incomes because it is easier. Both of them agree about including <emphasis role="strong">costs related to operational and marketing plan</emphasis>.</para>
<para>The operational plan is related to design, and it controls the process of production. Companies have the responsibility of ensuring that operations and production are efficient and effective in terms of meeting customer requirements. Therefore, companies should include the cost of <emphasis role="strong"><emphasis>manpower</emphasis></emphasis> to be engaged in production. This section includes the <emphasis role="strong"><emphasis>amortization</emphasis></emphasis> of machines and <emphasis role="strong"><emphasis>spare parts</emphasis></emphasis> to build the product as well as <emphasis role="strong"><emphasis>technology</emphasis></emphasis> needs by manpower. Computers, laptops, mobiles, data center, servers, software licenses are examples about what companies can include.</para>
<para>Companies should also include <emphasis role="strong"><emphasis>Infrastructure</emphasis></emphasis> expenses such as stationaries, office utilities, office rent, power and industrial warehouse among others. To complete all of the expenses, companies should not forget <emphasis role="strong"><emphasis>marketing, travel, and accommodation</emphasis></emphasis>issues. Product or service could need its own marketing strategy. The purpose of advertising is to inform, persuade, and remind customers about your companys products.</para>
<para>Other issue is that some companies can decide in order not to invest in marketing. Travel and accommodation issues are costs to attribute when commercial workers are selling out of company office.</para>
<para>Eventually, companies should add <emphasis role="strong"><emphasis>maintenance</emphasis></emphasis> cost. The principal common items could be:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Travel expenses and accommodation;</para></listitem>
<listitem>
<para>Maintenance workforce;</para></listitem>
<listitem>
<para>Warranty repairs;</para></listitem>
<listitem>
<para>Spare parts;</para></listitem>
<listitem>
<para>Call centre.</para></listitem>
</itemizedlist>
<para>Furthermore, companies should include <emphasis role="strong">costs related to development advanced maintenance applications and associated services</emphasis> in the incomes area. <emphasis role="strong">Hardware items</emphasis> such as IoT sensors to analyse maintenance information should also be added.</para>
<para>Costs projections for the future PMM Business Model tables includes five projection years and software and hardware costs associated to advanced maintenance.</para>
<para><link linkend="T8-10">Tables <xref linkend="T8-10" remap="8.10"/></link> to <link linkend="T8-12"><xref linkend="T8-12" remap="8.12"/></link> allow to provide the description of costs for different kinds of companies focused on future PMM Business Models. The only two differences between the table of structure cost for current Business Model and the future PMM Business Model are the five year projections and hardware/software investments in maintenance.</para>
<para>In natural evolution of future PMM Business Model, the costs could be higher at least for the first or second year due to hardware and software investment. As a consequence for following years other costs associated in maintenance should decrease.</para>
<table-wrap position="float" id="T8-10">
<label><link linkend="T8-10">Table <xref linkend="T8-10" remap="8.10"/></link></label>
<caption><para>Structure cost template for technological provider business model projection</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<tbody>
<tr style="border-bottom: 0.1em solid;">
<td valign="bottom" align="left">REVENUE STREAMS PROJECTION</td>
<td valign="bottom" align="left">UNIT PRICE (Per Client/Asset)</td>
<td valign="bottom" align="left">YEAR 1</td>
<td valign="bottom" align="left">YEAR 2</td>
<td valign="bottom" align="left">YEAR 3</td>
<td valign="bottom" align="left">YEAR 4</td>
<td valign="bottom" align="left">YEAR 5</td>
</tr>
<tr>
<td valign="top" align="left">Product license</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">Consultancy</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">Customization</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">Others(specify)</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">TOTAL</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td></tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T8-11">
<label><link linkend="T8-11">Table <xref linkend="T8-11" remap="8.11"/></link></label>
<caption><para>Structure costs Template for current business model and industrial use case</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left">COST STRUCTURE</td>
<td valign="top" align="left">LAST YEAR</td></tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">MANPOWER</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Factory workers</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Engineering and design</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Commercial workers</td>
<td valign="top" align="left"></td></tr>
<tr>
<td valign="top" align="left">TECHNOLOGY</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Hardware inside the product</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Software inside the product</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Data centre, server</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;IT Support</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Computers, laptops, mobiles</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Software licensing</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">TRAVEL EXPENSES AND ACCOMODATION</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;To Sell</td>
<td valign="top" align="left"></td></tr>
<tr>
<td valign="top" align="left">INFRASTRUCTURE</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Industrial warehouse, business office</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Power, water, stationery,</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Administration</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">MARKETING</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">LOGISTICS</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Logistics and distribution</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">AMORTIZATION</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">PURCHASING</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Spare parts - products</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">MAINTENANCE</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Travel expenses and accommodation</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Spare parts</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Warranty repairs</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Maintenance workers</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Call centre</td>
<td valign="top" align="left"></td>
</tr>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Others (specify)</td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">TOTAL</td>
<td valign="top" align="left"></td></tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T8-12">
<label><link linkend="T8-12">Table <xref linkend="T8-12" remap="8.12"/></link></label>
<caption><para>Structure costs template for future PMM Business Model and industrial use case</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left">COST STRUCTURE</td>
<td valign="top" align="left">YEAR 1</td>
<td valign="top" align="left">YEAR 2</td>
<td valign="top" align="left">YEAR 3</td>
<td valign="top" align="left">YEAR 4</td>
<td valign="top" align="left">YEAR 5</td></tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">MANPOWER</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Factory workers</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Engineering and design</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Commercial workers</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td></tr>
<tr>
<td valign="top" align="left">TECHNOLOGY</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Hardware</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td></tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Software</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Data centre, server</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;IT Support</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Computers, laptops, mobiles</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Software licensing</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">TRAVEL EXPENSES AND ACCOMODATION</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;To sell</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td></tr>
<tr>
<td valign="top" align="left">INFRAESTRUCTURE</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Industrial warehouse, business office</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Power, water, stationery,</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Administration</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">MARKETING</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">LOGISTICS</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Logistics and distribution</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">AMORTIZATION</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">PURCHASING</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Spare parts - Products</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">MAINTENANCE</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">Software Development</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Sensor and hardware</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Warranty repairs</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Maintenance workers</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Spare Parts</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Travel expenses and accommodation</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;Others (specify)</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">TOTAL</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td></tr>
</tbody>
</table>
</table-wrap>
<para>This is because there is a product with additional maintenance services, especially predictive maintenance, which implies a reduction of overhead costs on it. There will probably be a decrease in repairs, spare parts, travels etc.</para>
<para>Finally, a sustainable business should have less costs than incomes. Depending on the type of business this should be true somewhere during the five years projection.</para>
<para><emphasis role="strong">Cost Structure for technological provider</emphasis> <emphasis role="strong">use case</emphasis> cannot take into account the same structure cost of the industrial provider. A typical software development company needs to have a new product:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Manpower with programmig and consultancy knowledge;</para></listitem>
<listitem>
<para>Platform maintenance costs;</para></listitem>
<listitem>
<para>Hardware as sensor to obtain data from machines;</para></listitem>
<listitem>
<para>Marketing expenses;</para></listitem>
<listitem>
<para>Other costs.</para></listitem>
</itemizedlist>
<table-wrap position="float" id="T8-13">
<label><link linkend="T8-13">Table <xref linkend="T8-13" remap="8.13"/></link></label>
<caption><para>Structure costs template for technological provider use case</para></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left">COST STRUCTURE</td>
<td valign="top" align="center">YEAR 1</td>
<td valign="top" align="center">YEAR 2</td>
<td valign="top" align="center">YEAR 3</td>
<td valign="top" align="center">YEAR 4</td>
<td valign="top" align="center">YEAR 5</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Software development</td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
</tr>
<tr>
<td valign="top" align="left">Hardware</td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
</tr>
<tr>
<td valign="top" align="left">Platform maintenance</td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
</tr>
<tr>
<td valign="top" align="left">Fixed costs</td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
</tr>
<tr style="border-bottom: 0.1em solid;">
<td valign="top" align="left">Others(specify)</td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
</tr>
<tr>
<td valign="top" align="left">TOTAL</td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td>
<td valign="top" align="center"></td></tr>
</tbody>
</table>
</table-wrap>
<para>Once companies filled the revenue streams and costs, either the current Business Model or the future PMM Business Model for an industrial use case or technological provider, it is time to address a detailed report. Next section describes the report schema to address.</para>
</section>
<section class="lev2" id="sec8-5-3">
<title>8.5.3 Schema of Economic Evaluation and Projection Report</title>
<para>Regardless of the type of company, industrial or technological provider, the schema to address is:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="strong">Introduction</emphasis></para>
<para>In this section each company will give a short description about their profile, the product and services they offer. Even, they could add some figures about annual turnover, market segmentation and objectives to address in MANTIS.</para>
<para>This section it is useful to put some key information to help understand the sequential sections;</para></listitem>
<listitem>
<para><emphasis role="strong">Value Proposition</emphasis></para>
<para>Companies have the oportunity to explain in more details their use case associated with maintenance and the MANTIS project as well as the PMM value proposition for a 5 five years projection.</para>
<para>The description of current and future value proposition incorporates new assets such as tecnology, services etc. which help to know deeply the business plan to address economic evaluation</para></listitem>
<listitem>
<para><emphasis role="strong">Economic Evaluation and Projection</emphasis></para>
<para>This is the most important and relevant section. Companies will show and describe numbers, figures, and some indicators associated with current Business Model and future PMM Business Model. Regarding current Business Model will be numbers only for one year, specially closed last year and the next 5 years in case of PMM Business Models.</para>
<para>After filling the revenue streams and costs with tables in a spreadsheet described in the previous sections, companies will summarize economic information in <link linkend="T8-14">Table <xref linkend="T8-14" remap="8.14"/></link>.</para></listitem>
</itemizedlist>
<table-wrap position="float" id="T8-14">
<label><link linkend="T8-14">Table <xref linkend="T8-14" remap="8.14"/></link></label>
<caption><para>Economic evaluation and projection template for industrial use case</para></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/ch08_Table_fig001.jpg"/>
</table-wrap>
<para>Technological provider companies will show only 5 years projections because the aim is to offer new service through software as service and consultancy. Therefore, it is a new business that companies will describe in the BMCs tool and summarize the economic information in <link linkend="T8-15">Table <xref linkend="T8-15" remap="8.15"/></link>.</para>
<table-wrap position="float" id="T8-15">
<label><link linkend="T8-15">Table <xref linkend="T8-15" remap="8.15"/></link></label>
<caption><para>Economic evaluation and projection template for technological provider use case</para></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/ch08_Table_fig002.jpg"/>
</table-wrap>
<para>Apart from the economic evaluation and description of most relevant figures and indicators (<link linkend="T8-14">Table <xref linkend="T8-14" remap="8.14"/></link>), the industrial use case companies will include a graphic to see the evolution in an easy and suitable format as can be seen in <link linkend="F8-17">Figure <xref linkend="F8-17" remap="8.17"/></link>.</para>
<fig id="F8-17" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-17">Figure <xref linkend="F8-17" remap="8.17"/></link></label>
<caption><para>Example economic evaluation projection.</para></caption>
<graphic xlink:href="graphics/ch08_fig019.jpg"/>
</fig>
</section>
</section>
<section class="lev1" id="sec8-6">
<title>8.6 Railways Use-Case Financial Business Model</title>
<para>Nowadays, the railway industry is in a position to exploit the opportunities associated to IIoT, enabling technologies under the paradigm of Industry 4.0.</para>
<para>According to [Roland Berger Study, 2017], railway sector supply, like other industries, sees four main levers driving the digital revolution as well as transforming existing business models: Interconnectivity [Fraga-Lamas et al., 2017], Digital Data [MARKETandMARKETS, 2016], Automation [Roland Berger Study, 2017] and Customer Interface.</para>
<fig id="F8-18" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-18">Figure <xref linkend="F8-18" remap="8.18"/></link></label>
<caption><para>Smart maintenance approach in Railway Sector.</para></caption>
<graphic xlink:href="graphics/ch08_fig020.jpg"/>
</fig>
<para>The efforts in railway industry are concentrated on digitization of train control and maintenance, with a focus on CBM-based PM. The sector focuses in the definition of the root cause of failures, limit view on asset health, all of them leading to taking decisions based on facts instead of the experience and gut feeling of senior technical experts. Moreover, infrastructure development and maintenance costs represent a large part of the entire rail company budget. For all, the railway sector pretends to add new services, such as integrated security, asset management, and predictive maintenance to improve timely decision-making for issues like safety, railway assets, productivity, scheduling, and system capacity.</para>
<para>For carrying out a strategy associated with smart maintenance, key enablers comprise CBM, data integration and asset management with the use of advanced technology such as IoT, Sensors, Big Data Analytics, and advanced visualization services (<link linkend="F8-18">Figure <xref linkend="F8-18" remap="8.18"/></link>).</para>
<para>PM models and approaches help to eliminate preventive and corrective interventions. The key intuition is suggesting the ideal time and measure for maintenance work using all the data. This way, diagnostic data are not only used as a control function, but also as a driver for maintenance activities shifting a paradigm and giving several advantages:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Change from a reactive maintenance mode to a proactive one;</para></listitem>
<listitem>
<para>Visibility on the health of the asset;</para></listitem>
<listitem>
<para>Delivering support to intervention teams, helping them take the right decisions (grind, stuff, mill or change a rail);</para></listitem>
<listitem>
<para>Reduce high costs by optimizing maintenance interventions;</para></listitem>
<listitem>
<para>Determining the root cause of failures;</para></listitem>
<listitem>
<para>Better usage of warranty recoveries;</para></listitem>
<listitem>
<para>Financial forecasting;</para></listitem>
<listitem>
<para>Create new services business models associated with new maintenance service approach.</para></listitem>
</itemizedlist>
<para>According with [MARKETandMARKETS, 2016] study the global smart railway market is estimated to grow from $10.50 bn of year 2016 to $20.58 bn by 2021, at a CAGR of 14.4%. In addition, according to the International Transport Forum of the Organisation for Economic Co-operation and Development (OECD), by 2050, passenger mobility will increase by 200&#x02013;300% and freight activity by as much as 150&#x02013;250% with respect to 2010 [Fraga-Lamas et al., 2017]. These figures show the impact for every component of the value chain of the railway industry.</para>
<para>Due to investing significantly into railway digitalization, according to [Roland Berger Study, 2017] 89% of the rail industry executives expect changes of their business model. The incorporation of digital competitive strategies allow for new advanced services such as digital products, remote train maintenance, mobile app, etc. Therefore, the railway industry has to adapt to new market situation.</para>
<section class="lev2" id="sec8-6-1">
<title>8.6.1 Financial Business Benefits Within a Specific Railway Maintenance Solution</title>
<para>In order to point out specific benefits in economic terms of Proactive Monitoring Maintenance, a railway use-case is used as an example. On first instance, the company has to define its short to long-term objectives. In the particular case of railways, the company wants to address a reliability test strategy, based on a mathematical and theoretical assumption, in order to validate the MTBF value (coming from predictive analysis) with a fixed confidence interval. The maintenance service cost could be optimized to save costs in work force with service intervals. On the other hand, exploitation results are mandatory to have new revenue streams in terms of advanced services.</para>
<para>To carry out these objectives the railway use-case had to change its actual maintenance business model. Generally, the typical maintenance service is reactive and the revenues are associated with maintenance contract. To implement the new strategy, the railway use-case implemented proactive and monitoring maintenance solutions as well as new advanced services. The requirements to implement were:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para><emphasis role="underline">Data acquisition:</emphasis> besides current diagnostic parameters (voltage, current, position at the end of manoeuvre and completion time of the last manoeuvre), the following key variables are monitored in addition to those mentioned above: distance between the point blades and thrust/load exerted by the point blades environmental condition (environmental temperature, humidity, temperature of the rail) and mechanical behaviour;</para></listitem>
<listitem>
<para><emphasis role="underline">Data processing:</emphasis> enabled by a failure preventive model, the system generates a set of alarms (sent to the maintenance staff) in order to avoid availability problems;</para></listitem>
<listitem>
<para><emphasis role="underline">Maintenance strategy optimization:</emphasis> based on data generated by the monitoring system, the system sends a set of recommendations to the maintenance responsible to facilitate maintenance operations and reduce down-times of the whole interlocking system.</para></listitem>
</itemizedlist>
<para>The implementation of this strategy influences the revenue streams and costs. The principal revenues are:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Increased total revenue regarding after sales services;</para></listitem>
<listitem>
<para>Optimized maintenance in service efficiency;</para></listitem>
<listitem>
<para>Saving costs in work force and maintenance service.</para></listitem>
</itemizedlist>
<para>Regarding cost structure:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Modular data acquisition system;</para></listitem>
<listitem>
<para>Monitoring system: IT infrastructure costs;</para></listitem>
<listitem>
<para>Maintenance staff costs;</para></listitem>
<listitem>
<para>Monitoring centre maintenance costs (call and data centre).</para></listitem>
</itemizedlist>
<para>Translating the revenues to economic terms, the traditional railway maintenance task involves 10% of the total costs, divided in the following items:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Maintenance workers &#x02013; 50%;</para></listitem>
<listitem>
<para>Services &#x02013; 14%;</para></listitem>
<listitem>
<para>Spare parts &#x02013; 13%;</para></listitem>
<listitem>
<para>Capital asset replacement plan &#x02013; 8%;</para></listitem>
<listitem>
<para>Warranty repairs &#x02013; 8%;</para></listitem>
<listitem>
<para>Administration costs &#x02013; 4%;</para></listitem>
<listitem>
<para>Call center &#x02013; 2%;</para></listitem>
<listitem>
<para>Shipment &#x02013; 1%.</para></listitem>
</itemizedlist>
<para>Thus, the highest value concerns <emphasis>labour costs</emphasis>. Labour force covers the costs of specialized technicians in charge of carrying out maintenance operations to the customer premises in accordance with the maintenance contract.</para>
<para>The <emphasis>services</emphasis> item include costs of supporting maintenance interventions, in particular: the field testing operations, potential remote support and training of the technicians. The third item covers the costs of the spare parts and related procurement costs.</para>
<para>As data disclose, the most relevant cost is manpower followed by services and spare parts associated with maintenance.</para>
<para>Therefore, over the following 5 years in future Proactive Monitoring Maintenance Business Model the principal aim is to reduce above all the cost of manpower with the improvements performed with the special data acquisition, data processing and maintenance strategy optimization.</para>
<table-wrap position="float" id="T8-16">
<label><link linkend="T8-16">Table <xref linkend="T8-16" remap="8.16"/></link></label>
<caption><para>Economic Projection related to current and future PMM Business Model</para></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/ch08_fig021.jpg"/>
</table-wrap>
<para>The <link linkend="T8-16">Table <xref linkend="T8-16" remap="8.16"/></link> shows a theoretical saving in maintenance costs of 25% at year 5.</para>
<para>This projection could be achievable due to the decrease of the labour costs (see maintenance workers) that impact less and less on overall maintenance costs thanks to the new solution developed: avoidance of false calls for railway equipment and optimisation of maintenance through increased efficiency of related intervention times.</para>
<para>The maintenance cost reduction progression and the remarkable indicators are shown in <link linkend="F8-19">Figure <xref linkend="F8-19" remap="8.19"/></link>.</para>
<fig id="F8-19" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F8-19">Figure <xref linkend="F8-19" remap="8.19"/></link></label>
<caption><para>Maintenance saving cost projection.</para></caption>
<graphic xlink:href="graphics/ch08_fig022.jpg"/>
</fig>
<para>Hardware and software investment have been estimated around 2%, including industrialization costs of the prototype solution.</para>
<para>This may be considered a negligible value compared to the cost of maintenance workers (around 50% of the maintenance costs).</para>
<para>Therefore, taking into account the estimated investment costs and related financial commitment, the projection of labour costs may be reconsidered, reaching 25% at year 5.</para>
<para>Regarding new revenue streams, advanced services can be defined. New maintenance software is able to monitor the railway assets and know in real-time the status and perform possible correction actions. The services would have the numbers in <link linkend="T8-17">Table <xref linkend="T8-17" remap="8.17"/></link> for 5 years, depending on client and market of each particular situation.</para>
<table-wrap position="float" id="T8-17">
<label><link linkend="T8-17">Table <xref linkend="T8-17" remap="8.17"/></link></label>
<caption><para>Advanced maintenance service program economic projection</para></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/ch08_fig022a.jpg"/>
</table-wrap>
<para>Railway industries as other industrial sectors have to adapt their business model to digitizing manufacturing and management processes, especially maintenance, to change from a reactive maintenance model to a proactive one.</para>
<para>This new strategy has some benefits such as visibility of health of assets, reducing costs by optimizing maintenance interventions, determining the root cause of failures, improving usage of warranty recoveries, etc.</para>
<para>The railway use case that was analyzed, has the objective to reduce the cost of maintenance manpower with the improvements performed, especially with regards to data acquisition, data processing and maintenance strategy optimization, under the rationale that it is the most important cost in maintenance (50%), with the next two items being maintenance services and spare parts, with 14% and 13% respectively. The 5 year projection considers a reduction of 25% costs in manpower.</para>
<para>Once the reduction of maintenance costs is achieved, in case of revenue streams, the objective would be to offer advanced services such as industrial asset monitoring to know in real-time the status of them and to perform possible corrections before downtime assets. The 5 year projections would include 70 assets to monitor with a profit of more than 600.000 by the end of year 5.</para>
</section>
</section>
<section class="lev1" id="sec8-7">
<title>8.7 Conclusions</title>
<para>Ever since asset failures have caused downtimes and extra costs, accidents or inefficiencies, businesses have supplied material and human resources to minimize their impact and avoid their re-occurrence. Current approaches try to preserve function and operability, optimize performance and increase asset lifespan with optimal investments. The current technical development in industry regarding information handling and digitalization leads to new ways of producing goods. The industry demands flexible, safe, environmental friendly and available production processes.</para>
<para>PMM in the industrial sector is key to improve competitiveness, productivity, reduction of downtime machine, interventions for remote machines and travels among others. Proactive maintenance solutions are common in processing industries like Oil and Gas, Wind, Utilities, and aerospace. Across the new PMM paradigm, industrial companies and technological providers can offer new services associated with smart products. With aims to establish proactive monitoring and maintenance, companies should address a new strategy based on servitization and service business model adding new technology and services related to industrial internet of thing.</para>
<para>The new PMM strategy should be analysed through 9 blocks related to the well-known BMCs as well as economic evaluations in order to project the future benefits of the business model. Thus, <emphasis role="strong">the value proposition</emphasis> with the introduction of integrated IT solution to monitor industrial assets with some different characteristics depending on the sector. <emphasis role="strong">Customer segments, relationships and channels</emphasis> are the business models blocks are the most relevants to reinforce and relate the companies activities with the end-user for sharing asset data and new revenue streams. <emphasis role="strong">Key activities, resources and partners</emphasis> blocks help the development of the principal IT-OT system with predictive advanced algorithms, sensors, cloud system, middleware or research, allowing future competitive advantages. Finally, <emphasis role="strong">cost structure and revenue streams</emphasis> blocks to recognise the traditional and new type of payment as software ad-hoc or as a service or for asset availability, among other features.</para>
<para>Throughout the execution of the MANTIS project, current and future PMM Business Models Canvases were analysed by industrial partners in production asset maintenance, vehicle management, energy production and health equipment maintenance. In addition, economic evaluations and projections have also been analysed.</para>
<para>In the transition from traditional to PMM business models, a financial tool is applied in MANTIS project. The principal items to take into account on behalf of the financial tool are <emphasis role="strong">revenue streams, costs and cost savings</emphasis>. Each item impacts on a segment of the company&#x02019;s strategy, referring, for example, revenue streams to more competitive product sales, maintenance contract sales, consultancy or new maintenance software services. On the other hand, the financial tool should reflect the costs and investments, such as specialized manpower, new technologies, new infrastructure, new marketing strategies, even amortizations, travel expenses and logistics. All the maintenance changes on the business model should translate to a reduction of operational costs and incremental revenues.</para>
<para>Economic evaluation and projection was developed in the MANTIS project, resulting in economic data from last year and 5 year projections comparing current and future PMM Business Models for industrial business cases. For technological providers, new business model with advanced maintenance software as value proposition and 5 years projections were analysed.</para>
<para>The present chapter also described the <emphasis role="strong">railway use-case for traditional to PMM business model</emphasis>. Its main objective is to reduce the cost of maintenance manpower by means of improvements such as implementing tools as data acquisition, data processing and maintenance strategy optimization. As it was seen, maintenance cost is the most important one, with an impact of around 50%, followed by maintenance services and spare parts with 14% and 13% respectively. <emphasis role="underline">With the incorporation of new PMM</emphasis> <emphasis role="underline">strategy, in the 5-year projection the railway use-case would have a reduction</emphasis> <emphasis role="underline">of 25% costs in manpower.</emphasis> Other use-cases analyzed within the MANTIS project, achieved an impact between 15% to 25% on machine downtime, warranty repairs, intervention costs, manpower, among others.</para>
<para>Once the reduction of maintenance costs is achieved, in case of revenue streams, the objective would be to offer advanced services such as industrial asset monitoring to know in real-time the status of them and to perform possible corrections before downtime assets. The railway use-case, during the 5 year projection, would include 70 assets to monitor with a profit of more than 600.000 by the end of year 5. Due to techniques such as the servization of maintenance, the railway use-case, as well as others analyzed in MANTIS project, would imply an increase of income between 10% to 20 % .</para>
<para>In conclusion, PMM is key to improve maintenance processes applied in industrial companies, which can be either production asset maintenance, vehicle management, energy production or health equipment companies. New service business models should carry out monitoring industrial asset with sensors and predictive technology. As a consequence, new smart products with advanced services could be offered to achieve the impact to reduce around 25% in maintenance costs and increase 20% of revenue streams.</para>
</section>
<section class="lev1" id="sec8-8">
<title>References</title>
<orderedlist numeration="arabic" continuation="restarts" spacing="normal">
<listitem>
<para>Anke, J., and Krenge, J. (2016). Prototyp eines Tools zur Abschtzung der Wirtschaftlichkeit von Smart Services fr vernetzte Produkte. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Anke%2C+J%2E%2C+and+Krenge%2C+J%2E+%282016%29%2E+Prototyp+eines+Tools+zur+Absch%5Bx00E4%3B%5Dtzung+der+Wirtschaftlichkeit+von+Smart+Services+f%5Bx00FC%3B%5Dr+vernetzte+Produkte%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Arnold Vogt (2016). Industrie 4.0 / IoT Vendor Benchmark 2017.</para></listitem>
<listitem>
<para>Baines, T.S., Lightfoot, H.W., Benedettini, O., and Kay, J.M. (2009). The servitization of manufacturing: A review of literature and reflection on future challenges. <emphasis>J. Manuf. Technol. Manag.</emphasis> 20, 547&#x02013;567. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Baines%2C+T%2ES%2E%2C+Lightfoot%2C+H%2EW%2E%2C+Benedettini%2C+O%2E%2C+and+Kay%2C+J%2EM%2E+%282009%29%2E+The+servitization+of+manufacturing%3A+A+review+of+literature+and+reflection+on+future+challenges%2E+J%2E+Manuf%2E+Technol%2E+Manag%2E+20%2C+547-567%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Boardman, A.E., Greenberg, D.H., Vining, A.R., and Weimer, D.L. (2017). Cost-benefit analysis: concepts and practice (Cambridge University Press). <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Boardman%2C+A%2EE%2E%2C+Greenberg%2C+D%2EH%2E%2C+Vining%2C+A%2ER%2E%2C+and+Weimer%2C+D%2EL%2E+%282017%29%2E+Cost-benefit+analysis%3A+concepts+and+practice+%28Cambridge+University+Press%29%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Brisk Insights (2016). Operational Predictive Maintenance Market By Component (Solutions, Services), By Deployement Type (Cloud Deployement, By On Premises Deployement), By Application (Automotive, Energy And Utilities, Helathcare, Manufacturing, Government &#x00026; Defense, Transport And Logistics), Industry Size, Growth, Share And Forecast To 2022.</para></listitem>
<listitem>
<para>Burke, P.Y. (2017). 74 &#x02013; Business Plan. In Technical Career Survival Handbook, (Academic Press), pp. 185&#x02013;186.</para></listitem>
<listitem>
<para>Capgemini Consulting (2014). Industry 4.0 &#x02013; The Capgemini Consulting View.</para></listitem>
<listitem>
<para>Cristi&#x000E1;n M. Lincovil B, and G. Ivonne Guti&#x000E9;rrez M (2006). OPTIMIZACIN ECON&#x000D3;MICA DE LA DISPONIBILIDAD.</para></listitem>
<listitem>
<para>Featherstone, S. (2015). 1 &#x02013; Creating a business plan. In A Complete Course in Canning and Related Processes (Fourteenth Edition), (Woodhead Publishing), pp. 3&#x02013;20. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Featherstone%2C+S%2E+%282015%29%2E+1+-+Creating+a+business+plan%2E+In+A+Complete+Course+in+Canning+and+Related+Processes+%28Fourteenth+Edition%29%2C+%28Woodhead+Publishing%29%2C+pp%2E+3-20%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Fraga-Lamas, P., Fern&#x000E1;ndez-Caram&#x000E9;s, T. M., &#x00026; Castedo, L. (2017). Towards the internet of smart trains: a review on industrial IoT-connected railways. Sensors, 17(6), 1457. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Fraga-Lamas%2C+P%2E%2C+Fern%E1ndez-Caram%E9s%2C+T%2E+M%2E%2C+%26+Castedo%2C+L%2E+%282017%29%2E+Towards+the+internet+of+smart+trains%3A+a+review+on+industrial+IoT-connected+railways%2E+Sensors%2C+17%286%29%2C+1457%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>GARTNER (2013). Forecast: The Internet of Things, Worldwide.</para></listitem>
<listitem>
<para>IBM (2013). Improving Operational and Financial Results through Predictive Maintenance. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=IBM+%282013%29%2E+Improving+Operational+and+Financial+Results+through+Predictive+Maintenance%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Jantunen, E. (2016). Appendix 21: Existing business models related to proactive Monitoring and maintenance (PMM).</para></listitem>
<listitem>
<para>de Jesus, D.M.F. (2012). Financial Projection Based on Business Model Canvas. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=de+Jesus%2C+D%2EM%2EF%2E+%282012%29%2E+Financial+Projection+Based+on+Business+Model+Canvas%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Kans, M. (2016). Service Management 4.0 and its applicability in the Swedish railway industry. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Kans%2C+M%2E+%282016%29%2E+Service+Management+4%2E0+and+its+applicability+in+the+Swedish+railway+industry%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Kans, M., and Ingwald, A. (2016). Business Model Development towards Service Management 4.0. Procedia CIRP 47, 489&#x02013;494. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Kans%2C+M%2E%2C+and+Ingwald%2C+A%2E+%282016%29%2E+Business+Model+Development+towards+Service+Management+4%2E0%2E+Procedia+CIRP+47%2C+489-494%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>MARKETandMARKETS (2016). Smart Railways Market by Solution (Passenger Information, Freight Information, Rail Communication, Advanced Security Monitoring, Rail Analytics), Component, Service (Professional, Managed), and Region - Global Forecast to 2021</para></listitem>
<listitem>
<para>M&#x000E1;rquez, A.C., de Le&#x000F3;n, P.M., and Herguedas, A.S. (2004). Ingenier&#x000ED;a de mantenimiento: t&#x000E9;cnicas y m&#x000E9;todos de aplicacin a la fase operativa de los equipos (Aenor). <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=M%E1rquez%2C+A%2EC%2E%2C+de+Le%F3n%2C+P%2EM%2E%2C+and+Herguedas%2C+A%2ES%2E+%282004%29%2E+Ingenier%EDa+de+mantenimiento%3A+t%E9cnicas+y+m%E9todos+de+aplicaci%5Bx00F3%3B%5Dn+a+la+fase+operativa+de+los+equipos+%28Aenor%29%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>McKinsey (2015). <emphasis>The Internet of Things: Mapping the Value Beyond the Hype.</emphasis></para></listitem>
<listitem>
<para>McKinsey &#x00026; Company, and VDMA (2016). How to succeed: Strategic options for European machinery.</para></listitem>
<listitem>
<para>Moen, E.F. (2014). Maintenance and barriers: Principles for barrier management in the petroleum industry will be more and more important and It is fundamental to understand the maintenance function in the barrier management. Institutt for produksjons-og kvalitetsteknikk. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Moen%2C+E%2EF%2E+%282014%29%2E+Maintenance+and+barriers%3A+Principles+for+barrier+management+in+the+petroleum+industry+will+be+more+and+more+important+and+It+is+fundamental+to+understand+the+maintenance+function+in+the+barrier+management%2E+Institutt+for+produksjons-og+kvalitetsteknikk%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Neely, A. (2007). The servitization of manufacturing: an analysis of global trends. In 14th European Operations Management Association Conference, (Turkey Ankara), pp. 1&#x02013;10. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Neely%2C+A%2E+%282007%29%2E+The+servitization+of+manufacturing%3A+an+analysis+of+global+trends%2E+In+14th+European+Operations+Management+Association+Conference%2C+%28Turkey+Ankara%29%2C+pp%2E+1-10%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Pauceanu, A.M. (2016). Chapter 4 &#x02013; Business Plan. In Entrepreneurship in the Gulf Cooperation Council, (Academic Press), pp. 79&#x02013;118. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Pauceanu%2C+A%2EM%2E+%282016%29%2E+Chapter+4+-+Business+Plan%2E+In+Entrepreneurship+in+the+Gulf+Cooperation+Council%2C+%28Academic+Press%29%2C+pp%2E+79-118%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Tanev, S., Rasmussen, E.S., and Hansen, K.R. (2016). 2 &#x02013; Business plan basics for engineers. In Start-Up Creation, (Woodhead Publishing), pp. 21&#x02013;37. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Tanev%2C+S%2E%2C+Rasmussen%2C+E%2ES%2E%2C+and+Hansen%2C+K%2ER%2E+%282016%29%2E+2+-+Business+plan+basics+for+engineers%2E+In+Start-Up+Creation%2C+%28Woodhead+Publishing%29%2C+pp%2E+21-37%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Roland Berger Study (2017). Rail supply digitization</para></listitem>
<listitem>
<para>Wurzbach, R. N. (2000). A web-based cost benefit analysis method for predictive maintenance. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Wurzbach%2C+R%2E+N%2E+%282000%29%2E+A+web-based+cost+benefit+analysis+method+for+predictive+maintenance%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Zolnowski, A. (2015). Analysis and Design of Service Business Models. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Zolnowski%2C+A%2E+%282015%29%2E+Analysis+and+Design+of+Service+Business+Models%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Zolnowski, A., Schmitt, A.K., and Bhmann, T. (2011). Understanding the impact of remote service technology on service business models in manufacturing: From improving after-sales services to building service ecosystems. (ECIS), p.</para></listitem>
<listitem>
<para>Zolnowski, A., Anke, J., and Gudat, J. (2017). Towards a Cost-Benefit-Analysis of Data-Driven Business Models. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Zolnowski%2C+A%2E%2C+Anke%2C+J%2E%2C+and+Gudat%2C+J%2E+%282017%29%2E+Towards+a+Cost-Benefit-Analysis+of+Data-Driven+Business+Models%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Zott, C., Amit, R., Massa, L., and others (2010). The business model: Theoretical roots, recent developments, and future research. IESE Bus. Sch.-Univ. Navar. 1&#x02013;43. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Zott%2C+C%2E%2C+Amit%2C+R%2E%2C+Massa%2C+L%2E%2C+and+others+%282010%29%2E+The+business+model%3A+Theoretical+roots%2C+recent+developments%2C+and+future+research%2E+IESE+Bus%2E+Sch%2E-Univ%2E+Navar%2E+1-43%2E" target="_blank">Google Scholar</ulink></para></listitem>
</orderedlist>
</section>
</chapter>
<chapter class="chapter" id="ch09" label="9" xreflabel="9">
<title>The Future of Maintenance</title>
<para><emphasis role="strong">Lambert Schomaker<superscript><emphasis role="strong">1</emphasis></superscript>, Michele Albano<superscript><emphasis role="strong">2</emphasis></superscript>, Erkki Jantunen<superscript><emphasis role="strong">3</emphasis></superscript> , and Luis Lino Ferreira<superscript><emphasis role="strong">2</emphasis></superscript></emphasis></para>
<para><superscript>1</superscript>University of Groningen, The Netherlands</para>
<para><superscript>2</superscript>ISEP, Polytechnic Institute of Porto, Porto, Portugal</para>
<para><superscript>3</superscript>VTT Technical Research Centre of Finland Ltd, Finland</para>
<para>In this book, a number of perspectives on predictive and proactive maintenance have been presented that were developed during the course of EU/ECSEL project MANTIS in the years 2015&#x02013;2018. At the start of the project, a number of developments heralded things to come: The Big Data and data science revolution, internet of things (IoT), advances in machine learning, improvements in wireless connectivity, sensor technologies and available computing power. At the level of software, cloud computing, software services, semantic interoperability and multi-tiered architectures all displayed a fast-moving field. This final chapter takes a step back and presents some views towards the future. Whether one deals with PM on a manufacturing process or a fleet of machines, in logistics or construction, the potential gains from improving the maintenance policies can be substantial. With a daily yield of 30% of a particular production process, an improvement of only 3 percentage points due to improved maintenance policies constitutes an improvement of 10% on the status quo. In maintenance services for customers of a leased fleet of machines, the statistical analysis of customer usage patterns allows a company to design services, adapted to the wear &#x00026; tear patterns that are typical for different customer groups and provide economically attractive solutions (e.g., &#x02018;bronze, silver and gold&#x02019; maintenance service levels). With this in mind, it is surprising that maintenance is sometimes considered as a liability. In the biological world, the praying mantis will clean its body and sensors, autonomously. For cyber-physical systems of the near future, one would hope that at least part of such behaviors is controlled on the basis of intrinsic feedback loops. This should ideally be realized in a cost-effective manner, i.e., with minimal human intervention at all levels of control, ranging from analytics to decision making, corrective interventions and preventive actions at the physical level. Only an integrated design of primary and secondary system functions of cyber-physical system will lead to efficient and resilient systems. When NASA rovers were sent to Mars, it quickly became clear that in spite of the impressive technological advances, a &#x02018;minor&#x02019; aspect was overlooked: Dust was accumulating on the solar panels, the camera lenses and the color-calibration disk for true-color adjustments [KinchMarsDust, 2007]. Unlike its biological counterparts on earth, e.g., insects, the Mars rover did not have actuators, neither for cleaning its essential photovoltaic energy-harvesting system, nor its sensors (<link linkend="F9-1">Figure <xref linkend="F9-1" remap="9.1"/></link>).</para>
<fig id="F9-1" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F9-1">Figure <xref linkend="F9-1" remap="9.1"/></link></label>
<caption><para>Unexpected dust accumulation on the color calibrator of a Mars rover. Without actuators for autonomous self maintenance, only the wind can help out [KinchMarsDust, 2007].</para></caption>
<graphic xlink:href="graphics/ch09_fig001.jpg"/>
</fig>
<para>Today, companies will need to decide what their basic policy is with regards to maintenance, a.k.a. that secondary but nevertheless essential process in working systems. Production tools and consumer products are increasingly designed on the basis of appearance and perceived simplicity. According to this philosophy, large components are often replaced as a whole and no attempt is made at maintenance &#x02018;below the hood&#x02019;. It is questionable whether this is sustainable in the long term, due to limited global resources. As an example, take a large data center. If the storage and server units are purchased and installed at one concentrated point in time, the sub systems will statistically fail after a few years and the total quality of service is jeopardized. Users of that data center will expect a reliable operation without interruptions. Should the company aim at a gross replacement of large groups of racks at the end of their life cycle, coincident with a long off-line period? Should one, alternatively, aim at gradual replacement of servers at the level of blade-server modules? Or should the investment be aimed at a solid frame architecture offering uninterrupted operations? The latter approach would entail a continuous robotic replacement of small disk units in a manner that is similar to a biological metabolic process, with an input buffer of spare disks and an output stream of defunct units. Is there a cost-effective combination of these different approaches? In order to make rational choices concerning the level of granularity and in order to learn effective operating policies as a company, the quantitative approach to maintenance as sketched out in this book is essential. In the remainder of the chapter, we will focus on the following provocative questions:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>Is it cybernetic or is it human?</para></listitem>
<listitem>
<para>Real-time communication in maintenance?</para></listitem>
<listitem>
<para>How to determine granularity in space and time?</para></listitem>
<listitem>
<para>Open or closed maintainability?</para></listitem>
<listitem>
<para>Insourcing or outsourcing?</para></listitem>
<listitem>
<para>Explicit modeling or data-driven pragmatics?</para></listitem>
<listitem>
<para>How to apply Virtual Reality and Augmented Reality?</para></listitem>
<listitem>
<para>Service robotics for maintenance?</para></listitem>
</itemizedlist>
<section class="lev1" id="sec9-1">
<title>9.1 Is it Cybernetic or Is it Human?</title>
<para>From the experiences in the project, it is clear that current practice is still lagging today&#x02019;s maintenance control capabilities. Although the amount of logged data can be quite impressive in the use cases, it is clear that the exploitation of the available information is limited. Maintenance-related decision processes are slow. They often require human-to-human interaction concerning the selection of relevant data from legacy data bases, sometimes with a complicated access-clearance procedure. As a result of preliminary analyses, usually more data are needed, requiring ever more human-to-human communication and negotiation. This predicament is exacerbated if the analytics is not performed by in-house data scientists but by external companies providing analytics services. The necessary data will usually exist, somewhere in a huge storage repository. However, even after it has been collected there will be a labor-intensive process of data cleaning, normalization and repackaging before it can be used by traditional statistics or modern machine learning off-the-shelf tools. In this process, additional human-to-human communication is needed ranging from database administrators to operators on the shop floor in order to actually understand the data and the underlying physical processes. In order to really close the maintenance-oriented feedback loop of a CPS, a number of steps need to be made:</para>
<itemizedlist mark="bullet" spacing="normal">
<listitem>
<para>A transition needs to be realized, from isolated ad-hoc problem analysis to continuous measurement processes and effective control policies, at a pace that fits the underlying process and is economically viable;</para></listitem>
<listitem>
<para>The connection between the target process itself and the analytics modules needs to be an information highway in itself, not an improvised ad-hoc connection. Standardisation is therefore a necessity at all levels, from network I/O to semantic interoperability. The Explanator concept for adding specific signal metadata to a .csv spreadsheet column (<link linkend="ch03">Chapter <xref linkend="ch03" remap="3"/></link>) is just one example of the additional scaffolding that is necessary;</para></listitem>
<listitem>
<para>The standardized data can be easily presented to in-house and external analytics consultants, allowing for selecting the best predictions or policy suggestions from several opinions or perspectives. If the packaging of analytics results is also standardized, even the selection process of finding the best solution can be automated.</para></listitem>
</itemizedlist>
<para>Closing this feedback loop into a fully autonomous cyber-physical system is not desirable at this stage, but a substantial reduction in human labor and an improvement in the quality of the decisions can be expected as a result of following these suggestions.</para>
</section>
<section class="lev1" id="sec9-2">
<title>9.2 Real-time Communication in Maintenance?</title>
<para>An effective condition maintenance procedure requires the acquisition of dozens of signals, some of which are related with the normal operation of the system, while other are very specific to the maintenance requirements. Some of these acquisitions are performed at very high rates (e.g., with an interval of 10 ms). Achieving these timing requirements in low-performance wireless networks like Zigbee or Bluetooth is a challenge that has to be faced when designing any maintenance-based processing system. Additionally, if we have several sources with the same requirements, the networks supporting them might be occupied close to their operating limit. Condition-based maintenance will in some cases also require true real-time detection of physical problems. This can be the case, for example, in cars, trains or industrial manufacturing machines. In such cases, malfunctions have to be detected within a tightly limited time frame in order either to avoid the problem, or put the system into a safe state. For different types of malfunctions, a characteristic limited deadline for reporting must be specified. Possible solutions to achieve these capabilities concern the use of multi-core computing platforms together with real-time operating systems that are also tightly integrated with the analysis tools. While finding an adequate real-time operating system can be an easy task, adapting and connecting data-analysis tools is still a challenge that needs to be addressed, both in relation to parallelization and to the technical details of the integration with such a real-time operating system. The work performed during the MANTIS project allowed the identification of two main challenges for future research: i) The realization of maintenance-related real-time communication capabilities, and ii) attaining real-time guarantees for (close to) real-time data analysis at the level of analytics.</para>
</section>
<section class="lev1" id="sec9-3">
<title>9.3 How to Determine Granularity in Space and Time?</title>
<para>Large, disruptive transformations in a system architecture are expensive, risky and time consuming. On the contrary, a system that is designed for maintenance and maintenance-related processes will allow for incremental improvements that do not disturb ongoing processes. Going back to the example of the data centers: If it holds a maintenance criterion such as &#x02018;duration of a disk hot swap should not take more than 2 minutes&#x02019;, this requirement may have a domino effect on the temporal granularity of other interventions. If other maintenance operations have a duration in the same order of magnitude, the total process is much easier to manage. In the current practice of data centers, users will often not even notice that maintenance on RAID disk systems is going on. In addition to the granularity in time, the scale of replaced system components (&#x02018;just a disk&#x02019; versus &#x02018;a complete rack&#x02019;) plays a role too. It is possible to design good systems with maintainability in mind, thus avoiding operating costs that can be foreseen and prevented. This can be realized by choosing the proper granularity for component replacement.</para>
</section>
<section class="lev1" id="sec9-4">
<title>9.4 Open or Closed Maintainability?</title>
<para>Depending on the application area, maintenance in manufacturing or maintenance of consumer products, there will be different company goals. The increased digitization has allowed some companies to create customer &#x02018;lock-in&#x02019; by means of specialized electronic tools for maintenance diagnostics. Sometimes there are contracts where a user is allowed to operate, but not maintain a particular product. Currently a countermovement is emerging in the United States under the &#x02018;Right to Repair Act&#x02019; in several states, for automobiles and agricultural equipment.</para>
<para>Rather than pursuing a conflict model, companies can create goodwill and customer involvement in maintenance issues by the proper use of big data on wear &#x00026; tear for different usage patterns and by offering customized maintenance services for individual users. In this manner, predictive maintenance is beneficial for both parties and companies keep a strong position thanks to their information advantage.</para>
</section>
<section class="lev1" id="sec9-5">
<title>9.5 Insourcing or Outsourcing?</title>
<para>The increased availability of analytics consultants with a statistical or machine learning background poses new dilemmas for companies. In recent years, there was a trend to outsource a number of activities and ICT services to external companies. It is debatable whether maintenance-related information processing is &#x02018;just ICT&#x02019;. The involved knowledge is highly sensitive and is the intellectual property of a particular company. Decisions will have to be made concerning the employment of external parties because maintenance-related topics are directly connected to the reputation of a product and the profit model that is in action. Evidently, legal mechanisms can be applied to mitigate the risk of allowing third parties access to a core process. On the other hand, predictive maintenance is so knowledge intensive and tightly coupled to the center of a company&#x02019;s activities that outsourcing should be avoided in some cases. Replacing a body part with an extraneous replacement may be acceptable in some areas, but would we outsource the brain, too? The amount of knowledge that needs to be shared with external analytics partners is usually also detailed and still requires substantial time investments, by both parties. Whereas the company sees an opportunity for success with limited loss of time, the analytics company expects to find low-hanging fruit: Both expectations are overly optimistic. There is no free lunch: Modern machine learning only works with enough high-quality data from the problem owner. At the same time, companies may have worked for decades on a particular maintenance problem such that it may be very difficult for newcomers to improve existing results.</para>
</section>
<section class="lev1" id="sec9-6">
<title>9.6 Explicit Modeling or Data-driven Pragmatics?</title>
<para>The project unrolled in a period where there are tremendous advances in artificial intelligence and machine learning. Almost every month there was a breakthrough in deep learning, considering hard, nonlinear problems, such as the game of Go. Although most advances have been made in the area of image processing, or problems where a 2D array of cells is given at the input, there are also advances in time-series processing using recurrent neural networks, such as the LSTMs. On the other end of the spectrum are the traditional modeling approaches, such as hidden-Markov models or regression statistics for modeling aspects of maintenance processes. In the traditional approach, modeling from interpretable models (white box modeling) is used to obtain a detailed insight in complex processes and the underlying causalities. In the relevant communities there is considerable mutual scepticism, not so say animosity, concerning &#x02018;the other&#x02019; (black box modeling) approach. The easy answer would be to say that time will tell. However, it is already becoming clear in what direction future developments may go. The new machine-learning tools allow to discuss regarding the existence of a particular I/O mapping with sufficient accuracy and reliability. From that point on, it depends on the actual goal of the analytics exercise. If a prediction with an acceptable error margin is at stake, it may not always be necessary &#x02013; for a company &#x02013; to understand all minute details and causes of this particular I/O relationship. Human-based analysis is costly and the traditional explicit-modeling approach is not always perfect, either. On the other hand, if a thorough understanding of an important and costly maintenance issue is needed, the investment in human-based research may be warranted. The new wave of machine learning has taught us at least one powerful lesson: Each model (including the handcrafted ones) can be viewed as just a stochastic sample from a universe <emphasis>&#x0210B;</emphasis> of possible model designs [Valiant, 1984]. There may exist several alternative models with similar accuracies and reliabilities: &#x0201C;All models are wrong, but some of them are more useful than others&#x0201D;. The challenge for a company then is to make effective use of these available variants for a particular problem. After all, the ultimate goal may be the economic yield, not the scientific understanding, per se.</para>
</section>
<section class="lev1" id="sec9-7">
<title>9.7 How to Apply Virtual Reality and Augmented Reality?</title>
<para>Current industry is one of the key domains where virtual, augmented and mixed reality can create a huge added value.</para>
<para>The most commonly known applications of Virtual Reality (VR) in condition monitoring and maintenance are training, actual maintenance, remote maintenance/condition monitoring and maintenance assembly. Virtual Reality generates a computerized environment of the real system to be used in the previously mentioned aspects (see <link linkend="F9-2">Figure <xref linkend="F9-2" remap="9.2"/></link>).</para>
<fig id="F9-2" position="float" xmlns:xlink="http://www.w3.org/1999/xlink">
<label><link linkend="F9-2">Figure <xref linkend="F9-2" remap="9.2"/></link></label>
<caption><para>Virtual Reality system for maintenance.</para></caption>
<graphic xlink:href="graphics/ch09_fig002.jpg"/>
</fig>
<para>VR could, for example, help the maintainer identify where exactly the problem is, from a whole system down to single component level, in a virtualized system so that the maintainer can reduce the asset&#x02019;s downtime.</para>
<para>Maintenance training has been used in several applications including aircrafts, automobiles, power plants and other process industry applications. In actual maintenance, VR has been used mostly in aircraft maintenance applications like military fighter planes because of the complexity of the maintenance process. Some other examples of remote maintenance applications for process industry and rotating machinery can also be found.</para>
<para>Augmented reality (AR) is a blend between VR and actual reality. AR applications contain basically the same repertoire than general VR solutions. AR allows for maintenance training, actual maintenance, service based mobile maintenance tools, remote maintenance, maintenance assembly and e-maintenance solutions. Some applications contain even simple diagnostics and utilization of the user&#x02019;s speech.</para>
<para>Maintenance training has been used in similar applications as in VR. Current maintenance covers solutions like aeroplanes, armoured military vehicles and different industrial solutions. Remote maintenance solutions are targeted mainly for small and medium size enterprises. Some examples can be found from service based mobile maintenance tools and e-maintenance solutions. More sophisticated solutions utilize even simple diagnostics and the user&#x02019;s speech to make decisions.</para>
</section>
<section class="lev1" id="sec9-8">
<title>9.8 Service Robotics for Maintenance?</title>
<para>Rapid developments are taking place not just in the area of machine learning. Robot technology is also advancing at a fast pace. There is a diversification of the application areas. Whereas in the eighties and nineties the focus was on statically located robot arms with strong force, high speed and accuracy, today&#x02019;s robots span a wider spectrum of implementations. Convenience of programming, man-machine collaboration and improved sensing are needed in the work place. In the Robocup@Home benchmark competition, the participants only hear a day in advance what the actual testing scenario will be (given a number of constraints on the robot design, in advance). Standardisation on the basis of ROS (Robot Operating System), and the Python programming ecosystem, have allowed fast prototyping. In light of these recent developments, it is amazing to see that current industrial robotics is still mechatronics based, very accurate but difficult and expensive to program. However, if one observes human operator activities along a production line, it is clear that part of this work also lends itself to possible robotisation. Evidently, the replacement of complex modules and tools in a manufacturing production line will still be a job for human operators. On the other hand, some tasks are repetetive and menial, involving visual or other types of inspection on known places along a production line, such as the occasional removal of dirt and providing oil in locations where friction starts to become a problem. The mean time between failure can be fairly short, for complicated production lines, while the human-based corrective actions sometimes are surprisingly trivial: A push here, a removal of a small obstruction, or simply cleaning an area with compressed air. Current robotics technology is well able to fulfill at least a part of these tasks, notably on a &#x02018;24/7&#x02019; basis. This prevents the buildup of large problems as a consequence of accumulated small problems. Under conditions of variable operating conditions, it would be too expensive to install dedicated hardware to solve a wide variety of the aforementioned problems. On the contrary, the solutions provided by service robotics are not hardware but behavior based, i.e., relying on trainable software. If service robots will be employed in a manufacturing setting, this will change but not obliterate the necessity of human interventions. Operators will become the robot trainers, providing the robots with a library of corrective behaviors for different problem locations along a production line. An important advantage is that such pragmatic knowledge is then shared among the members of a team of maintenance robots. The presence of such robots also provides an incentive for making the maintenance-related knowledge more explicit within the company. Interesting new research topics will evolve at the level of dynamic tool use by robots, using standard clip-on tools. An image emerges of large cyber-physical facilities with mobile maintenance robots supporting static production robots and maintaining active legacy production machines of a different era that are not yet at their end-of-life state from an economical point of view.</para>
</section>
<section class="lev1" id="sec9-9">
<title>9.9 How will the Maintenance Practices Change</title>
<para>Most of the maintenance strategies that are followed today to some extent have existed for a long time e.g., the first edition of Maintenance Engineering Handbook was printed in 1957 [Higgings, 1987], the first edition of Moubray&#x02019;s Reliablity-centred Maintenance [Moubray, 2007] was published in 1991, Jardine &#x00026; Tsang came up with Maintenance, Replacement, and Reliability [Jardine, 2006], Theory and Applications in 2006, Crespo&#x02019;s The Maintenance Management Framework [Crespo, 2007] was published in 2007, and E-maintenance came out in 2010 [Holmberg, 2010]. Also, the key standards have existed for a long time. In spite of the existence of these theories and methods, the everyday maintenance practices vary a lot. The so called corrective maintenance is still used to a great extent although it has been proven to be in many cases costly and ineffective. The main reasons for this situation are the lack of data that would help to understand the need for maintenance, and the lack of ways to define what would be the best practice from financial point of view.</para>
<para>The book in hand discusses widely the first aspect i.e., how to get data and how to turn that data into meaningful information about the current state of the machine i.e., is maintenance needed and when should that action be taken when considered from technical point of view. It is clearly shown what the Cyber-Physical Systems are capable of and how they can help in the definition of the condition of machinery and thus basically enable Condition-Based Maintenance i.e., enable maintenance based on need and not e.g., calendar. The dramatic reduction of cost of sensors and the dramatic development of processing power and wireless communication are the key elements in enabling IoT 4.0 and thus Maintenance 4.0 [Jantunen, 2018].</para>
<para>The second aspect i.e., the lack of ways to define what would be the best practices is not covered quite to the same extent. However, the increase of reliable data will enable the full use of the maintenance strategies named in the beginning of this section. For example Reliability-centred Maintenance (RCM) has been considered very hard to use due to the enormous manual work in collecting meaningful data. Even when the studies have been made in cases where it has been considered financially justified, the results have been criticized or at least doubted due the number of assumptions that have been made and the limitations in the amount of the data. Now this will change dramatically as the data will be available and, assuming it is managed in proper ways as discussed in chapter 3 of this book, and the basis for following the above described strategies exists. In fact it can be claimed that the most dramatic change will take place when the above described data will be used by Computerized Maintenance Management Systems (CMMS) to answer the what if question i.e., make comparisons between choices on strategic level and on practical level. The point is that collected data make it possible to run numerous scenarios comparing the financial results of different kind of approaches, e.g., conservative maintenance where maintenance actions are carried out early in order to guarantee high availability or more risky approach where maintenance actions are carried out closer to the end of the life of the components. In theory this means that in the future maintenance policies can be based purely on financial issues and not feelings, opinions, and partial information.</para>
<para>In a futuristic scenario, the maintenance can be automatic and without human intervention, managed by the CMMS that manages work orders and spare parts and at the same time, much of the practical work is actually carried out by robots as described earlier in this chapter. The reality today is that we are very far from the above described scenario. Today manufacturing companies are very keen on developing maintenance service businesses for the equipment they have sold. The main idea behind this is that it is realized that providing maintenance services to customers can stabilize the business against the variation of sales that are varying depending on local and world economy. Another scenario for manufacturing companies is that they would sell not the production machinery but the capability to produce at certain level. Even though there was a boom in providing maintenance services, quite a lot of work is needed in order to reach the scenario described above. Unfortunately, today the level from where the manufacturing companies start their development for the development of maintenance services varies a lot. Consequently, it can be expected that there will be great success stories when the development will follow efficient maintenance strategies, but there is also the risk that someone might have too high hopes regarding how quick and easy the development can be.</para>
</section>
<section class="lev1" id="sec9-10">
<title>9.10 Conclusion</title>
<para>The evolution from human-paced, ad-hoc, intuition-based maintenance to integrated, quantitative and autonomous maintenance in cyber-physical systems will still require considerable effort in the coming years. The MANTIS project is an important step in this direction, which could not have been realized by a single organization.</para>
</section>
<section class="lev1" id="sec9-11">
<title>References</title>
<orderedlist numeration="arabic" continuation="restarts" spacing="normal">
<listitem>
<para>Crespo M&#x00B4;arquez, A. (2007). The Maintenance Management Framework. London: Springer-Verlag London Limited.</para></listitem>
<listitem>
<para>Higgings, L. R. (Ed.). (1987). Maintenance Engineering Handbook (4th ed.). New York: McGraw-Hill Book Company.</para></listitem>
<listitem>
<para>Holmberg, K.;Adgar, A.;Arnaiz, A.;Jantunen, E.;Mascolo, J.; Mekid, S. (2010). E-maintenance (1st p.). London: Springer-Verlag London Limited. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Holmberg%2C+K%2E%3BAdgar%2C+A%2E%3BArnaiz%2C+A%2E%3BJantunen%2C+E%2E%3BMascolo%2C+J%2E%3B+Mekid%2C+S%2E+%282010%29%2E+E-maintenance+%281st+p%2E%29%2E+London%3A+Springer-Verlag+London+Limited%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Jantunen, E.; Di Orio, G.; Heged&#x00A8;us, C.; Varga, P.; Moldov&#x00B4;an, I.; Larrinaga, F.; Becker, M.; Albano, M.; Mal&#x00B4;o, P. (2018). Maintenance 4.0 World of Integrated Information. I-ESA 2018: 9th Conference on Interoperability for Enterprise Systems and Applications. Berlin: Fraunhofer IPK.</para></listitem>
<listitem>
<para>Jardine, A. K.; Tsang, A. H. (2006). Maintenance, Replacement and Reliability, Theory and Applications. Boca Raton: CRC Press, Taylor &#x00026; Francis Group. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Jardine%2C+A%2E+K%2E%3B+Tsang%2C+A%2E+H%2E+%282006%29%2E+Maintenance%2C+Replacement+and+Reliability%2C+Theory+and+Applications%2E+Boca+Raton%3A+CRC+Press%2C+Taylor+%26+Francis+Group%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Kinch, K. M., Sohl-Dickstein, J., Bell, J. F., Johnson, J. R., Goetz, W. and Landis, G. A. (2007). Dust deposition on the Mars Exploration Rover Panoramic Camera (Pancam) calibration targets. <emphasis>Journal of Geophysical Research: Planets</emphasis>, 112(E6). <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Kinch%2C+K%2E+M%2E%2C+Sohl-Dickstein%2C+J%2E%2C+Bell%2C+J%2E+F%2E%2C+Johnson%2C+J%2E+R%2E%2C+Goetz%2C+W%2E+and+Landis%2C+G%2E+A%2E+%282007%29%2E+Dust+deposition+on+the+Mars+Exploration+Rover+Panoramic+Camera+%28Pancam%29+calibration+targets%2E+Journal+of+Geophysical+Research%3A+Planets%2C+112%28E6%29%2E" target="_blank">Google Scholar</ulink></para></listitem>
<listitem>
<para>Moubray, J. (2007). Reliability-centred Maintenance (2nd p.). Amsterdam: Butterworth-Heinemann, Elsevier.</para></listitem>
<listitem>
<para>Valiant, L. G. (1984). A theory of the learnable. <emphasis>Communications of the ACM</emphasis>, 27(11), pp. 1134&#x02013;1142. <ulink url="https://scholar.google.com/scholar?hl=en&amp;q=Valiant%2C+L%2E+G%2E+%281984%29%2E+A+theory+of+the+learnable%2E+Communications+of+the+ACM%2C+27%2811%29%2C+pp%2E+1134-1142%2E" target="_blank">Google Scholar</ulink></para>
</listitem>
</orderedlist>
</section>
</chapter>
<chapter class="nosec" id="ch10">
<title>About the Editors</title>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/ch10_michele.jpg"/></para>
<para><emphasis role="strong">Dr. Michele Albano</emphasis> is Research Associate in the CISTER Research Unit of the Polytechnic of Porto (Portugal), and Adjunct Professor at Polytechnic Institute of Cavado and Ave (Portugal). His research focuses on Cyber Physical Systems and Communication Middleware, in particular for the application areas of industrial maintenance, smart grids, and vehicular networks. He is a Founding Member of the Technical Committee on Green Communications and Computing (TCGCC).</para>
<para>Michele received his BSc degree in Physics in 2003, and his BSc, MSc and PhD degrees in Computer Science in 2004, 2006 and 2010 respectively, all of them from the University of Pisa, Italy. He was visiting researcher at Universidad de Malaga in 2007, at University of New York at Stony Brook in 2009, he was a postdoctoral researcher at the Instituto de Telecomunica&#x000E7;&#x000F5;es in Aveiro, Portugal, from 2010 to 2012, and before being a researcher he worked as a software engineer and wireless technology specialist in private companies in the period from 2001 to 2006. In the context of the MANTIS project, he was Work Package leader for WP3 &#x0201C;Smart sensing and data acquisition technologies&#x0201D; and for WP8 &#x0201C;Dissemination of knowledge and exploitation&#x0201D;.</para>
<para>Michele is on the editorial board of the International Journal of Social Technologies and of the Transactions on Emerging Telecommunication Technologies since 2011, and in 2015 he has been appointed as Editor in Chief for the Journal of Green Engineering.</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/ch10_Edward_ja_pokat.jpg"/></para>
<para><emphasis role="strong">Dr. Erkki Jantunen</emphasis> is principal scientist at VTT Technical Research Centre of Finland Ltd. Between 1978 and 1990 he worked in the shipbuilding industry in structural, vibration and hydrodynamic fields. Since 1990 he has been employed by VTT having various project responsibilities related to maintenance, condition monitoring and diagnosis of rotating machinery. He has been a member of the editorial board and acted as a reviewer of a number of scientific journals. He has been project manager of many research projects. He is the author and co-author of several books and more than 150 research papers in the field of condition monitoring, diagnosis and prognosis and e-maintenance. He has a position as a visiting professor at the University of Sunderland.</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/ch10_Gregor.jpg"/></para>
<para><emphasis role="strong">Dr. Gregor Papa</emphasis> is a Senior Researcher and a Head of Computer Systems Department at the Jo&#x0017E;ef Stefan Institute, Slovenia, and an Associate Professor at the Jo&#x0017E;ef Stefan International Postgraduate School, Slovenia. His research interests include meta-heuristic optimisation methods and hardware implementations of high-complexity algorithms.</para>
<para>Gregor received the PhD degree in Electrical engineering from the University of Ljubljana, Slovenia, in 2002. His work is published in international journals and conference proceedings. He led and participated in several national and European projects. He is a member of the IEEE and ACM.</para>
<para>Gregor is a member of the Editorial Board of the Automatika journal (Taylor &#x00026; Francis) for the field &#x0201C;Applied Computational Intelligence&#x0201D;. He is a Consultant at the Slovenian Strategic research and innovation partnership for Smart cities and communities. In the context of the MANTIS project, he was a Work Package leader of &#x0201C;HMI design and development&#x0201D;.</para>
<para><graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="graphics/ch10_Urko.jpg"/></para>
<para><emphasis role="strong">Dr. Urko Zurutuza</emphasis> has coordinated the MANTIS European project from 2015 to 2018, related to Proactive Maintenance, and composed of 47 partners and an overall budget of around 30M&#x20AC;. Urko has been the manager of Data Analysis and Cybersecurity Research Group at Mondragon University since 2015. He has been responsible for coordinating the research activities in the fields of Intelligent Systems for Advanced Manufacturing, Intelligent Systems for Industrial Processes, Intelligent Systems for Cybersecurity and Intelligent Systems for the Health Industry.</para>
<para>Urko has participated in 30+ public funding research projects, authored 40+ publications, and has been member of 50+ conference program committees. He has coordinated research projects at regional, national and European levels, and has experience in FP6, FP7, H2020, ECSEL and AAL Joint Programmes.</para>
<para>Urko has taught subjects related to cybersecurity, data analysis, advanced manufacturing and research methodology at Mondragon University, in different Master and Doctoral degrees.</para>
</chapter>
</book>
