<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article
  PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "https://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article article-type="research-article" dtd-version="1.1" specific-use="sps-1.9" xml:lang="en" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
	<front>
		<journal-meta>
			<journal-id journal-id-type="publisher-id">iei</journal-id>
			<journal-title-group>
				<journal-title>Ingeniería e Investigación</journal-title>
				<abbrev-journal-title abbrev-type="publisher">Ing. Investig.</abbrev-journal-title>
			</journal-title-group>
			<issn pub-type="ppub">0120-5609</issn>
			<publisher>
				<publisher-name>Facultad de Ingeniería, Universidad Nacional de Colombia.</publisher-name>
			</publisher>
		</journal-meta>
		<article-meta>
                    <article-id pub-id-type="other">200</article-id>
			<article-id pub-id-type="doi">10.15446/ing.investig.v41n3.79308</article-id>
			<article-categories>
				<subj-group subj-group-type="heading">
					<subject>Original articles</subject>
				</subj-group>
			</article-categories>
			<title-group>
				<article-title>Full Model Selection Problem and Pipelines for Time-Series Databases: Contrasting Population-Based and Single-point Search Metaheuristics</article-title>
				<trans-title-group xml:lang="es">
					<trans-title>Problema de selección de modelo completo y tuberías para bases de datos de series de tiempo: contrastando metaheurísticas basadas en población y de un solo punto de búsqueda</trans-title>
				</trans-title-group>
			</title-group>
			<contrib-group>
				<contrib contrib-type="author">
					<contrib-id contrib-id-type="orcid">0000-0002-1831-6148</contrib-id>
					<name>
						<surname>Pérez-Castro</surname>
						<given-names>Nancy</given-names>
					</name>
					<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
				</contrib>
				<contrib contrib-type="author">
					<contrib-id contrib-id-type="orcid">0000-0002-0935-7642</contrib-id>
					<name>
						<surname>Acosta-Mesa</surname>
						<given-names>Héctor Gabriel</given-names>
					</name>
					<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
				</contrib>
				<contrib contrib-type="author">
					<contrib-id contrib-id-type="orcid">0000-0002-1565-5267</contrib-id>
					<name>
						<surname>Mezura-Montes</surname>
						<given-names>Efrén</given-names>
					</name>
					<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
				</contrib>
				<contrib contrib-type="author">
					<contrib-id contrib-id-type="orcid">0000-0002-0708-9875</contrib-id>
					<name>
						<surname>Cruz-Ramírez</surname>
						<given-names>Nicandro</given-names>
					</name>
					<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
				</contrib>
			</contrib-group>
			<aff id="aff1">
				<label>1 </label>
				<institution content-type="original">Ph.D. Artificial Intelligence, University of Veracruz Artificial Intelligence Research Institute, Mexico. Affiliation: Graduate of the PhD in Artificial Intelligence, University of Veracruz Artificial Intelligence Research Institute, Mexico. E-mail: naperez@uv.mx</institution>
				<institution content-type="normalized">Universidad Autónoma de Veracruz Villa Rica</institution>
				<institution content-type="orgname">University of Veracruz</institution>
				<country country="MX">Mexico</country>
				<email>naperez@uv.mx</email>
			</aff>
			<aff id="aff2">
				<label>2</label>
				<institution content-type="original"> Ph.D. Artificial Intelligence, University of Sheffield, Sheffield, UK. Affiliation: Research Professor, University of Veracruz Artificial Intelligence Research Institute, Mexico. E-mail: heacosta@uv.mx.</institution>
				<institution content-type="normalized">Universidad Autónoma de Veracruz Villa Rica</institution>
				<institution content-type="orgname">University of Sheffield</institution>
				<addr-line>
					<city>Sheffield</city>
				</addr-line>
				<country country="MX">Mexico</country>
				<email>heacosta@uv.mx</email>
			</aff>
			<aff id="aff3">
				<label>3</label>
				<institution content-type="original"> Ph.D. Computer Science, Center for Research and Advanced Studies of the National Polytechnic Institute (CINVESTAV-IPN), Mexico. Affiliation: Research Professor, University of Veracruz Artificial Intelligence Research Institute, Mexico. E-mail: emezura@uv.mx</institution>
				<institution content-type="normalized">Universidad Autónoma de Veracruz Villa Rica</institution>
				<institution content-type="orgname">University of Veracruz</institution>
				<country country="MX">Mexico</country>
				<email>emezura@uv.mx</email>
			</aff>
			<aff id="aff4">
				<label>4</label>
				<institution content-type="original"> Ph.D. Artificial Intelligence, University of Sheffield, Sheffield, UK. Affiliation: Research Professor, University of Veracruz Artificial Intelligence Research Institute, Mexico. E-mail: ncruz@uv.mx</institution>
				<institution content-type="normalized">Universidad Autónoma de Veracruz Villa Rica</institution>
				<institution content-type="orgname">University of Veracruz</institution>
				<country country="MX">Mexico</country>
				<email>ncruz@uv.mx</email>
			</aff>
			<pub-date date-type="pub" publication-format="electronic">
				<day>20</day>
				<month>08</month>
				<year>2021</year>
			</pub-date>
			<pub-date date-type="collection" publication-format="electronic">
				<season>Sep-Dec</season>
				<year>2021</year>
			</pub-date>
			<volume>41</volume>
			<issue>3</issue>
			<elocation-id>e200</elocation-id>
			<history>
				<date date-type="received">
					<day>25</day>
					<month>04</month>
					<year>2019</year>
				</date>
				<date date-type="accepted">
					<day>18</day>
					<month>03</month>
					<year>2021</year>
				</date>
			</history>
			<permissions>
				<license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/" xml:lang="en">
					<license-p>This is an open-access article distributed under the terms of the Creative Commons Attribution License</license-p>
				</license>
			</permissions>
			<abstract>
				<title>ABSTRACT</title>
				<p>The increasing production of temporal data, especially time series, has motivated valuable knowledge to understand phenomena or for decision-making. As the availability of algorithms to process data increases, the problem of choosing the most suitable one becomes more prevalent. This problem is known as the Full Model Selection (FMS), which consists of finding an appropriate set of methods and hyperparameter optimization to perform a set of structured tasks as a pipeline. Multiple approaches (based on metaheuristics) have been proposed to address this problem, in which automated pipelines are built for multitasking without much dependence on user knowledge. Most of these approaches propose pipelines to process non-temporal data. Motivated by this, this paper proposes an architecture for finding optimized pipelines for time-series tasks. A micro-differential evolution algorithm (,u-DE, population-based metaheuristic) with different variants and continuous encoding is compared against a local search (LS, single-point search) with binary and mixed encoding. Multiple experiments are carried out to analyze the performance of each approach in ten time-series databases. The final results suggest that the ,u-DE approach with <italic>rand/1/bin</italic> variant is useful to find competitive pipelines without sacrificing performance, whereas a local search with binary encoding achieves the lowest misclassification error rates but has the highest computational cost during the training stage.</p>
			</abstract>
			<trans-abstract xml:lang="es">
				<title>RESUMEN</title>
				<p>La creciente producción de datos temporales, especialmente de series de tiempo, ha motivado la extracción analítica de conocimiento valioso para comprender fenómenos o para la toma de decisiones. A medida que aumenta la disponibilidad de algoritmos para procesar datos, el problema de elegir el más adecuado se vuelve más frecuente. Este problema se conoce como la Selección del Modelo Completo (SMC), que consiste en encontrar un conjunto apropiado de métodos y la optimización de hiperparámetros para realizar un conjunto de tareas estructuradas como una tubería. Se han propuesto múltiples enfoques (basados en metaheurísticas) para abordar este problema, en los que se construyen tuberías automatizadas para realizar múltiples tareas sin mucha dependencia del conocimiento del usuario. La mayoría de estos enfoques proponen tuberías para procesar datos no temporales. Motivado por esto, este artículo propone una arquitectura para encontrar tuberías optimizadas para tareas de series de tiempo. El algoritmo de micro-Evolución Diferencial (,u-ED, metaheurística basada en población) con diferentes variantes y codificación continua, es comparado contra una búsqueda local (BL, búsqueda de un solo punto) con codificación binaria y mixta. Se realizan múltiples experimentos para analizar el rendimiento de cada enfoque en diez bases de datos de series de tiempo. Los resultados finales sugieren que el enfoque de ,u-ED con la variante <italic>rand/1/bin</italic> es útil para encontrar tuberías competitivas sin sacrificar el rendimiento, mientras que la BL con codificación binaria logra las tasas de error de clasificación incorrecta más bajas, pero tiene el costo computacional más alto durante la etapa de entrenamiento.</p>
			</trans-abstract>
			<kwd-group xml:lang="en">
				<title>Keywords:</title>
				<kwd>full model selection</kwd>
				<kwd>time series</kwd>
				<kwd>metaheuristics</kwd>
			</kwd-group>
			<kwd-group xml:lang="es">
				<title>Palabras clave:</title>
				<kwd>selección del modelo completo</kwd>
				<kwd>series de tiempo</kwd>
				<kwd>metaheurísticas</kwd>
			</kwd-group>
			<counts>
				<fig-count count="18"/>
				<table-count count="7"/>
				<equation-count count="7"/>
				<ref-count count="56"/>
				<page-count count="17"/>
			</counts>
		</article-meta>
	</front>
	<body>
		<sec sec-type="intro">
			<title>Introduction</title>
			<p>In recent years, the ability to generate and store data has far outpaced the capability to analyze and exploit it (<xref ref-type="bibr" rid="B45">Rydning 2018</xref>). According to <xref ref-type="bibr" rid="B20">Gantza and Reisel (2012)</xref>, just 3% of global data are currently tagged and ready for manipulation, and only 0,5% of this is used for analysis, at least in 2012.</p>
			<p>Therefore, the interest in analyzing and extracting useful information to understand phenomena or for decision-making has brought the attention of practitioners and the research community. The increasing production of temporal data, especially time series, has motivated the analysis for extracting valuable knowledge through knowledge discovery in databases (KDD) processes and data mining (DM) techniques (<xref ref-type="bibr" rid="B50">Sun, Yang, Liu, Chen, Rao, and Bai 2019</xref>, <xref ref-type="bibr" rid="B9">Boulle, Dallas, Nakatsukasa, and Samaddar 2020</xref>).</p>
			<p>Time series are an important class of temporal data objects, and they can be easily obtained from scientific research (<xref ref-type="bibr" rid="B19">Fu 2011</xref>) and other domains such as medicine, engineering, earth and planetary sciences, physics and astronomy, mathematics, environmental sciences, biochemistry, genetic and molecular biology, agricultural and biological sciences, among others.</p>
			<p>
				<xref ref-type="fig" rid="f1">Figure 1</xref> shows a scientific document analysis by subject areas where time series have been used, especially in classification tasks during the last seven years, obtained from the Elsevier-Scopus database, where 7 973 articles were considered.</p>
			<p>
				<fig id="f1">
					<label>Figure 1</label>
					<caption>
						<title>Analysis of the time-series usage that has been reported in scientific documents in the last seven years. </title>
					</caption>
					<graphic xlink:href="0120-5609-iei-41-03-e200-gf1.gif"/>
					<attrib>Source: Authors</attrib>
				</fig>
			</p>
			<p>Time series <italic>T = (t</italic>
 <sub>1</sub>
 <italic>,...,t</italic>
 <sub>
 <italic>n</italic>
</sub> 
 <italic>)</italic> ∈ℛ is the result of the observation of the underlying procedure in which a set of values is collected through measurements made into uniformly spaced time-instants. Therefore, a time series can be defined as an ordered sequence of <italic>n</italic> real-valued variables (<xref ref-type="bibr" rid="B18">Esling and Agon 2012</xref>, <xref ref-type="bibr" rid="B27">Jastrzebska 2019</xref>).</p>
			<p>A wide variety of DM techniques has been proposed and applied to dealing diverse tasks in domains where time series can be involved (<xref ref-type="bibr" rid="B24">Gong, Chen, Yuan, and Yao2019</xref>, <xref ref-type="bibr" rid="B27">Jastrzebska 2019</xref>, <xref ref-type="bibr" rid="B2">Ali, Alqahtani, Jones, and Xie 2019</xref>).</p>
			<p>However, classical DM techniques often perform poorly in the presence of time-series data, because most of them treat time-series as unrelated data, thus resulting in inaccurate or inconsistent models (<xref ref-type="bibr" rid="B38">Rashid and Hossain 2012</xref>).</p>
			<p>To overcome the disadvantages of traditional DM techniques with time series, a set of techniques has been proposed which are part of Temporal Data Mining (TDM). TDM has a huge array of techniques for tackling tasks such as query by content, clustering, classification, segmentation, and others (<xref ref-type="bibr" rid="B55">Yang 2017</xref>).</p>
			<p>When time series data are involved in the data mining process, the quality of the mined data can depend on two important issues: the first is the choice of the appropriate algorithm for a given task, while the second is the proper hyper-parameter selection that may produce a relatively good performance.</p>
			<p>Both issues are known as algorithm selection (AS), and model selection (MS), and these are often solved separately. Nevertheless, there are some proposals which have addressed both AS and MS at the same time under the issue known as full model selection (FMS) (<xref ref-type="bibr" rid="B17">Escalante, Montes, and Sucar 2009</xref>).</p>
			<p>Therefore, FMS consists of finding an appropriate set of methods and their hyperparameter optimization for multitasking. This combination can be represented as a kind of pipeline, characterized by avoiding the dependency on user knowledge (<xref ref-type="bibr" rid="B26">Hutter, Kotthoff, and Vanschoren 2019</xref>).</p>
			<p>Multiple approaches have been proposed to find automated pipelines according to the hyperparameter optimization process (<xref ref-type="bibr" rid="B56">Yu and Zhu 2020</xref>). These approaches can be categorized into three main classes: a) approaches based on exhaustive traditional search (<xref ref-type="bibr" rid="B6">Bergstraand Bengio 2012</xref>), b) approaches based on Bayesian optimization (<xref ref-type="bibr" rid="B49">Shahriari, Swersky, Wang, Adams, and de Freitas 2016</xref>), and c) approaches based on metaheuristics (<xref ref-type="bibr" rid="B26">Hutter <italic>et al.</italic> 2019</xref>).</p>
			<p>The first class of these approaches can be impractical and costly because the search focuses on exhaustive exploration defined for a particular block of the pipeline. In contrast to exhaustive approaches, Bayesian approaches keep track of past evaluation results, which they use to find better model settings than random search in fewer iterations. The major drawback of Bayesian optimization approaches is that inference time grows cubically in the number of observations.</p>
			<p>Metaheuristics represent a flexible option that has been increasingly used to build optimized pipelines. Population-based metaheuristics such as evolutionary or swarm intelligence algorithms have been adopted to propose an automatic framework that finds streamlined pipelines (<xref ref-type="bibr" rid="B51">Sun, Pfahringer, and Mayo 2013</xref>, <xref ref-type="bibr" rid="B32">Olson, Urbanowicz, Andrews, Lavender, Kidd, and Moore 2016</xref>, <xref ref-type="bibr" rid="B13">de Sá, Pinto, Oliveira, and Pappa 2017</xref>).</p>
			<p>Most of the population-based metaheuristic approaches have focused on building pipelines for databases in which the temporary factor is not considered. Therefore, approaches dealing with FMS while involving the building of time series pipelines are scarce.</p>
			<p>Single-point search, a part of metaheuristics, has been used for search optimized structures or hyperparameter selection (<xref ref-type="bibr" rid="B3">Aly, Guadagni, and Dugan 2019</xref>). Local search is an example of a single-point search that has turned out to be a practical option to solve complex problems despite being the most straightforward.</p>
			<p>In this paper, an architecture is proposed for finding an optimized pipeline for time series databases in which the FMS problem is related. It is empirically studied from two points of view: the first, from a population-based approach, where μ-DE is used as a search engine; and the second, from a single-point search, where a local search is adopted.</p>
			<p>The main objectives of this work are to empirically study the proposed architecture, varying the search engine and solution encoding; and to offer an alternative that automatically assists the selection of an optimized pipeline for time series database tasks, <italic>i.e.,</italic> to solve the FMS problem for time series.</p>
			<sec>
				<title>Related works</title>
				<p>From the literature review, it is essential to note that FMS is not a new trend. Since the 90s, solutions have emerged to deal with the issue of selecting an algorithm from a portfolio of options in order to carry out a single task (<xref ref-type="bibr" rid="B40">Rice 1976</xref>).</p>
				<p>Subsequently, the need arises to incorporate more tasks into said selection (multi-task) and deal with hyperparameter optimization, resulting in machine learning pipelines (<xref ref-type="bibr" rid="B26">Hutter <italic>et al.</italic> 2019</xref>).</p>
				<p>Nowadays, learning pipelines are developed to be truly usable by a non-expert. Against this background, a need for automated machine learning (AutoML, a recently coined term) systems can be used to handle various tasks and solve the FMS problem, a challenging and time-consuming process.</p>
				<p>Grid search, random search, Bayesian optimization, and metaheuristics are four conventional approaches to building AutoML systems for diverse applications (<xref ref-type="bibr" rid="B6">Bergstra and Bengio 2012</xref>). Grid search and random search are traditional hyperparameter optimization methods that could prove impractical to explore high-dimensional spaces at a high computational cost.</p>
				<p>Bayesian optimization has been effective in this realm and has even outperformed manual hyperparameter tuning by expert practitioners. Auto-WEKA (<xref ref-type="bibr" rid="B25">Hall <italic>et al.</italic> 2009</xref>), mlr (<xref ref-type="bibr" rid="B7">Bischl <italic>et al.</italic> 2016</xref>) and auto-SKLearn (<xref ref-type="bibr" rid="B36">Pedregosa <italic>et al.</italic> 2011</xref>) are approaches based on Bayesian optimization, and their prime objective is to find the best combination between complete learning pipelines and their respective parameters.</p>
				<p>Both approaches follow a hierarchical method that first chooses a particular algorithm and, only after this step, optimizes its parameters. Thus, algorithms may be left out which, with the right hyperparameters, could generate better results than the selected ones.</p>
				<p>On the other hand, metaheuristics, especially evolutionary and swarm intelligence algorithms, have gained a particular interest in the research community by allowing the construction of machine learning pipelines that can be complex and extensive.</p>
				<p>In the rest of this section, a set of metaheuristics-based approaches for AutoML are described.</p>
				<sec>
					<title><italic>Metaheuristics-based approaches</italic></title>
					<p>In 2009, <xref ref-type="bibr" rid="B17">Escalante <italic>etal.</italic> (2009)</xref> proposed a machine learning pipeline that included selecting a preprocessing algorithm, a feature selection algorithm, a classifier and, all their hyper-parameters. Their approach used a modified Particle Swarm Optimization (PSO) to deal with the limited configuration space and was called PSMS system. Although the authors found that they could apply their method to different datasets without domain knowledge, most of the datasets used had unrelated attributes. In order to avoid overfitting, the authors proposed using k-cross-validation, and then the approach was extended with a custom assembling strategy that combined the best solutions from multiple generations (<xref ref-type="bibr" rid="B16">Escalante, Montes, and Sucar 2010</xref>).</p>
					<p>Later, Sun <italic>et al.</italic> extended the idea of PSMS and proposed the unification of the PSO algorithm and the Genetic Algorithm (GA) (2013). This approach was called GPS (which stands for GA-PSO-FMS). A GA were was used to optimize the pipeline structure, while the PSO for the hyperparameter optimization of each pipeline. The pipeline proposed by the authors included selecting from a pool of methods such as data sampling, data cleansing, feature transformation, feature selection, and classification. The datasets used for evaluating GPS were characterized by a high number of instances, thus causing an increase in the computational cost during the loss function evaluation. Therefore, the authors proposed the use of an internal binary tree structure to speed up the GPS system.</p>
					<p>Another interesting line of research is the application of multi-objective evolutionary algorithms. One of these approaches is the Multi-objective Support Vector Machine Model Selection (MOSVMMS) (<xref ref-type="bibr" rid="B41">Rosales-Pérez, Escalante, Gonzalez, Reyes-Garcia, and Coello-Coello 2013</xref>), where the search is guided by a Non-dominated Sorted Genetic Algorithm-II (NSGA-II).</p>
					<p>The authors built a pipeline formed by feature selection, pre-processing, and classification tasks focused only in the SVM classifier. The models were evaluated under bias and variance trade-off as prime objective functions. This approach was only tested on thirteen binary classification problems. Two extensions of this approach were reported, the first called Multi-Objective Model Type Selection (MOMTS), where a multi-objective evolutionary algorithm based on decomposition (MOEA/D) was used instead of the NSGA-II (<xref ref-type="bibr" rid="B43">Rosales-Pérez, Gonzalez, Coello-Coello, Escalante, and Reyes-Garcia 2014</xref>). MOMTS focused only on selecting classification models without other involved tasks. However, the authors explored the idea of measuring complexity models through the Vapnik-Chervonenkis dimension, which could have a high computational cost as the dimension of the datasets grows.</p>
					<p>For that reason, the second extension proposed by Rosales-Pérez, Gonzalez, Coello, Escalante, and Reyes-Garcia was the Surrogate Assisted Multi-Objective Model Selection (SAMOMS) (2015), in which a pipeline structure is considered, (preprocessing, feature selection, and classification). They proposed a surrogate assistant to speed up the fitness evaluation.</p>
					<p>The Tree-Based Pipeline Optimization Tool (TPOT) is an open-source software package for configuring pipelines in a more flexible manner (<xref ref-type="bibr" rid="B32">Olson <italic>et al.</italic> 2016</xref>). TPOT uses a genetic programming algorithm for optimizing structures and hyperparameters. The main operator included in TPOT has supervised classification, feature preprocessing operators, and feature selection operators, all ofthem taken from scikit-learn. The main drawback of TPOT is considering unconstrained search, where resources can be spent on generating and evaluating invalid solutions.</p>
					<p>So far, these approaches do not present evidence of the treatment of time-series databases. Most of them use a fixed pipeline length in sequential steps. TPOT is, to date, the approach that stands out for optimizing the design of pipes. Early efforts for an approach that suggests automated pipelines for time series can be found a previous work proposed by <xref ref-type="bibr" rid="B37">Pérez-Castro, Acosta-Mesa, Mezura-Montes, and Cruz-Ramírez (2015)</xref>. The authors proposed using a micro version of differential evolution to solve the FMS problem, and they suggested optimized pipelines. In that work, smoothing, time series representation, and classification through the k-nearest neighbor algorithm are only considered. This work is an extension of the work mentioned above.</p>
				</sec>
			</sec>
			<sec>
				<title>FMS problem in time-series databases</title>
				<p>The FMS term, conceived by <xref ref-type="bibr" rid="B17">Escalante <italic>et al.</italic> (2009)</xref>, consists of selecting a combination of suitable methods to obtain a learning pipeline for a particular database with a low generalization error.</p>
				<p>In this paper, the FMS problem in time series databases is tackled as a single-objective optimization problem, defined by <xref ref-type="disp-formula" rid="e3">Equation (1)</xref>, based on the definition made by <xref ref-type="bibr" rid="B14">Díaz-Pacheco, Gonzalez-Bernal, Reyes-García, and Escalante-Balderas (2018)</xref>, which consists of searching for suitable pipeline composed by a smoothing <italic>s*</italic>
 <sub>
 <italic>λ</italic>
</sub> 
 <italic>Є S,</italic> a time series representation <italic>r*</italic>
 <sub>
 <italic>λ</italic>
</sub> 
 <italic>Є R,</italic> a numerosity reduction <italic>e*</italic>
 <sub>
 <italic>λ</italic>
</sub> 
 <italic>Є E,</italic> and classification method <italic>c*</italic>
 <sub>
 <italic>λ</italic>
</sub> 
 <italic>Є C</italic> with their related hyper-parameter setting A from the corresponding domain space ∧.</p>
				<p>For each pipeline, a loss function ℒ is estimated over a labeled time-series database <inline-formula id="e1">
						<inline-graphic xlink:href="0120-5609-iei-41-03-e200-ie1.gif"/>
					</inline-formula><italic>,</italic> where for <inline-formula id="e2">
						<inline-graphic xlink:href="0120-5609-iei-41-03-e200-ie2.gif"/>
					</inline-formula><italic>,</italic> which denotes an ordered sequence of n real-valued variables (univariate time series), and y<sub>i</sub> e Y for the corresponding label value.</p>
				<p>In order to build pipelines with a low generalization error, database D is divided into <italic>k</italic> disjoint partitions <italic>(D</italic>
 <sup>
 <italic>(i)</italic>
</sup> 
 <sub>
 <italic>t</italic>
</sub> and <italic>D</italic>
 <sup>
 <italic>(i)</italic>
</sup> 
 <sub>
 <italic>v</italic> 
</sub> for i = 1,2,...,k).</p>
				<p>
					<disp-formula id="e3">
						<graphic xlink:href="0120-5609-iei-41-03-e200-e3.gif"/>
					</disp-formula>
				</p>
				<p>Where, <italic>S</italic> is the set of available smoothing methods; <italic>R</italic> is the set of available time series representation methods; <italic>E</italic> is the set of available numerosity reduction methods; C is the set of available classifiers; λ is a vector of hyperparameters; <italic>D</italic>
 <sub>
 <italic>t</italic>
</sub> is a training data partition; <italic>D</italic>
 <sub>
 <italic>v</italic>
</sub> is a validation partition; ℒ is a loss function computed on the validation set; and <italic>argmin</italic> returns the lowest values estimated by the loss function.</p>
			</sec>
		</sec>
		<sec sec-type="methods">
			<title>Methodology overview</title>
			<sec>
				<title><italic>Materials</italic></title>
				<sec>
					<title><italic>Benchmark databases</italic></title>
					<p>In this article, a part of the well-known collection of univariate time series databases is used (<xref ref-type="bibr" rid="B29">Keogh <italic>et al.</italic> 2011</xref>). The essential characteristics of those databases are summarized in <xref ref-type="table" rid="t1">Table 1</xref>.</p>
					<p>
						<table-wrap id="t1">
							<label>Table 1</label>
							<caption>
								<title>Time-series databases description</title>
							</caption>
							<graphic xlink:href="0120-5609-iei-41-03-e200-gt1.gif"/>
							<table-wrap-foot>
								<fn id="TFN1">
									<p>Source: Authors</p>
								</fn>
							</table-wrap-foot>
						</table-wrap>
					</p>
					<p>A brief description of each database is presented below:</p>
					<p>
						<list list-type="bullet">
							<list-item>
								<p><italic>Beef:</italic> This database consists of five classes of beef spectrograms acquired from raw samples, cooked using two different cooking regimes. Each beef class represents a differing degree of contamination with offal (<xref ref-type="bibr" rid="B1">Al-Jowder, Kemsley, and Wilson 2002</xref>).</p>
							</list-item>
							<list-item>
								<p><italic>CBF:</italic> Cylinder-Bell-Funnel is a simulated database where each class is standard normal noise plus an offset term that differs for each category (<xref ref-type="bibr" rid="B46">Saito 2000</xref>).</p>
							</list-item>
							<list-item>
								<p><italic>Coffee:</italic> The coffee database consists of two spectrograms class that distinguish between Robusta and Arabica coffee beans (<xref ref-type="bibr" rid="B4">Bagnall, Davis, Hills, and Lines 2012</xref>).</p>
							</list-item>
							<list-item>
								<p><italic>ECG200:</italic> The electrocardiogram (ECG) database contains the measurements recorded by one electrode during one heartbeat. The two classes correspond to a normal heartbeat and a myocardial infarction, respectively (<xref ref-type="bibr" rid="B33">Olszewski 2001</xref>).</p>
							</list-item>
							<list-item>
								<p><italic>FaceFour:</italic> This database was built from face profile images. Each time series was obtained by converting a local (outer) angle at every point x of the face profile contour, starting from the head profile's neck area (<xref ref-type="bibr" rid="B29">Keogh <italic>et al.</italic> 2011</xref>).</p>
							</list-item>
							<list-item>
								<p><italic>Gun-Point:</italic> This database was obtained from motions with hands involving one female actor and one male actor. Two classes were identified: Gun-Draw (actors point the gun at a target for approximately one second) and Gun-Point (actors point with their index fingers to a goal for about one second). Each time series corresponds to the centroid of the actor's right hands in the x-axis (<xref ref-type="bibr" rid="B39">Ratanamahatana and Keogh 2005</xref>).</p>
							</list-item>
							<list-item>
								<p><italic>Lightning-2 and Lightning-7:</italic> The FORTE satellite detects transient electromagnetic events associated with lightning using a suite of optical and radio-frequency (RF) instruments. Data is collected with a sample rate of 50 MHz for 800 microseconds that are transformed into spectrograms, which are collapsed in frequency to produce a power density time series, with 3 181 samples in each time series. These are then smoothed to produce time series of length 637 and 319 (<xref ref-type="bibr" rid="B15">Eads <italic>et al.</italic> 2002</xref>).</p>
							</list-item>
							<list-item>
								<p><italic>OliveOil:</italic> This is another example of the food spectro-graphs used in chemometrics to classify food types. Each class of this database corresponds to virgin olive oils originating from four European producing countries.</p>
							</list-item>
							<list-item>
								<p><italic>Trace:</italic> It is a synthetic database created by Davide Roverso and designed to simulate instrumentation failures in a nuclear power plant. All instances are linearly interpolated to have the same length of 275 data point (<xref ref-type="bibr" rid="B44">Roverso 2000</xref>).</p>
							</list-item>
						</list>
					</p>
					<p>It can be seen that most databases describe real phenomena. A behavior analysis of time-series databases was carried out. This analysis consisted of observing three characteristics: a) class separation (CS), b) noise level (NL), and c) the similarity between the training and testing sets (SBS). The mean, median, and average of these per class for each database were computed and plotted from raw databases. From the visualization, the three characteristics of the above-listed were ranked. CS 1 means non-separable, and a value of 3 means easily separable. NL can take values between 1-5, where 1 means low noise and five high noise. SBM was measured in a range of 1 to 3, where 1 represents low similarity, and 3 means high similarity. The results of this analysis are summarized in <xref ref-type="table" rid="t2">Table 2</xref>.</p>
					<p>
						<table-wrap id="t2">
							<label>Table 2</label>
							<caption>
								<title>Characteristics of visual analysis in time-series databases</title>
							</caption>
							<graphic xlink:href="0120-5609-iei-41-03-e200-gt2.gif"/>
							<table-wrap-foot>
								<fn id="TFN2">
									<p>Note: Acronyms: CS (Class Separation); NL (Noise Level); SBS (similarity between the training and testing sets) </p>
								</fn>
								<fn id="TFN3">
									<p>Source: Authors</p>
								</fn>
							</table-wrap-foot>
						</table-wrap>
					</p>
					<p>It is important to note that the data was not pre-processed for the experimental stage.</p>
				</sec>
				<sec>
					<title><italic>Pipeline tasks: available methods</italic></title>
					<p>In this work, four main tasks are considered to build a learning pipeline for time series which involves solving the FMS problem.</p>
					<p>
						<list list-type="order">
							<list-item>
								<p><italic>Smoothing:</italic> It is usually used to soften out the irregular roughness to see a clearer signal. This task does not provide a model, but it can be a promising first step in describing various series components (<xref ref-type="bibr" rid="B23">Giron-Sierra 2018</xref>). It is common to use the term filter to describe the smoothing procedure. Moving Average (<xref ref-type="bibr" rid="B5">Baijal, Singh, Rani, and Agarwal 2016</xref>), the <xref ref-type="bibr" rid="B48">Savitzky-Golay filter (Savitzky and Golay 1964</xref>), and Local Regression (with and without weights) are considered with related hyper-parameters (<xref ref-type="bibr" rid="B12">Cleveland and Loader 1996</xref>).</p>
							</list-item>
							<list-item>
								<p><italic>Time-series representation:</italic> This task consists of transforming the time series to another domain to reduce dimensionality, followed by an indexing mechanism. Piecewise Aggregate Approximation (PAA) (<xref ref-type="bibr" rid="B28">Keogh, Chakrabarti, Pazzani, and Mehrotra 2001</xref>), Symbolic Aggregate approximation (SAX) (<xref ref-type="bibr" rid="B30">Lin, Keogh, Wei, and Lonardi 2007</xref>) and Principal Component Analysis (PCA) are available methods (<xref ref-type="bibr" rid="B34">Page, Lischeid, Epting, and Huggenberger 2012</xref>).</p>
							</list-item>
							<list-item>
								<p><italic>Numerosity reduction:</italic> It is a procedure used to reduce data volume by using suitable forms of data representation. The Monte Carlo 1 (MC1) and INSIGHT approaches were included (<xref ref-type="bibr" rid="B21">Garcia, Derrac, Cano, and Herrera 2012</xref>, <xref ref-type="bibr" rid="B10">Buza, Nanopoulos, and Schmidt-Thieme 2011</xref>). </p>
							</list-item>
							<list-item>
								<p><italic>Clasification:</italic> It is a machine learning supervised task that consists of identifying the category to which a new observation belongs, based on a training set of data containing examples whose category membership is known. Some classifiers were considered, such as k-nearest neighbors, Naive Bayes, among others (<xref ref-type="bibr" rid="B8">Bishop 2006</xref>).</p>
							</list-item>
						</list>
					</p>
					<p>
						<xref ref-type="table" rid="t3">Table 3</xref> shows a summary of the available methods for each pipeline tasks and their related hyperparameters.</p>
					<p>
						<table-wrap id="t3">
							<label>Table 3</label>
							<caption>
								<title>Description of available methods for each pipeline task and the description of their hyperparameters (values in square brackets indicate lower and upper limits)</title>
							</caption>
							<graphic xlink:href="0120-5609-iei-41-03-e200-gt3.gif"/>
							<table-wrap-foot>
								<fn id="TFN4">
									<p>Note: Acronyms: S (Smoothing); R (Representation); NR (Numerosity Reduction); C (Classification); TLS (Time-series length); SG (Savitzky-Golay Filter); MA (Moving Average); LRw (Local Regression-lowess); LRe (Local Regression-loess); SAX(Aggregate approximation); PAA (Piecewise Aggregate Approximation); PCA (Principal Component Analysis); INSIGHT (Instance Selection based on Graph-coverage and Hubness for Time-series); MC1 (Monte Carlo 1); KNN-ED (K-Nearest Neighbor-Euclidean Distance); KNN-LBDTW (K-Nearest Neighbor-Lower Bounding Dynamic Time Warping); NB (Naive Bayes); DT (Decision Tree); AB (AdaBoost); SVM (Support Vector Machine) </p>
								</fn>
								<fn id="TFN5">
									<p>Source: Authors</p>
								</fn>
							</table-wrap-foot>
						</table-wrap>
					</p>
				</sec>
				<sec>
					<title><italic>Encoding pipeline solutions</italic></title>
					<p>A candidate solution for the learning pipeline for time-series databases is represented as a vector in this work. Each vector can be formed by continuous values, binary values, or mixed values (continuous and discrete values).</p>
					<p><italic>Continuous encoding:</italic> Each potential solution is encoded as a continuous vector which is formed as in <xref ref-type="disp-formula" rid="e4">Equation (2)</xref>.</p>
					<p>
						<disp-formula id="e4">
							<graphic xlink:href="0120-5609-iei-41-03-e200-e4.gif"/>
						</disp-formula>
					</p>
					<p>Where depicts each position within a particular vector; and <italic>X</italic>
 <sub>
 <italic>j,s</italic>
</sub> 
 <italic>Є</italic> [1;4], <italic>X</italic>
 <sub>
 <italic>j,r</italic>
</sub> 
 <italic>Є</italic> [1;3], <italic>X</italic>
 <sub>
 <italic>j,e</italic>
</sub> 
 <italic>Є</italic> [1;2], and <italic>x</italic>
 <sub>
 <italic>j</italic>
</sub> 
 <italic>,</italic>
 <sub>
 <italic>c</italic>
</sub> 
 <italic>Є</italic> [1;6] represent the ID of the selected smoothing, time-series representation, numerosity reduction, and classification available methods, respectively.</p>
					<p><italic>h</italic>
 <sub>
 <italic>j,1</italic>
</sub> 
 <italic>...,ns, h</italic>
 <sub>
 <italic>j,1,…,</italic>
</sub> 
 <italic>nr, h</italic>
 <sub>
 <italic>j,1,…,ne</italic>
</sub> 
 <italic>,</italic> and <italic>h</italic>
 <sub>
 <italic>j</italic>
</sub> 
 <italic>,</italic>
 <sub>
 <italic>1,…,nc</italic>
</sub> encode the set of hyperparameters related to the overall available methods, where ns, nr, ne, and <italic>nc</italic> represent the number of hyperparameters per type of task into the learning pipeline that has different limits. Each position can take random continuous values according to Equation (3), which determines a value between the lower and upper bounds of each hyperparameter, described in <xref ref-type="table" rid="t3">Table 3</xref>.</p>
					<p>
						<disp-formula id="e5">
							<graphic xlink:href="0120-5609-iei-41-03-e200-e5.gif"/>
						</disp-formula>
					</p>
					<p>In <xref ref-type="disp-formula" rid="e5">Equation (3)</xref>, <italic>lb</italic>
 <sub>
 <italic>i</italic>
</sub> is a lower bound, <italic>ubi</italic> an upper bound, and <italic>rand</italic> represents a random value between 0 and 1. <xref ref-type="fig" rid="f3">Figure 3</xref> shows an example of the structure of a continuous vector solution. Black boxes represent the positions that encode the selected method according to the task in the learning pipeline (smoothing, time-series representation, numerosity reduction, and classification), and the white boxes encode their related hyperparameters. For continuous encoding, vectors of 24 dimensions are considered to represent a learning pipeline, which is equivalent to a candidate solution of FMS problem for time-series databases.</p>
					<p>
						<fig id="f2">
							<label>Figure 2</label>
							<caption>
								<title>Graphical representation of a solution encoding used in contentious or mixed encoding. </title>
							</caption>
							<graphic xlink:href="0120-5609-iei-41-03-e200-gf2.gif"/>
							<attrib>Source: Authors</attrib>
						</fig>
					</p>
					<p><italic>Mixed encoding:</italic> Mixed encoding consists of a vector of 24 dimensions, as shown in <xref ref-type="fig" rid="f2">Figure 2</xref>. However, for this option, both continuous and discrete values are permitted.</p>
					<p>Continuous values are generated by <xref ref-type="disp-formula" rid="e5">Equation (3)</xref>, according to the limits of each position. In contrast, discrete values are generated by <italic>randi(),</italic> the function of MATLAB language that returns integer values drawn from a discrete uniform distribution, where limits are also respected.</p>
					<p><italic>Binary encoding:</italic> Binary encoding consists of a vector formed by binary values (0 or 1). These values can be grouped into binary strings that represent continuous or discrete values.</p>
					<p>The length of a particular binary string depends on the boundary of values to be expressed. Binary string length <italic>lj</italic> is computed with <xref ref-type="disp-formula" rid="e6">Equation (4)</xref>, where <italic>int</italic> expresses a integer value, <italic>log</italic>
 <sub>
 <italic>2</italic>
</sub> is the log base 2, <italic>ub</italic> the upper boundary, <italic>lb</italic> the lower boundary, and <italic>precision</italic> is a constant that means the number of decimal places to encode.</p>
					<p>
						<disp-formula id="e6">
							<graphic xlink:href="0120-5609-iei-41-03-e200-e6.gif"/>
						</disp-formula>
					</p>
					<p>Then, the overall binary vector length <italic>bvl</italic> to encode a potential pipeline solution is the concatenation of each binary string. <xref ref-type="disp-formula" rid="e7">Equation (5)</xref> states how it is computed, where D is the amount of continuous or discrete values that can be encoded as binary strings, and <italic>l</italic>
 <sub>
 <italic>j</italic>
</sub> is the maximum length of these binary strings.</p>
					<p>
						<disp-formula id="e7">
							<graphic xlink:href="0120-5609-iei-41-03-e200-e7.gif"/>
						</disp-formula>
					</p>
					<p>If a mixed vector structure is considered containing 24 values that represent a potential pipeline solution, which respects the boundaries of the values presented in <xref ref-type="table" rid="t3">Table 3</xref>, then vector with a length of 169 positions is required. It can be seen in <xref ref-type="fig" rid="f3">Figure 3</xref> that the first three binary values correspond to a binary string representing integer values between 1 and 4 that are the number of available smoothing methods. The next binary strings encode the rest of the values.</p>
					<p>
						<fig id="f3">
							<label>Figure 3</label>
							<caption>
								<title>Graphical representation of a binary encoding. </title>
							</caption>
							<graphic xlink:href="0120-5609-iei-41-03-e200-gf3.gif"/>
							<attrib>Source: Authors</attrib>
						</fig>
					</p>
					<p>A decoding process is needed to compute the quality of each binary encode solution. Decoding is performed for each binary string x<sub>s</sub> of complete binary vector according to <xref ref-type="disp-formula" rid="e8">Equation (6)</xref>, where <italic>lbj</italic> is the lower boundary used for this binary string, <italic>ub</italic>
 <sub>
 <italic>j</italic>
</sub> is the upper boundary used for this binary string, x<sub>int</sub> is the result of traditional binary to decimal conversion, and <italic>l</italic>
 <sub>
 <italic>j</italic>
</sub> is the binary string length performed obtained from (4) </p>
					<p>
						<disp-formula id="e8">
							<graphic xlink:href="0120-5609-iei-41-03-e200-e8.gif"/>
						</disp-formula>
					</p>
				</sec>
				<sec>
					<title><italic>Fitness function</italic></title>
					<p>The Cross Validation Error Rate (CVER) is used as the fitness function <italic>f</italic>
 <sub>
 <italic>x</italic>
</sub> to evaluate the quality of a learning pipeline under a particular time-series database. <xref ref-type="disp-formula" rid="e9">Equation (7)</xref> describes f<sub>x</sub>, where <italic>a</italic> represents the portion of instances in the time-series database that was incorrectly classified, and <italic>b</italic> is the total number of instances in such database. k depicts the number of stratified subsamples (folds) chosen randomly but with roughly equal size in the cross validation method that is adopted to avoid over-fitting.</p>
					<p>
						<disp-formula id="e9">
							<graphic xlink:href="0120-5609-iei-41-03-e200-e9.gif"/>
						</disp-formula>
					</p>
				</sec>
			</sec>
			<sec>
				<title><italic>Methods: search engines</italic></title>
				<sec>
					<title><italic>Micro Differential Evolution (μ-DE)</italic></title>
					<p>Population-based metaheuristics such as evolutionary algorithms have a reduced population version that has proven to be efficient for solving large scale optimization problems (<xref ref-type="bibr" rid="B31">Olguín-Carbajal <italic>et al.</italic> 2019</xref>, <xref ref-type="bibr" rid="B47">Salehinejad, Rahnamayan, and Tizhoosh 2017</xref>). The reduced population versions usually are denoted with the prefix u Besides the small population, μ algorithms are characterized by a restart mechanism to avoid stagnation.</p>
					<p>The μ-DE cycle and conventional operations, based on the scaled difference between two vectors of a population set, remain the same as in the classical DE. Usually, the population size in μ -DE can take a value between four and six vectors (<xref ref-type="bibr" rid="B53">Viveros Jiménez, Mezura Montes, and Gelbukh 2012</xref>, <xref ref-type="bibr" rid="B11">Caraffini, Neri, and Poikolainen 2013</xref>). Regarding the restart mechanism, μ -DE requires randomly replacing the N worst vectors each R generations. In this paper, the u-DE proposed by <xref ref-type="bibr" rid="B35">Parsopoulos (2009)</xref> is used as a population-based metaheuristic.</p>
					<p>
						<xref ref-type="fig" rid="ch1">Algorithm 1</xref> summarizes the main steps of the adopted u-DE. Step six shows the mutation and combination process, for that, different variants such as <italic>rand/1/bin, rand/1/exp, best/1/bin,</italic> and <italic>best/1/exp</italic> are used in the experimentation.</p>
					<p>
						<fig id="ch1">
							<graphic xlink:href="0120-5609-iei-41-03-e200-gch1.gif"/>
						</fig>
					</p>
					<p>This results in four versions to solve FMS problem, called P-DEMS1 (Population-Differential Evolution Model Selection 1, using <italic>rand/1/bin),</italic> P-DEMS2 (Population-Differential Evolution Model Selection 2, using <italic>rand/1/exp),</italic> P-DEMS3 (Population-Differential Evolution Model Selection 3, using <italic>best/1/bin),</italic> and P-DEMS4 (Population-Differential Evolution Model Selection 4, using <italic>best/1/exp).</italic></p>
				</sec>
				<sec>
					<title><italic>Local search (LS)</italic></title>
					<p>LS, a single-point optimization metaheuristic, is considered to be the oldest and most straightforward method (<xref ref-type="bibr" rid="B52">Talbi 2009</xref>). However, it has recently been used to train complex structures of neural networks and examine their hyperparam-eters for successful image classification (<xref ref-type="bibr" rid="B3">Aly <italic>et al.</italic> 2019</xref>). The LS algorithm used in this paper is briefly described in <xref ref-type="fig" rid="ch2">Algorithm 2</xref>.</p>
					<p>
						<fig id="ch2">
							<graphic xlink:href="0120-5609-iei-41-03-e200-gch2.gif"/>
						</fig>
					</p>
					<p>For each iteration of the LS, a single solution s is replaced by a neighbor as long as the objective function is improved. Otherwise, the original solution is preserved. The search stops when all candidate neighbors are worse than the current solution, meaning that a local optimum is reached.</p>
					<p>Step five (<xref ref-type="fig" rid="f2">Algorithm 2</xref>) corresponds to the operator that generates the N neighbors of a slightly varied solution, according to the type of solution encoding. The neighbors are generated based on nvar Є [1; D] modifications that equivalent to the selected random positions. For example, <xref ref-type="fig" rid="f4">Figure 4</xref>a shows a binary vector where nvar = 3. Thus, 3 positions are switched (0 instead 1 or vice-versa).</p>
					<p>
						<fig id="f4">
							<label>Figure 4</label>
							<caption>
								<title>Examples of neighbors generated by LS. (a) Using a binary encode. (b) Using mixed encode. </title>
							</caption>
							<graphic xlink:href="0120-5609-iei-41-03-e200-gf4.gif"/>
							<attrib>Source: Authors</attrib>
						</fig>
					</p>
					<p>On the other hand, when mixed encoding is used, the nvar selected values are replaced with new values that are within boundaries of their corresponding variables (<xref ref-type="fig" rid="f4">Figure 4</xref>b).</p>
					<p>Two versions of LS are adopted as search engines: S-LSMS1 (Single-Local Search Model Selection, where binary encoding is used) and S-LSMS2 (Single-Local Search Model Selection, where mixed encoding is used).</p>
				</sec>
			</sec>
			<sec>
				<title><italic>Methodology architecture</italic></title>
				<p>In this section, the general architecture adopted for evaluating both population-based and single-point search approaches for solving the FMS problem to find a suitable learning pipeline for time-series databases is described.</p>
				<p>
					<fig id="f5">
						<label>Figure 5</label>
						<caption>
							<title>General methodology for FMS approaches in time series. </title>
						</caption>
						<graphic xlink:href="0120-5609-iei-41-03-e200-gf5.gif"/>
						<attrib>Source: Authors</attrib>
					</fig>
				</p>
				<p>The architecture can be summarized into four main steps (<xref ref-type="fig" rid="f5">Figure 5</xref>).</p>
				<p>
					<list list-type="order">
						<list-item>
							<p>A training time-series database is considered as input data.</p>
						</list-item>
						<list-item>
							<p>The training database is split into k stratified subsets (each subset contains approximately the same percentage of samples of each target class as the complete set) that are available during the search process.</p>
						</list-item>
						<list-item>
							<p>This step consists of the search process guided by the metaheuristics, either the population-based or single-point versions. Regarding population-based options (based on Algorithm 1), these generate random solutions according to continuous encoding. The overall population is evaluated thought the fitness function (Equation 6) under the stratified subsets generated in the second step. The solutions evolve throughout an established number of iterations, and, in the end, the best solution is obtained. Regarding single-point search (based on Algorithm 2), a unique solution is generated (binary or mixed encoding) which improves throughout the iterations. In the end, the best solution is also obtained.</p>
						</list-item>
						<list-item>
							<p>The final best solution found in the search process is evaluated with the test database.</p>
						</list-item>
					</list>
				</p>
			</sec>
			<sec>
				<title>Experiments and results</title>
				<p>This section presents a set of experiments where the PM (population-based metaheuristics) versions and SM (single-point-based metaheuristics) are used as the search engines to solve the FMS problem and find a suitable pipeline for time-series databases.</p>
				<p>The experimentation is presented in five subsections: (1) a comparison ofthefinal statistical results of each metaheuristic, (2) a convergence plot analysis, (3) a diversity analysis of the PM versions, (4) an analysis of the final obtained models, and (5) a frequency analysis of the methods' usage. Each metaheuristic was evaluated in the ten time-series databases described in <xref ref-type="table" rid="t1">Table 1</xref>. Considering the high computation time required by the approaches, five independent runs for each metaheuristic were carried out. The termination condition was 3 000 evaluations. The configuration used by each involved metaheuristics is described in <xref ref-type="table" rid="t4">Table 4</xref>, based on (<xref ref-type="bibr" rid="B53">Viveros-Jiménez <italic>et al.</italic> 2012</xref>, <xref ref-type="bibr" rid="B17">Escalante <italic>et al.</italic> 2009</xref>).</p>
				<p>
					<table-wrap id="t4">
						<label>Table 4</label>
						<caption>
							<title>Experimental settings for each metaheuristic approach</title>
						</caption>
						<graphic xlink:href="0120-5609-iei-41-03-e200-gt4.gif"/>
						<table-wrap-foot>
							<fn id="TFN6">
								<p>Note: Acronyms: PM (Population-based Metaheuristic); SM (Single-point-based Metaheuristics); f-DE (micro Differential Evolution); LS (Local Search); I (Iterations); NP (Population Size); CR (Crossover Rate); F (Scaled Factor); N (Number of restart solutions); R (Replacement generation); Nk (Number of neighbors) </p>
							</fn>
							<fn id="TFN7">
								<p>Source: Authors</p>
							</fn>
						</table-wrap-foot>
					</table-wrap>
				</p>
				<sec>
					<title><italic>Final Statistical Results</italic></title>
					<p>
						<xref ref-type="table" rid="t5">Table 5</xref> shows the final numerical results of CVER obtained by the six metaheuristics versions. The reported values correspond to the average of five trials evaluated in the testing set of each database.</p>
					<p>
						<table-wrap id="t5">
							<label>Table 5</label>
							<caption>
								<title>Comparison of averaging performance among the six metaheuristics for each database</title>
							</caption>
							<graphic xlink:href="0120-5609-iei-41-03-e200-gt5.gif"/>
							<table-wrap-foot>
								<fn id="TFN8">
									<p>Note: Acronyms: P-DEMS1 (Population-Differential Evolution Model Selection 1 with <italic>rand/1/bin);</italic> P-DEMS2 (Population-Differential Evolution Model Selection 2 with <italic>rand/1/exp);</italic> P-DEMS3 (Population-Differential Evolution Model Selection 3 with <italic>best/1/bin),</italic> and P-DEMS4 (Population-Differential Evolution Model Selection 4 with <italic>best/1/exp);</italic> S-LSMS1 (Single-Local Search Model Selection with binary encoding); S-LSMS2 (Single-Local Search Model Selection with mixed encoding). Values to the right of ± represent the standard deviation, the values in parentheses represent the ranks computed by the Friedman test, and values in parentheses to the left mean the lowest values found or the best ranking. </p>
								</fn>
								<fn id="TFN9">
									<p>Source: Authors</p>
								</fn>
							</table-wrap-foot>
						</table-wrap>
					</p>
					<p>Due to the fact that the samples have a non-normal distribution, and multiple comparisons are needed, the non-parametric Friedman test was used (<xref ref-type="bibr" rid="B22">García, Fernández, Luengo, and Herrera 2010</xref>). The related samples are the performances of the metaheuristics measured across the same data sets. The Friedman tests evaluates the following null hypothesis: all methods obtain similar results with non-significant differences.</p>
					<p>In the Friedman test, numerical results are converted to ranks. Thus, it ranks the metaheuristics for each problem separately. The best performing metaheuristic should have rank 1, the second best, rank 2, etc., as shown in <xref ref-type="table" rid="t5">Table 5</xref>. When there are ties, average ranks are computed. With six compared metaheuristics and ten databases, the p-value computed by the Friedman test was 0,183, which means that the null hypothesis is accepted. Thus, there are no significant differences found among the compared metaheuristics.</p>
					<p>However, according to the average rank shown in <xref ref-type="table" rid="t5">Table 5</xref>, the S-LSMS1 (SM with binary representation) was the highest rank in most of the databases. It was followed by the P-DEMS1 (PM based on μ-DE, where the base vector is randomly chosen, and a binomial crossover is used).</p>
					<p>To enhance statistical validation, the Tukey <italic>post-hoc</italic> test based on the Friedman results was applied by using the best and median values obtained over the five runs for each metaheuristic over the whole databases. <xref ref-type="fig" rid="f6">Figures 6</xref> and <xref ref-type="fig" rid="f7">7</xref> show the results of this test, where the x-axis exhibits the confidence interval of mean ranks (given by the Friedman test) and the y-axis shows the name of each metaheuristic compared. Using the best and median values, the test yielded a p-value = 0,005 and a p-value = 0,250, respectively. In the case of <xref ref-type="fig" rid="f6">Figure 6</xref>, there was a significant difference between S-LSMS1 and P-DEMS4.</p>
					<p>
						<fig id="f6">
							<label>Figure 6</label>
							<caption>
								<title>Tukey <italic>post-hoc</italic> test using the best values over the whole set of databases </title>
							</caption>
							<graphic xlink:href="0120-5609-iei-41-03-e200-gf6.gif"/>
							<attrib>Source: Authors</attrib>
						</fig>
					</p>
					<p>
						<fig id="f7">
							<label>Figure 7</label>
							<caption>
								<title>Tukey post-hoc test using the median values over the whole set of databases. </title>
							</caption>
							<graphic xlink:href="0120-5609-iei-41-03-e200-gf7.gif"/>
							<attrib>Source: Authors</attrib>
						</fig>
					</p>
					<p>Meanwhile, in <xref ref-type="fig" rid="f7">Figure 7</xref>, there are no significant differences between the metaheuristics. Finally, a pairwise comparison was conducted to determine which of the metaheuristics exhibit a different performance against a selected control metaheuristic, namely S-MSLS1 because it was the best ranked.</p>
					<p>The non-parametric 95% confidence Wilcoxon rank sum test was applied to the numerical results of the six metaheuristics for each database. <xref ref-type="table" rid="t4">Table 4</xref> shows the numerical results of the pairwise comparison. The metaheuristics are sorted according to the average rank provided by the Friedman test.</p>
					<p>The results in <xref ref-type="table" rid="t6">Table 6</xref> show that the S-LSMS1 technique was able to provide the most competitive results among the compared metaheuristics. S-LSMS1 outperformed P-DEMS1 in two (out often) databases Lightning-7 and Trace, while P-DEMS1 outperformed S-LSMS1 in Coffee and GunPoint. S-LSMS1 outperformed P-DEMS2 in four databases (Beef, ECG200, Lightning-7, and Trace), while P-DEMS2 outperformed S-LSMS1 in just the Coffee database.</p>
					<p>
						<table-wrap id="t6">
							<label>Table 6</label>
							<caption>
								<title>Comparison between the control metaheuristic S-LSMS1 and the rest of metaheuristics</title>
							</caption>
							<graphic xlink:href="0120-5609-iei-41-03-e200-gt6.gif"/>
							<table-wrap-foot>
								<fn id="TFN10">
									<p>Note: Acronyms: P-DEMS1 (Population-Differential Evolution Model Selection 1 with <italic>rand/1/bin);</italic> P-DEMS2 (Population-Differential Evolution Model Selection 2 with <italic>rand/1/exp);</italic> P-DEMS3 (Population-Differential Evolution Model Selection 3 with <italic>best/1/bin),</italic> and P-DEMS4 (Population-Differential Evolution Model Selection 4 with <italic>best/1/exp);</italic> S-LSMS1 (Single-Local Search Model Selection with binary encoding); S-LSMS2 (Single-Local Search Model Selection with mixed encoding). (-) means that there was a significant difference favoring the control metaheuristic. (+) implies that there was a significant difference favoring the compared metaheuristic. (=) means that no significant difference was observed between the compared metaheuristics. Values in parentheses to the left mean the best values found.</p>
								</fn>
								<fn id="TFN11">
									<p>Source: Authors</p>
								</fn>
							</table-wrap-foot>
						</table-wrap>
					</p>
					<p>S-LSMS1 outperformed S-LSMS2 in Beef and Lightning-7, and was beaten in the Coffee database. S-LSMS1 outperformed P-DEMS4 in four databases (Beef, ECG200, Lightning-7, and Trace) and was outperformed in just one (Coffee). In summary, S-LSMS1 was able to obtain the best numerical values at least in eight often databases (Beef, CBF, ECG200, Face-Four, Lightning-2, Lightning-7, OliveOil, and Trace). Finally, P-DEMS3 was outperformed by S-LSMS1 in three databases (ECG200, Lightning-7, and Trace) and outperformed in just one (Coffee).</p>
				</sec>
			</sec>
			<sec>
				<title><italic>Analysis of convergence plots</italic></title>
				<p>To further understand the behavior of each compared meta-heuristic, the convergence plots of a set of representative databases are analyzed.</p>
				<p>The average of five independent runs for each database is plotted. From <xref ref-type="fig" rid="f8">Figures 8</xref> to <xref ref-type="fig" rid="f13">13</xref>, convergence plots for Beef, CBF, Gun Point, Lightning-2, OliveOil, and Trace are shown. The x-axis represents the number of performing iterations for each metaheuristic, and the y-axis represents the fitness function value obtained for each iteration.</p>
				<p>
					<fig id="f8">
						<label>Figure 8</label>
						<caption>
							<title>Convergence plots comparison for the Beef database </title>
						</caption>
						<graphic xlink:href="0120-5609-iei-41-03-e200-gf8.gif"/>
						<attrib>Source: Authors</attrib>
					</fig>
				</p>
				<p>
					<fig id="f9">
						<label>Figure 9</label>
						<caption>
							<title>Convergence plots comparison for the CBF database. </title>
						</caption>
						<graphic xlink:href="0120-5609-iei-41-03-e200-gf9.gif"/>
						<attrib>Source: Authors</attrib>
					</fig>
				</p>
				<p>
					<fig id="f10">
						<label>Figure 10</label>
						<caption>
							<title>Convergence plots comparison for the Gun-Point database.</title>
						</caption>
						<graphic xlink:href="0120-5609-iei-41-03-e200-gf10.gif"/>
						<attrib>Source: Authors</attrib>
					</fig>
				</p>
				<p>
					<fig id="f11">
						<label>Figure 11</label>
						<caption>
							<title>Convergence plots comparison for the Lightning-2 database. </title>
						</caption>
						<graphic xlink:href="0120-5609-iei-41-03-e200-gf11.gif"/>
						<attrib>Source: Authors</attrib>
					</fig>
				</p>
				<p>
					<fig id="f12">
						<label>Figure 12</label>
						<caption>
							<title>Convergence plots comparison for the OliveOil database </title>
						</caption>
						<graphic xlink:href="0120-5609-iei-41-03-e200-gf12.gif"/>
						<attrib>Source: Authors</attrib>
					</fig>
				</p>
				<p>
					<fig id="f13">
						<label>Figure 13</label>
						<caption>
							<title>Convergence plots comparison for the Trace database. </title>
						</caption>
						<graphic xlink:href="0120-5609-iei-41-03-e200-gf13.gif"/>
						<attrib>Source: Authors</attrib>
					</fig>
				</p>
				<p>The x-axis was plotted in the logarithmic scale for a better display of the results. The results suggest that, in the case of the ,u-DE, to obtain a fast and competitive solution, the best option is P-DEMS1, which uses a random base vector and binomial crossover.</p>
				<p>However, in cases such as in the Trace database, PDEMS1 was trapped in local optima. Regarding SMs, S-LSMS2 (mixed representation) achieves fast convergence with respect to S-LSM1 (binary representation), but the first is usually caught in local optima, e.g., Beef, CBF, Lightning-2, or OliveOil, while S-LSM2 finds better values.</p>
				<p>Finally, an important finding is that P-DEMS1 had a faster fitness improvement in early iterations, <italic>i.e.,</italic> before 100 iterations in most databases. However, S-LSMS1 was capable of finding competitive final results at the end of the search process.</p>
			</sec>
			<sec>
				<title><italic>Diversity analysis of population-based metaheuristics</italic></title>
				<p>According to <xref ref-type="bibr" rid="B54">Yang, Li, Cai, and Guan (2015)</xref>, the population diversity has a strong influence on the performance of evolutionary algorithms. Therefore, a brief analysis of population diversity in PM versions is presented. The diversity measure is based on the distance between vectors in the variable space. For each iteration, a centroid is computed in the current population.</p>
				<p>Then, the Euclidean distance is calculated between each vector of the population and the centroid vector. With the aim of measuring the individuals' dispersion, the standard deviation over the whole distances at the current population is computed. The diversity measure was computed for each PM over the five independent runs per each database.</p>
				<p>
					<xref ref-type="fig" rid="f14">Figure 14</xref> shows the averaging diversity measure of each PM over the ten databases. A high diversity in P-DEMS 1 is observed against the other f-DE versions. It can be said that the use of a random base vector instead of the best one, as well as the binomial instead of the exponential crossover, favors a better diversity maintenance.</p>
				<p>
					<fig id="f14">
						<label>Figure 14</label>
						<caption>
							<title>Average diversity measure in population-based metaheuristics.</title>
						</caption>
						<graphic xlink:href="0120-5609-iei-41-03-e200-gf14.gif"/>
						<attrib>Source: Authors</attrib>
					</fig>
				</p>
				<sec>
					<title><italic>Analysis of final pipeline-models</italic></title>
					<p>
						<xref ref-type="table" rid="t7">Table 7</xref> shows the best pipelines suggested by each compared approach for each database. The third column details the pipeline models. Despite the fact that differences were observed in the solution models, there are interesting similarities.</p>
					<p>
						<table-wrap id="t7">
							<label>Table 7</label>
							<caption>
								<title>Best final pipelines obtained by each approach for all databases</title>
							</caption>
							<graphic xlink:href="0120-5609-iei-41-03-e200-gt7.gif"/>
							<table-wrap-foot>
								<fn id="TFN12">
									<p>Note: Acronyms: S (Smoothing); R (Representation); NR (Numerosity Reduction); C (Classification); TLS (Time-series length); SG (Savitzky-Golay Filter); MA (Moving Average); LRw (Local Regression-lowess); LRe (Local Regression-loess); SAX(Aggregate approximation); PAA (Piecewise Aggregate Approximation); PCA (Principal Component Analysis); INSIGHT (Instance Selection based on Graph-coverage and Hubness for Time-series); MC1 (Monte Carlo 1); KNN-ED (K-Nearest Neighbor-Euclidean Distance); KNN-LBDTW (K-Nearest Neighbor-Lower Bounding Dynamic Time Warping); NB (Naive Bayes); DT (Decision Tree); AB (AdaBoost); SVM (Support Vector Machine); P-DEMS1 (Population-Differential Evolution Model Selection 1 with rand/1/bin); P-DEMS2 (Population-Differential Evolution Model Selection 2 with rand/1/exp); P-DEMS3 (Population-Differential Evolution Model Selection 3 with best/1/bin), and P-DEMS4 (Population-Differential Evolution Model Selection 4 with best/1/exp); S-LSMS1 (Single-Local Search Model Selection with binary encoding); S-LSMS2 (Single-Local Search Model Selection with mixed encoding). Symbols in parentheses mean the spent runtime during training and the subscripts represent the performance of the pipeline in terms of classification error. (|) means high runtime &gt; 933 minutes, (•) means medium runtime &gt; 272 and &lt; 873 minutes, (J.) means low runtime &lt; 272 minutes. Subscripts next to symbols: g means good, r means regular, and b means bad performance. </p>
								</fn>
								<fn id="TFN13">
									<p>Source: Authors</p>
								</fn>
							</table-wrap-foot>
						</table-wrap>
					</p>
					<p>Regarding the smoothing task, Moving Average was the most preferred. PAA was the most commonly used and suggested method for time series representation, while INSIGHT was the most popular numerosity reduction technique.</p>
					<p>As for the classification task, the decision tree and the Ad-aBoost (with decision trees as the weak learners) appeared as the most suitable. From the resulting final models, it can be seen that there were some evaluated databases with different models with similar performance values.</p>
					<p>Databases such as Beef, Gun-Point, and Lightning-7 were detected as possible multimodal problems. They reported more diversification in the selected methods and their related hyperparameters.</p>
					<p>Runtime varies considerably due to the different features of the temporal databases and the selected methods for carrying out a specific sub-task. Overall, P-DEMS3 reported the lowest runtime computational cost, while S-LSMS1 was the most expensive approach.</p>
					<p>However, S-LSMS1 reported the best performance during training and competitive results in the testing phase. <xref ref-type="fig" rid="f15">Figure 15</xref> shows a graphical example of the suggested pipeline that was applied to the CBF database. It can be seen that the average behavior of the original CBF database remains after the processing originated by the applied pipeline. A significant dimensionality reduction was observed.</p>
					<p>
						<fig id="f15">
							<label>Figure 15</label>
							<caption>
								<title>Frequency analysis of included into pipeline task by metaheuristics.</title>
							</caption>
							<graphic xlink:href="0120-5609-iei-41-03-e200-gf15.gif"/>
							<attrib>Note: Acronyms: P-Metaheuristics (Population-based Metaheuristics); S-Metaheuristics (Single point search Metaheuristics); SG (Savitzky-Golay Filter); MA (Moving Average); LRw (Local Regression-lowess); LRe (Local Regression-loess); SAX(Aggregate approXimation); PAA (Piecewise Aggregate Approximation); PCA (Principal Component Analysis); INSIGHT (Instance Selection based on Graph-coverage and Hubness for Time-series); MC1 (Monte Carlo 1); KNN1 (K-Nearest Neighbor-Euclidean Distance); KNN2 (K-Nearest Neighbor-Lower Bounding Dynamic Time Warping); NB (Naive Bayes); DT (Decision Tree); AB (AdaBoost); SVM (Support Vector Machine). Source: Authors</attrib>
						</fig>
					</p>
				</sec>
				<sec>
					<title><italic>Frequency analysis of considered method by metaheuristics</italic></title>
					<p>In order to enhance the analysis of the preferred solutions, a selection frequency analysis of the methods considered by the approaches for the FMS problem in time-series databases was made. <xref ref-type="fig" rid="f16">Figure 16</xref> shows the average frequency, for each recognized method, computed from five trials over all databases for each metaheuristic.</p>
					<p>
						<fig id="f16">
							<label>Figure 16</label>
							<caption>
								<title>Example of pipeline-model applied to CBF database. Model: S:MA{span = 67}, R:PAA{nseg=21}. R:MC1{%ins=0.81403,nitera=400}, C:AB{n = 436}. </title>
							</caption>
							<graphic xlink:href="0120-5609-iei-41-03-e200-gf16.gif"/>
							<attrib>Note: (1) The average behavior of the original testing database is plotted that is compared to the average smoothed testing test. (2) The averaged smoothed testing database is plotted before the time-series representation process. (3) The averaged smoothed testing database is plotted after the time-series representation and numerosity reduction processes were applied. Source: Authors</attrib>
						</fig>
					</p>
					<p>The frequency results for population-baed meteheuristics were based on 601 200 evaluated pipeline-models, while single-point search metaheuristics were based on 300 600 models. Regarding the smoothing options, Moving Average method, was the most solicited by population-based meteheuristics, while the Savitzky-Golay filter was the most preferred by single-point search metaheuristics.</p>
					<p>For time-series representation, the PAA method was the most preferred for both population-based and single-based metaheuristics. INSIGHT was the most selected numerosity reduction method. Regarding the classifiers, it can be confirmed that Adaboost was the most suitable classifier, while KNN1 was the less preferred.</p>
				</sec>
			</sec>
		</sec>
		<sec sec-type="conclusions">
			<title>Conclusions and future work</title>
			<p>In this paper, a comparison study between two metaheuristic approaches to deal with FMS and pipelines building for time-series databases was presented. The first approach was based on the micro-version of a differential evolution algorithm, named as ¡u-DEMS in this work, from which four variants were tested based on <italic>rand/1/bin</italic> (P-DEMS1), <italic>rand/1/exp</italic> (P-DEMS2), <italic>best/1/bin</italic> (P-DEMS3), and <italic>best/1/exp</italic> (PDEMS4).</p>
			<p>The second approach focused on evaluating local search behavior S-LSMS, the most straightforward single-point search metaheuristics. Two versions were assessed, one of them with binary encoding and the second one with mixed encoding.</p>
			<p>Six complete pipeline-model search options were evaluated, out of which four are P-DEMS variants and two are S-LSMS variants. Each of the variants was evaluated in ten different time-series databases.</p>
			<p>The set of experiments was divided into five parts: the statistical analysis of the numerical results, the analysis of the convergence graphs, a diversity analysis focused only on the population variants, the analysis of the final pipeline models, and the study of the selection frequency of the methods involved. From these experiments, some important conclusions and findings are listed below:</p>
			<p>Statistical analysis suggests that S-LSMS1 (binary encoding version) is the best option when working with time-series databases that have high dimensionality, are noisy, and whose number of classes is higher than two, such as Lightning-2, Lightning-7, and OliveOil. S-LSMS1 has the advantage of being simple in its structure, and only requires two parameters to be set. However, it has the disadvantage of having a high computational cost.</p>
			<p>On the other hand, if the database is dichotomous, the noise is moderate, and its dimensionality length is below an approximate value of 350. Therefore, population-based metaheuristics P-DEMS1, which uses <italic>rand/1/bin,</italic> turns out to be the best option. Besides, it achieved competitive results around the first 100 iterations.</p>
			<p>Regarding the exploration capacity, it was observed that the population-based metaheuristic P-DEMS1 with the<italic>rand/1/bin</italic> variant provides a better diversity of pipeline models.</p>
			<p>With respect to the final pipeline-models, it can be seen that, for most of the databases, a complete model was found which contained the most straightforward methods for the tasks of smoothing, dimensionality reduction, and number reduction. These methods are Moving Average, PAA, and INSIGHT, respectively. On the side of the classification task, AdaBoost was the most common method.</p>
			<p>An important finding was discovering different complete pipeline model configurations with similar performance for the same database. Therefore, some temporary databases can be seen as a multi modal problem.</p>
			<p>As part of future work, a complexity measure could be considered as a fitness function to then tackle the FMS problem as a multi-objective problem. Additionally, a mechanism to build more flexible pipelines where the length and order can incorporated, in addition to searching for a way to fairly compare it against other state-of-the-art approaches.</p>
		</sec>
	</body>
	<back>
		<ack>
			<title>Acknowledgements</title>
			<p>The authors would like to acknowledge support from the Mexican National Council for Science and Technology(CONA-CyT) through scholarship number 259655 and project No. 220522.</p>
		</ack>
		<ref-list>
			<title>References</title>
			<ref id="B1">
				<mixed-citation>Al-Jowder, O., Kemsley, E., and Wilson, R. H. (2002). Detection of adulteration in cooked meat products by mid-infrared spectroscopy. <italic>Journal of Agricultural and Food Chemistry</italic>, 50(6), 1325-1329. 10.1021/jf0108967</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Al-Jowder</surname>
							<given-names>O.</given-names>
						</name>
						<name>
							<surname>Kemsley</surname>
							<given-names>E.</given-names>
						</name>
						<name>
							<surname>Wilson</surname>
							<given-names>R. H</given-names>
						</name>
					</person-group>
					<year>2002</year>
					<article-title>Detection of adulteration in cooked meat products by mid-infrared spectroscopy</article-title>
					<source>Journal of Agricultural and Food Chemistry</source>
					<volume>50</volume>
					<issue>6</issue>
					<fpage>1325</fpage>
					<lpage>1329</lpage>
					<pub-id pub-id-type="doi">10.1021/jf0108967</pub-id>
				</element-citation>
			</ref>
			<ref id="B2">
				<mixed-citation>Ali, M., Alqahtani, A., Jones, M. W., and Xie, X. (2019). Clustering and classification for time series data in visual analytics: A survey. <italic>IEEE Access</italic>, 7, 181314-181338. 10.1109/AC-CESS.2019.2958551</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Ali</surname>
							<given-names>M.</given-names>
						</name>
						<name>
							<surname>Alqahtani</surname>
							<given-names>A.</given-names>
						</name>
						<name>
							<surname>Jones</surname>
							<given-names>M. W.</given-names>
						</name>
						<name>
							<surname>Xie</surname>
							<given-names>X</given-names>
						</name>
					</person-group>
					<year>2019</year>
					<article-title>Clustering and classification for time series data in visual analytics: A survey</article-title>
					<source>IEEE Access</source>
					<issue>7</issue>
					<fpage>181314</fpage>
					<lpage>181338</lpage>
					<pub-id pub-id-type="doi">10.1109/AC-CESS.2019.2958551</pub-id>
				</element-citation>
			</ref>
			<ref id="B3">
				<mixed-citation>Aly, A., Guadagni, G., and Dugan, J. B. (2019). Derivative-free optimization of neural networks using local search. In IEEE (Eds.) <italic>2019 IEEE 10th Annual Ubiquitous Computing, Electronics Mobile Communication Conference (UEMCON)</italic> (pp. 0293-0299). New York, NY: IEEE. 10.1109/UEM-CON47517.2019.8993007</mixed-citation>
				<element-citation publication-type="confproc">
					<person-group person-group-type="author">
						<name>
							<surname>Aly</surname>
							<given-names>A.</given-names>
						</name>
						<name>
							<surname>Guadagni</surname>
							<given-names>G.</given-names>
						</name>
						<name>
							<surname>Dugan</surname>
							<given-names>J. B</given-names>
						</name>
					</person-group>
					<year>2019</year>
					<source>Derivative-free optimization of neural networks using local search</source>
					<conf-name>IEEE (Eds.) 2019 IEEE 10th Annual Ubiquitous Computing, Electronics Mobile Communication Conference (UEMCON)</conf-name>
					<fpage>0293</fpage>
					<lpage>0299</lpage>
					<publisher-loc>New York, NY</publisher-loc>
					<publisher-name>IEEE</publisher-name>
					<pub-id pub-id-type="doi">10.1109/UEM-CON47517.2019.8993007</pub-id>
				</element-citation>
			</ref>
			<ref id="B4">
				<mixed-citation>Bagnall, A., Davis, L., Hills, J., and Lines, J. (2012). Transformation based ensembles for time series classification. In SIAM (Eds.) <italic>Proceedings of the 2012 SIAM international conference on data mining</italic> (pp. 307-318). Philadelphia, PA: Society for Industrial and Applied Mathematics. 10.1137/1.9781611972825.27</mixed-citation>
				<element-citation publication-type="confproc">
					<person-group person-group-type="author">
						<name>
							<surname>Bagnall</surname>
							<given-names>A.</given-names>
						</name>
						<name>
							<surname>Davis</surname>
							<given-names>L.</given-names>
						</name>
						<name>
							<surname>Hills</surname>
							<given-names>J.</given-names>
						</name>
						<name>
							<surname>Lines</surname>
							<given-names>J</given-names>
						</name>
					</person-group>
					<year>2012</year>
					<source>Transformation based ensembles for time series classification</source>
					<conf-name>SIAM (Eds.) Proceedings of the 2012 SIAM international conference on data mining</conf-name>
					<fpage>307</fpage>
					<lpage>318</lpage>
					<publisher-loc>Philadelphia, PA</publisher-loc>
					<publisher-name>Society for Industrial and Applied Mathematics</publisher-name>
					<pub-id pub-id-type="doi">10.1137/1.9781611972825.27</pub-id>
				</element-citation>
			</ref>
			<ref id="B5">
				<mixed-citation>Baijal, S., Singh, S., Rani, A., and Agarwal, S. (2016). Performance evaluation of S-Golay and MA filter on the basis of white and flicker noise. In <italic>Proceedings of Second International Symposium on Signal Processing and Intelligent Recognition Systems (SIRS-2015)</italic> (pp. 245-255). New York, NY: Springer. 10.1007/978-3-319-28658-721</mixed-citation>
				<element-citation publication-type="confproc">
					<person-group person-group-type="author">
						<name>
							<surname>Baijal</surname>
							<given-names>S.</given-names>
						</name>
						<name>
							<surname>Singh</surname>
							<given-names>S.</given-names>
						</name>
						<name>
							<surname>Rani</surname>
							<given-names>A.</given-names>
						</name>
						<name>
							<surname>Agarwal</surname>
							<given-names>S</given-names>
						</name>
					</person-group>
					<year>2016</year>
					<source>Performance evaluation of S-Golay and MA filter on the basis of white and flicker noise</source>
					<conf-name>Proceedings of Second International Symposium on Signal Processing and Intelligent Recognition Systems (SIRS-2015)</conf-name>
					<fpage>245</fpage>
					<lpage>255</lpage>
					<publisher-loc>New York, NY</publisher-loc>
					<publisher-name>Springer</publisher-name>
					<pub-id pub-id-type="doi">10.1007/978-3-319-28658-721</pub-id>
				</element-citation>
			</ref>
			<ref id="B6">
				<mixed-citation>Bergstra, J. and Bengio, Y. (2012). Random search for hyper-parameter optimization. <italic>The Journal of Machine Learning Research</italic>, 13(2), 281-305. <ext-link ext-link-type="uri" xlink:href="https://www.jmlr.org/papers/volume13/bergstra12a/ber\gstra12a">https://www.jmlr.org/papers/volume13/bergstra12a/ber\gstra12a</ext-link>
				</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Bergstra</surname>
							<given-names>J.</given-names>
						</name>
						<name>
							<surname>Bengio</surname>
							<given-names>Y</given-names>
						</name>
					</person-group>
					<year>2012</year>
					<article-title>Random search for hyper-parameter optimization</article-title>
					<source>The Journal of Machine Learning Research</source>
					<volume>13</volume>
					<issue>2</issue>
					<fpage>281</fpage>
					<lpage>305</lpage>
					<ext-link ext-link-type="uri" xlink:href="https://www.jmlr.org/papers/volume13/bergstra12a/ber\gstra12a">https://www.jmlr.org/papers/volume13/bergstra12a/ber\gstra12a</ext-link>
				</element-citation>
			</ref>
			<ref id="B7">
				<mixed-citation>Bischl, B., Lang, M., Kotthoff, L., Schiffner, J., Richter, J., Studerus, E., Casalicchio, G., and Jones, Z. M. (2016). mlr: Machine learning in R. <italic>The Journal of Machine Learning Research</italic>, 17(170), 1-5. <ext-link ext-link-type="uri" xlink:href="http://jmlr.org/papers/v17/15-066.html">http://jmlr.org/papers/v17/15-066.html</ext-link>
				</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Bischl</surname>
							<given-names>B.</given-names>
						</name>
						<name>
							<surname>Lang</surname>
							<given-names>M.</given-names>
						</name>
						<name>
							<surname>Kotthoff</surname>
							<given-names>L.</given-names>
						</name>
						<name>
							<surname>Schiffner</surname>
							<given-names>J.</given-names>
						</name>
						<name>
							<surname>Richter</surname>
							<given-names>J.</given-names>
						</name>
						<name>
							<surname>Studerus</surname>
							<given-names>E.</given-names>
						</name>
						<name>
							<surname>Casalicchio</surname>
							<given-names>G.</given-names>
						</name>
						<name>
							<surname>Jones</surname>
							<given-names>Z. M</given-names>
						</name>
					</person-group>
					<year>2016</year>
					<article-title>mlr: Machine learning in R</article-title>
					<source>The Journal of Machine Learning Research</source>
					<volume>17</volume>
					<issue>170</issue>
					<fpage>1</fpage>
					<lpage>5</lpage>
					<ext-link ext-link-type="uri" xlink:href="http://jmlr.org/papers/v17/15-066.html">http://jmlr.org/papers/v17/15-066.html</ext-link>
				</element-citation>
			</ref>
			<ref id="B8">
				<mixed-citation>Bishop, C. M. (2006). <italic>Pattern recognition and machine learning</italic>. New York, NY: Springer .</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Bishop</surname>
							<given-names>C. M</given-names>
						</name>
					</person-group>
					<year>2006</year>
					<source>Pattern recognition and machine learning</source>
					<publisher-loc>New York, NY</publisher-loc>
					<publisher-name>Springer</publisher-name>
				</element-citation>
			</ref>
			<ref id="B9">
				<mixed-citation>Boullé, N., Dallas, V., Nakatsukasa, Y., and Samaddar, D. (2020). Classification of chaotic time series with deep learning. <italic>Physica D: Nonlinear Phenomena</italic>, <italic>403</italic>, 132261. 10.1016/j.physd.2019.132261</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Boullé</surname>
							<given-names>N.</given-names>
						</name>
						<name>
							<surname>Dallas</surname>
							<given-names>V.</given-names>
						</name>
						<name>
							<surname>Nakatsukasa</surname>
							<given-names>Y.</given-names>
						</name>
						<name>
							<surname>Samaddar</surname>
							<given-names>D</given-names>
						</name>
					</person-group>
					<year>2020</year>
					<article-title>Classification of chaotic time series with deep learning</article-title>
					<source>Physica D: Nonlinear Phenomena</source>
					<issue>403</issue>
					<fpage>132261</fpage>
					<lpage>132261</lpage>
					<pub-id pub-id-type="doi">10.1016/j.physd.2019.132261</pub-id>
				</element-citation>
			</ref>
			<ref id="B10">
				<mixed-citation>Buza, K., Nanopoulos, A., and Schmidt-Thieme, L. (2011). Insight: Efficient and effective instance selection for time-series classification. In Huang, J. Z., Cao, L., and Srivastava, J. (Eds.) <italic>Pacific-Asia Conference on Knowledge Discovery and Data Mining</italic> (pp. 149-160). Heidelberg/Berlin, Germany: Springer.</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Buza</surname>
							<given-names>K.</given-names>
						</name>
						<name>
							<surname>Nanopoulos</surname>
							<given-names>A.</given-names>
						</name>
						<name>
							<surname>Schmidt-Thieme</surname>
							<given-names>L</given-names>
						</name>
					</person-group>
					<year>2011</year>
					<chapter-title>Insight: Efficient and effective instance selection for time-series classification</chapter-title>
					<person-group person-group-type="author">
						<name>
							<surname>Huang</surname>
							<given-names>J. Z.</given-names>
						</name>
						<name>
							<surname>Cao</surname>
							<given-names>L.</given-names>
						</name>
						<name>
							<surname>Srivastava</surname>
							<given-names>J</given-names>
						</name>
					</person-group>
					<source>Pacific-Asia Conference on Knowledge Discovery and Data Mining</source>
					<fpage>149</fpage>
					<lpage>160</lpage>
					<publisher-loc>Heidelberg/Berlin, Germany</publisher-loc>
					<publisher-name>Springer</publisher-name>
				</element-citation>
			</ref>
			<ref id="B11">
				<mixed-citation>Caraffini, F., Neri, F., and Poikolainen, I. (2013). Micro-differential evolution with extra moves along the axes. In IEEE (Eds.) <italic>2013 IEEE Symposium on Differential Evolution (SDE)</italic> (pp. 46-53). New York, NY: IEEE . 10.1109/SDE.2013.6601441</mixed-citation>
				<element-citation publication-type="confproc">
					<person-group person-group-type="author">
						<name>
							<surname>Caraffini</surname>
							<given-names>F.</given-names>
						</name>
						<name>
							<surname>Neri</surname>
							<given-names>F.</given-names>
						</name>
						<name>
							<surname>Poikolainen</surname>
							<given-names>I.</given-names>
						</name>
					</person-group>
					<year>2013</year>
					<source>Micro-differential evolution with extra moves along the axes</source>
					<conf-name>IEEE (Eds.) 2013 IEEE Symposium on Differential Evolution (SDE)</conf-name>
					<fpage>46</fpage>
					<lpage>53</lpage>
					<publisher-loc>New York, NY</publisher-loc>
					<publisher-name>IEEE</publisher-name>
					<pub-id pub-id-type="doi">10.1109/SDE.2013.6601441</pub-id>
				</element-citation>
			</ref>
			<ref id="B12">
				<mixed-citation>Cleveland, W. S. and Loader, C. (1996). Smoothing by local regression: Principles and methods. In Hardle, W., and Scmiek, M. G. (Eds.) <italic>Statistical Theory and Computational Aspects of Smoothing</italic> (pp. 10-49). Heidelberg, Germany: Physica-Verlag HD. 10.1007/978-3-642-48425-4_2</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Cleveland</surname>
							<given-names>W. S.</given-names>
						</name>
						<name>
							<surname>Loader</surname>
							<given-names>C</given-names>
						</name>
					</person-group>
					<year>1996</year>
					<chapter-title>Smoothing by local regression: Principles and methods</chapter-title>
					<person-group person-group-type="author">
						<name>
							<surname>Hardle</surname>
							<given-names>W.</given-names>
						</name>
						<name>
							<surname>Scmiek</surname>
							<given-names>M. G</given-names>
						</name>
					</person-group>
					<source>Statistical Theory and Computational Aspects of Smoothing</source>
					<fpage>10</fpage>
					<lpage>49</lpage>
					<publisher-loc>Heidelberg, Germany</publisher-loc>
					<publisher-name>Physica-Verlag</publisher-name>
					<pub-id pub-id-type="doi">10.1007/978-3-642-48425-4_2</pub-id>
				</element-citation>
			</ref>
			<ref id="B13">
				<mixed-citation>de Sá, A. G. C., Pinto, W. J. G. S., Oliveira, L. O. V. B., and Pappa, G. L. (2017). RECIPE: A grammar-based framework for automatically evolving classification pipelines. In McDermott, J., Castelli, M., Sekanina, L., Haasdijk, E., and García-Sánchez, P. (Eds.) <italic>European Conference on Genetic Programming</italic> (pp. 246-261), Springer International Publishing, Cham. 10.1007/978-3-319-55696-316</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>de Sá</surname>
							<given-names>A. G. C.</given-names>
						</name>
						<name>
							<surname>Pinto</surname>
							<given-names>W. J. G. S.</given-names>
						</name>
						<name>
							<surname>Oliveira</surname>
							<given-names>L. O. V. B.</given-names>
						</name>
						<name>
							<surname>Pappa</surname>
							<given-names>G. L</given-names>
						</name>
					</person-group>
					<year>2017</year>
					<chapter-title>RECIPE: A grammar-based framework for automatically evolving classification pipelines</chapter-title>
					<person-group person-group-type="author">
						<name>
							<surname>McDermott</surname>
							<given-names>J.</given-names>
						</name>
						<name>
							<surname>Castelli</surname>
							<given-names>M.</given-names>
						</name>
						<name>
							<surname>Sekanina</surname>
							<given-names>L.</given-names>
						</name>
						<name>
							<surname>Haasdijk</surname>
							<given-names>E.</given-names>
						</name>
						<name>
							<surname>García-Sánchez</surname>
							<given-names>P</given-names>
						</name>
					</person-group>
					<source>European Conference on Genetic Programming</source>
					<fpage>246</fpage>
					<lpage>261</lpage>
					<publisher-name>Springer International Publishing, Cham</publisher-name>
					<pub-id pub-id-type="doi">10.1007/978-3-319-55696-316</pub-id>
				</element-citation>
			</ref>
			<ref id="B14">
				<mixed-citation>Díaz-Pacheco, A., Gonzalez-Bernal, J. A., Reyes-García, C. A., and Escalante-Balderas, H.J. (2018). Full model selection in big data. In Castro, F., Miranda-Jiménez, S., and González-Mendoza, M. (Eds.) <italic>Advances in Soft Computing</italic> (pp. 279-289). Springer International Publishing, Cham. 10.1007/978-3-030-02837-423</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Díaz-Pacheco</surname>
							<given-names>A.</given-names>
						</name>
						<name>
							<surname>Gonzalez-Bernal</surname>
							<given-names>J. A.</given-names>
						</name>
						<name>
							<surname>Reyes-García</surname>
							<given-names>C. A.</given-names>
						</name>
						<name>
							<surname>Escalante-Balderas</surname>
							<given-names>H.J</given-names>
						</name>
					</person-group>
					<year>2018</year>
					<chapter-title>Full model selection in big data</chapter-title>
					<person-group person-group-type="author">
						<name>
							<surname>Castro</surname>
							<given-names>F.</given-names>
						</name>
						<name>
							<surname>Miranda-Jiménez</surname>
							<given-names>S.</given-names>
						</name>
						<name>
							<surname>González-Mendoza</surname>
							<given-names>M</given-names>
						</name>
					</person-group>
					<source>Advances in Soft Computing</source>
					<fpage>279</fpage>
					<lpage>289</lpage>
					<publisher-name>Springer International Publishing, Cham</publisher-name>
					<pub-id pub-id-type="doi">10.1007/978-3-030-02837-423</pub-id>
				</element-citation>
			</ref>
			<ref id="B15">
				<mixed-citation>Eads, D. R., Hill, D., Davis, S., Perkins, S. J., Ma, J., Porter, R. B., and Theiler, J. P. (2002). Genetic algorithms and support vector machines for time series classification. In Bosacchi, B., Fogel, D. B., and Bezdek, J. C. (Eds.) <italic>Applications and Science of Neural Networks, Fuzzy Systems, and Evolutionary Computation V</italic> (vol. 4787, pp. 74-85). Bellingham, WA: International Society for Optics and Photonics. 10.1117/12.453526</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Eads</surname>
							<given-names>D. R.</given-names>
						</name>
						<name>
							<surname>Hill</surname>
							<given-names>D.</given-names>
						</name>
						<name>
							<surname>Davis</surname>
							<given-names>S.</given-names>
						</name>
						<name>
							<surname>Perkins</surname>
							<given-names>S. J.</given-names>
						</name>
						<name>
							<surname>Ma</surname>
							<given-names>J.</given-names>
						</name>
						<name>
							<surname>Porter</surname>
							<given-names>R. B.</given-names>
						</name>
						<name>
							<surname>Theiler</surname>
							<given-names>J. P</given-names>
						</name>
					</person-group>
					<year>2002</year>
					<chapter-title>Genetic algorithms and support vector machines for time series classification</chapter-title>
					<person-group person-group-type="author">
						<name>
							<surname>Bosacchi</surname>
							<given-names>B.</given-names>
						</name>
						<name>
							<surname>Fogel</surname>
							<given-names>D. B.</given-names>
						</name>
						<name>
							<surname>Bezdek</surname>
							<given-names>J. C</given-names>
						</name>
					</person-group>
					<source>Applications and Science of Neural Networks, Fuzzy Systems, and Evolutionary Computation V</source>
					<volume>4787</volume>
					<fpage>74</fpage>
					<lpage>85</lpage>
					<publisher-loc>Bellingham, WA</publisher-loc>
					<publisher-name>International Society for Optics and Photonics</publisher-name>
					<pub-id pub-id-type="doi">10.1117/12.453526</pub-id>
				</element-citation>
			</ref>
			<ref id="B16">
				<mixed-citation>Escalante, H. J., Montes, M., and Sucar, E. (2010). Ensemble particle swarm model selection. In IEEE (Eds.) The <italic>2010 International Joint Conference on Neural Networks (IJCNN)</italic> (pp. 1-8). New York, NY: IEEE . 10.1109/IJCNN.2010.5596915</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Escalante</surname>
							<given-names>H. J.</given-names>
						</name>
						<name>
							<surname>Montes</surname>
							<given-names>M.</given-names>
						</name>
						<name>
							<surname>Sucar</surname>
							<given-names>E</given-names>
						</name>
					</person-group>
					<year>2010</year>
					<chapter-title>Ensemble particle swarm model selection</chapter-title>
					<person-group person-group-type="author">
						<collab>IEEE</collab>
					</person-group>
					<source>The <italic>2010 International Joint Conference on Neural Networks (IJCNN)</italic></source>
					<fpage>1</fpage>
					<lpage>8</lpage>
					<publisher-loc>New York, NY</publisher-loc>
					<publisher-name>IEEE</publisher-name>
					<pub-id pub-id-type="doi">10.1109/IJCNN.2010.5596915</pub-id>
				</element-citation>
			</ref>
			<ref id="B17">
				<mixed-citation>Escalante, H. J., Montes, M., and Sucar, L. E. (2009). Particle swarm model selection. <italic>Journal of Machine Learning Research</italic>, 10(2), 405-440. <ext-link ext-link-type="uri" xlink:href="http://jmlr.org/papers/v10/escalante09a.html">http://jmlr.org/papers/v10/escalante09a.html</ext-link>
				</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Escalante</surname>
							<given-names>H. J.</given-names>
						</name>
						<name>
							<surname>Montes</surname>
							<given-names>M.</given-names>
						</name>
						<name>
							<surname>Sucar</surname>
							<given-names>L. E</given-names>
						</name>
					</person-group>
					<year>2009</year>
					<article-title>Particle swarm model selection</article-title>
					<source>Journal of Machine Learning Research</source>
					<volume>10</volume>
					<issue>2</issue>
					<fpage>405</fpage>
					<lpage>440</lpage>
					<ext-link ext-link-type="uri" xlink:href="http://jmlr.org/papers/v10/escalante09a.html">http://jmlr.org/papers/v10/escalante09a.html</ext-link>
				</element-citation>
			</ref>
			<ref id="B18">
				<mixed-citation>Esling, P. and Agon, C. (2012). Time-series data mining. <italic>ACM Computing Surveys (CSUR)</italic>, 45(1), 1-12. 10.1145/2379776.2379788</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Esling</surname>
							<given-names>P.</given-names>
						</name>
						<name>
							<surname>Agon</surname>
							<given-names>C</given-names>
						</name>
					</person-group>
					<year>2012</year>
					<article-title>Time-series data mining</article-title>
					<source>ACM Computing Surveys (CSUR)</source>
					<volume>45</volume>
					<issue>1</issue>
					<fpage>1</fpage>
					<lpage>12</lpage>
					<pub-id pub-id-type="doi">10.1145/2379776.2379788</pub-id>
				</element-citation>
			</ref>
			<ref id="B19">
				<mixed-citation>Fu, T.-c. (2011). A review on time series data mining. <italic>Engineering Applications of Artificial Intelligence</italic>, 24(1), 164-181. 10.1016/j.engappai.2010.09.007</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Fu</surname>
							<given-names>T.-c</given-names>
						</name>
					</person-group>
					<year>2011</year>
					<article-title>A review on time series data mining</article-title>
					<source>Engineering Applications of Artificial Intelligence</source>
					<volume>24</volume>
					<issue>1</issue>
					<fpage>164</fpage>
					<lpage>181</lpage>
					<pub-id pub-id-type="doi">10.1016/j.engappai.2010.09.007</pub-id>
				</element-citation>
			</ref>
			<ref id="B20">
				<mixed-citation>Gantza, J. and Reisel, D. (2012). The digital universe in 2020: Big data, bigger digital shadows, and biggest growth in the far east. <italic>IDC iView: IDC Analyze the Future</italic>, 2007(2012), 1-16. <ext-link ext-link-type="uri" xlink:href="https://www.speicherguide.de/download/dokus/IDC-Digital-Universe-Studie-iView-11.12.pdf">https://www.speicherguide.de/download/dokus/IDC-Digital-Universe-Studie-iView-11.12.pdf</ext-link>
				</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Gantza</surname>
							<given-names>J.</given-names>
						</name>
						<name>
							<surname>Reisel</surname>
							<given-names>D</given-names>
						</name>
					</person-group>
					<year>2012</year>
					<article-title>The digital universe in 2020: Big data, bigger digital shadows, and biggest growth in the far east</article-title>
					<source>IDC iView: IDC Analyze the Future</source>
					<year>2007</year>
					<volume>2012</volume>
					<fpage>1</fpage>
					<lpage>16</lpage>
					<ext-link ext-link-type="uri" xlink:href="https://www.speicherguide.de/download/dokus/IDC-Digital-Universe-Studie-iView-11.12.pdf">https://www.speicherguide.de/download/dokus/IDC-Digital-Universe-Studie-iView-11.12.pdf</ext-link>
				</element-citation>
			</ref>
			<ref id="B21">
				<mixed-citation>Garcia, S., Derrac, J., Cano, J., and Herrera, F. (2012). Prototype selection for nearest neighbor classification: Taxonomy and empirical study. <italic>IEEE transactions on pattern analysis and machine intelligence</italic>, 34(3), 417-435. 10.1109/TPAMI.2011.142</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Garcia</surname>
							<given-names>S.</given-names>
						</name>
						<name>
							<surname>Derrac</surname>
							<given-names>J.</given-names>
						</name>
						<name>
							<surname>Cano</surname>
							<given-names>J.</given-names>
						</name>
						<name>
							<surname>Herrera</surname>
							<given-names>F</given-names>
						</name>
					</person-group>
					<year>2012</year>
					<article-title>Prototype selection for nearest neighbor classification: Taxonomy and empirical study</article-title>
					<source>IEEE transactions on pattern analysis and machine intelligence</source>
					<volume>34</volume>
					<issue>3</issue>
					<fpage>417</fpage>
					<lpage>435</lpage>
					<pub-id pub-id-type="doi">10.1109/TPAMI.2011.142</pub-id>
				</element-citation>
			</ref>
			<ref id="B22">
				<mixed-citation>García, S., Fernández, A., Luengo, J., and Herrera, F. (2010). Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: Experimental analysis of power. <italic>Information Sciences</italic>, 180(10), 2044-2064. 10.1016/j.ins.2009.12.010</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>García</surname>
							<given-names>S.</given-names>
						</name>
						<name>
							<surname>Fernández</surname>
							<given-names>A.</given-names>
						</name>
						<name>
							<surname>Luengo</surname>
							<given-names>J.</given-names>
						</name>
						<name>
							<surname>Herrera</surname>
							<given-names>F</given-names>
						</name>
					</person-group>
					<year>2010</year>
					<article-title>Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: Experimental analysis of power</article-title>
					<source>Information Sciences</source>
					<volume>180</volume>
					<issue>10</issue>
					<fpage>2044</fpage>
					<lpage>2064</lpage>
					<pub-id pub-id-type="doi">10.1016/j.ins.2009.12.010</pub-id>
				</element-citation>
			</ref>
			<ref id="B23">
				<mixed-citation>Giron-Sierra, J. (2018). <italic>Digital Signal Processing with Matlab Examples, Volume 3: Model-Based Actions and Sparse Representation</italic>. Singapore: Springer Singapore.</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Giron-Sierra</surname>
							<given-names>J</given-names>
						</name>
					</person-group>
					<year>2018</year>
					<source>Digital Signal Processing with Matlab Examples, Volume 3: Model-Based Actions and Sparse Representation</source>
					<publisher-loc>Singapore</publisher-loc>
					<publisher-name>Springer Singapore</publisher-name>
				</element-citation>
			</ref>
			<ref id="B24">
				<mixed-citation>Gong, Z., Chen, H., Yuan, B., and Yao, X. (2019). Multiobjective learning in the model space for time series classification. <italic>IEEE Transactions on Cybernetics</italic>, 49(3), 918-932. 10.1109/TCYB.2018.2789422</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Gong</surname>
							<given-names>Z.</given-names>
						</name>
						<name>
							<surname>Chen</surname>
							<given-names>H.</given-names>
						</name>
						<name>
							<surname>Yuan</surname>
							<given-names>B.</given-names>
						</name>
						<name>
							<surname>Yao</surname>
							<given-names>X.</given-names>
						</name>
					</person-group>
					<year>2019</year>
					<article-title>Multiobjective learning in the model space for time series classification</article-title>
					<source>IEEE Transactions on Cybernetics</source>
					<volume>49</volume>
					<issue>3</issue>
					<fpage>918</fpage>
					<lpage>932</lpage>
					<pub-id pub-id-type="doi">10.1109/TCYB.2018.2789422</pub-id>
				</element-citation>
			</ref>
			<ref id="B25">
				<mixed-citation>Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P., and Witten, I. H. (2009). The weka data mining software: An update. <italic>ACM SIGKDD Explorations Newsletter</italic>, <italic>11</italic> (1), 10-18. 10.1145/1656274.1656278</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Hall</surname>
							<given-names>M.</given-names>
						</name>
						<name>
							<surname>Frank</surname>
							<given-names>E.</given-names>
						</name>
						<name>
							<surname>Holmes</surname>
							<given-names>G.</given-names>
						</name>
						<name>
							<surname>Pfahringer</surname>
							<given-names>B.</given-names>
						</name>
						<name>
							<surname>Reutemann</surname>
							<given-names>P.</given-names>
						</name>
						<name>
							<surname>Witten</surname>
							<given-names>I. H</given-names>
						</name>
					</person-group>
					<year>2009</year>
					<article-title>The weka data mining software: An update</article-title>
					<source>ACM SIGKDD Explorations Newsletter</source>
					<volume>11</volume>
					<issue>1</issue>
					<fpage>10</fpage>
					<lpage>18</lpage>
					<pub-id pub-id-type="doi">10.1145/1656274.1656278</pub-id>
				</element-citation>
			</ref>
			<ref id="B26">
				<mixed-citation>Hutter, F., Kotthoff, L., and Vanschoren, J. (2019). <italic>Automated Machine Learning: Methods, Systems, Challenges</italic>. New York, NY: Springer . 10.1007/978-3-030-05318-5</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Hutter</surname>
							<given-names>F.</given-names>
						</name>
						<name>
							<surname>Kotthoff</surname>
							<given-names>L.</given-names>
						</name>
						<name>
							<surname>Vanschoren</surname>
							<given-names>J</given-names>
						</name>
					</person-group>
					<year>2019</year>
					<source>Automated Machine Learning: Methods, Systems, Challenges</source>
					<publisher-loc>New York, NY</publisher-loc>
					<publisher-name>Springer</publisher-name>
					<pub-id pub-id-type="doi">10.1007/978-3-030-05318-5</pub-id>
				</element-citation>
			</ref>
			<ref id="B27">
				<mixed-citation>Jastrzebska, A. (2019). Time series classification through visual pattern recognition. <italic>Journal of King Saud University - Computer and Information Sciences</italic>. 10.1016/j.jksuci.2019.12.012</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Jastrzebska</surname>
							<given-names>A.</given-names>
						</name>
					</person-group>
					<year>2019</year>
					<article-title>Time series classification through visual pattern recognition</article-title>
					<source>Journal of King Saud University - Computer and Information Sciences</source>
					<pub-id pub-id-type="doi">10.1016/j.jksuci.2019.12.012</pub-id>
				</element-citation>
			</ref>
			<ref id="B28">
				<mixed-citation>Keogh, E., Chakrabarti, K., Pazzani, M., and Mehrotra, S. (2001). Dimensionality reduction for fast similarity search in large time series databases. <italic>Knowledge and Information Systems</italic>, 3(3), 263-286. 10.1007/PL00011669</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Keogh</surname>
							<given-names>E.</given-names>
						</name>
						<name>
							<surname>Chakrabarti</surname>
							<given-names>K.</given-names>
						</name>
						<name>
							<surname>Pazzani</surname>
							<given-names>M.</given-names>
						</name>
						<name>
							<surname>Mehrotra</surname>
							<given-names>S</given-names>
						</name>
					</person-group>
					<year>2001</year>
					<article-title>Dimensionality reduction for fast similarity search in large time series databases</article-title>
					<source>Knowledge and Information Systems</source>
					<volume>3</volume>
					<issue>3</issue>
					<fpage>263</fpage>
					<lpage>286</lpage>
					<pub-id pub-id-type="doi">10.1007/PL00011669</pub-id>
				</element-citation>
			</ref>
			<ref id="B29">
				<mixed-citation>Keogh, E., Zhu, Q., Hu, B., Hao, Y., Xi, X., Wei, L., and Ratanama-hatana, C. A. (2011). The UCR Time Series Classification/Clustering Homepage. <ext-link ext-link-type="uri" xlink:href="https://www.cs.ucr.edu/~eamonn/time_series_data/">https://www.cs.ucr.edu/~eamonn/time_series_data/</ext-link>
				</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Keogh</surname>
							<given-names>E.</given-names>
						</name>
						<name>
							<surname>Zhu</surname>
							<given-names>Q.</given-names>
						</name>
						<name>
							<surname>Hu</surname>
							<given-names>B.</given-names>
						</name>
						<name>
							<surname>Hao</surname>
							<given-names>Y.</given-names>
						</name>
						<name>
							<surname>Xi</surname>
							<given-names>X.</given-names>
						</name>
						<name>
							<surname>Wei</surname>
							<given-names>L.</given-names>
						</name>
						<name>
							<surname>Ratanama-hatana</surname>
							<given-names>C. A</given-names>
						</name>
					</person-group>
					<year>2011</year>
					<source>The UCR Time Series Classification/Clustering Homepage</source>
					<ext-link ext-link-type="uri" xlink:href="https://www.cs.ucr.edu/~eamonn/time_series_data/">https://www.cs.ucr.edu/~eamonn/time_series_data/</ext-link>
				</element-citation>
			</ref>
			<ref id="B30">
				<mixed-citation>Lin, J., Keogh, E., Wei, L., and Lonardi, S. (2007). Experiencing sax: a novel symbolic representation of time series. <italic>Data Mining and Knowledge Discovery</italic>, 15(2), 107-144. 10.1007/s10618-007-0064-z</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Lin</surname>
							<given-names>J.</given-names>
						</name>
						<name>
							<surname>Keogh</surname>
							<given-names>E.</given-names>
						</name>
						<name>
							<surname>Wei</surname>
							<given-names>L.</given-names>
						</name>
						<name>
							<surname>Lonardi</surname>
							<given-names>S</given-names>
						</name>
					</person-group>
					<year>2007</year>
					<article-title>Experiencing sax: a novel symbolic representation of time series.</article-title>
					<source>Data Mining and Knowledge Discovery</source>
					<volume>15</volume>
					<issue>2</issue>
					<fpage>107</fpage>
					<lpage>144</lpage>
					<pub-id pub-id-type="doi">10.1007/s10618-007-0064-z</pub-id>
				</element-citation>
			</ref>
			<ref id="B31">
				<mixed-citation>Olguín-Carbajal, M., Herrera-Lozada, J. C., Sandoval-Gutierrez, J., Vasquez-Gomez, J. I., Serrano-Talamantes, J. F., Chavez-Estrada, F. A., Rivera-Zarate, I., and Hernandez-Boláos, M. (2019). A micro-differential evolution algorithm for continuous complex functions. <italic>IEEE Access</italic>, 7, 172783-172795. 10.1109/ACCESS.2019.2954296</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Olguín-Carbajal</surname>
							<given-names>M.</given-names>
						</name>
						<name>
							<surname>Herrera-Lozada</surname>
							<given-names>J. C.</given-names>
						</name>
						<name>
							<surname>Sandoval-Gutierrez</surname>
							<given-names>J.</given-names>
						</name>
						<name>
							<surname>Vasquez-Gomez</surname>
							<given-names>J. I.</given-names>
						</name>
						<name>
							<surname>Serrano-Talamantes</surname>
							<given-names>J. F.</given-names>
						</name>
						<name>
							<surname>Chavez-Estrada</surname>
							<given-names>F. A.</given-names>
						</name>
						<name>
							<surname>Rivera-Zarate</surname>
							<given-names>I.</given-names>
						</name>
						<name>
							<surname>Hernandez-Boláos</surname>
							<given-names>M</given-names>
						</name>
					</person-group>
					<year>2019</year>
					<article-title>A micro-differential evolution algorithm for continuous complex functions</article-title>
					<source>IEEE Access</source>
					<issue>7</issue>
					<fpage>172783</fpage>
					<lpage>172795</lpage>
					<pub-id pub-id-type="doi">10.1109/ACCESS.2019.2954296</pub-id>
				</element-citation>
			</ref>
			<ref id="B32">
				<mixed-citation>Olson, R. S., Urbanowicz, R. J., Andrews, P. C., Lavender, N. A., Kidd, L. C., and Moore, J. H. (2016). Automating biomedical data science through tree-based pipeline optimization. In Squillero, G., and Burelli, P. (Eds.) <italic>European Conference on the Applications of Evolutionary Computation</italic> (pp. 123-137). Cham, Germany: Springer. <ext-link ext-link-type="uri" xlink:href="10.1007/9783-319-31204-09">10.1007/9783-319-31204-09</ext-link>
				</mixed-citation>
				<element-citation publication-type="confproc">
					<person-group person-group-type="author">
						<name>
							<surname>Olson</surname>
							<given-names>R. S.</given-names>
						</name>
						<name>
							<surname>Urbanowicz</surname>
							<given-names>R. J.</given-names>
						</name>
						<name>
							<surname>Andrews</surname>
							<given-names>P. C.</given-names>
						</name>
						<name>
							<surname>Lavender</surname>
							<given-names>N. A.</given-names>
						</name>
						<name>
							<surname>Kidd</surname>
							<given-names>L. C.</given-names>
						</name>
						<name>
							<surname>Moore</surname>
							<given-names>J. H</given-names>
						</name>
					</person-group>
					<year>2016</year>
					<source>Automating biomedical data science through tree-based pipeline optimization</source>
					<person-group person-group-type="author">
						<name>
							<surname>Squillero</surname>
							<given-names>G.</given-names>
						</name>
						<name>
							<surname>Burelli</surname>
							<given-names>P</given-names>
						</name>
					</person-group>
					<conf-name>European Conference on the Applications of Evolutionary Computation</conf-name>
					<fpage>123</fpage>
					<lpage>137</lpage>
					<publisher-loc>Cham, Germany</publisher-loc>
					<publisher-name>Springer</publisher-name>
					<ext-link ext-link-type="uri" xlink:href="10.1007/9783-319-31204-09">10.1007/9783-319-31204-09</ext-link>
				</element-citation>
			</ref>
			<ref id="B33">
				<mixed-citation>Olszewski, R. T. (2001). Generalized feature extraction for structural pattern recognition in time-series data (Doctoral thesis, Carnegie Mellon University, Pittsburgh, PA). <ext-link ext-link-type="uri" xlink:href="https://apps.dtic.mil/sti/pdfs/ADA457624.pdf">https://apps.dtic.mil/sti/pdfs/ADA457624.pdf</ext-link>
				</mixed-citation>
				<element-citation publication-type="thesis">
					<person-group person-group-type="author">
						<name>
							<surname>Olszewski</surname>
							<given-names>R. T</given-names>
						</name>
					</person-group>
					<year>2001</year>
					<source>Generalized feature extraction for structural pattern recognition in time-series data</source>
					<comment content-type="degree">Doctoral thesis</comment>
					<publisher-name>Carnegie Mellon University</publisher-name>
					<publisher-loc>Pittsburgh, PA</publisher-loc>
					<publisher-loc>Pittsburgh, PA</publisher-loc>
					<ext-link ext-link-type="uri" xlink:href="https://apps.dtic.mil/sti/pdfs/ADA457624.pdf">https://apps.dtic.mil/sti/pdfs/ADA457624.pdf</ext-link>
				</element-citation>
			</ref>
			<ref id="B34">
				<mixed-citation>Page, R. M., Lischeid, G., Epting, J., and Huggenberger, P. (2012). Principal component analysis of time series for identifying indicator variables for riverine groundwater extraction management. <italic>Journal of Hydrology</italic>, <italic>432</italic>, 137-144. 10.1016/j.jhydrol.2012.02.025</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Page</surname>
							<given-names>R. M.</given-names>
						</name>
						<name>
							<surname>Lischeid</surname>
							<given-names>G.</given-names>
						</name>
						<name>
							<surname>Epting</surname>
							<given-names>J.</given-names>
						</name>
						<name>
							<surname>Huggenberger</surname>
							<given-names>P</given-names>
						</name>
					</person-group>
					<year>2012</year>
					<article-title>Principal component analysis of time series for identifying indicator variables for riverine groundwater extraction management</article-title>
					<source>Journal of Hydrology</source>
					<issue>432</issue>
					<fpage>137</fpage>
					<lpage>144</lpage>
					<pub-id pub-id-type="doi">10.1016/j.jhydrol.2012.02.025</pub-id>
				</element-citation>
			</ref>
			<ref id="B35">
				<mixed-citation>Parsopoulos, K. E. (2009). Cooperative micro-differential evolution for high-dimensional problems. In ACM (Eds.) <italic>GECCO '09: Proceedings of the 11th Annual Conference on Genetic and Evolutionary Computation</italic> (pp. 531-538). New York, NY: ACM. 10.1145/1569901.1569975</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Parsopoulos</surname>
							<given-names>K. E</given-names>
						</name>
					</person-group>
					<year>2009</year>
					<chapter-title>Cooperative micro-differential evolution for high-dimensional problems</chapter-title>
					<person-group person-group-type="author">
						<collab>ACM</collab>
					</person-group>
					<source>GECCO '09: Proceedings of the 11th Annual Conference on Genetic and Evolutionary Computation</source>
					<fpage>531</fpage>
					<lpage>538</lpage>
					<publisher-loc>New York, NY</publisher-loc>
					<publisher-name>ACM</publisher-name>
					<pub-id pub-id-type="doi">10.1145/1569901.1569975</pub-id>
				</element-citation>
			</ref>
			<ref id="B36">
				<mixed-citation>Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., Dubourg, V., Vanderplas, J., Passos, A., and Cournapeau, D. (2011). Scikit-learn: Machine learning in python. <italic>The Journal of Machine Learning Research</italic>, <italic>12</italic>, 2825-2830.</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Pedregosa</surname>
							<given-names>F.</given-names>
						</name>
						<name>
							<surname>Varoquaux</surname>
							<given-names>G.</given-names>
						</name>
						<name>
							<surname>Gramfort</surname>
							<given-names>A.</given-names>
						</name>
						<name>
							<surname>Michel</surname>
							<given-names>V.</given-names>
						</name>
						<name>
							<surname>Thirion</surname>
							<given-names>B.</given-names>
						</name>
						<name>
							<surname>Grisel</surname>
							<given-names>O.</given-names>
						</name>
						<name>
							<surname>Blondel</surname>
							<given-names>M.</given-names>
						</name>
						<name>
							<surname>Prettenhofer</surname>
							<given-names>P.</given-names>
						</name>
						<name>
							<surname>Weiss</surname>
							<given-names>R.</given-names>
						</name>
						<name>
							<surname>Dubourg</surname>
							<given-names>V.</given-names>
						</name>
						<name>
							<surname>Vanderplas</surname>
							<given-names>J.</given-names>
						</name>
						<name>
							<surname>Passos</surname>
							<given-names>A.</given-names>
						</name>
						<name>
							<surname>Cournapeau</surname>
							<given-names>D</given-names>
						</name>
					</person-group>
					<year>2011</year>
					<article-title>Scikit-learn: Machine learning in python</article-title>
					<source>The Journal of Machine Learning Research</source>
					<issue>12</issue>
					<fpage>2825</fpage>
					<lpage>2830</lpage>
				</element-citation>
			</ref>
			<ref id="B37">
				<mixed-citation>Pérez-Castro, N., Acosta-Mesa, H., Mezura-Montes, E., and Cruz-Ramírez, N. (2015). Towards the full model selection in temporal databases by using micro-differential evolution. an empirical study. In IEEE (Eds.) <italic>2015 IEEE International Autumn Meeting on Power, Electronics and Computing (ROPEC)</italic> (pp. 1-6). New York, NY: IEEE . 10.1109/ROPEC.2015.7395161</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Pérez-Castro</surname>
							<given-names>N.</given-names>
						</name>
						<name>
							<surname>Acosta-Mesa</surname>
							<given-names>H.</given-names>
						</name>
						<name>
							<surname>Mezura-Montes</surname>
							<given-names>E.</given-names>
						</name>
						<name>
							<surname>Cruz-Ramírez</surname>
							<given-names>N</given-names>
						</name>
					</person-group>
					<year>2015</year>
					<source>Towards the full model selection in temporal databases by using micro-differential evolution. an empirical study</source>
					<person-group person-group-type="author">
						<collab>IEEE</collab>
					</person-group>
					<source>2015 IEEE International Autumn Meeting on Power, Electronics and Computing (ROPEC)</source>
					<fpage>1</fpage>
					<lpage>6</lpage>
					<publisher-loc>New York, NY</publisher-loc>
					<publisher-name>IEEE</publisher-name>
					<pub-id pub-id-type="doi">10.1109/ROPEC.2015.7395161</pub-id>
				</element-citation>
			</ref>
			<ref id="B38">
				<mixed-citation>Rashid, A. and Hossain, M. A. (2012) Challenging issues of spatio-temporal data mining. <italic>Computer Engineering and Intelligent Systems</italic>, 3(4), 55-63. <ext-link ext-link-type="uri" xlink:href="https://www.iiste.org/Journals/index.php/CEIS/article/view/1484">https://www.iiste.org/Journals/index.php/CEIS/article/view/1484</ext-link>
				</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Rashid</surname>
							<given-names>A.</given-names>
						</name>
						<name>
							<surname>Hossain</surname>
							<given-names>M. A</given-names>
						</name>
					</person-group>
					<year>2012</year>
					<article-title>Challenging issues of spatio-temporal data mining</article-title>
					<source>Computer Engineering and Intelligent Systems</source>
					<volume>3</volume>
					<issue>4</issue>
					<fpage>55</fpage>
					<lpage>63</lpage>
					<ext-link ext-link-type="uri" xlink:href="https://www.iiste.org/Journals/index.php/CEIS/article/view/1484">https://www.iiste.org/Journals/index.php/CEIS/article/view/1484</ext-link>
				</element-citation>
			</ref>
			<ref id="B39">
				<mixed-citation>Ratanamahatana, C. A. and Keogh, E. (2005). Three myths about dynamic time warping data mining. In SIAM (Eds.) <italic>Proceedings of the 2005 SIAM International Conference on Data Mining</italic> (pp. 506-510). Philadelphia, PA: Society for Industrial and Applied Mathematics . 10.1137/1.9781611972757.50</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Ratanamahatana</surname>
							<given-names>C. A.</given-names>
						</name>
						<name>
							<surname>Keogh</surname>
							<given-names>E</given-names>
						</name>
					</person-group>
					<year>2005</year>
					<source>Three myths about dynamic time warping data mining</source>
					<person-group person-group-type="author">
						<collab>SIAM</collab>
					</person-group>
					<source>Proceedings of the 2005 SIAM International Conference on Data Mining</source>
					<fpage>506</fpage>
					<lpage>510</lpage>
					<publisher-loc>Philadelphia, PA</publisher-loc>
					<publisher-name>Society for Industrial and Applied Mathematics</publisher-name>
					<pub-id pub-id-type="doi">10.1137/1.9781611972757.50</pub-id>
				</element-citation>
			</ref>
			<ref id="B40">
				<mixed-citation>Rice, J. R. (1976). The algorithm selection problem. In Rubinoff, M. and Yovits, M. C. (Eds.) <italic>Advances in computers</italic> (vol. 15, pp. 65-118). Amsterdam, Netherlands: Elsevier. 10.1016/S0065-2458(08)60520-3</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Rice</surname>
							<given-names>J. R</given-names>
						</name>
					</person-group>
					<year>1976</year>
					<chapter-title>The algorithm selection problem</chapter-title>
					<person-group person-group-type="author">
						<name>
							<surname>Rubinoff</surname>
							<given-names>M.</given-names>
						</name>
						<name>
							<surname>Yovits</surname>
							<given-names>M. C</given-names>
						</name>
					</person-group>
					<source>Advances in computers</source>
					<volume>15</volume>
					<fpage>65</fpage>
					<lpage>118</lpage>
					<publisher-loc>Amsterdam, Netherlands</publisher-loc>
					<publisher-name>Elsevier</publisher-name>
					<pub-id pub-id-type="doi">10.1016/S0065-2458(08)60520-3</pub-id>
				</element-citation>
			</ref>
			<ref id="B41">
				<mixed-citation>Rosales-Pérez, A., Escalante, H. J., Gonzalez, J. A., Reyes-Garcia, C. A., and Coello-Coello, C. A. (2013). Bias and variance multi-objective optimization for support vector machines model selection. In Sanches, J. A. M., Micó, L., and Cardoso, J. S. (Eds.) <italic>Iberian Conference on Pattern Recognition and Image Analysis</italic> (pp. 108-116). Berlin/Heidelberg, Germany: Springer. 10.1007/978-3-642-38628-212</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Rosales-Pérez</surname>
							<given-names>A.</given-names>
						</name>
						<name>
							<surname>Escalante</surname>
							<given-names>H. J.</given-names>
						</name>
						<name>
							<surname>Gonzalez</surname>
							<given-names>J. A.</given-names>
						</name>
						<name>
							<surname>Reyes-Garcia</surname>
							<given-names>C. A.</given-names>
						</name>
						<name>
							<surname>Coello-Coello</surname>
							<given-names>C. A</given-names>
						</name>
					</person-group>
					<year>2013</year>
					<source>Bias and variance multi-objective optimization for support vector machines model selection</source>
					<person-group person-group-type="author">
						<name>
							<surname>Sanches</surname>
							<given-names>J. A. M.</given-names>
						</name>
						<name>
							<surname>Micó</surname>
							<given-names>L.</given-names>
						</name>
						<name>
							<surname>Cardoso</surname>
							<given-names>J. S</given-names>
						</name>
					</person-group>
					<source>Iberian Conference on Pattern Recognition and Image Analysis</source>
					<fpage>108</fpage>
					<lpage>116</lpage>
					<publisher-loc>Berlin/Heidelberg, Germany</publisher-loc>
					<publisher-name>Springer</publisher-name>
					<pub-id pub-id-type="doi">10.1007/978-3-642-38628-212</pub-id>
				</element-citation>
			</ref>
			<ref id="B42">
				<mixed-citation>Rosales-Pérez, A., Gonzalez, J. A., Coello-Coello, C. A., Escalante, H. J., and Reyes-Garcia, C. A. (2015). Surrogate-assisted multi-objective model selection for support vector machines. <italic>Neurocomputing</italic>, <italic>150</italic>, 163-172. 10.1016/j.neucom.2014.08.075</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Rosales-Pérez</surname>
							<given-names>A.</given-names>
						</name>
						<name>
							<surname>Gonzalez</surname>
							<given-names>J. A.</given-names>
						</name>
						<name>
							<surname>Coello-Coello</surname>
							<given-names>C. A.</given-names>
						</name>
						<name>
							<surname>Escalante</surname>
							<given-names>H. J.</given-names>
						</name>
						<name>
							<surname>Reyes-Garcia</surname>
							<given-names>C. A</given-names>
						</name>
					</person-group>
					<year>2015</year>
					<article-title>Surrogate-assisted multi-objective model selection for support vector machines</article-title>
					<source>Neurocomputing</source>
					<issue>150</issue>
					<fpage>163</fpage>
					<lpage>172</lpage>
					<pub-id pub-id-type="doi">10.1016/j.neucom.2014.08.075</pub-id>
				</element-citation>
			</ref>
			<ref id="B43">
				<mixed-citation>Rosales-Pérez, A., Gonzalez, J. A., Coello-Coello, C. A., Escalante, H. J., and Reyes-Garcia, C. A. (2014). Multi-objective model type selection. <italic>Neurocomputing</italic>, <italic>146</italic>, 83-94. 10.1016/j.neucom.2014.05.077</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Rosales-Pérez</surname>
							<given-names>A.</given-names>
						</name>
						<name>
							<surname>Gonzalez</surname>
							<given-names>J. A.</given-names>
						</name>
						<name>
							<surname>Coello-Coello</surname>
							<given-names>C. A.</given-names>
						</name>
						<name>
							<surname>Escalante</surname>
							<given-names>H. J.</given-names>
						</name>
						<name>
							<surname>Reyes-Garcia</surname>
							<given-names>C. A</given-names>
						</name>
					</person-group>
					<year>2014</year>
					<article-title>Multi-objective model type selection</article-title>
					<source>Neurocomputing</source>
					<issue>146</issue>
					<fpage>83</fpage>
					<lpage>94</lpage>
					<pub-id pub-id-type="doi">10.1016/j.neucom.2014.05.077</pub-id>
				</element-citation>
			</ref>
			<ref id="B44">
				<mixed-citation>Roverso, D. (2000). Multivariate temporal classification by windowed wavelet decomposition and recurrent neural networks. In ANS (Eds.) <italic>3rdANS international topical meeting on nuclear plant instrumentation, control and human-machine interface</italic> (vol. 20, pp. 527-538). La Grange Park, IL: American Nuclear Society.</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Roverso</surname>
							<given-names>D</given-names>
						</name>
					</person-group>
					<year>2000</year>
					<source>Multivariate temporal classification by windowed wavelet decomposition and recurrent neural networks</source>
					<person-group person-group-type="author">
						<collab>ANS</collab>
					</person-group>
					<source>3rdANS international topical meeting on nuclear plant instrumentation, control and human-machine interface</source>
					<volume>20</volume>
					<fpage>527</fpage>
					<lpage>538</lpage>
					<publisher-loc>La Grange Park, IL</publisher-loc>
					<publisher-name>American Nuclear Society</publisher-name>
				</element-citation>
			</ref>
			<ref id="B45">
				<mixed-citation>Rydning, D. R.-J. G.-J. (2018). The digitization of the world from edge to core. <ext-link ext-link-type="uri" xlink:href="http://cloudcode.me/media/1014/idc.pdf">http://cloudcode.me/media/1014/idc.pdf</ext-link>
				</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Rydning</surname>
							<given-names>D. R.-J. G.-J</given-names>
						</name>
					</person-group>
					<year>2018</year>
					<source>The digitization of the world from edge to core</source>
					<ext-link ext-link-type="uri" xlink:href="http://cloudcode.me/media/1014/idc.pdf">http://cloudcode.me/media/1014/idc.pdf</ext-link>
				</element-citation>
			</ref>
			<ref id="B46">
				<mixed-citation>Saito, N. (2000). Local feature extraction and its applications using a library of bases. In Coifman, R. (Ed.) <italic>Topics in Analysis and Its Applications: Selected Theses</italic> (pp. 269-451). 10.1142/97898128133050005</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Saito</surname>
							<given-names>N</given-names>
						</name>
					</person-group>
					<year>2000</year>
					<chapter-title>Local feature extraction and its applications using a library of bases</chapter-title>
					<person-group person-group-type="author">
						<name>
							<surname>Coifman</surname>
							<given-names>R</given-names>
						</name>
					</person-group>
					<source>Topics in Analysis and Its Applications: Selected Theses</source>
					<fpage>269</fpage>
					<lpage>451</lpage>
					<pub-id pub-id-type="doi">10.1142/97898128133050005</pub-id>
				</element-citation>
			</ref>
			<ref id="B47">
				<mixed-citation>Salehinejad, H., Rahnamayan, S., and Tizhoosh, H. R. (2017). Micro-differential evolution: Diversity enhancement and a comparative study. <italic>Applied Soft Computing</italic>, <italic>52</italic>, 812-833. 10.1016/j.asoc.2016.09.042</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Salehinejad</surname>
							<given-names>H.</given-names>
						</name>
						<name>
							<surname>Rahnamayan</surname>
							<given-names>S.</given-names>
						</name>
						<name>
							<surname>Tizhoosh</surname>
							<given-names>H. R</given-names>
						</name>
					</person-group>
					<year>2017</year>
					<article-title>Micro-differential evolution: Diversity enhancement and a comparative study</article-title>
					<source>Applied Soft Computing</source>
					<issue>52</issue>
					<fpage>812</fpage>
					<lpage>833</lpage>
					<pub-id pub-id-type="doi">10.1016/j.asoc.2016.09.042</pub-id>
				</element-citation>
			</ref>
			<ref id="B48">
				<mixed-citation>Savitzky, A. and Golay, M. J. E. (1964). Smoothing and differentiation of data by simplified least squares procedures. <italic>Analytical Chemistry</italic>, 36(8), 1627-1639. 10.1021/ac60214a047</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Savitzky</surname>
							<given-names>A.</given-names>
						</name>
						<name>
							<surname>Golay</surname>
							<given-names>M. J. E</given-names>
						</name>
					</person-group>
					<year>1964</year>
					<article-title>Smoothing and differentiation of data by simplified least squares procedures</article-title>
					<source>Analytical Chemistry</source>
					<volume>36</volume>
					<issue>8</issue>
					<fpage>1627</fpage>
					<lpage>1639</lpage>
					<pub-id pub-id-type="doi">10.1021/ac60214a047</pub-id>
				</element-citation>
			</ref>
			<ref id="B49">
				<mixed-citation>Shahriari, B., Swersky, K., Wang, Z., Adams, R. P., and de Freitas, N. (2016). Taking the human out of the loop: A review of bayesian optimization. <italic>Proceedings of the IEEE</italic>, 104(1), 148-175. 10.1109/JPROC.2015.2494218</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Shahriari</surname>
							<given-names>B.</given-names>
						</name>
						<name>
							<surname>Swersky</surname>
							<given-names>K.</given-names>
						</name>
						<name>
							<surname>Wang</surname>
							<given-names>Z.</given-names>
						</name>
						<name>
							<surname>Adams</surname>
							<given-names>R. P.</given-names>
						</name>
						<name>
							<surname>de Freitas</surname>
							<given-names>N</given-names>
						</name>
					</person-group>
					<year>2016</year>
					<article-title>Taking the human out of the loop: A review of bayesian optimization</article-title>
					<source>Proceedings of the IEEE</source>
					<volume>104</volume>
					<issue>1</issue>
					<fpage>148</fpage>
					<lpage>175</lpage>
					<pub-id pub-id-type="doi">10.1109/JPROC.2015.2494218</pub-id>
				</element-citation>
			</ref>
			<ref id="B50">
				<mixed-citation>Sun, J., Yang, Y., Liu, Y., Chen, C., Rao, W., and Bai, Y. (2019). Univariate time series classification using information geometry. <italic>Pattern Recognition</italic>, <italic>95</italic>, 24-35. 10.1016/j.patcog.2019.05.0406</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Sun</surname>
							<given-names>J.</given-names>
						</name>
						<name>
							<surname>Yang</surname>
							<given-names>Y.</given-names>
						</name>
						<name>
							<surname>Liu</surname>
							<given-names>Y.</given-names>
						</name>
						<name>
							<surname>Chen</surname>
							<given-names>C.</given-names>
						</name>
						<name>
							<surname>Rao</surname>
							<given-names>W.</given-names>
						</name>
						<name>
							<surname>Bai</surname>
							<given-names>Y</given-names>
						</name>
					</person-group>
					<year>2019</year>
					<article-title>Univariate time series classification using information geometry</article-title>
					<source>Pattern Recognition</source>
					<issue>95</issue>
					<fpage>24</fpage>
					<lpage>35</lpage>
					<pub-id pub-id-type="doi">10.1016/j.patcog.2019.05.0406</pub-id>
				</element-citation>
			</ref>
			<ref id="B51">
				<mixed-citation>Sun, Q., Pfahringer, B., and Mayo, M. (2013). Towards a framework for designing full model selection and optimization systems. In Zhou, Z.-H., Roli, F., and Kittler, J. (Eds.) <italic>International Workshop on Multiple Classifier Systems</italic> (pp. 259-270). Springer, Berlin, Heidelberg. 10.1016/j.patcog.2019.05.040</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Sun</surname>
							<given-names>Q.</given-names>
						</name>
						<name>
							<surname>Pfahringer</surname>
							<given-names>B.</given-names>
						</name>
						<name>
							<surname>Mayo</surname>
							<given-names>M</given-names>
						</name>
					</person-group>
					<year>2013</year>
					<chapter-title>Towards a framework for designing full model selection and optimization systems</chapter-title>
					<person-group person-group-type="author">
						<name>
							<surname>Zhou</surname>
							<given-names>Z.-H.</given-names>
						</name>
						<name>
							<surname>Roli</surname>
							<given-names>F.</given-names>
						</name>
						<name>
							<surname>Kittler</surname>
							<given-names>J</given-names>
						</name>
					</person-group>
					<source>International Workshop on Multiple Classifier Systems</source>
					<fpage>259</fpage>
					<lpage>270</lpage>
					<publisher-loc>Springer, Berlin</publisher-loc>
					<publisher-name>Heidelberg</publisher-name>
					<pub-id pub-id-type="doi">10.1016/j.patcog.2019.05.040</pub-id>
				</element-citation>
			</ref>
			<ref id="B52">
				<mixed-citation>Talbi, E. (2009). <italic>Metaheuristics: From Design to Implementation</italic>. John Wiley &amp; Sons.</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Talbi</surname>
							<given-names>E</given-names>
						</name>
					</person-group>
					<year>2009</year>
					<source>Metaheuristics: From Design to Implementation</source>
					<publisher-name>John Wiley &amp; Sons</publisher-name>
				</element-citation>
			</ref>
			<ref id="B53">
				<mixed-citation>Viveros-Jiménez, F., Mezura-Montes, E., and Gelbukh, A. (2012). Empirical analysis of a micro-evolutionary algorithm for numerical optimization. <italic>International Journal of Physical Sciences</italic>, 7(8), 1235-1258. 10.5897/IJPS11.303</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Viveros-Jiménez</surname>
							<given-names>F.</given-names>
						</name>
						<name>
							<surname>Mezura-Montes</surname>
							<given-names>E.</given-names>
						</name>
						<name>
							<surname>Gelbukh</surname>
							<given-names>A</given-names>
						</name>
					</person-group>
					<year>2012</year>
					<article-title>Empirical analysis of a micro-evolutionary algorithm for numerical optimization</article-title>
					<source>International Journal of Physical Sciences</source>
					<volume>7</volume>
					<issue>8</issue>
					<fpage>1235</fpage>
					<lpage>1258</lpage>
					<pub-id pub-id-type="doi">10.5897/IJPS11.303</pub-id>
				</element-citation>
			</ref>
			<ref id="B54">
				<mixed-citation>Yang, M., Li, C., Cai, Z., and Guan, J. (2015). Differential evolution with auto-enhanced population diversity. <italic>IEEE transactions on cybernetics</italic>, 45(2), 302-315. 10.1109/TCYB.2014.2339495</mixed-citation>
				<element-citation publication-type="journal">
					<person-group person-group-type="author">
						<name>
							<surname>Yang</surname>
							<given-names>M.</given-names>
						</name>
						<name>
							<surname>Li</surname>
							<given-names>C.</given-names>
						</name>
						<name>
							<surname>Cai</surname>
							<given-names>Z.</given-names>
						</name>
						<name>
							<surname>Guan</surname>
							<given-names>J</given-names>
						</name>
					</person-group>
					<year>2015</year>
					<article-title>Differential evolution with auto-enhanced population diversity</article-title>
					<source>IEEE transactions on cybernetics</source>
					<volume>45</volume>
					<issue>2</issue>
					<fpage>302</fpage>
					<lpage>315</lpage>
					<pub-id pub-id-type="doi">10.1109/TCYB.2014.2339495</pub-id>
				</element-citation>
			</ref>
			<ref id="B55">
				<mixed-citation>Yang, Y. (2017). Chapter 2 - temporal data mining. In Y. Yang (Ed.) <italic>Temporal Data Miningvia unsupervised Ensemble Learning</italic> (pp. 9-18). Amsterdam, Netherlands: Elsevier . 10.1016/B978-0-12-811654-8.00002-6</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Yang</surname>
							<given-names>Y</given-names>
						</name>
					</person-group>
					<year>2017</year>
					<chapter-title>Chapter 2 - temporal data mining</chapter-title>
					<person-group person-group-type="author">Y. Yang</person-group>
					<source>Temporal Data Miningvia unsupervised Ensemble Learning</source>
					<fpage>9</fpage>
					<lpage>18</lpage>
					<publisher-loc>Amsterdam, Netherlands</publisher-loc>
					<publisher-name>Elsevier</publisher-name>
					<pub-id pub-id-type="doi">10.1016/B978-0-12-811654-8.00002-6</pub-id>
				</element-citation>
			</ref>
			<ref id="B56">
				<mixed-citation>Yu, T. and Zhu, H. (2020). Hyper-parameter optimization: A review of algorithms and applications. <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/pdf/2003.05689.pdf">https://arxiv.org/pdf/2003.05689.pdf</ext-link>
				</mixed-citation>
				<element-citation publication-type="book">
					<person-group person-group-type="author">
						<name>
							<surname>Yu</surname>
							<given-names>T.</given-names>
						</name>
						<name>
							<surname>Zhu</surname>
							<given-names>H</given-names>
						</name>
					</person-group>
					<year>2020</year>
					<source>Hyper-parameter optimization: A review of algorithms and applications</source>
					<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/pdf/2003.05689.pdf">https://arxiv.org/pdf/2003.05689.pdf</ext-link>
				</element-citation>
			</ref>
		</ref-list>
		<fn-group>
			<fn fn-type="other" id="fn1">
				<label>How to cite: </label>
				<p>Pérez-Castro, N., Acosta-Mesa, H. G., Mezura-Montes, E., Cruz-Ramírez, N. (2021). Full Model Selection Problem and Pipelines for Time-Series Databases: Contrasting Population-Based and Single-Point Search Metaheuristics. <italic>Ingeniería e Investigación, 41(3),</italic> e79308. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.15446/ing.investig.v41n3.79308">https://doi.org/10.15446/ing.investig.v41n3.79308</ext-link>
				</p>
			</fn>
		</fn-group>
	</back>
</article>