Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

CrisisDPS: Crisis Data Processing Services

2019, International Conference on Information Systems for Crisis Response and Management (ISCRAM)

Over the last few years, extensive research has been conducted to develop technologies to support humanitarian aid tasks. However, many technologies are still limited as they require both manual and automatic approaches, and more importantly, are not ready to be integrated into the disaster response workflows. To tackle this limitation, we develop automatic data processing services that are freely and publicly available, and made to be simple, efficient, and accessible to non-experts. Our services take textual messages (e.g., tweets, Facebook posts, SMS) as input to determine (i) which disaster type the message belongs to, (ii) whether it is informative or not, and (iii) what type of humanitarian information it conveys. We built our services upon machine learning classifiers that are obtained from large-scale comparative experiments utilizing both classical and deep learning algorithms. Our services outperform state-of-the-art publicly available tools in terms of classification accuracy.

F. Alam et al. Crisis Data Processing Services CrisisDPS: Crisis Data Processing Services Firoj Alam∗ Muhammad Imran Qatar Computing Research Institute Hamad Bin Khalifa University Doha, Qatar falam@hbku.edu.qa Qatar Computing Research Institute Hamad Bin Khalifa University Doha, Qatar mimran@hbku.edu.qa Ferda Ofli Qatar Computing Research Institute Hamad Bin Khalifa University Doha, Qatar fofli@hbku.edu.qa ABSTRACT Over the last few years, extensive research has been conducted to develop technologies to support humanitarian aid tasks. However, many technologies are still limited as they require both manual and automatic approaches, and more importantly, are not ready to be integrated into the disaster response workflows. To tackle this limitation, we develop automatic data processing services that are freely and publicly available, and made to be simple, efficient, and accessible to non-experts. Our services take textual messages (e.g., tweets, Facebook posts, SMS) as input to determine (i) which disaster type the message belongs to, (ii) whether it is informative or not, and (iii) what type of humanitarian information it conveys. We built our services upon machine learning classifiers that are obtained from large-scale comparative experiments utilizing both classical and deep learning algorithms. Our services outperform state-of-the-art publicly available tools in terms of classification accuracy. Keywords Social media, humanitarian data processing, text classification, application programming interfaces, data processing services. INTRODUCTION In the last decade, we have witnessed a significant increase in the use of Information and Communications Technologies (ICT), mobile devices, and sensors during time-critical events such as natural or human-induced disasters (Wattegama 2014; Imran, Castillo, Diaz, et al. 2015). This increase in the number of data sources results in producing more and more data each day. These data sources include social media platforms (e.g., Facebook and Twitter), WhatsApp groups, SMS communications, web blogs, news articles, RSS feeds, and many more. The combination of easy-to-use technologies and data sources generate vast amounts of data at high velocity, i.e., thousands of documents appear each second. Extracting useful information from these sources is highly important for humanitarian organizations, government agencies, and public administrative authorities to make timely decisions and to launch relief efforts during emergency situations (Starbird et al. 2010; Vieweg, Hughes, et al. 2010). The information needs of these stakeholders vary depending on their role, responsibilities, and the situation they are dealing with (Vieweg, Castillo, et al. 2014). Often extracting such information becomes difficult for non-technical domain experts from these high volume data sources. It is reported in the literature that manual analysis of such high-volume and high-velocity social media data streams is impossible (Hiltz et al. 2014; Ludwig et al. 2015). ∗ corresponding author CoRe Paper – Social Media in Crises and Conflicts Proceedings of the 16th ISCRAM Conference - València, Spain May 2019 Zeno Franco, José J. González and José H. Canós, eds. F. Alam et al. Crisis Data Processing Services Dealing with such large data sources and extracting useful information from them involve many challenges such as parsing unstructured and brief content, filtering out irrelevant and noisy content, handling information overload, among others. Over the last few years, many Artificial Intelligence (AI) techniques and computational methods have been proposed to process social media data for disaster response and management. These techniques aim to solve various challenges ranging from information filtering, overload, and categorization to summarization (Castillo 2016; Rudra et al. 2018; Imran, Castillo, Diaz, et al. 2015). In addition, there have been publicly available tools that collect social media data and automatically categorize them during disaster events (Imran, Castillo, Lucas, et al. 2014; Burel, Saif, Fernandez, et al. 2017; Meier 2012; Okolloh 2009). The common features of these systems include classifying tweets, grouping them into clusters, visualizing them into timelines, geotagging them onto maps, and visualizing topics, sentiment, and concepts over time. We elaborate further on these systems in the Related Work section. In this paper, as an attempt to shed more light in the direction motivated by the aforementioned studies, we develop automatic data processing services to assist both technical and non-technical end-users including crisis managers, officials from formal humanitarian organizations, first responders, volunteering organizations, non-governmental organizations (NGOs), and researchers in the crisis informatics community in their humanitarian tasks. Specifically, we provide the following services, which are also the contributions of this paper: • Disaster type classification: The aim of this service is to determine whether a text message (e.g., a tweet or a Facebook post or an SMS) is about one of the six disaster types, namely earthquake, fire, flood, hurricane, bombing, and shooting. • Informativeness classification: This service determines whether a text message is informative for humanitarian aid purposes or not. Informative messages contain useful information for response organizations to plan and launch relief efforts. • Humanitarian information type classification: The aim of this service is to classify whether a text message is about one of the following ten humanitarian information types: affected individual, caution and advice, displaced and evacuations, donation and volunteering, infrastructure and utilities damage, injured or dead people, missing and found people, response efforts, and sympathy and support. More details about these humanitarian information types are given in the later sections. • Crisis Data Processing Services (CrisisDPS): To provide the aforementioned three classification services to end-users, we develop an end-to-end system that offers various RESTful Application Program Interfaces (APIs). Specifically, we provide separate APIs for (i) single item processing and (ii) batch processing in addition to (iii) a file processing service. • State-of-the-art classification models: We develop state-of-the-art supervised machine learning models to support the aforementioned three classification services. To develop the underlying machine learning models, we have conducted extensive experiments using both classical and deep learning algorithms. The classical algorithms include Naïve Bayes (NB) (McCallum, Nigam, et al. 1998), Support Vector Machines (SVM) (Hearst et al. 1998) and Random Forests (RF) (Liaw, Wiener, et al. 2002) whereas the deep learning algorithms include Convolutional Neural Networks (CNN) (LeCun et al. 1989), and Long Short-Term Memory Neural Networks (LSTM) (Hochreiter and Schmidhuber 1997). We have also collected and combined many publicly available datasets such as CrisisLex1, CrisisNLP2, among others. While consolidating these datasets, one of the challenges was the mismatch between class labels across datasets. We manually mapped the mismatches and selected the class labels that can effectively serve humanitarian purposes. More details can be found in the Datasets section. We make all of our services publicly available to potential end-users at no cost. The rest of the paper is organized as follows. In the next section, we present a literature review. Next, we describe our data processing services. After that, we present the details of the datasets that we used to develop our machine learning models. Then, we present our experiments, results and discussion. Finally, we conclude the paper in the last section. 1http://crisislex.org/ 2http://crisisnlp.qcri.org/ CoRe Paper – Social Media in Crises and Conflicts Proceedings of the 16th ISCRAM Conference - València, Spain May 2019 Zeno Franco, José J. González and José H. Canós, eds. F. Alam et al. Crisis Data Processing Services RELATED WORK There has been a significant progress in crisis computing research with a particular focus on mainly facilitating humanitarian organizations in their relief efforts and supporting their decision-making processes. One of the key research activities has been analyzing social media content. There are many studies that leverage social media such as Twitter, Facebook and YouTube for curating, analyzing, and summarizing crisis-related information in order to extract situational awareness and actionable information (Zade et al. 2018; Imran, Castillo, Lucas, et al. 2014; Vieweg, Hughes, et al. 2010; Terpstra et al. 2012; Tsou et al. 2017). Among different social media platforms, Twitter has been used extensively due to the fact that information appears timely and can be accessed and processed effectively, which are the required key features for the humanitarian organizations. Apart from research studies, many systems have been developed over time, some of which are publicly and freely available. We briefly cover some of these research studies and tools in this section. Below we provide a list of available tools and systems that utilize social media data to extract crisis-related information. • Ushahidi (Meier 2012; Okolloh 2009)3: Ushahidi is a platform that started its journey back in 2008 as a free and open-source platform that allowed interested individuals and groups to create live, interactive maps. Initially, it aimed to visualize post-election messages on a map. Its current functionality includes mapping and visualization tools to create real-time, dynamic, and multifaceted crisis maps. It allows for visualizing how a crisis situation is evolving over time. A notable deployment is 2010 Haiti earthquake (Goolsby 2010). It is currently providing one-month-free and premium supports to its users. • TweetTracker (Kumar et al. 2011)4: TweetTracker is another application aimed at supporting Humanitarian Aid and Disaster Relief (HADR) responders to monitor and analyze social media content during crisis situations. The functionality of this tool includes analyzing location, real-time trending, data filtering, and historical analysis. • Tweedr5: Tweedr is another tool that aimed to extract actionable information during natural disasters from Twitter. The functionality of the tool includes clustering and classification. • Artificial Intelligence for Disaster Response (AIDR) (Imran, Castillo, Lucas, et al. 2014)6: AIDR is an active system that facilitates users (e.g., humanitarian organizations) to collect tweets during disaster events and create classifiers on-the-fly by annotating incoming Twitter data according to the categories of their own choices. It is freely available with minimal usability effort. • Twitris (Purohit and Sheth 2013)7: Twitris application is also targeted for Twitter content analysis. The functionality of this system includes collecting, aggregating, and analyzing tweets to give deeper insights as well as facilitate the research and development process. • SensePlace2 (MacEachren et al. 2011)8: SensePlace2 application is focused more on extracting spatiotemporal information with analytic capabilities to deal with large volumes of tweets. • Enhanced Messaging for the Emergency Response Sector (EMERSE) (Caragea et al. 2011): This system has been developed to analyze tweets and SMS messages. The major functionality of this system includes Twitter crawler, machine translation system, and classification modules while also supporting users with an iPhone application. • Emergency Situation Awareness (ESA) (Yin et al. 2012)9: ESA is aimed at enhancing situational awareness with respect to crises induced by natural hazards, particularly earthquakes. They focus on spatio-temporal information and functionalities include event detection, text classification, online clustering, and geotagging. • Twicident (Abel et al. 2012)10: Twicident is a framework and system that automatically track and filter information, which is relevant for the real-world incidents or crises. 3https://www.ushahidi.com/ 4http://tweettracker.fulton.asu.edu/ 5https://github.com/dssg/tweedr/ 6http://aidr.qcri.org/ 7http://twitris.knoesis.org/ 8https://www.geovista.psu.edu/SensePlace2/ 9https://esa.csiro.au/aus/index.html 10http://www.wis.ewi.tudelft.nl/twitcident/ CoRe Paper – Social Media in Crises and Conflicts Proceedings of the 16th ISCRAM Conference - València, Spain May 2019 Zeno Franco, José J. González and José H. Canós, eds. F. Alam et al. Crisis Data Processing Services • Pipeline for Post-Crisis Twitter Data Acquisition (Kejriwal and Gu 2018): This is an ongoing effort that aims to develop a system that requires minimal supervision in an active learning framework. • Crisis Event Extraction Service (CREES) (Burel and Alani 2018)11: CREES is an open-source web API to automatically classify social media posts during crisis situations. The system is developed for three different tasks, where the classification models are based on deep learning algorithms. One can deploy the system on their own server to use the API functionality. This system is currently integrated into the Ushahidi platform as part of the COMRADES project12. • Botometer (Varol et al. 2017)13: Botometer is an application aimed to detect accounts that are controlled by automated tools, termed as social bots. Such an application is important to detect and filter fake or manipulated content on social media. As for the crisis-related tweet classification research, current literature explores both classical algorithms (e.g., Maximum Entropy, Logistic Regression (LR), Random Forest (RF), Naïve Bayes (NB) classifier, Support Vector Machines (SVM), and Conditional Random Fields (CRFs)) and deep learning techniques (e.g., Convolutional Neural Networks (CNN), Recurrent Neural Networks (RNN), and Long Short-term Memory Neural Networks (LSTM)). In (Imran, S. Elbassuoni, et al. 2013; Imran, S. M. Elbassuoni, et al. 2013), authors report different datasets and NB- and CRF-based classifiers to extract actionable information during disaster events. The feature sets used for developing the classifiers include word unigrams, bigrams, Part-of-Speech (POS) tags, among others. In (Nguyen et al. 2017), authors report a comparative study using classical and deep learning-based algorithms, including SVM, LR, RF, and CNN. Their experimental setting also consists of in-domain and out-domain configurations. Across all experimental configurations, CNN outperforms other algorithms. A similar comparative study has been conducted in (Burel, Saif, Fernandez, et al. 2017; Burel and Alani 2018), where the authors explore different types of features. Their study includes three classification tasks such as (i) related vs. not-related, (ii) event types, and (iii) information types. Their findings suggest that SVM and CNN provide very competitive results. The study of Aipe et al. 2018 explores CNN with other tweet-centric (e.g., user mentions, hashtags, and their combinations) features. They report results that outperform state-of-the-art results, which used a similar dataset such as CrisisNLP14. However, different than our work, they map class labels of CrisisNLP with a set of their own classes. In another study, Neppalli et al. 2018 compare NB, CNN and Recurrent Neural Network (RNN) with different feature combinations and suggest that CNN outperforms the other models. For a more extensive literature review on this subject, we refer the reader to the study conducted by Imran, Castillo, Diaz, et al. 2015. SYSTEM AND DATA PROCESSING SERVICES In this work, we present CrisisDPS, a system to process crisis-related data to support multiple humanitarian tasks. The system is developed to facilitate both technical and non-technical users in the domains of crisis informatics and humanitarian aid working at various levels in an organizational hierarchy, e.g., in an Incident-Command System (ICS)15, Public Information Officers (PIOs), Safety officers, logistics sections, etc. As most humanitarian organizations already have their information systems, our goal is to provide them with ready-to-use data processing services that can be easily integrated into their existing systems, for example, through a new widget backed by one of the CrisisDPS services. Figure 1 depicts various functionalities of the CrisisDPS system at a high-level. The system mainly provides three types of data processing services, namely (i) disaster type classification, (ii) informativeness classification, and (iii) humanitarian information type classification. We describe each of these services in detail later in this section. Each classification service offers three types of functionality: (i) single-item processing API, (ii) batch processing API, and (iii) file processing service. Furthermore, we also offer a web demo service to test the system online. Data Processing Services The three types of data processing services provided by the CrisisDPS system are described below. 1. Disaster type classification: This service aims to classify a given text message into one of the six disaster types, namely, earthquake, fire, flood, hurricane, bombing, and shooting. We also offer a general category called none, 11https://evhart.github.io/crees/ 12https://www.comrades-project.eu/ 13https://botometer.iuni.iu.edu/#!/api 14http://crisisnlp.qcri.org/ 15https://en.wikipedia.org/wiki/IncidentC ommandS ystem CoRe Paper – Social Media in Crises and Conflicts Proceedings of the 16th ISCRAM Conference - València, Spain May 2019 Zeno Franco, José J. González and José H. Canós, eds. F. Alam et al. Crisis Data Processing Services Single Item File Processing Processing API Batch Processing API (.txt, .tsv,.csv,.json) API Services Client-side Classifiers Server-side Disaster type Informative Humanitarian Figure 1. CrisisDPS: Crisis Data Processing Services which represents everything else that does not belong to any of the six disaster types. The motivation behind this service is to determine whether a message belongs to a particular disaster event or not before dispatching it for further processing. Moreover, this service could be useful for decision-makers in an emergency department who are dealing with multiple disasters at a time helping them to distinguish disaster-specific messages. 2. Informativeness classification: To determine whether a text message contains some useful information for disaster managers or not, we provide the “informativeness” service. Given a text message, the service classifies it either as informative or not informative. We consider a message as informative if it contains some useful information for humanitarian aid; such as disaster-related warnings, reports about injured, dead, or affected people, rescue requests, volunteering or donation offers, reports of damaged houses, roads, etc. 3. Humanitarian information type classification: Given a message is informative, which means it contains some useful information for humanitarian aid, the next task is to determine what kind of useful information it conveys. For this purpose, we provide humanitarian information type services. These services correspond to a set of automatic classification services to classify a given text message into one of the ten humanitarian information types described below: 1. Affected individual: messages that contain information about affected people due to a disaster event; 2. Caution and advice: messages that report warnings, cautions, and give advice to people in the disaster area; 3. Displaced and evacuations: messages that report about displaced people or evacuations due to the disaster event; 4. Donation and volunteering: messages that request for donations, e.g., food, water, shelter, or offer help or volunteering services; 5. Infrastructure and utilities damage: messages that report damages to built structures such as buildings, roads, and bridges; 6. Injured or dead people: messages that report injured or dead people due to the disaster event; 7. Missing and found people: messages that report missing or found people due to the disaster event; 8. Requests or needs: messages that contain requests or urgent needs of affected people; 9. Response efforts: messages that report ongoing response efforts by humanitarian organizations, NGOs, and volunteers; 10. Sympathy and support: messages that convey thoughts, prayers, sympathy, and support to the victims of the disaster; CoRe Paper – Social Media in Crises and Conflicts Proceedings of the 16th ISCRAM Conference - València, Spain May 2019 Zeno Franco, José J. González and José H. Canós, eds. F. Alam et al. Crisis Data Processing Services 11. Not informative: messages that do not belong to any of the previous categories or do not contain any useful information. All three types of classification services are backed by the state-of-the-art machine learning classifiers. Specifically, we used supervised machine learning approaches to train the classifiers. More details about these classifiers and the datasets used for training them are given in the next section. CrisisDPS APIs The CrisisDPS system provides three types of RESTful APIs for each one of the three classification services. Next, we describe these APIs: 1. Single-item processing API: This programmatic API provides support to classify a single item (e.g., one tweet, or a Facebook message) at a time. Given an item, one can specify one or more classification services to classify it. For example, given a tweet, we want to determine whether it is an earthquake or a hurricane tweet and whether it contains some useful information or not. Moreover, if it contains useful information, what type of information it contains. 2. Batch processing API: To process more than one items in a batch, the system provides a batch processing programmatic API. For example, given a set of 10,000 tweets, the task is to classify each tweet to determine their informativeness or what humanitarian information they report. 3. File processing service: For non-technical end-users who can not use our programmatic APIs, we provide a file processing service through a web interface. Figure 2 shows the file uploading interface. Given a file containing a list of messages, tweets, or Facebook posts, this API can classify each message using one or more of the classification services that the end-user selects. Currently, we support four file types: TXT (text files), JSON (JavaScript Object Notation), TSV (Tab-separated values), and CSV (Comma-separated values). Non-technical end-users can easily use the file processing interface to facilitate their humanitarian tasks. Figure 2. Web interface of the file processing service. The APIs are implemented in the Java language using the Jersey framework16, which is an open source and production quality framework for developing RESTful web services in Java. It is combined with Spring17 and Hibernate18 frameworks to communicate with databases. The single-item and batch processing APIs only offer programmatic access to use the classification services. However, the file processing API offers a web interface. All the services, APIs, and technical details regarding how to use these APIs with examples are provided on the CrisisDPS website: http://crisisdps.qcri.org/. Figure 3 shows a screenshot of the demo page with example tweets and outputs of the selected classification services. DATASETS For our task, we have selected almost all of the publicly available datasets including CrisisLex, CrisisNLP, among others. One of the difficulties that arises while combining different datasets is the discrepancies in the class labels. For example, in one dataset, the negative class is labeled as not-related while in another one as not-informative. Even though these two labels are semantically the same in theory, they are still two different labels in practice. 16https://jersey.github.io. Currently, we use Jersey, in future we might consider to use Spring REST. 17https://spring.io/ 18http://hibernate.org/ CoRe Paper – Social Media in Crises and Conflicts Proceedings of the 16th ISCRAM Conference - València, Spain May 2019 Zeno Franco, José J. González and José H. Canós, eds. F. Alam et al. Crisis Data Processing Services Figure 3. Example usages of the demo page. Therefore, for our task we mapped the semantically similar categories to have a unique class label. In addition, we have also filtered out instances with low frequencies and less relevance to the humanitarian tasks at hand. For example, we removed the instances with class labels such as animal management, non government, traditional media, terrorism related, and terrorism related information. We also removed non-English tweets. For this study, we aim to develop three different classifiers to serve three different classification tasks: (i) disaster type classification, (ii) informativeness classification, and (iii) humanitarian information type classification. To prepare the data for the informativeness task, we selected all the tweets with informative/related class labels and the tweets with humanitarian categories as informative and considered the rest as not-informative. For the humanitarian information type classification task, we selected data that are labeled with humanitarian categories, which resulted in ten categories as listed in Table 3. In this dataset we have also included instances with not-informative label, however, we have only selected a random sample from the whole set of not-informative instances. For the disaster type classification task, we selected tweets with the following six disaster types: bombing, earthquake, fire, flood, hurricane, and shooting. In addition, we have also included none category as the “background” class. That is, the tweets that do not belong to any of the mentioned six disaster types are considered as belonging to the none category. For the sake of completeness, we provide below brief descriptions of the datasets consolidated for the experiments. • AIDR19 is a publicly available system that has been developed to support humanitarian organizations and research community. The system has a functionality that allows users to annotate data while collecting tweets. Over time, AIDR has collected data from many events both labeled and unlabeled. We only selected manually labeled data that are relevant to this study. • CrisisLex20 comprises two different datasets, i.e., CrisisLexT26 and CrisisLexT6, that have been used and reported in the literature for crisis informatics research. The CrisisLexT26 is one of the largest datasets consisting of 26 different crisis events that took place in 2012 and 2013 (Olteanu et al. 2014). This dataset was prepared to explore two different crisis dimensions. First dimension is the disaster type (natural vs. humaninduced), their sub-type (e.g. meteorological, hydrological, etc.), temporal characteristics (instantaneous vs. progressive), and geographic spread (focalized vs. diffused). Second dimension is the content type in which several type of categories and sub-categories were identified, including informativeness (informative vs. not-informative), information type (six different subcategories), and source of information (i.e., eyewitness, government, NGOs, Business, Media, and Outsiders). The second dataset, CrisisLexT6, consists of six disasters that took place between October 2012 and July 2013 in USA, Canada, and Australia. Crisis keywords and locations have been used during the data collection process (Olteanu et al. 2014). Approximately 10K tweets has been annotated using Figure Eight21 (also known as CrowdFlower) from each crisis event. Annotation of this dataset include related vs. not-related. 19http://aidr.qcri.org/ 20https://crisislex.org 21https://www.figure-eight.com/ CoRe Paper – Social Media in Crises and Conflicts Proceedings of the 16th ISCRAM Conference - València, Spain May 2019 Zeno Franco, José J. González and José H. Canós, eds. F. Alam et al. Crisis Data Processing Services • CrisisMMD22 is a multimodal dataset consisting of tweets and associated images (Alam, Ofli, et al. 2018). Tweets have been collected from seven natural disaster that took place in 2017. The annotations include three tasks: (i) informative vs. not-informative, (ii) humanitarian categories (eight classes), and (iii) damage severity (three classes). The third annotation task, i.e., damage severity, was applied only on images. The annotation for text and images was run independently for each event using Figure Eight. • CrisisNLP23 consists of ∼50K tweets, which have been collected from 19 different disasters that took place between 2013 and 2015. The dataset was annotated in two steps. First set of annotations was obtained using Stand-By-Task-Force (SBTF)24 volunteers. The second set of annotations was curated using Figure Eight. • Disaster Response Data25 contain 30K tweets collected during disasters such as an earthquake in Haiti, 2010; an earthquake in Chile, 2010; floods in Pakistan, 2010; Hurricane Sandy in USA, 2012, and news articles. The annotations include 36 different categories. • Disasters on Social Media26 dataset consists of 10K tweets collected and annotated with labels related vs. not-related to the disasters. • SWDM27 consists of two data collections. The Joplin collection contains tweets collected during the tornado that struck Joplin, Missouri in May 22, 2011. The Sandy collection contains tweets collected during Hurricane Sandy, that hit Northeastern US on Oct 29, 2012. The Joplin dataset consists 4,400 labeled instances and the Sandy dataset consists of 2,000 labeled instances (Imran, S. Elbassuoni, et al. 2013). After consolidating data from all these collections, we prepared three datasets for three different classification tasks. Tables 2, 3 and 1 present the class distributions of our datasets for disaster type, informativeness, and humanitarian information type, respectively. We split each dataset into three subsets using 70%, 10% and 20% as training, development and test, respectively. The training set is used to learn the model, the development (dev) set is used for parameter tuning, and the test set is used for the model evaluation. Table 1. Data split for disaster type classification. Classes Train (70%) Dev (10%) Test (20%) Total Class Dist. Bombing Earthquake Fire Flood Hurricane Shooting None 378 9,088 2,185 10,156 23,491 527 10,306 53 1,285 309 1,436 3,322 74 1,457 109 2,611 628 2,917 6,746 153 2,961 540 12,984 3,122 14,509 33,559 754 14,724 0.007 0.162 0.039 0.181 0.418 0.009 0.184 Total 56,131 7,936 16,125 80,192 Table 2. Data split for informativeness classification. Classes Train (70%) Dev (10%) Test (20%) Total Class Dist. Informative Not-informative 131,028 74,104 18,531 10,480 37,624 21,280 187,183 105,864 0.639 0.361 Total 205,132 29,011 58,904 293,047 EXPERIMENTS, RESULTS, AND DISCUSSION In this section, we describe the details of our extensive experiments, report our results, and provide a discussion. 22http://crisisnlp.qcri.org/ 23http://crisisnlp.qcri.org/ 24www.standbytaskforce.org 25https://www.figure-eight.com/dataset/combined-disaster-response-data/ 26https://data.world/crowdflower/disasters-on-social-media 27http://crisisnlp.qcri.org/ CoRe Paper – Social Media in Crises and Conflicts Proceedings of the 16th ISCRAM Conference - València, Spain May 2019 Zeno Franco, José J. González and José H. Canós, eds. F. Alam et al. Crisis Data Processing Services Table 3. Data split for humanitarian information type classification. Classes Affected individual Caution and advice Displaced and evacuations Donation and volunteering Infrastructure and utilities damage Injured or dead people Missing and found people Requests or needs Response efforts Sympathy and support Not-informative Total Train (70%) Dev (10%) Test (20%) Total Class Dist. 2,536 2,644 519 5,677 4,619 2,771 460 5,088 1,857 4,438 4,200 358 374 73 802 653 392 65 719 263 627 594 729 760 150 1,631 1,327 796 133 1,462 534 1,275 1,206 3,623 3,778 742 8,110 6,599 3,959 658 7,269 2,654 6,340 6,000 0.073 0.076 0.015 0.163 0.133 0.080 0.013 0.146 0.053 0.127 0.121 34,809 4,920 10,003 49,732 Preprocessing The tweet texts are noisy, usually consisting of many symbols, emoticons, and invisible characters. Therefore, we preprocessed them to use in classification experiments. The preprocessing part includes removal of stop words, non-ASCII characters, punctuations (replaced with whitespace), numbers, URLs, and hashtags. Classification Experiments For this study, we have conducted classification experiments using both classical and deep learning algorithms. As for the classical models, we used the three most popular machine learning algorithms, i.e., (i) Naïve Bayes (NB), (ii) Random Forest (RF), and (iii) Support Vector Machines (SVM). The usefulness of these algorithms have been reported in several studies in crisis computing literature (Burel, Saif, and Alani 2017; Neppalli et al. 2018; Imran, Mitra, et al. 2016). As for the deep learning algorithms, we used a Convolutional Neural Network (CNN), which has also been reported as successful in many studies for disaster tweets classification (Nguyen et al. 2017; Burel and Alani 2018). • The NB algorithm (McCallum, Nigam, et al. 1998) is a simple probabilistic method that calculates a set of probabilities by counting the frequency and combinations of values in a given dataset. The algorithm uses Bayes theorem and stands on the assumption that all attributes are independent given the value of the class variable, which is rarely true in real applications, but it still performs well and learns rapidly in many supervised tasks (Neppalli et al. 2018; Imran, Mitra, et al. 2016). • The SVM algorithm (Platt 1998) is based on the Structural Risk Minimization principle from computational learning theory. The algorithm is established as universal learners and well-known for its ability to learn independently of the dimensionality of the feature space. These properties of the algorithm make it one of the most popular supervised classification methods. • Unlike the previous two classical algorithms, the RF algorithm (Liaw, Wiener, et al. 2002) is an ensemble learning algorithm, comprising multiple decision trees where each tree contributes to the classification decision with a single vote. A final class label is assigned based on majority voting. RF reduces variances in the classification by randomizing features and training instances. • In Figure 4, we present the architecture of our CNN model. The input to the network is a tweet t = (w1, . . . , wn ) containing words each coming from a finite vocabulary V that we obtained from our training dataset. The first layer of the network maps each of these words into a distributed representation Rd by looking up a shared embedding matrix E ∈ R | V |×d . We initialized E using a publicly available pre-trained embedding model designed using disaster related tweets28 (Alam, Joty, et al. 2018). The output of the look-up layer is a matrix X ∈ Rn×d , which is passed through a number of convolutional and pooling layers to learn higher-level feature representations. A convolution operation applies a filter u ∈ Rk.d to a window of k vectors to produce a new feature ht = f (u.Xt:t+k−1 ), where Xt:t+k−1 is the concatenation of k look-up vectors, and f is a nonlinear 28http://crisisnlp.qcri.org/data/lrec2016/crisisNLP_word2vec_model_v1.2.zip CoRe Paper – Social Media in Crises and Conflicts Proceedings of the 16th ISCRAM Conference - València, Spain May 2019 Zeno Franco, José J. González and José H. Canós, eds. F. Alam et al. Crisis Data Processing Services activation. We used rectified linear units (ReLU) as a nonlinear activation function. We apply this filter to each possible k-length windows in X to generate a feature map, h j = [h1, . . . , hn+k−1 ]. This process is repeated N times with N different filters to get N different feature maps. We then apply a max-pooling operation, m = [µ p (h1 ), · · · , µ p (h N )], where µ p (h j ) refers to the max operation applied to each window of p features in the feature map hi . Intuitively, the convolution operation composes local features into higher-level representations in the feature maps, and max-pooling extracts the most important aspects of each feature map while reducing the output dimensionality. Since each convolution-pooling operation is performed independently, the features extracted become invariant in order. In order to incorporate order information between the pooled features, we use a fully-connected (dense) layer z = f (Vm), where V is the weight matrix. From which we have another dense layer zc = f (Vc z) where Vc is the corresponding weight matrix. The activations zc are used for classification. Formally, the classification layer defines a Softmax   exp WkT zc   p(y = k |t, θ) = Í (1) Tz exp W ′ c ′ k k where Wk are the class weights and θ = {V, Vc, W } defines the relevant parameters. Figure 4. The architecture of the CNN. To train the classifiers using these classical algorithms, we converted the preprocessed tweets into bag-of-ngrams vectors weighted with logarithmic term frequencies (tf) multiplied with inverse document frequencies (idf), as shown in Equation 2.   of tweets t f × idf = log(1 + fi j ) × log number ofnumber (2) tweets that include word i where fi j is the frequency for word i in tweet j. To take advantage of the contextual benefits of n-grams, we extracted tri-gram features. Because this resulted in a large dictionary and we filtered out lower frequency features by preserving 40K most frequent n-grams. For the experiments with SVM and its linear kernel we tune the c parameter, and with RF we tune the number of trees on the development set. We trained the CNN models using the Adadelta optimizer (Zeiler 2012). The learning rate was set to 0.01 when optimizing on the classification loss. The maximum number of epochs was set to 1,000, and dropout (Srivastava et al. 2014) rate of 0.02 was used to avoid overfitting. We did early stopping based on the accuracy on the development set with a patience of 50. We used 100, 150, 200, and 250 filters with the corresponding window size of 2, 3, 4, and 5 respectively. We used a pooling length of 2. We did not tune any hyperparameter (e.g., the size of hidden layers, filter size, dropout rate) in the experimental setting. We also applied batch normalization due to its success reported in the literature (Ioffe and Szegedy 2015). Table 4 summarizes the results obtained by different classification algorithms described above. We provide weighted average precision (P), recall (R) and F1-measure (F1) for each task. The rationale behind choosing the weighted metric is that it takes into account the class imbalance problem. The performance of the NB model is lower than other models. Whereas RF and SVM model provides competitive results across different tasks. For the informativeness task RF outperforms SVM, whereas for humanitarian and disaster type tasks SVM outperforms RF. We obtain higher results consistently across all the tasks using CNN. CoRe Paper – Social Media in Crises and Conflicts Proceedings of the 16th ISCRAM Conference - València, Spain May 2019 Zeno Franco, José J. González and José H. Canós, eds. F. Alam et al. Crisis Data Processing Services Table 4. Classification performance comparison of all the classifiers on three different tasks. Disaster Type NB SVM RF CNN Informativeness Humanitarian Acc P R F1 Acc P R F1 Acc P R F1 0.73 0.93 0.91 0.93 0.82 0.93 0.92 0.93 0.73 0.93 0.91 0.93 0.71 0.93 0.91 0.93 0.85 0.90 0.93 0.93 0.85 0.90 0.93 0.93 0.85 0.90 0.93 0.93 0.85 0.90 0.93 0.93 0.65 0.78 0.76 0.78 0.77 0.78 0.77 0.78 0.65 0.78 0.76 0.78 0.69 0.78 0.76 0.78 Table 5. Mapping of the class labels between CREES and CrisisDPS. Disaster Type Informativeness CREES CrisisDPS CREES CrisisDPS Bombings Collapse Crash Derailment Earthquake Explosion Fire Floods Haze Meteorite None Shootings Typhoon Wildfire Bombing None None None Earthquake None Fire Flood None None None Shooting Hurricane Fire Related Not-related Informative Not-informative Humanitarian Information Type CREES CrisisDPS Affected individuals Caution and advice Donations and volunteering Infrastructure and utilities Not applicable Not labeled Other useful information Sympathy and support Affected individual Caution and advice Donation and volunteering Infrastructure and utilities damage Not-informative X X Sympathy and support X indicates the CREES class labels that were not considered in this work. Baseline Comparisons To compute the baseline results we used the publicly available deep learning models29 reported in (Burel and Alani 2018). The CREES API consists of three models for three different tasks such as (i) related vs. not-related, (ii) event type (11 class labels), and (iii) information type (i.e., humanitarian categories). To evaluate the performance of the CREES models and have the baseline results for our study, we used our test set and mapped the class labels as shown in Table 5. Table 6. Comparative results for disaster type classification. CREES Classes CrisisDPS P R F1 P R F1 Bombing Earthquake Fire Flood Hurricane Shooting None 0.54 0.96 0.85 0.35 0.98 0.97 0.45 0.83 0.66 0.88 0.94 0.1 0.96 0.73 0.66 0.78 0.86 0.51 0.18 0.97 0.55 0.94 0.92 0.95 0.94 0.93 1.00 0.94 0.93 0.94 0.93 0.93 0.96 0.99 0.87 0.94 0.93 0.94 0.93 0.95 0.99 0.91 W/Avg 0.76 0.50 0.45 0.93 0.93 0.93 In Tables 6, 7, and 8, we provide class-wise precision (P), recall (R) and F1-measure (F1) as well as their weighted average both for the baseline CREES models and the proposed CrisisDPS models. We observed that there is a significant performance difference across all classification tasks compared to the results reported in the study of CREES (Burel and Alani 2018). For the humanitarian categories, the CREES model is designed using eight class 29https://github.com/evhart/crees CoRe Paper – Social Media in Crises and Conflicts Proceedings of the 16th ISCRAM Conference - València, Spain May 2019 Zeno Franco, José J. González and José H. Canós, eds. F. Alam et al. Crisis Data Processing Services labels, whereas our models is designed with eleven class labels. Hence, in Table 8, we do not have results for some class labels using the CREES model. In the Table 8, the weighted average results for CrisisDPS are computed using only the subset of categories available in CREES for a fair comparison. Table 7. Comparative results for informativeness classification. CREES Classes CrisisDPS P R F1 P R F1 Informative Not-informative 0.88 0.76 0.86 0.79 0.87 0.78 0.95 0.91 0.95 0.90 0.95 0.91 W/Avg 0.84 0.84 0.84 0.93 0.93 0.93 Table 8. Comparative results for humanitarian information type classification. CREES Classes CrisisDPS P R F1 P R F1 Affected individual Caution and advice Displaced and evacuations Donation and volunteering Infrastructure and utilities damage Injured or dead people Missing and found people Requests or needs Response efforts Sympathy and support Not-informative 0.28 0.51 0.54 0.78 0.47 0.87 0.70 0.58 0.70 0.32 0.84 0.02 0.40 0.54 0.61 0.45 0.60 0.04 0.72 0.72 0.65 0.77 0.77 0.86 0.67 0.86 0.75 0.80 0.74 0.69 0.74 0.54 0.80 0.75 0.86 0.44 0.89 0.77 0.79 0.74 0.70 0.73 0.59 0.79 0.76 0.86 0.53 0.88 0.76 0.80 0.74 W/Avg 0.41 0.36 0.31 0.76* 0.76* 0.76* * computed using only the subset of categories available in CREES. Discussion and Future Directions We have provided comparative results between classical and deep learning models. We have also compared the performance of our best model with publicly available models. Even though we have only provided an end-to-end comparison with the CREES models in this study, our models also outperform some other existing works such as (Nguyen et al. 2017; Neppalli et al. 2018). To estimate the load of our system, we plan to have an extensive study in the future and plan to deploy the system accordingly. We also plan to make the system more scalable. Other future work includes integrating image processing modules such as finding an informative image, identifying their humanitarian information type, and assessing the damage content. Even though we developed the classification models using only social media data such as tweets, the CrisisDPS system can easily be extended by integrating other models. The benefit of CrisisDPS is that it provides services with RESTful APIs by taking the burden of running them on a local server. Such facilities can be helpful for humanitarian organizations and crisis computing research community to develop their own applications and systems using our services. Our goal is to make the CrisisDPS freely available, however, an authentication system will be implemented to control the rate limit (e.g., data processing limit per day per user) and to prevent potential abuse of the system so that every user can get equal benefit from the system. CONCLUSIONS In this study, we introduced CrisisDPS, a system for automatic data processing services, which comprises state-ofthe-art information classification models that enable humanitarian organizations and crisis informatics research community to develop their own applications and systems. Currently, CrisisDPS provide services that can facilitate three important humanitarian tasks: (i) determine which disaster type a text message belongs to, (ii) determine CoRe Paper – Social Media in Crises and Conflicts Proceedings of the 16th ISCRAM Conference - València, Spain May 2019 Zeno Franco, José J. González and José H. Canós, eds. F. Alam et al. Crisis Data Processing Services whether the message is informative for disaster response, and (iii) determine whether it manifests about any humanitarian information type. Such information is highly relevant for humanitarian organizations if processed timely and effectively. The classification models that we developed for these three tasks outperform existing publicly available models. We aim to provide as much flexibility as possible to the end-users who are willing to use our system. For example, our current implementation supports single item and batch processing through different APIs, file processing for different formats, and an online demo page for quick testing purposes. We plan to improve our services based on a thorough user study in the future. We also plan to make the system more scalable and extend it with automatic image processing capabilities. REFERENCES Abel, F., Hauff, C., Houben, G.-J., Stronkman, R., and Tao, K. (2012). “Semantics+ filtering+ search= twitcident. exploring information in social web streams”. In: Proc. of the 23rd ACM HT conference. ACM, pp. 285–294. Aipe, A., Mukuntha, N., Ekbal, A., and Kurohashi, S. (2018). “Deep Learning Approach towards Multi-label Classification of Crisis Related Tweets”. In: Proc. of the 15th ISCRAM, 2018. Alam, F., Joty, S. R., and Imran, M. (2018). “Graph Based Semi-supervised Learning with Convolution Neural Networks to Classify Crisis Related Tweets”. In: Proc. of the 12th ICWSM, 2018. AAAI press. Alam, F., Ofli, F., and Imran, M. (2018). “CrisisMMD: Multimodal twitter datasets from natural disasters”. English. In: Proc. of the 12th ICWSM, 2018. AAAI press, pp. 465–473. Burel, G. and Alani, H. (2018). “Crisis Event Extraction Service (CREES)-Automatic Detection and Classification of Crisis-related Content on Social Media”. In: Proc. of the 15th ISCRAM, 2018. Burel, G., Saif, H., and Alani, H. (2017). “Semantic wide and deep learning for detecting crisis-information categories on social media”. In: International Semantic Web Conference. Springer, pp. 138–155. Burel, G., Saif, H., Fernandez, M., and Alani, H. (2017). “On Semantics and Deep Learning for Event Detection in Crisis Situations”. In: Workshop on Semantic Deep Learning (SemDeep), at ESWC 2017. Caragea, C., McNeese, N., Jaiswal, A., Traylor, G., Kim, H.-W., Mitra, P., Wu, D., Tapia, A. H., Giles, L., Jansen, B. J., et al. (2011). “Classifying text messages for the haiti earthquake”. In: Proc. of the 8th ISCRAM, 2011. Castillo, C. (2016). Big Crisis Data. Cambridge, UK: Cambridge University Press. Goolsby, R. (2010). “Social media as crisis platform: The future of community maps/crisis maps”. In: ACM Transactions on Intelligent Systems and Technology (TIST) 1.1, p. 7. Hearst, M. A., Dumais, S. T., Osuna, E., Platt, J., and Scholkopf, B. (1998). “Support vector machines”. In: IEEE Intelligent Systems and their applications 13.4, pp. 18–28. Hiltz, S. R., Kushma, J., and Plotnick, L. (2014). “Use of Social Media by U.S. Public Sector Emergency Managers: Barriers and Wish Lists”. In: University Park, Pennsylvania, USA. Hochreiter, S. and Schmidhuber, J. (1997). “Long short-term memory”. In: Neural computation 9.8, pp. 1735–1780. Imran, M., Castillo, C., Diaz, F., and Vieweg, S. (2015). “Processing social media messages in mass emergency: A survey”. In: ACM Computing Surveys 47.4, p. 67. Imran, M., Castillo, C., Lucas, J., Meier, P., and Vieweg, S. (2014). “AIDR: Artificial intelligence for disaster response”. In: Proc. of the ACM Conference on WWW, pp. 159–162. Imran, M., Elbassuoni, S. M., Castillo, C., Diaz, F., and Meier, P. (2013). “Extracting information nuggets from disaster-related messages in social media”. In: Proc. of the 12th ISCRAM. Imran, M., Elbassuoni, S., Castillo, C., Diaz, F., and Meier, P. (2013). “Practical extraction of disaster-relevant information from social media”. In: Proc. of the 22nd WWW. ACM, pp. 1021–1024. Imran, M., Mitra, P., and Castillo, C. (2016). “Twitter as a Lifeline: Human-annotated Twitter Corpora for NLP of Crisis-related Messages”. In: Proc. of the LREC, 2016. Paris, France: ELRA. Ioffe, S. and Szegedy, C. (2015). “Batch normalization: Accelerating deep network training by reducing internal covariate shift”. In: arXiv preprint arXiv:1502.03167. Kejriwal, M. and Gu, Y. (2018). “A Pipeline for Post-Crisis Twitter Data Acquisition”. In: arXiv preprint arXiv:1801.05881. Kumar, S., Barbier, G., Abbasi, M. A., and Liu, H. (2011). “TweetTracker: An Analysis Tool for Humanitarian and Disaster Relief.” In: Proc. of the 5th ICWSM, 2011. AAAI press. CoRe Paper – Social Media in Crises and Conflicts Proceedings of the 16th ISCRAM Conference - València, Spain May 2019 Zeno Franco, José J. González and José H. Canós, eds. F. Alam et al. Crisis Data Processing Services LeCun, Y., Boser, B., Denker, J. S., Henderson, D., Howard, R. E., Hubbard, W., and Jackel, L. D. (1989). “Backpropagation applied to handwritten zip code recognition”. In: Neural computation 1.4, pp. 541–551. Liaw, A., Wiener, M., et al. (2002). “Classification and regression by Random Forest”. In: R news 2.3, pp. 18–22. Ludwig, T., Reuter, C., Siebigteroth, T., and Pipek, V. (2015). “CrowdMonitor: Mobile crowd sensing for assessing physical and digital activities of citizens during emergencies”. In: Proc. of the 33rd Annual ACM Conference on Human Factors in Computing Systems. ACM, pp. 4083–4092. MacEachren, A. M., Jaiswal, A., Robinson, A. C., Pezanowski, S., Savelyev, A., Mitra, P., Zhang, X., and Blanford, J. (2011). “Senseplace2: Geotwitter analytics support for situational awareness”. In: Proc. of the IEEE conference on VAST, 2011. IEEE, pp. 181–190. McCallum, A., Nigam, K., et al. (1998). “A comparison of event models for naive bayes text classification”. In: AAAI-98 workshop on learning for text categorization. Vol. 752. 1. Citeseer, pp. 41–48. Meier, P. (2012). “Crisis mapping in action: How open source software and global volunteer networks are changing the world, one map at a time”. In: Journal of Map & Geography Libraries 8.2, pp. 89–100. Neppalli, V. K., Caragea, C., and Caragea, D. (2018). “Deep Neural Networks versus Naïve Bayes Classifiers for Identifying Informative Tweets during Disasters”. In: Proc. of the 15th ISCRAM, 2018. Nguyen, D. T., Al-Mannai, K., Joty, S. R., Sajjad, H., Imran, M., and Mitra, P. (2017). “Robust Classification of Crisis-Related Data on Social Networks Using Convolutional Neural Networks.” In: Proc. of the 11th ICWSM, 2017. AAAI press, pp. 632–635. Okolloh, O. (2009). “Ushahidi, or ‘testimony’: Web 2.0 tools for crowdsourcing crisis information”. In: Participatory learning and action 59.1, pp. 65–70. Olteanu, A., Castillo, C., Diaz, F., and Vieweg, S. (2014). “CrisisLex: A Lexicon for Collecting and Filtering Microblogged Communications in Crises.” In: Proc. of the 8th ICWSM, 2014. AAAI press. Platt, J. (1998). Sequential minimal optimization: A fast algorithm for training support vector machines. Tech. rep. Redmond, USA.: Microsoft. Purohit, H. and Sheth, A. P. (2013). “Twitris v3: From Citizen Sensing to Analysis, Coordination and Action.” In: Proc. of the 7th ICWSM, 2013. AAAI press. Rudra, K., Goyal, P., Ganguly, N., Mitra, P., and Imran, M. (2018). “Identifying sub-events and summarizing disaster-related information from microblogs”. In: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval. ACM, pp. 265–274. Srivastava, N., Hinton, G. E., Krizhevsky, A., Sutskever, I., and Salakhutdinov, R. (2014). “Dropout: a simple way to prevent neural networks from overfitting.” In: Journal of MLR 15.1, pp. 1929–1958. Starbird, K., Palen, L., Hughes, A. L., and Vieweg, S. (2010). “Chatter on the red: what hazards threat reveals about the social life of microblogged information”. In: ACM Conference on Computer Supported Cooperative Work, pp. 241–250. Terpstra, T., De Vries, A., Stronkman, R., and Paradies, G. (2012). Towards a realtime Twitter analysis during crises for operational crisis management. British Columbia, Canada: Simon Fraser University Burnaby. Tsou, M.-H., Jung, C.-T., Allen, C., Yang, J.-A., Han, S. Y., Spitzberg, B. H., and Dozier, J. (2017). “Building a Real-Time Geo-Targeted Event Observation (Geo) Viewer for Disaster Management and Situation Awareness”. In: Proc. of the International Cartographic Conference. Springer, pp. 85–98. Varol, O., Ferrara, E., Davis, C., Menczer, F., and Flammini, A. (2017). “Online Human-Bot Interactions: Detection, Estimation, and Characterization”. In: Proc. of the 11th ICWSM, 2018. AAAI press. Vieweg, S., Castillo, C., and Imran, M. (2014). “Integrating social media communications into the rapid assessment of sudden onset disasters”. In: International Conference on Social Informatics. Springer, pp. 444–461. Vieweg, S., Hughes, A. L., Starbird, K., and Palen, L. (2010). “Microblogging during two natural hazards events: what twitter may contribute to situational awareness”. In: SIGCHI Conference on Human Factors in Computing Systems. ACM, pp. 1079–1088. Wattegama, C. (2014). “ICT for disaster management”. In: UNDP-APDIP. Yin, J., Lampert, A., Cameron, M., Robinson, B., and Power, R. (2012). “Using social media to enhance emergency situation awareness”. In: IEEE Intelligent Systems 27.6, pp. 52–59. CoRe Paper – Social Media in Crises and Conflicts Proceedings of the 16th ISCRAM Conference - València, Spain May 2019 Zeno Franco, José J. González and José H. Canós, eds. F. Alam et al. Crisis Data Processing Services Zade, H., Shah, K., Rangarajan, V., Kshirsagar, P., Imran, M., and Starbird, K. (2018). “From Situational Awareness to Actionability: Towards Improving the Utility of Social Media Data for Crisis Response”. In: Proceedings of the ACM on Human-Computer Interaction 2.CSCW, p. 195. Zeiler, M. D. (2012). “ADADELTA: an adaptive learning rate method”. In: arXiv preprint arXiv:1212.5701. CoRe Paper – Social Media in Crises and Conflicts Proceedings of the 16th ISCRAM Conference - València, Spain May 2019 Zeno Franco, José J. González and José H. Canós, eds.