DaSCI Webinars


DaSCI Seminars
Are talks by an outstanding invited researcher who presents recent disruptive advances on AI. The seminars will last about 1 hour and 30 minutes (45min. Speaker + 30 min. for questions)
Graph Mining with Graph Neural Networks
Lecturer: Bryan Perozzi is a Research Scientist in Google Research’s Algorithms and Optimization group, where he routinely analyzes some of the world’s largest (and perhaps most interesting) graphs. Bryan’s research focuses on developing techniques for learning expressive representations of relational data with neural networks. These scalable algorithms are useful for prediction tasks (classification/regression), pattern discovery, and anomaly detection in large networked data sets. Bryan is an author of 30+ peer-reviewed papers at leading conferences in machine learning and data mining (such as NeurIPS, ICML, KDD, and WWW). His doctoral work on learning network representations was awarded the prestigious SIGKDD Dissertation Award. Bryan received his Ph.D. in Computer Science from Stony Brook University in 2016, and his M.S. from the Johns Hopkins University in 2011.
Date:17/05/2021
Abstract: How can neural networks best model data which doesn’t have a fixed structure? In this talk, I will discuss graph neural networks (GNNs), a very active area of current research in machine learning aimed at answering this interesting (and practical) question. After reviewing the basics of GNNs, I’ll discuss some challenges applying these methods in industry, and some of the methods we’ve developed for addressing these challenges.
Detecting the “Fake News” Before It Was Even Written, Media Literacy, and Flattening the Curve of the COVID-19 Infodemic
Lecturer: Dr. Preslav Nakov is a Principal Scientist at the Qatar Computing Research Institute (QCRI), HBKU , where he leads the Tanbih mega-project (developed in collaboration with MIT ), which aims to limit the effect of “fake news”, propaganda and media bias by making users aware of what they are reading, thus promoting media literacy and critical thinking. He received his PhD degree in Computer Science from the University of California at Berkeley, supported by a Fulbright grant. Dr. Preslav Nakov is President of ACL SIGLEX , Secretary of ACL SIGSLAV , and a member of the EACL advisory board. He is also member of the editorial board of a number of journals including Computational Linguistics, TACL , CS&L, NLE , AI Communications, and Frontiers in AI. He authored a Morgan & Claypool book on Semantic Relations between Nominals and two books on computer algorithms. He published 250+ research papers, and he was named among the top 2% of the world’s most-cited in the career achievement category, part of a global list compiled by Stanford University. He received a Best Long Paper Award at CIKM ‘2020, a Best Demo Paper Award (Honorable Mention) at ACL ‘2020, a Best Task Paper Award (Honorable Mention) at SemEval’2020, a Best Poster Award at SocInfo’2019, and the Young Researcher Award at RANLP ‘2011. He was also the first to receive the Bulgarian President’s John Atanasoff award, named after the inventor of the first automatic electronic digital computer. Dr. Nakov’s research was featured by over 100 news outlets, including Forbes, Boston Globe, Aljazeera, DefenseOne, Business Insider, MIT Technology Review, Science Daily, Popular Science, Fast Company, The Register, WIRED , and Engadget, among others.
Date:19/04/2021
Abstract: Given the recent proliferation of disinformation online, there has been growing research interest in automatically debunking rumors, false claims, and “fake news”. A number of fact-checking initiatives have been launched so far, both manual and automatic, but the whole enterprise remains in a state of crisis: by the time a claim is finally fact-checked, it could have reached millions of users, and the harm caused could hardly be undone.
An arguably more promising direction is to focus on analyzing entire news outlets, which can be done in advance; then, we could fact-check the news before it was even written: by checking how trustworthy the outlet that has published it is (which is what journalists actually do). We will show how we do this in the Tanbih news aggregator (http://www.tanbih.org/), which aims to limit the impact of “fake news”, propaganda and media bias by making users aware of what they are reading, thus promoting media literacy and critical thinking, which are arguably the best way to address disinformation in the long run. In particular, we develop media profiles that show the general factuality of reporting, the degree of propagandistic content, hyper-partisanship, leading political ideology, general frame of reporting, stance with respect to various claims and topics, as well as audience reach and audience bias in social media.
Another important observation is that the term “fake news” misleads people to focus exclusively on factuality, and to ignore the other half of the problem: the potential malicious intent. Thus, we detect the use of specific propaganda techniques in text, e.g., appeal to emotions, fear, prejudices, logical fallacies, etc. We will show how we do this in the Prta system (https://www.tanbih.org/prta), another media literacy tool, which got the Best Demo Award (Honorable Mention) at ACL -2020; an associated shared task got the Best task award (Honorable Mention) at SemEval-2020.
Finally, at the time of COVID -19, the problem of disinformation online got elevated to a whole new level as the first global infodemic. While fighting this infodemic is typically thought of in terms of factuality, the problem is much broader as malicious content includes not only “fake news”, rumors, and conspiracy theories, but also promotion of fake cures, panic, racism, xenophobia, and mistrust in the authorities, among others. Thus, we argue for the need of a holistic approach combining the perspectives of journalists, fact-checkers, policymakers, social media platforms, and society as a whole, and we present our recent research in that direction (https://mt.qcri.org/covid19disinformationdetector/).
Efficient Deep Learning
Lecturer: Marco Pedersoli is Assistant Professor at ETS Montreal. He obtained his PhD in computer science in 2012 at the Autonomous University of Barcelona and the Computer Vision Center of Barcelona. Then, he was a postdoctoral fellow in computer vision and machine learning at KU Leuven with Prof. Tuytelaars and later at INRIA Grenoble with Drs. Verbeek and Schmid. At ETS Montreal he is a member of LIVIA and he is co-chairing an industrial Chair on Embedded Neural Networks for Connected Building Control. His research is mostly applied on visual recognition, the automatic interpretation and understanding of images and videos. His specific focus is on reducing the complexity and the amount of annotation required for deep learning algorithms such as convolutional and recurrent neural networks. Prof. Pedersoli has authored more than 40 publications in top-tier international conferences and journals in computer vision and machine learning.
Date:12/04/2021
Abstract: In the last 10 years deep learning (DL) models have shown great progress in many different fields, from Computer Vision to Natural Language Processing. However, DL methods require great computational resources (i.e. GPUs or TPUs) and very large datasets, which also makes the training phase very long and painful. Thus, there is a strong need for reducing the computational cost of DL methods both in training as well as in deployment. In this talk, I present the most common families of approaches used to reduce the requirements of DL methods in terms of Memory and Computation for both training and deployment, and show how a reduction of the model footprint does not always produce a corresponding speed-up. Finally, I will present some recent results that suggest that large DL models are important mostly for facilitating the model training, and when that is finished, we can deploy a much smaller and faster model with almost no loss in accuracy.
How to do Data Science without writing code
Lecturer: Victoriano Izquierdo (1990) is a software engineer from Granada who is the co-founder and CEO of Graphext. Graphext is developing an advanced data analysis software built upon the last advances in data science and artificial intelligence for supporting small and big companies to address hard problems using data.
Date: 15/02/2021
Abstract: How to do Data Science without writing code
Variational Autoencoders for Audio, Visual and Audio-Visual Learning
Lecturer: Xavier Alameda-Pineda is a (tenured) Research Scientist at Inria, in the Perception Group. He obtained the M.Sc. (equivalent) in Mathematics in 2008, in Telecommunications in 2009 from BarcelonaTech and in Computer Science in 2010 from Université Grenoble-Alpes (UGA). He then worked towards his Ph.D. in Mathematics and Computer Science, and obtained it in 2013, from UGA. After a two-year post-doc period at the Multimodal Human Understanding Group, at University of Trento, he was appointed with his current position. Xavier is an active member of SIGMM, and a senior member of IEEE and a member of ELLIS. He is co-chairing the “Audio-visual machine perception and interaction for companion robots” chair of the Multidisciplinary Institute of Artificial Intelligence. Xavier is the Coordinator of the H2020 Project SPRING: Socially Pertinent Robots in Gerontological Healthcare. Xavier’s research interests are in combining machine learning, computer vision and audio processing for scene and behavior analysis and human-robot interaction. More info at xavirema.eu
Date: 01/02/2021
Abstract: Since their introduction, Variational Autoencoders (VAE) demonstrated great performance in key unsupervised feature representation applications, specifically in visual and auditory representation. In this seminar, the global methodology of variational auto-encoders will be presented, along with applications in learning with audio and visual data. Special emphasis will be put in discussing the use of VAE for audio-visual learning, showcasing its interest with the task of audio-visual speech enhancement.
Variational Autoencoders for Audio, Visual and Audio-Visual Learning – Recording
Five Sources of Biases and Ethical Issues in NLP, and What to Do about Them
Lecturer: Dirk Hovy is associate professor of computer science at Bocconi University in Milan, Italy. Before that, he was faculty and a postdoc in Copenhagen, got a PhD from USC, and a linguistics masters in Germany. He is interested in the interaction between language, society, and machine learning, or what language can tell us about society, and what computers can tell us about language. He has authored over 60 articles on these topics, including 3 best paper awards. He has organized one conference and several workshops (on abusive language, ethics in NLP, and computational social science). Outside of work, Dirk enjoys cooking, running, and leather-crafting. For updated information, see http://www.dirkhovy.com
Date: 11/01/2021
Abstract: Never before was it so easy to write a powerful NLP system, never before did it have such a potential impact. However, these systems are now increasingly used in applications they were not intended for, by people who treat them as interchangeable black boxes. The results can be simple performance drops, but also systematic biases against various user groups. In this talk, I will discuss several types of biases that affect NLP models (based on Shah et al. 2020 and Hovy & Spruit, 2016), what their sources are, and potential counter measures.
Five Sources of Biases and Ethical Issues in NLP, and What to Do about Them – Recording
Image and Video Generation using Deep Learning
Lecturer: Stéphane Lathuilière is an associate professor (maître de conférence) at Telecom Paris, France, in the multimedia team. Until October 2019, he was a post-doctoral fellow at the University of Trento (Italy) in the Multimedia and Human Understanding Group, led by Prof. Nicu Sebe and Prof. Elisa Ricci. He received the M.Sc. degree in applied mathematics and computer science from ENSIMAG, Grenoble Institute of Technology (Grenoble INP), France, in 2014. He completed his master thesis at the International Research Institute MICA (Hanoi, Vietnam). He worked towards his Ph.D. in mathematics and computer science in the Perception Team at Inria under the supervision of Dr. Radu Horaud, and obtained it from Université Grenoble Alpes (France) in 2018. His research interests cover machine learning for computer vision problems (eg. domain adaptation, continual learning) and deep models for image and video generation. He regularly publishes papers in the most prestigious computer vision conferences (CVPR, ICCV, ECCV, NeurIPS) and top journals (IEEE TPAMI).
Date: 14/12/2020
Abstract: Generating realistic images and videos has countless applications in different areas, ranging from photography technologies to e-commerce business. Recently, deep generative approaches have emerged as effective techniques for generation tasks. In this talk, we will first present the problem of pose-guided person image generation. Specifically, given an image of a person and a target pose, a new image of that person in the target pose is synthesized. We will show that important body-pose changes affect generation quality and that specific feature map deformations lead to better images. Then, we will present our recent framework for video generation. More precisely, our approach generates videos where an object in a source image is animated according to the motion of a driving video. In this task, we employ a motion representation based on keypoints that are learned in a self-supervised fashion. Therefore, our approach can animate any arbitrary object without using annotation or prior information about the specific object to animate.
Aggregating Weak Annotations from Crowds
Lecturer: Edwin Simpson, is a lecturer (equivalent to assistant professor) at the University of Bristol, working on interactive natural language processing. His research focusses on learning from small and unreliable data, including user feedback, and adapts Bayesian approaches to topics such as argumentation, summarisation and sequence labelling. Previously, he was a post-doc at TU Darmstadt, Germany and completed his PhD at the University of Oxford on Bayesian methods for aggregating crowdsourced data.
Date: 09/11/2020
Abstract: Current machine learning methods are data hungry. Crowdsourcing is a common solution to acquiring annotated data at large scale for a modest price. However, the quality of the annotations is highly variable and annotators do not always agree on the correct label for each data point. This talk presents techniques for aggregating crowdsourced annotations using preference learning and classifier combination to estimate gold-standard rankings and labels, which can be used as training data for ML models. We apply approximate Bayesian approaches to handle noise, small amounts of data per annotator, and provide a basis for active learning. While these techniques are applicable to any kind of data, we demonstrate their effectiveness for natural language processing tasks.
Reinforcement Learning
Lecturer: Sergio Guadarrama is a senior software engineer at Google Brain. He focuses on reinforcement learning and neural networks. His research focuses on robust, scalable and efficient reinforcement learning. He is currently the leader of the TF-Agents project and a lead developer of TensorFlow (co-creator of TF-Slim). Before joining Google, he was a researcher at the University of California, Berkeley, where he worked with Professor Lotfi Zadeh and Professor Trevor Darrell. He received his B.A. and Ph.D. from the Universidad Politécnica de Madrid.
Date: 26/10/2020
Abstract: Learning by reinforcement (RL) is a type of machine learning where the objective is to learn to solve a task through interaction with the environment, maximizing the expected return. Unlike supervised learning, the solution requires making multiple decisions in a sequential way and the reinforcement occurs through rewards. The two main components are the environment, which represents the problem to be solved, and the agent, which represents the learning algorithm.
DaSCI Lectures
Are talks by a DaSCI senior researcher who presents the latest advances on DaSCI consolidated research lines.The seminars will last about 1 hour and 30 minutes (45min. Speaker + 30 min. for questions)
New pooling layer structures using linear combinations of increasing functions and grouping functions
Lecturer: Humberto Bustince Sola is a full professor of Computer Science and Artificial Intelligence at the Public University of Navarra and honorary professor at the University of Nottingham since 2017. He is the main researcher of the Research Group on Artificial Intelligence and Approximate Reasoning, whose research lines are both theoretical (data fusion functions, information and comparison measures, fuzzy sets and their extensions) and applied (Deep learning, image processing, classification, machine learning, data mining, big data or the computational brain). He has led 13 research projects funded by national and regional governments, and two excellence networks on soft computing. He has been the main researcher in projects with companies and entities such as Caja de Ahorros de Navarra, INCITA, Gamesa Tracasa or the Servicio Navarro de Salud. He has taken part in two international projects. He has authored or coauthored more than 300 works, according to Web of Science, including around 160 in Q1 journals. He was a highly cited researcher among the top 1%most relevant scientists in the world in 2018, according to Clarivate Analytics. He collaborates with first line research groups from countries such as United Kingdom, Belgium, Australia, the Czech Republic, Slovakia, Canada or Brasil. He is editor in chief of the Mathware&Soft Computing online magazine of the European Society of Fuzzy Logic and technologies and of the Axioms journal. Associated editor of the IEEE Transactions on Fuzzy Systems journal and member of the editorial boards of the journals Fuzzy Sets and Systems, Information Fusion, International Journal of Computational Intelligence Systems and Journal of Intelligent & Fuzzy Systems. Moreover, he is a coauthor of a book about averaging functions, and has been the co-editor of several books. He has been in charge of organizing several first level international conferences such as EUROFUSE 2009 and AGOP 2013. He is Senior Member of IEEE y Fellow of the International Fuzzy Systems Association (IFSA). Member of the Basque Academy of Sciences, Arts and Literature, Jakiunde, since 2018. He has advised 11 Ph.D thesis.
Date:13/09/2021
Abstract: In this talk, we start making a revision of the main concepts of aggregation functions and how these functions have been applied to artificial intelligence. We also discuss how some application in artificial intelligence have led to consider specific classes of aggregations as well as wider families of increasing functions. As an example of the applicability of these developments, we consider the case of convolutional neural networks. Traditional convolutional neural networks make use of the maximum or arithmetic mean in order to reduce the features extracted by convolutional layers. In this work we replace this downsampling process, known as pooling operation, by several alternative functions. We consider linear combinations of order statistics and generalizations of the Sugeno integral, extending the later’s domain to the whole real line and setting the theoretical base for their application. We apply these new functions to three different architectures of increasing complexity, showing that the best pooling function is architecture dependent and should be fine tuned similarly to other model hyper parameters. However, we also empirically prove over multiple datasets that linear combinations outperform traditional pooling functions in most cases, and that combinations with either the Sugeno integral or one of its generalizations usually yield the best results, proving a strong candidate to apply in most architectures.
New pooling layer structures using linear combinations of increasing functions and grouping functions – Recording
Bayesian Modelling and Inference with Applications to Image Recovery and Classification
Lecturer: Rafael Molina received the degree in Mathematics (Statistics) and the Ph.D. degree in optimal design in linear models from the University of Granada, Granada, Spain, in 1979 and 1983, respectively. In 2000, he became Professor of Computer Science and Artificial Intelligence at the University of Granada. His research interest focuses mainly on using Bayesian modelling and inference in image restoration (applications to astronomy and medicine), super-resolution of images and video, active learning, supervised and unsupervised learning and crowdsourcing.
Date: 7/06/2021
Abstract: A fundamental principle of the Bayesian philosophy is to regard all parameters and unobservable variables of a given problem as unknown stochastic quantities. The inference goal is to calculate or approximate the distribution of all the unknowns given the observations.
Variational Bayesian (VB) inference is a family of deterministic probability distribution approximation procedures that offers distinct advantages over alternative approaches based on stochastic sampling and those providing only point estimates. VB inference is flexible to be applied in different practical problems, yet it is broad enough to subsume as its special cases several alternative inference approaches including Maximum A Posteriori (MAP) and the Expectation-Maximization (EM) algorithm. VB inference and Expectation Propagation are both variational methods that minimize functionals based on the Kullback-Leibler (KL) divergence. Connections between VB and marginalization-based Loopy Belief Propagation (LBP) can also be easily established.
In this talk, I provide a personal overview of Bayesian modeling and inference methods for image recovery (regression) and classification problems. I will place emphasis on the relationship between the KL-divergence and the Evidence Lower Bound (ELBO), the pros and cons of Variational Bayesian (VB) methods, their connections to other inference methods, and the use of local variational bounds. The talk will also include a (brief) description of some VB applications: blind image and color deconvolution, super-resolution, (deep) Gaussian processes, activation uncertainty in neural networks , histological image classification, multiple instance learning, and crowdsourcing in medicine and the LIGO problem.
Bayesian Modelling and Inference with Applications to Image Recovery and Classification – Recording
The Artificial Intelligence of the new Agriculture
Lecturer: Salvador Gutiérrez
Date: 05/07/2021
Abstract: In agriculture, the aim is to reduce costs and environmental impact, improve sustainability and increase crop quality and yield. In order to develop useful applications for farmers, crop information is needed that can be used to make better decisions. New advances in non-invasive sensor technologies allow the acquisition of large amounts of data from the field. And given the revolution brought about by artificial intelligence, the combination of artificial intelligence with data from multiple sensors allows the extraction of useful information for the farmer. This presentation will present the current advances in AI and sensors in agriculture, mainly in viticulture, and future lines and challenges will be discussed.
The Artificial Intelligence of the new Agriculture – Recording
Artificial Intelligence from the point of view of ethics and law
First Lecturer (20 min): Francisco Lara
Date: 10/05/2021
Title & Abstract: “The Ethics of the Artificial Intelligence”. In my talk I will briefly review the main questions being asked by authors working on ethical issues in AI, from the more speculative (about the possibilities and effects of superintelligence or the status of robots) to the concrete and immediate (the bias or lack of transparency of algorithms). I will also talk a bit about EthAI+, the project I am currently coordinating, which aims to theorise about a virtual assistant that would improve the ethical abilities of human beings.
Second Lecturer (20 min) : Javier Valls
Date: 10/05/2021
Title & Abstract: “Law implications in the Artificial Intelligence”. The impact of technology on society and the legal consequences that can be derived from it, both positive and negative, will be explained. To this end, some systems of human rights impact of artificial intelligence will be analysed. Finally, we will look at what an artificial intelligence risk system is.
Artificial Intelligence from the point of view of ethics and law – Recording
Development of Community Finding Algorithms based on Evolutionary Strategies
Lecturer: David Camacho
Date: 22/03/2021
Abstract: The seminar will provide a brief introduction to Community Finding Algorithms, widely studied in the area of Graph Network Analysis and Degree-Based Computing, where the problem of detecting this type of node groups, or clusters, when modelled as a temporal problem will be explored in depth. A brief introduction will be given to traditional detection methods (static in nature), and to more current methods that attempt to address the problem from a dynamic perspective. In particular, it will briefly describe how bio-inspired strategies, namely evolutionary algorithms (single and multi-objective) are being used to try to find stable and quality communities over time. This talk will briefly present the design, implementation and empirical analysis of a new multi-objective genetic algorithm that combines an immigrant-based scheme with local search strategies for dynamic community detection.
Development of Community Finding Algorithms based on Evolutionary Strategies – Recording
Artificial intelligence, refugees and border security. Ethical implications of technological and political worlds
Lecturer: Ana Valdivia
Date: 08/02/2021
Abstract: Over the last decade, a large number of people are on the move due to conflict, instability, climatic emergency consequences, and other economic reasons. In Europe, the so-called refugee crisis has become a testing ground to explore the use of artificial intelligence for law enforcement and border security. Interoperable databases, facial recognition and fingerprints registration, iris data collection, lie detectors, and other forms of data-driven risk assessments now all form part of the European border policies for refugees.
In this webinar, we will explore which socio-technical systems are applied nowadays at European borders by analysing technical specifications. After that, we will discuss the ethical impact and human rights violation that this situation is causing. It is now necessary that computer scientist and data engineers recognise how technology might perpetuate harms, and collaborate with academics from other disciplines to mitigate discrimination.
Artificial Intelligence, Migrants, and Border Security – Recording
EXplainable Artificial Intelligence (XAI): Concepts, Taxonomies, Opportunities and Challenges toward Responsible AI
Lecturer: Natalia Díaz Rodríguez
Date: 25/01/2021
Abstract: The overview presented in this article examines the existing literature and contributions already done in the field of XAI, including a prospect toward what is yet to be reached. For this purpose we summarize previous efforts made to define explainability in Machine Learning, establishing a novel definition of explainable Machine Learning that covers such prior conceptual propositions with a major focus on the audience for which the explainability is sought. Departing from this definition, we propose and discuss about a taxonomy of recent contributions related to the explainability of different Machine Learning models, including those aimed at explaining Deep Learning methods for which a second dedicated taxonomy is built and examined in detail. This critical literature analysis serves as the motivating background for a series of challenges faced by XAI, such as the interesting crossroads of data fusion and explainability. Our prospects lead toward the concept of Responsible Artificial Intelligence, namely, a methodology for the large-scale implementation of AI methods in real organizations with fairness, model explainability and accountability at its core. Our ultimate goal is to provide newcomers to the field of XAI with a thorough taxonomy that can serve as reference material in order to stimulate future research advances, but also to encourage experts and professionals from other disciplines to embrace the benefits of AI in their activity sectors, without any prior bias for its lack of interpretability.
Autoencoders: an Overview and Applications
Lecturer: David Charte
Date: 21/12/2020
Abstract: In this talk, we motivate the need for representation learning techniques, especially those based in artificial neural networks. We arrive to a definition of autoencoders which then are further developed in a step-by-step example. Next, several applications of autoencoders are described and illustrated with case studies as well as uses in the literature. Last, some comments on the current situation and possible future trends are provided.
AI Ethics: encompassing the facets of FATE. Fairness, transparency and auditability
Lecturer: José Daniel Pascual Triana
Date: 23/11/2020
Abstract: Artificial Intelligence Ethics is the field that strives to apply the ethical and moral principles of humans to the development and operation of machine learning. This includes, amongst other topics, bias reduction to enforce parity, transparency and model auditing.
Due to the sheer amount of data that is currently generated and used, as well as the increased awareness of the population and the evolving legislation to keep up with the times, AI Ethics’ relevance keeps rising as a means to maintain the trust in the analysis and treatment of data. In this seminar, a taxonomy of AI Ethics will be presented, current techniques to promote it will be shown and the usefulness of several tools for data and model treatment will be discussed.
AI Ethics: encompassing the facets of FATE-Recording (in Spanish)
DaSCI Readings
Are short talks by a DaSCI PhD Student who presents recents results on the different DaSCI research lines. Two presentations per day. Each presentation will be approximately 30 minutes long, followed by 15 minutes for questions
Descriptive analysis of breast cancer using data mining
Lecturer: Manuel Trasierras Fresco
Date: 28/06/2121
Abstract: This work presents an approach based on emerging pattern mining to analyse cancer through genomic data. Unlike existing approaches, mainly focused on predictive purposes, this proposal aims to improve the understanding of cancer descriptively, not requiring either any prior knowledge or hypothesis to be validated. Additionally, it enables to consider high-order relationships, so both direct and indirect gene relationships related to different functional pathways in the disease can be retrieved. The prime hypothesis is that splitting genomic cancer data into two subsets, that is, cases and controls, will allow us to determine which genes, and their expressions, are associated with the disease. The possibilities of the proposal are demonstrated by analysing a set of paired breast cancer samples in RNA-Seq format. Some of the extracted insights were already described in the related literature as good cancer biomarkers, while others could describe new functional relationships between different genes.
PAF-ND: addressing multi-class imbalance learning with Nested Dichotomies
Lecturer: José Alberto Fernández Sánchez
Date: 28/06/2021
Abstract: Multi-class classification tasks add additional difficulties to the binary classification problem from several sides. Among them, the possibility of obtaining a homogeneous distribution of the classes involved is often one of the most recurring issues in real world problems. This issue leads to what are known as imbalanced learning scenarios. In this work, we explore a method that improves the predictive ability of models when using a type of decomposition strategy known as Nested Dichotomies. Nested Dichotomies is a solution that hierarchically decomposes the classes of the problem and uses an inference method based on probabilities. The method presented here attempts to modify the probability estimates of these models within the hierarchy towards a more equitable classification of the classes by means of Bézier curves.
Reducing Data Complexity using Autoencoders with Class-informed Loss Functions
Lecturer: David Charte
Date: 15/03/2021
Abstract: The data we currently use for knowledge extraction can show different kinds of complexity: class overlap, complex boundaries, dimensionality, etc. This work proposes and evaluates three autoencoder-based models which help reduce complexity by learning from class labels. We also check which complexity measures are better predictors of classification performance.
Multi-step Histogram Based Outlier Scores for Unsupervised Anomaly Detection: ArcelorMittal Engineering Dataset Case of Study
Lecturer: Ignacio Aguilera
Date: 15/03/2021
Abstract: Multi-step Histogram Based Outlier Scores for Unsupervised Anomaly Detection: ArcelorMittal Engineering Dataset Case of Study.
Abstract: Anomaly detection is the task of detecting samples that behave differently from the rest of the data or that include abnormal values. Unsupervised anomaly detection is the most extended scenario, which means that the algorithms cannot train with a labeled input and do not know the anomaly behavior beforehand. Histogram-based methods are one of the most popular and widely used approaches, remarking a good performance and a low runtime. Despite the good performance, histogram-based anomaly detectors are not capable of processing data flows while updating their knowledge and deal with a high amount of samples. In this paper we propose a new histogram-based approach for addressing the aforementioned problems introducing the ability of updating the information inside a histogram. We have applied these strategies to design a new algorithm called Multi-step Histogram Based Outlier Scores (MHBOS), including five new histogram update mechanisms. MHBOS has been validated using the ODDS Library as a general case of use. A real engineering problem provided by the multinational company ArcelorMittal has been used to further validate the performance in a real scenario. The results have shown the performance and validity of MHBOS as well as the proposed strategies in terms of performance and computing times.
StyleGAN: Background and evolution
Lecturer: Guillermo Gómez Trenado
Date: 22/02/2021
Abstract: The work developed by Tero Karras and his team at Nvidia has been the state-of-the-art in GAN for image generation since 2017. In this DaSCI reading we’ll use this results to discuss different aspects of GAN, the iterative process by which the authors detected and corrected the limitations of their work, the technological solutions that allowed such results and the difficulties that we may find if we face related tasks.
Action Recognition for Anomaly Detection using Transfer Learning and Weak Supervision
Lecturer: Francisco Luque
Date: 18/01/2021
Abstract: Automatic video surveillance is an emerging research area, where a huge number of publications are appearing everyday. Particularly, action anomaly detection is a fairly relevant task nowadays. The mainstream approach to the problem using deep models consists in transfer learning from action recognition and weakly supervised fine-tuning for anomaly detection. The objective of the current study is to identify the key aspects of this approaches, and assess the importance of each decision on the training process. To this end, we propose a specific pipeline, where a model is defined by three key aspects: the action recognition model, the pretraining dataset and the weakly supervised fine-tuning policy. Furthermore, we perform extensive experiments to validate the impact of each of the previous aspects in the final solution.
Fuzzy Monitoring of In-bed Postural Changes for the Prevention of Pressure Ulcers using Inertial Sensors Attached to Clothing
Lecturer: Edna Rocío Bernal Monroy
Date: 18/01/2021
Abstract: Postural changes while maintaining a correct body position are the most efficient method of preventing pressure ulcers. However, executing a protocol ofpostural changes over a long period of time is an arduous task for caregivers.To address this problem, we propose a fuzzy monitoring system for posturalchanges which recognizes in-bed postures by means of micro inertial sensors attached to patients’ clothes. First, we integrate a data-driven model to classifyin-bed postures from the micro inertial sensors which are located in the socksand t-shirt of the patient. Second, a knowledge-based fuzzy model computes thepriority of postural changes for body zones based on expert-defined protocols.Results show encouraging performance in the classification of in-bed posturesand high adaptability of the knowledge-based fuzzy approach.
COVID-19 study based on chest X-rays of patients
Lecturer: Anabel Gómez
Date: 06/11/2020
Abstract: COVID-19 is becoming one of the most infectious diseases of the 21st century. Due to the importance of its early detection, new ways to detect it are emerging. In this study, we focus on its detection using chest X-rays, pointing out the main problems of the most used data sets for this purpose. We propose a new data set and a new methodology that allows us to detect cases of COVID-19 with an accuracy of 76.18%, which is higher than the accuracies obtained by experts.
Image inpainting using non-adversarial networks. Towards a deeper semantic understanding of images
Lecturer: Guillermo Gómez
Date: 06/11/2020
Abstract: In this study we explore the problem of image inpainting from a non-adversarial perspective. Can we use general generative models to solve problems other than those for which it was trained to? Do models acquire a deeper and transferable knowledge about the nature of the images they generate? We propose a novel methodology for the image inpainting problem using the implicit knowledge acquired in non-adversarial generative models.
Sentiment Analysis based Multi-person Multi-criteria Decision Making (SA-MpMcDM) Methodology
Lecturer: Cristina Zuheros
Date: 30/11/2020
Abstract: Traditional decision making models are limited by pre-defined numerical and linguistic terms. We present the SA-MpMcDM methodology, which allows experts to evaluate through unlimited natural language and even through numerical ratings. We propose a deep learning model to extract the expert knowledge from the evaluations. We evaluate the methodology in a real case study, which we collect into the TripR-2020 dataset
MonuMAI: Architectural information extraction of monuments through Deep Learning techniques
Lecturer: Alberto Castillo
Date: 30/11/2020
Abstract: An important part of art history can be discovered through the visual information in monument facades. However, the analysis of this visual information, i.e, morphology and architectural elements, requires high expert knowledge. An automatic system for identifying the architectural style or detecting the architectural elements of a monument based on one image will certainly help improving our knowledge in art and history.
The aim of this seminary is to introduce the MonuMAI (Monument with Mathematics and Artificial Intelligence) framework published in the related work [1]. In particular, we designed MonuMAI dataset considering the proposed architectural styles taxonomy, developed MonuMAI deep learning pipeline, and built citizen science based MonuMAI mobile app that uses the proposed deep learning pipeline and dataset for performing in real life conditions.
[1] Lamas, Alberto & Tabik, Siham & Cruz, Policarpo & Montes, Rosana & Martínez-Sevilla, Álvaro & Cruz, Teresa & Herrera, Francisco. (2020) MonuMAI: Dataset, deep learning pipeline and citizen science based app for monumental heritage taxonomy and classification. Neurocomputing. doi.org/10.1016/j.neucom.2020.09.041