On Cooperative Reinforcement Learning with Homogeneous Agents

Friday, September 10, 2021 - 02:20 pm
Storey Innovation Center 1400


Live Virtual Meeting Link

Meeting Location:
Storey Innovation Center 1400
 

BIO: Dr. Qi Zhang is an assistant professor of the Computer Science and Engineering department and the Artificial Intelligence Institute at the University of South Carolina. He got his Ph.D. from the Computer Science and Engineering department at the University of Michigan. His research aims for solutions for coordinating systems of decision-making agents operating in uncertain, dynamic environments. As hand-engineered solutions for such environments often fall short, He uses ideas from planning and reinforcement learning to develop and analyze algorithms that autonomously coordinate agents in an effective, trustworthy, and communication-efficient manner. In particular, He has been working on social commitments for trustworthy coordination, communication learning, and language emergence among coordinated agents and applications of (multi-agent) reinforcement learning such as intelligent transportation systems, dialogue systems, etc.

ABSTRACT: This talk will present two recent works on training homogeneous reinforcement learning (RL) agents in two distinct scenarios, respectively. The first scenario considers training a group of homogeneous agents that will be deployed in isolation to perform a single-agent RL task, which finds applications in ensemble RL. The first work develops effective techniques for training such as an ensemble of deep Q-learning agents, which help achieve state-of-the-art policy distillation performance in Atari games and continuous control tasks. The second scenario considers training a group of homogeneous agents to cooperatively perform a multi-agent RL task such as team sports. The second work develops novel techniques that exploit the homogeneity to train the agents in a distributed and communication-efficient manner.

talk poster

 

Supporting Special Purposes in General-Purpose Memory Hierarchies

Friday, September 3, 2021 - 02:20 pm
Storey Innovation Center 1400

See https://www.icaslab.com/Seminar

 

Meeting Link:
https://teams.microsoft.com/l/meetup-join/19%3ameeting_YWQ4NzcyYjMtYmZjNC00MTNjLTk4NTItYjRkOTFmYjk3NTRk%40thread.v2/0?context=%7b%22Tid%22%3a%224b2a4b19-d135-420e-8bb2-b1cd238998cc%22%2c%22Oid%22%3a%225fc2170a-7068-4a33-9021-df11b94ba696%22%7d

BIO: Dr. Xiaochen Guo is an associate professor in the Department of Electrical and Computer Engineering at Lehigh University. Dr. Guo received her Ph.D. degree in Electrical and Computer Engineering from the University of Rochester, and received the IBM Ph.D. Fellowship twice. Dr. Guo’s research interests are in the broad area of computer architecture, with an emphasis on leveraging emerging technologies to build energy-efficient microprocessors and memory systems. Dr. Guo is an IEEE senior member and a recipient of the National Science Foundation CAREER Award, the P. C. Rossin Assistant Professorship, and the Lawrence Berkeley National Laboratory Computing Sciences Research Pathways Fellowship.

ABSTRACT: General-purpose computing systems employ memory hierarchies to provide the appearance of a single large, fast, and coherent memory for general applications with good locality. However, conventional memory hierarchies cannot provide sufficient isolation for security workloads, support richer semantics, or hide memory latency for irregular memory accesses. This talk will present two of our recent works aiming to address these special needs in important workloads. In the first work, we propose to add a virtually addressed, set-associative scratchpad (SPX64) to a general-purpose CPU to support isolation and hash lookups in security and persistent applications. The proposed scratchpad is placed alongside of a traditional cache, and is able to avoid many of the programming challenges associated with traditional scratchpads without sacrificing generality. SPX64 delivers increased security and improves performance. In the second work, a software-assisted hardware prefetcher is proposed, which focuses on repeating irregular memory access patterns for data structures that cannot benefit from conventional memory hierarchies and hardware prefetchers. The key idea is to provide a programming interface to record cache miss sequence on the first appearance of a memory access pattern and prefetch through replaying the pattern on the following repetitions. By leveraging the programmer knowledge, the proposed Record-and-Replay (RnR) prefetcher can achieve over 95% prefetching accuracy and miss coverage.

talk poster

Towards More Trustworthy Deep Learning

Thursday, August 12, 2021 - 11:00 am
online

DISSERTATION DEFENSE

Department of Computer Science and Engineering

University of South Carolina

 Towards More Trustworthy Deep Learning: Accurate, Resilient, and Explainable Countermeasures Against Adversarial Examples

Author : Fei Zuo

Advisor : Dr. Qiang Zeng

Date : Aug 12, 2021

Time : 11:00am

Place : Virtual Defense

                 

Abstract

Despite the great achievements made by neural networks on tasks such as image classification, they are brittle and vulnerable to adversarial example (AE) attacks. Along with the prevalence of deep learning techniques, the threat of AEs attracts increasingly attentions since it may lead to serious consequences in some vital applications such as disease diagnosis. 

To defeat attacks based on AEs, both detection and defensive techniques attract the research community’s attention. While many countermeasures against AEs have been proposed, recent studies show that the existing detection methods usually goes ineffective when facing adaptive AEs. In this work, we exploit AEs by identifying their noticeable characteristics.

 

First, we noticed that L2 adversarial perturbations are among the most effective but difficult-to-detect attacks. How to detect adaptive L2 AEs is still an open question. At the same time, we find that, by randomly erasing some pixels in an L2 AE and then restoring it with an inpainting technique, the AE, before and after the steps, tends to have different classification results, while a benign sample does not show this symptom. We thus propose a novel AE detection technique, Erase-and-Restore (E&R), that exploits the intriguing sensitivity of L2 attacks. Comprehensive experiments conducted on standard image datasets show that the proposed detector is effective and accurate. More importantly, our approach demonstrates strong resilience to adaptive attacks. We also interpret the detection technique through both visualization and quantification.

 

Second, previous work considers that it is challenging to properly alleviate the effect of the heavy corruptions caused by L0 attacks. However, we argue that the uncontrollable heavy perturbation is

an inherent limitation of L0 AEs, and thwart such attacks. We thus propose a novel AE detector by converting the detection problem into a comparison problem. In addition, we show that the pre-processing technique used for detection can also work as an effective defense, which has a high probability of removing the adversarial influence of L0 perturbations. Thus, our system demonstrates not only high AE detection accuracies, but also a notable capability to correct the classification results.

 

Finally, we propose a comprehensive AE detector which systematically combines the two detection methods to thwart all categories of widely discussed AEs, i.e., L0, L2, and L∞ attacks. By acquiring the both strengths from its assembly components, the new hybrid AE detector is not only able to distinguish various kinds of  

AEs, but also has a very low false positive rate on benign images. More significantly, through exploiting the noticeable characteristics of AEs, the proposed detector is highly resilient to adaptive attack, filling a critical gap in AE detection.

Empowering neurodevelopmental studies through benchmarking and modeling

Tuesday, April 27, 2021 - 11:00 am

When: Tuesday, April 27, 11:00-12:00

https://zoom.us/j/98717098004?pwd=T3lPSGZ2K1pKTllQaWhRMDBtNVcrUT09

Speaker: Dr. Christian O’Reilly, McGill University, Canada

Talk abstract: Modeling is the bedrock on which science and technology have been built. Nowadays, almost every part of manufactured objects – may it be a supercomputer or a simple light bulb -- is modeled and simulated for us to gain a comprehensive understanding of how it works and how it will react under different conditions. Compared to human-made objects, our ability to get a grip on complex biological systems such as the brain has been hindered by these systems being black boxes which inner workings were mostly unknown. As we gain more insights on the mechanisms at play, our capacity to model and simulate these systems increases and further shed light on their remaining mysteries. In parallel, as the advances in medicine and science provide us with a finer appreciation of these biological systems, it also generates more intricate challenges. Tackling these new problems often requires integrating many sources of knowledge across fields and scales, from slow-evolving social factors to millisecond molecular interactions. Understanding complex multi-factorial and multidimensional neurodevelopmental issues like those present in the autistic spectrum disorder is such a problem. In this context, setting up a solid analytical framework empowered by modeling and simulation is even more important. In the first half of this talk I will go over some of my experiences in analyzing and modeling neuronal systems at different scales, from the macroscopic whole-brain scale to the microscopic cellular scale. Then, in the second part, building on these experiences I will make a case for the importance of systematically benchmarking the different aspects of the brain across scales and integrating such knowledge into analytical tools that we can use for scientific discoveries and clinical decisions. Speaker bio (short): Christian O’Reilly (Google Scholar) received his B.Ing (electrical eng.; 2007), his M.Sc.A. (biomedical eng.; 2011), and his Ph.D. (biomedical eng.; 2012) from the École Polytechnique de Montréal where he worked under the mentoring of Pr. R. Plamondon to apply pattern recognition and machine learning to predict brain stroke risks. He was later a postdoctoral fellow in Pr. T. Nielsen’s laboratory at the Center for Advanced Research in Sleep Medicine of the Hôpital du Sacré-Coeur/Université de Montréal (2012-2014) and then a NSERC postdoctoral fellow at McGill's Brain Imaging Center (2014-2015) where he worked in Pr. Baillet’s laboratory on characterizing EEG sleep transients, their sources, and their functional connectivity. During this period, he also was a visiting scholar in Pr. K. Friston's laboratory at the University College of London to study effective connectivity during sleep transients using dynamic causal modeling, an approach based on the Bayesian inversion of neural mass models. He later took on a 6-month fellowship with the Pr. M. Elsabbagh on functional connectivity in autism after which he moved to Switzerland to work for the Blue Brain project (Pr. S. Hill; EPFL; 2015-2018) where he led efforts on large-scale biophysically detailed modeling of the thalamocortical loop. Since 2020, he resumed his collaboration with the Dr. Elsabbagh as a research associate at the Azrieli Centre for Autism Research (McGill) where he is studying brain connectivity in autism and related neurodevelopmental disorders.

Multi-Robot Coordination with Environmental Disturbances

Wednesday, April 7, 2021 - 12:00 pm
DISSERTATION DEFENSE Department of Computer Science and Engineering University of South Carolina Author : Adem Coskun Advisor : Dr. Marco Valtorta Date : April 7, 2021 Time : 12:00 pm - 02:00 pm Place : Virtual Defense (link below) Link : Microsoft Teams: https://teams.microsoft.com/l/meetup-join/19%3ameeting_YTJkMzc0MjEtNDg3… Abstract Multi-robot systems are increasingly deployed in environments where they interact with humans. From the perspective of a robot, such interaction could be considered a disturbance that causes a well-planned trajectory to fail. This dissertation addresses the problem of multi-robot coordination in scenarios where the robots may experience unexpected delays in their movements. Prior work by Cap Gregoire, and Frazzoli introduced a control law, called RMTRACK, which enables robots in such scenarios to execute pre-planned paths in spite of disturbances that affect the execution speed of each robot while guaranteeing that each robot can reach its goal without collisions and without deadlocks. We extend that approach to handle scenarios in which the disturbance probabilities are unknown when execution starts and are non-uniform across the environment. The key idea is to "repair" a plan on-the-fly, by swapping the order in which a pair of robots passes through a mutual collision region (i.e. a coordination space obstacle), when making such a change is expected to improve the overall performance of the system. We introduce a technique based on Gaussian processes to estimate future disturbances, and propose two algorithms for testing, at appropriate times, whether a swap of a given obstacle would be beneficial. Tests in simulation demonstrate that our algorithms achieve significantly smaller average travel time than RMTRACK at only a modest computational expense. However, deadlock may arise when rearranging the order in which robots pass collision regions and other obstacles. We provide a precise definition of deadlock using a graphical representation and prove some of its important properties. We show how to exploit the representation to detect the possibility of deadlock and to characterize conditions under which deadlock may not occur. We provide experiments in simulated environments that illustrate the potential usefulness of our theory of deadlock.

Computational and Causal Approaches on Social Media and Multimodal Sensing Data: Examining Wellbeing in Situated Contexts

Friday, April 2, 2021 - 11:00 am
Time: Apr 2, 2021 11:00 AM Eastern Time (US and Canada) https://zoom.us/j/8440139296?pwd=MU9EWkJ5VEMyNzlneHI3Q1NTN0JxZz09 Abstract: A core aspect of our social lives is often embedded in the communities that we are situated in, such as our workplaces, neighborhoods, localities, and school/ college campuses. The inter-connectedness and inter-dependencies of our interactions, experiences, and concerns intertwine our situated context with our wellbeing. A better understanding of our wellbeing and psychosocial dynamics will help us devise strategies to address our wellbeing through proactive and tailored support strategies. However, existing methodologies to assess wellbeing suffer from limitations of scale and timeliness. Parallelly, given its ubiquity and widespread use, social media can be considered a “passive sensor” that can act as a complementary source of unobtrusive, real-time, and naturalistic data to infer wellbeing. In this talk, Koustuv Saha, from Georgia Tech, will present computational and causal approaches for leveraging social media in concert with complementary multimodal data to examine wellbeing. He will show how theory-driven computational methods can be applied on unique social media and complementary multimodal data to capture attributes of human behavior and psychosocial dynamics in situated communities, particularly college campuses and workplaces. Further, he will dive deep into drawing meaning out of online-inferences about offline metrics. Finally, this talk will propel the vision towards human-centered technologies tailored to situations, demands, and needs, facilitating technology-supported remote functioning, evaluating the prospective utility of social platforms for wellbeing, and understanding the harms/benefits of computational and data-driven assessments. Bio: Koustuv Saha is a doctoral candidate in Computer Science in the School of Interactive Computing at Georgia Tech. His research interest is in Social Computing and Computational Social Science. In his research, he adopts machine learning, natural language, and causal inference analysis to examine human behavior and wellbeing using social media and online data, along with complementary multimodal sensing data. His work has been published at several high prestige venues, including CHI, CSCW, ICWSM, IMWUT, JMIR, TBM, ACII, FAT*, PervasiveHealth, and WebSci, among others. He has been recognized as Foley Scholar, a recipient of the Foley Scholarship Award, GVU Center’s highest recognition for student excellence in research contributions to computing. He is a recipient of the Snap Research Fellowship, a finalist of the Symantec Graduate Fellowship, and his research has won the Outstanding Study Design Award at ICWSM 2019. His research has been covered at prestigious media outlets, including the New York Times, CBC Radio, NBC, 11Alive, the Hill, and the Commonwealth Times. During his Ph.D., he has had research internships at Snap Research, Microsoft Research, Max Planck Institute, and Fred Hutch Cancer Research. Earlier, he completed his B.Tech (Hons.) in Computer Science and Engineering from the Indian Institute of Technology (IIT) Kharagpur. He was also awarded the NTSE Scholarship by the Govt. of India, and he holds an overall industry research experience of five years. More about Koustuv can be found out at https://koustuv.com.

Regularized Deep Network Learning for Multi-label Visual Recognition

Wednesday, March 31, 2021 - 10:00 am
DISSERTATION DEFENSE Department of Computer Science and Engineering University of South Carolina Author : Hao Guo Advisor : Dr. Song Wang Date : March 31, 2021 Time : 10:00am - 12:00 pm Place : Virtual Defense (link below) Link : https://bluejeans.com/140824998 Abstract This dissertation is focused on the task of multi-label visual recognition, a fundamental task of computer vision. It aims to tell the presence of multiple visual classes from the input image, where the visual classes, such as objects, scenes, attributes, etc., are usually defined as image labels. Due to the prosperous deep networks, this task has been widely studied and significantly improved in recent years. However, it remains a challenging task due to appearance complexity of multiple visual contents co-occurring in one image. This research explores to regularize the deep network learning for multi-label visual recognition. First, an attention concentration method is proposed to refine the deep network learning for human attribute recognition, i.e., a challenging instance of multi-label visual recognition. Here the visual attention of deep networks, in terms of attention maps, is an imitation of human attention in visual recognition. Derived by the deep network with only label-level supervision, attention maps interpretively highlight areas indicating the most relevant regions that contribute most to the final network prediction. Based on the observation that human attributes are usually depicted by local image regions, the added attention concentration enhances the deep network learning for human attribute recognition by forcing the recognition on compact attribute-relevant regions. Second, inspired by the consistent relevance between a visual class and an image region, an attention consistency strategy is explored and enforced during deep network learning for human attribute recognition. Specifically, two kinds of attention consistency are studied in this dissertation, including the equivariance under spatial transforms, such as flipping, scaling and rotation, and the invariance between different networks for recognizing the same attribute from the same image. These two kinds of attention consistency are formulated as a unified attention consistency loss and combined with the traditional classification loss for network learning. Experiments on public datasets verify its effectiveness by achieving new state-of-the-art performance for human attribute recognition. Finally, to address the long-tailed category distribution of multi-label visual recognition, the collaborative learning between using uniform and re-balanced samplings is proposed for regularizing the network training. While the uniform sampling leads to relatively low performance on tail categories, re-balanced sampling can improve the performance on tail classes, but may also hurt the performance on head classes in network training due to label co-occurrence. This research proposes a new approach to train on both class-biased samplings in a collaborative way, resulting in performance improvement for both head and tail classes. Based on a two-branch network taking the uniform sampling and re-balanced sampling as the inputs, respectively, a cross-branch loss enforces consistency when the same input goes through the two branches. The experimental results demonstrate that the proposed method significantly outperforms existing state-of-the-art methods on long-tailed multi-label visual recognition.

Deep Learning Based Sound Event Detection and Classification

Monday, March 29, 2021 - 01:00 pm
DISSERTATION DEFENSE Department of Computer Science and Engineering University of South Carolina Author : Alireza Nasiri Advisor : Dr. Jianjun Hu Date : March 29, 2021 Time : 1:00 - 3:00 pm Place : Virtual Defense (link below) Zoom link: https://us02web.zoom.us/j/83125251774?pwd=NDEwK3M4b3NuT0djQ25BMlQ2cGtuZ… Abstract Hearing sense has an important role in our daily lives. During the recent years, there has been many studies to transfer this capability to the computers. In this dissertation, we design and implement deep learning based algorithms to improve the ability of the computers in recognizing the different sound events. In the first topic, we investigate sound event detection, which identifies the time boundaries of the sound events in addition to the type of the events. For sound event detection, we propose a new method, AudioMask, to benefit from the object-detection techniques in computer vision. In this method, we convert the question of identifying time boundaries for sound events, into the problem of identifying objects in images by treating the spectrograms of the sound as images. AudioMask first applies Mask R-CNN, an algorithm for detecting objects in images, to the log-scaled mel-spectrograms of the sound files. Then we use a frame-based sound event classifier trained independently from Mask R-CNN, to analyze each individual frame in the candidate segments. Our experiments show that, this approach has promising results and can successfully identify the exact time boundaries of the sound events. In the second topic, we present SoundCLR, a supervised contrastive learning based method for effective environmental sound classification with state-of-the-art performance, which works by learning representations that disentangle the samples of each class from those of other classes. We also exploit transfer learning and strong data augmentation to improve the results. Our extensive benchmark experiments show that our hybrid deep network models trained with combined contrastive and cross-entropy loss achieved the state-of-the-art performance on three benchmark datasets ESC-10, ESC-50, and US8K with validation accuracies of 99.75%, 93.4%, and 86.49% respectively. The ensemble version of our models also outperforms other top ensemble methods. Finally, we analyze the acoustic emissions that are generated during the degradation process of SiC composites. The aim here is to identify the state of the degradation in the material, by classifying its emitted acoustic signals. As our baseline, we use random forest method on expert-defined features. Also we propose a deep neural network of convolutional layers to identify the patterns in the raw sound signals. Our experiments show that both of our methods are reliably capable of identifying the degradation state of the composite, and in average, the convolutional model significantly outperforms the random forest technique.

From the Lab to Community: AI for Document Understanding and Public Health

Monday, March 29, 2021 - 11:00 am
Topic: Seminar: Muhammad Rahman Time: Mar 29, 2021 11:00 AM Eastern Time (US and Canada) Join Zoom Meeting https://zoom.us/j/97536411087?pwd=RExTdkVQcEg4OERFMUJhWm5rQThndz09 Title: Abstract: Artificial intelligence (AI) has made incredible scientific and technological contributions in many areas including business, healthcare and psychology. Due to the multidisciplinary nature and the ability to revolutionize, almost every field has started welcoming AI. The last decade is the witness of progresses of AI and machine learning, and their applications. In this talk, I will present my work that used AI and machine learning to solve interesting research challenges. The first part of my talk will describe an AI-powered framework that I have developed for large document understanding. The research contributed by modeling and extracting the logical and semantic structure of electronic documents using machine learning techniques. In the second part of my talk, I will present an ongoing work that uses computational technology to design a study for measuring COVID-19 effects on people with substance use disorders. I will conclude the talk by introducing few other AI-powered initiatives in mental health, substance use and addiction that I am currently working on. Bio: Dr. Muhammad Rahman is a Postdoctoral Researcher at National Institutes of Health (NIH). Before that, he was a Postdoctoral Fellow in the Center for Language and Speech Processing (CLSP) research lab at Johns Hopkins University. He obtained his Ph.D. in computer science from the University of Maryland, Baltimore County. His research is at the intersection of artificial intelligence (AI), machine learning, natural language processing, mental health, addiction and public health. Dr. Rahman’s current research mostly focuses on the real-world applications of advanced AI and machine learning techniques in addiction, mental health and behavioral psychology. As a part of NIH, he is working on designing and developing real-time digital intervention techniques to support substance use disorders and mental illness patients. During his Ph.D., Dr. Rahman worked on large document understanding that automatically identifies different sections of documents and understands their purpose within the document. He also had research internships at AT&T Labs and eBay Research where he worked on large scale industrial research projects. https://irp.drugabuse.gov/staff-members/muhammad-mahbubur-rahman-ph-d/