Search results
(41 - 60 of 72)
Pages
- Title
- Extreme Fine-grained Parallelism On Modern Many-Core Architectures
- Creator
- Nookala, Poornima
- Date
- 2022
- Description
-
Processors with 100s of threads of execution and GPUs with 1000s of cores are among the state-of-the-art in high-end computing systems. This...
Show moreProcessors with 100s of threads of execution and GPUs with 1000s of cores are among the state-of-the-art in high-end computing systems. This transition to many-core computing has required the community to develop new algorithms to overcome significant latency bottlenecks through massive concurrency. Implementing efficient parallel runtimes that can scale up to hundreds of threads with extremely fine-grained tasks (less than 100 microseconds) remains a challenge. We propose XQueue, a novel lockless concurrent queueing system that can scale up to hundreds of threads. We integrate XQueue into LLVM OpenMP and implement X-OpenMP, a library for lightweight tasking on modern many-core systems with hundreds of cores. We show that it is possible to implement a parallel execution model using lock-less techniques for enabling applications to strongly scale on many-core architectures. While the fork-join model is suitable for on-node parallelism, the use of joins and synchronization induces artificial dependencies which can lead to under utilization of resources. Data-flow based parallelism is crucial to overcome the limitations of fork-join parallelism by specifying dependencies at a finer granularity. It is also crucial for parallel runtime systems to support heterogeneous platforms to better utilize the hardware resources that are available in modern day supercomputers. The existing parallel programming environments that support distributed memory either discover the DAG entirely on all processes which limits the scalability or introduce explicit communications which increases the complexity of programming. We implement Template Task Graph (TTG), a novel programming model and its C++ implementation by marrying the ideas of control and data flowgraph programming. TTG can address the issues of performance portability without sacrificing scalability or programmability by providing higher-level abstractions than conventionally provided by task-centric programming systems, but without impeding the ability of these runtimes to manage task creation and execution as well as data and resource management efficiently. TTG implementation currently supports distributed memory execution over 2 different task runtimes PaRSEC and MADNESS.
Show less
- Title
- Towards a Secure and Resilient Smart Grid Cyberinfrastructure Using Software-Defined Networking
- Creator
- Qu, Yanfeng
- Date
- 2022
- Description
-
To enhance the cyber-resilience and security of the smart grid against malicious attacks and system errors, we present software-defined...
Show moreTo enhance the cyber-resilience and security of the smart grid against malicious attacks and system errors, we present software-defined networking (SDN)-based communication architecture design for smart grid operation. Our design utilizes SDN technology, which improves network manageability, and provides application-oriented visibility and direct programmability, to deploy the multiple SDN-aware applications to enhance grid security and resilience including optimization-based network management to recover Phasor Measurement Unit (PMU) network connectivity and restore power system observability; Flow-based anomaly detection and optimization-based network management to mitigate Manipulation of demand of IoT (MadIoT) attack. We also developed a prototype system in a cyber-physical testbed and conducted extensive evaluation experiments using the IEEE 30-bus system, IEEE 118-bus system, and IIT campus microgrid.
Show less
- Title
- PROGRAM SURVIVABILITY THROUGH K-VARIANT ARCHITECTURE
- Creator
- BEKIROGLU, BERK
- Date
- 2021
- Description
-
Numerous software systems, particularly mission and safety-critical systems, require a high level of security during their execution....
Show moreNumerous software systems, particularly mission and safety-critical systems, require a high level of security during their execution. Enhancing software security through architecture is a highly effective method of defending against cyberattacks. The N-version is a software architecture that was developed to increase the security of software systems. In the N-version architecture, functionally equivalent versions of a program run concurrently to complete a mission or task. Each version is developed independently by a different team using only the software specifications in common. As a result, each version is expected to contain unique vulnerabilities. Due to the high cost of developing and maintaining an N-version system, this architecture is typically used only in high-budget projects requiring a high-security level. The K-variant, an alternative architecture for enhancing system security, is explained and analyzed in this thesis. In contrast to the N-version architecture, each variant is automatically generated using source-to-source program transformation techniques. Automation significantly reduces the cost of developing variants in the K-variant architecture. The K-variant architecture can help protect systems from memory exploitation attacks. Various attack strategies can be used against K-variant systems in order to increase the likelihood of a successful attack. Various attack strategies are proposed and investigated in this thesis. Furthermore, experimental studies are being conducted to investigate various defense mechanisms against proposed attack strategies. The effectiveness of each defense mechanism against various attack strategies is evaluated by using a metric of the probability of an unsuccessful attack. Additionally, various source code program transformation techniques for generating new variants in the K-variant architecture have been proposed and investigated experimentally. This thesis also describes a machine learning technique for estimating the survivability of K-variant systems under various attack types and defense strategies. To make the design of K-variant systems easier, a neural network model is proposed. With the developed tool that utilizes the neural network model, fast and accurate predictions about the survivability of K-variant systems can be obtained.
Show less
- Title
- Workload Interference Analysis and Mitigation on Dragonfly Class Networks
- Creator
- Kang, Yao
- Date
- 2022
- Description
-
Dragonfly class of networks are promising interconnect topologies that support current and next-generation high-performance computing (HPC)...
Show moreDragonfly class of networks are promising interconnect topologies that support current and next-generation high-performance computing (HPC) systems. Serving as the "central nervous system", Dragonfly tightly couples tens of thousands of compute nodes together by providing high-bandwidth, low-latency data exchange for exascale computing capability. Dragonfly can support unprecedented system scale at a reasonable cost thanks to its hierarchical architecture. In Dragonfly systems, network resources such as routers and links are arranged into identical groups.Groups are all-to-all connected through global links, and routers within groups are connected via local links. In contrast to the fully connected inter-group topology, connections for the routers within groups are designed according to the system requirement. For example, the one-dimensional all-to-all connection is favored for higher network bandwidth, a two-dimensional grid arrangement can be constructed to support larger system size, and a tree structure router connection is built for the extreme system scale. The hierarchical design with groups enables the topology to support unprecedented system size while maintaining a low-diameter network. Packets can be minimally delivered by simply traversing the network hierarchy between groups through global links and reaching their destinations through local links. In case of network congestion, packets can be non-minimally forwarded through any intermediate group to increase the system throughput. As a result, all network resources are shared such that links and routers are not dedicated to any node pair. While link utilization is increased, shared network resources lead to inevitable network contention among different traffic flows, especially for the systems that hold multiple workloads at the same time. This network contention is observed as the workload interference that causes degraded system performance with delayed workload execution time. In this thesis, we first model and analyze the workload interference effect on Dragonfly+ topology through extensive system simulation.Based on the comprehensive interference study, we propose Q-adaptive routing, a multi-agent reinforcement learning based solution for Dragonfly systems. Compared with the existing routing solutions, the proposed Q-adaptive routing can learn to forward packets more efficiently with smaller packet latency and higher system throughput. Next, we demonstrate that intelligent routing algorithms such as Q-adaptive routing can greatly mitigate workload interference and optimize the overall system performance. Subsequently, we propose a dynamic job placement strategy for workload interference prevention. When combined with Q-adaptive routing, dynamic job placement gives users the flexibility to either reduce workload interference from communication intensive applications or protect target applications for higher performance stability.
Show less
- Title
- Technological Consciousness in Midwestern American Farming: From Party Lines to Autonomous Tractors
- Creator
- Sziron, Mónika
- Date
- 2022
- Description
-
This dissertation is primarily concerned with understanding the current conceptions, perceptions, and ethical concerns of artificial...
Show moreThis dissertation is primarily concerned with understanding the current conceptions, perceptions, and ethical concerns of artificial intelligence in Midwestern agriculture. Using the theory of technological consciousness as a backdrop for understanding the relationship between Midwestern agriculture and technology, in chapter two this dissertation first provides a narrative review of major technological developments throughout history in Midwestern farming and how the human experience in farming is influenced by technology throughout history. This history provides context for the current state of Midwestern agriculture, which is now increasingly entangled with artificial intelligence. The theory behind artificial intelligence ethics and general trends in artificial intelligence are discussed in chapter three. To understand present conceptions, perceptions, and ethical concerns of artificial intelligence for Midwestern farmers, a pilot survey was dispersed to farmers and pilot media content analysis was conducted on Midwestern agriculture publications. The results from this pilot survey and pilot media content analysis are discussed in chapter four. Chapter five delves into theory and how the human experience with technology has evolved over time and its effects on the human experience today. This chapter also provides theoretical insights for the future of farming with artificial intelligence. The dissertation concludes with reviewing the ethical concerns relating to artificial intelligence in agriculture for Midwestern farmers, provides recommendations for developers of agriculture technology, and highlights the new partnership between farmers and computer scientists and how this partnership will lead the way in the future of Midwestern farming.
Show less
- Title
- PIMMINER: A HIGH-PERFORMANCE PIM ARCHITECTURE-AWARE GRAPH MINING FRAMEWORK
- Creator
- Su, Jiya
- Date
- 2022
- Description
-
Graph mining applications, such as subgraph pattern matching and mining, are widely used in real-world domains such as bioinformatics, social...
Show moreGraph mining applications, such as subgraph pattern matching and mining, are widely used in real-world domains such as bioinformatics, social network analysis, and computer vision. Such applications are considered as a new class of data-intensive applications that generate massive irregular computation workloads and memory accesses, which degrade the performance and scalability significantly. Leveraging emerging hardware, such as process-in-memory (PIM) technology, could potentially accelerate such applications. In this paper, we propose PIMMiner, a high-performance PIM architecture graph mining framework. We first identify that current PIM architecture cannot be fully utilized by graph mining applications. Next, we propose a set of optimizations that enhance the locality, and internal bandwidth utilization and reduce remote bank accesses and load imbalance through cohesive algorithm and architecture co-designs. We compare PIMMiner with several state-of-the-art graph mining frameworks and show that PIMMiner is able to outperform all of them significantly.
Show less
- Title
- Quantum Computation for the Understanding of Mass: Simulating Quantum Field Theories
- Creator
- Rivero Ramírez, Pedro
- Date
- 2021
- Description
-
This thesis demonstrates the production of hadron mass on a quantum computer. Working in the Nambu–Jona-Lasinio model in 1+1 dimensions and 2...
Show moreThis thesis demonstrates the production of hadron mass on a quantum computer. Working in the Nambu–Jona-Lasinio model in 1+1 dimensions and 2 flavors, I show a separation of the contribution of quark masses and interactions to the mass. Along the way I develop a new tool called Quantum Sampling Regression (QSR) that allows for an optimal sampling of low qubit quantum computers when using hybrid variational eigenvalue solving techniques. I demonstrate the regime where QSR dominates the current standard Variational Eigensolver Technique, and benchmark it by improving the calculation of deuteron binding energy. Finally, I developed QRAND — a multiprotocol and multiplatform quantum random number generation framework — in support of the quantum computing community.
Show less
- Title
- Distribution-aware Visual Semantic Understanding
- Creator
- Chen, Ying
- Date
- 2021
- Description
-
Understanding visual semantics, including change detection and semantic segmentation, is an essential task in many computer vision and image...
Show moreUnderstanding visual semantics, including change detection and semantic segmentation, is an essential task in many computer vision and image processing applications. Examples of visual semantics understanding in images include land cover monitoring, urban expansion evaluation, autonomous driving, and scene understanding. The goal is to locate and recognize appropriate pixel-wise semantic labels in images. Classical computer vision algorithms involve sophisticated semi-heuristic pre-processing steps and potentially manual interaction. In this thesis, I propose and evaluate end-to-end deep neural approaches for processing images which achieve better performance compared with existing approaches. Supervised semantic segmentation has been widely studied and achieved great success with deep learning. However, existing deep learning methods typically suffer from generalization issues where a well-trained model may not work well on unseen samples from a different dataset. This is due to a distribution change or domain shift between the training and test sets that can degrade performance. Providing more labeled samples covering many possible variations can further improve the generalization of models, but acquiring labeled data is typically time-consuming, labor-intensive and requires domain knowledge. To tackle this label scarcity bottleneck for supervised learning, we propose to apply unsupervised domain adaptation, semi-supervised learning, and semi-supervised domain adaptation for neural semantic segmentation. The motivation behind unsupervised domain adaptation for semantic segmentation is to transfer learned knowledge from one or more source domains with sufficient labeled samples to a different but relevant target domain where labeled data is sparse or non-existent. The adaptation algorithm tends to learn a common representation space where the distributions over both source and target domains are matched. In this way, we expect a classifier working well in the source domain to generalize well to the target domain. More specifically, we try to learn class-aware source-target domain distribution differences, and transfer the knowledge learned from labeled synthetic data on the source domain to the unlabeled real data on the target domain. Different from domain adaptation, semi-supervised semantic segmentation aims at utilizing a large amount of unlabeled data to improve semantic classification trained on a small amount of labeled data from the same distribution. Specifically, supervised semantic segmentation is trained together with an unsupervised model by applying perturbations on encoded states of the network instead of the input, or using mask-based data augmentation techniques to encourage consistent predictions over mixed samples. In this way, learned representation which capture many kinds of unseen variations in unlabeled data, benefit the supervised semantic classifier. We propose a mask-based data augmentation semi-supervised learning network to utilize structure information from a variety of unlabeled examples to improve the learning on a limited number of labeled examples.Both unsupervised domain adaptation (UDA) with full source supervision but without target supervision and semi-supervised learning (SSL) with partial supervision have shown to be able to address the generalization problem to some extent. While such methods are effective at aligning different feature distributions, their inability to efficiently exploit unlabeled data leads to intra-domain discrepancy in the target domain, where the target domain is separated into two unaligned sub-distributions due to source-aligned and target-aligned data. That is, enforcing partial alignment between full labeled source data and a few labeled target data does not guarantee that the remaining unlabeled target samples will be aligned with source feature clusters, thus leaving them unaligned. Hence, I propose methods for incorporating the advantages of both UDA and SSL, termed semi-supervised domain adaptation (SSDA), with a goal to align cross-domain features as well as addressing the intra-domain discrepancy within the target domain. I propose a simple yet effective semi-supervised domain adaptation approach by utilizing a two-step domain adaptation addressing both cross-domain and intra-domain shifts.
Show less
- Title
- Improving Utility and Efficiency for Privacy Preserving Data Analysis
- Creator
- Liu, Bingyu
- Date
- 2022
- Description
-
In recent decades, the smart cities are incorporating with Internet-of-Things (IoT) infrastructures for improving the citizens’ quality of...
Show moreIn recent decades, the smart cities are incorporating with Internet-of-Things (IoT) infrastructures for improving the citizens’ quality of life by leveraging information/data. The huge amount of data is extracted and generated from the devices (e.g., mobile applications, GPS navigation systems, urban traffic cameras, etc.), or city sectors such as Intelligent Transportation Systems (ITS), Resource Allocation, Utilities, Crime Detection, Hospitals, and other community services.This dissertation aims to systematically research the Data Analysis in IoT System, which mainly consists of two aspects: Utility and Efficiency. First, ITS as a representative system in IoT in the smart city, I present the work on privacy preserving for the trajectories data, which is achieved by the differential privacy technique with a novel sanitation framework. Moreover, I have studied the resource allocation problem in two different approaches: Cryptographic computation and Hardware en- claves with the utility and efficiency accordingly. For the Cryptographic computation approach, I utilize Secure Multi-party Computation (SMC) technique for achieving the privacy-aware divisible double auction without a mediator. Besides, I also pro- pose a hardware-based solution Trusted Execution Environment (TEE) for performance improvement. At the same time, integrity and confidentiality are also able to be guaranteed. The proposed hybridized Trusted Execution Environment (TEE)- Blockchain System is designed for securely executing smart contract. Finally, I have studied the Cryptographic Video DNN Inference for the smart city surveillance, which privately inferring videos (e.g., action recognition, and video, and classification) on 3D spatial-temporal features with the C3D and I3D pre-trained DNN models with high performance. This dissertation proposes the privacy preserving frameworks and mechanisms are able to be applied efficiently for IoT in the real-world.
Show less
- Title
- Understanding and Combating Filter Bubbles in News Recommender Systems
- Creator
- Liu, Ping
- Date
- 2022
- Description
-
Algorithmic personalization of news and social media content aims to improve user experience. However, there is evidence that this filtering...
Show moreAlgorithmic personalization of news and social media content aims to improve user experience. However, there is evidence that this filtering can have the unintended side effect of creating homogeneous ``filter bubbles'' in which users are over-exposed to ideas that conform with their pre-existing perceptions and beliefs. In this thesis, I investigate this phenomenon in political news recommendation algorithms, which have important implications for civil discourse.I first collect and curate a collection of over 900K news articles from over 40 sources. The dataset was annotated in the topic and partisan leaning dimensions by conducting an initial pilot study and later via Amazon Mturk. This dataset is studied and used consistently throughout this thesis. In the first part of the thesis, I conduct simulation studies to investigate how different algorithmic strategies affect filter bubble formation. Drawing on Pew studies of political typologies, we identify heterogeneous effects based on the user's pre-existing preferences. For example, I find that i) users with more extreme preferences are shown less diverse content but have higher click-through rates than users with less extreme preferences, ii) content-based and collaborative-filtering recommenders result in markedly different filter bubbles, and iii) when users have divergent views on different topics, recommenders tend to have a homogenization effect.Secondly, I conduct a content analysis of the news to understand language usage among and across various topics and political stances. I examine words and phrases used by the liberal media and by the conservative media on each topic. I first study what differentiates the liberal media from the conservative media on each topic. I then study common phrases that are used by the liberals and the conservatives on different topics. For example, I examine which phrases are shared by the liberal articles on guns and conservative articles on abortion. Finally, I compare and visualize these words using different clustering algorithms and supervised classification methods.In the last chapter, I conduct an extensive user study to find possible solutions to combat the filter bubbles in the political news recommender systems. I designed a self-contained website that enables a content-based news recommender system and indexed 40,000 U.S.~political articles. I recruited over 800 U.S.~participants from Amazon Mechanical Turk (approved by IRB). The qualified participants are split into control and treatment groups. The users in the treatment group are provided transparency and interaction mechanisms, which grant them more control over the recommendations. Our results show that providing interaction and transparency a) increases click-through rates, b) has the potential to reduce the filter bubbles, and c) raises more awareness about filter bubbles.
Show less
- Title
- Enhancing Explanation Generation in the CaJaDE system using Interactive User Feedback
- Creator
- Lee, Juseung
- Date
- 2022
- Description
-
In today’s data-driven world, it is becoming increasingly difficult to interpret and understand query results after going through several...
Show moreIn today’s data-driven world, it is becoming increasingly difficult to interpret and understand query results after going through several manipulation steps, especially on a large database. There is a need for automated techniques that explain query results in a meaningful way. A recent study, CaJaDE(Context-Aware Join-Augmented Deep Explanations), presents a novel approach to generating explanations of query results including crucial contextual information. However, it becomes difficult to interpret explanations since the search space increases exponentially.In this thesis, we propose a new approach that introduces a user interaction model for a purpose of enhancing the generation of explanations in the CaJaDE system. We implemented a user interaction model that consists of three modules: User Selection, Recommendation Score, and User Rating. With these modules, our approach guides a user while exploring relevant join graphs, and lets them be involved in the decision-making process while generating join graphs. We demonstrate through performance experiments and user study that our approach is an effective method for users to understand explanations.
Show less
- Title
- Intelligent Job Scheduling on High Performance Computing Systems
- Creator
- Fan, Yuping
- Date
- 2021
- Description
-
Job scheduler is a crucial component in high-performance computing (HPC) systems. It sorts and allocates jobs according to site policies and...
Show moreJob scheduler is a crucial component in high-performance computing (HPC) systems. It sorts and allocates jobs according to site policies and resource availability. It plays an important role in the efficient use of system resources and users satisfaction. Existing HPC job schedulers typically leverage simple heuristics to schedule jobs. However, the rapid growth in system infrastructure and the introduction of diverse workloads pose serious challenges to the traditional heuristic approaches. First, the current approaches concentrate on CPU footprint and ignore the performance of other resources. Second, the scheduling policies are manually designed and only consider some isolated job information, such as job size and runtime estimate. Such a manual design process prevents the schedulers from making informative decisions by extracting the abundant environment information (i.e., system and queue information). Moreover, they can hardly adapt to workload changes, leading to degraded scheduling performance. These challenges call for a new job scheduling framework that can extract useful information from diverse workloads and the increasingly complicated system environment, and finally make well-informed scheduling decisions in real time.In this work, we propose an intelligent HPC job scheduling framework to address these emerging challenges. Our research takes advantage of advanced machine learning and optimization methods to extract useful workload- and system-specific information and to further educate the framework to make efficient scheduling decisions under various system configurations and diverse workloads. The framework contains four major efforts. First, we focus on providing more accurate job runtime estimations. Estimated job runtime is one of the most important factors affecting scheduling decisions. However, user provided runtime estimates are highly inaccurate and existing solutions are prone to underestimation which causes jobs to be killed. We leverage and enhance a machine learning method called Tobit model to improve the accuracy of job runtime estimates at the same time reduce underestimation rate. More importantly, using TRIP’s improved job runtime estimates boosts scheduling performance by up to 45%. Second, we conduct research on multi-resource scheduling. HPC systems are undergoing significant changes in recent years. New hardware devices, such as GPU and burst buffer, have been integrated into production HPC systems, which significantly expands the schedulable resources. Unfortunately, the current production schedulers allocate jobs solely based on CPU footprint, which severely hurts system performance. In our work, we propose a framework taking all scalable resources into consideration by transforming this problem into multi-objective optimization (MOO) problem and rapid solving it via genetic algorithm. Next, we leverage reinforcement learning (RL) to automatically learn efficient workload- and system-specific scheduling policies. Existing HPC schedulers either use generalized and simple heuristics or optimization methods that ignore workload and system characteristics. To overcome this issue, we design a new scheduling agent DRAS to automatically learn efficient scheduling policies. DRAS leverages the advance in deep reinforcement learning and incorporates the key features of HPC scheduling in the form of a hierarchical neural network structure. We develop a three-phase training process to help DRAS effectively learn the scheduling environment (i.e., the system and its workloads) and to rapidly converge to an optimal policy. Finally, we explore the problem of scheduling mixed workloads, i.e., rigid, malleable and on-demand workloads, on a single HPC system. Traditionally, rigid jobs are the main tenants of HPC systems. In recent years, malleable applications, i.e., jobs that can change sizes before and during execution, are emerging on HPC systems. In addition, dedicated clusters were the main platforms to run on-demand jobs, i.e., jobs needed to be completed in the shortest time possible. As the sizes of on-demand jobs are growing, HPC systems become more cost-efficient platforms for on-demand jobs. However, existing studies do not consider the problem of scheduling all three types of workloads. In our work, we propose six mechanisms, which combine checkpointing, shrink, expansion techniques, to schedule the mixed workloads on one HPC system.
Show less
- Title
- Algorithms for Discrete Data in Statistics and Operations Research
- Creator
- Schwartz, William K.
- Date
- 2021
- Description
-
This thesis develops mathematical background for the design of algorithms for discrete-data problems, two in statistics and one in operations...
Show moreThis thesis develops mathematical background for the design of algorithms for discrete-data problems, two in statistics and one in operations research. Chapter 1 gives some background on what chapters 2 to 4 have in common. It also defines some basic terminology that the other chapters use.Chapter 2 offers a general approach to modeling longitudinal network data, including exponential random graph models (ERGMs), that vary according to certain discrete-time Markov chains (The abstract of chapter 2 borrows heavily from the abstract of Schwartz et al., 2021). It connects conditional and Markovian exponential families, permutation- uniform Markov chains, various (temporal) ERGMs, and statistical considerations such as dyadic independence and exchangeability. Markovian exponential families are explored in depth to prove that they and only they have exponential family finite sample distributions with the same parameter as that of the transition probabilities. Many new statistical and algebraic properties of permutation-uniform Markov chains are derived. We introduce exponential random ?-multigraph models, motivated by our result on replacing ? observations of a permutation-uniform Markov chain of graphs with a single observation of a corresponding multigraph. Our approach simplifies analysis of some network and autoregressive models from the literature. Removing models’ temporal dependence but not interpretability permitted us to offer closed-form expressions for maximum likelihood estimators that previously did not have closed-form expression available. Chapter 3 designs novel, exact, conditional tests of statistical goodness-of-fit for mixed membership stochastic block models (MMSBMs) of networks, both directed and undirected. The tests employ a ?²-like statistic from which we define p-values for the general null hypothesis that the observed network’s distribution is in the MMSBM as well as for the simple null hypothesis that the distribution is in the MMSBM with specified parameters. For both tests the alternative hypothesis is that the distribution is unconstrained, and they both assume we have observed the block assignments. As exact tests that avoid asymptotic arguments, they are suitable for both small and large networks. Further we provide and analyze a Monte Carlo algorithm to compute the p-value for the simple null hypothesis. In addition to our rigorous results, simulations demonstrate the validity of the test and the convergence of the algorithm. As a conditional test, it requires the algorithm sample the fiber of a sufficient statistic. In contrast to the Markov chain Monte Carlo samplers common in the literature, our algorithm is an exact simulation, so it is faster, more accurate, and easier to implement. Computing the p-value for the general null hypothesis remains an open problem because it depends on an intractable optimization problem. We discuss the two schools of thought evident in the literature on how to deal with such problems, and we recommend a future research program to bridge the gap those two schools. Chapter 4 investigates an auctioneer’s revenue maximization problem in combinatorial auctions. In combinatorial auctions bidders express demand for discrete packages of multiple units of multiple, indivisible goods. The auctioneer’s NP-complete winner determination problem (WDP) is to fit these packages together within the available supply to maximize the bids’ sum. To shorten the path practitioners traverse from from legalese auction rules to computer code, we offer a new wdp formalism to reflect how government auctioneers sell billions of dollars of radio-spectrum licenses in combinatorial auctions today. It models common tie-breaking rules by maximizing a sum of bid vectors lexicographically. After a novel pre-solving technique based on package bids’ marginal values, we develop an algorithm for the WDP. In developing the algorithm’s branch-and-bound part adapted to lexicographic maximization, we discover a partial explanation of why classical WDP has been successful in using the linear programming relaxation: it equals the Lagrangian dual. We adapt the relaxation to lexicographic maximization. The algorithm’s dynamic-programming part retrieves already computed partial solutions from a novel data structure suited specifically to our WDP formalism. Finally we show that the data structure can “warm start” a popular algorithm for solving for opportunity-cost prices.
Show less
- Title
- TOPICDP – ENSURING DIFFERENTIAL PRIVACY FOR TOPIC MINING
- Creator
- Sharma, Jayashree
- Date
- 2021
- Description
-
Topic mining enables applications to recognize patterns and draw insights from text data, which can be used for applications such as sentiment...
Show moreTopic mining enables applications to recognize patterns and draw insights from text data, which can be used for applications such as sentiment analysis, building of recommender systems and classifiers. The text data can be a set of documents or emails or product feedback and reviews. Each document is analysed using probabilistic models and statistical analysis to discover patterns that reflects underlying topics.TopicDP is a differentially private topic mining technique, which injects well-calibrated Gaussian noise into the matrix output of the topic mining model generated from LDA algorithm. This method ensures differential privacy and good utility of the topic mining model. We derive smooth sensitivity for the Gaussian mechanism via sensitivity sampling, which resses the major challenges of high sensitivity in case of topic mining for differential privacy. Furthermore, we theoretically prove the differential privacy guarantee and utility error bounds of TopicDP. Finally, we conduct extensive experiments on two real-word text datasets (Enron email and Amazon Product Reviews), and the experimental results demonstrate that TopicDP can generate better privacy preserving performance for topic mining as compared against other state-of-the-art differential privacy mechanisms.
Show less
- Title
- SOLID-STATE SMART PLUG DEVICE
- Creator
- Deng, Zhixi
- Date
- 2022
- Description
-
Electrical faults are a leading cause of residential fire, and flexible power cords are particularly susceptible to metal or insulation...
Show moreElectrical faults are a leading cause of residential fire, and flexible power cords are particularly susceptible to metal or insulation degradation that may lead to a variety of electrical faults. Smart Plugs are a type of plug-in device controlling electrical loads via wireless communication for consumer market. However, there is lack of circuit protection features in existing Smart Plug products. Moreover, there is no previous product or research on Smart Plug with circuit protection features. This thesis introduces a new Smart Plug 2.0 concept which offers all-in-one protection against over-current, arc, and ground faults in addition to the smart features in Smart Plug products. It aims at preventing fire and shock hazards caused by degraded or damaged power cords and electrical connections in homes and offices. It offers microsecond-scale time resolution to detect and respond to a fault condition, and significantly reduces the electrothermal stress on household electrical wires and loads. A new arc fault detection method is developed using machine learning models based on load current di/dt events. The Smart Plug 2.0 concept has been validated experimentally. A 120V/10A solid-state Smart Plug 2.0 prototype using power MOSEFTs is designed and tested. It has experimentally demonstrated the comprehensive protection features against all types of electrical faults.
Show less
- Title
- ENGAGEMENT FRAMEWORK FOR SERIOUS THERAPEUTIC GAMES FOR HEALTH
- Creator
- Damarjian, Alex G.
- Date
- 2022
- Description
-
The conventional treatment of amblyopia in pediatric patients routinely experience low patient compliance due toits limitations. Therapeutic...
Show moreThe conventional treatment of amblyopia in pediatric patients routinely experience low patient compliance due toits limitations. Therapeutic games that utilize VR technology have the potential to open new avenues of medical research and treatment. A review of the prevailing literature shows the effectiveness of VR based games for therapeutic applications and the potential for increased patient compliance. A strong component of the literature is grounded in the medical humanities, specifically the way in which thought patterns, cognitive development, and perceived social rejection affect patient engagement and treatment efficacy. In order to increase the effectiveness of therapeutic games and streamline their development, a new framework has been created using existing research into therapeutic games. This framework ensures that all therapeutic games meet certain criteria within ethics, immersion, active learning, universal accessibility, aesthetics, and medicine. When applied to game development, specifically virtual and extended reality games, it can be used to transform existing therapeutic or diagnostic models into games operating as health care tools. The result is a more effective, lower cost, more accessible treatment option with increased patient compliance and greater overall outcomes.
Show less
- Title
- MEASUREMENT OF ELECTRON NEUTRINO AND ANTINEUTRINO APPEARANCE WITH THE NOνA EXPERIMENT
- Creator
- Yu, Shiqi
- Date
- 2020
- Description
-
As a long-baseline neutrino oscillation experiment, the NuMI Off-axis $\nu_e$ Appearance (NOvA) experiment aims at studying neutrino physics...
Show moreAs a long-baseline neutrino oscillation experiment, the NuMI Off-axis $\nu_e$ Appearance (NOvA) experiment aims at studying neutrino physics by measuring neutrino oscillation parameters using the neutrino flux from the Main Injector (NuMI) beam. It has two functionally identical detectors. The near detector is onsite at Fermi National Accelerator Laboratory. The far detector is 810 km away from the source of neutrinos and antineutrinos, at Ash River, Minnesota. At the near detector, muon neutrinos or antineutrinos, before significant oscillations take place, are used to correct the Monte Carlo simulation. At the far detector, the neutrino and antineutrino fluxes after significant oscillations have happened are measured and analyzed to study neutrino oscillation. The NOvA experiment is sensitive to the values of $\sin^2\theta_{23}$, $\Delta m^2_{32}$, and $\delta_{CP}$. The latest values from the NOvA 2020 analysis are as follows: $\sin^2\theta_{23}=0.57^{+0.03}_{-0.04}$, $\Delta m^2_{32}=(2.41\pm0.07)\times10^{-3}$ eV$^2$/c$^4$, and $\delta_{CP}=0.82\pi$ with a wide 1$\sigma$ interval of uncertainty. My study is focused on the neutrino oscillation analysis with NOvA, including detector light model tuning, particle classification with convolutional neural network, electron neutrino and antineutrino energy reconstruction, and oscillation background estimation. Most of my studies have been used in the latest NOvA publication and the NOvA 2020 analysis.
Show less
- Title
- Towards Trustworthy Multiagent and Machine Learning Systems
- Creator
- Xie, Shangyu
- Date
- 2022
- Description
-
This dissertation aims to systematically research the "trustworthy" Multiagent and Machine Learning systems in the context of the Internet of...
Show moreThis dissertation aims to systematically research the "trustworthy" Multiagent and Machine Learning systems in the context of the Internet of Things (IoT) system, which mainly consists of two aspects: data privacy and robustness. Specifically, data privacy concerns about the protection of the data in one given system, i.e., the data identified to be sensitive or private cannot be disclosed directly to others; robustness refers to the ability of the system to defend/mitigate the potential attacks/threats, i.e., maintaining the stable and normal operation of one system.Starting from the smart grid, a representative multiagent system in the IoT, I demonstrate two works on improving data privacy and robustness in aspects of different applications, load balancing and energy trading, which integrates secure multiparty computation (SMC) protocols for normal computation to ensure data privacy. More significantly, the schemes can be readily extended to other applications in IoT, e.g., connected vehicles, mobile sensing systems.For the machine learning, I have studied two main areas, i.e., computer vision and natural language processing with the privacy and robustness correspondingly. I first present the comprehensive robustness evaluation study of the DNN-based video recognition systems with two novel proposed attacks in both test and training phase, i.e., adversarial and poisoning attacks. Besides, I also propose the adaptive defenses to fully evaluate such two attacks, which can thus further advance the robustness of system. I also propose the privacy evaluation for the language systems and show the practice to reveal and address the privacy risks in the language models. Finally, I demonstrate a private and efficient data computation framework with the cloud computing technology to provide more robust and private IoT systems.
Show less
- Title
- Deep Learning Methods For Wireless Networks Optimization
- Creator
- Zhang, Shuai
- Date
- 2022
- Description
-
The resurgence of deep learning techniques has brought forth fundamental changes to how hard problems could be solved. It used to be held that...
Show moreThe resurgence of deep learning techniques has brought forth fundamental changes to how hard problems could be solved. It used to be held that the solutions to complex wireless network problems require accurate mathematical modeling of the network operation, but now the success of deep learning has shown that a data-driven method could generate powerful and useful representations such that the problem could be solved efficiently with surprisingly competent performance. Network researchers have recognized this and started to capitalize on the learning methods’ prowess. But most works follow the existing black-box learning paradigms without much accommodation to the nature and essence of the underlying network problems. This thesis focuses on a particular type of classical problem: multiple commodity flow scheduling in an interference-limited environment. Though it does not permit efficient exact algorithms due to its NP-hard complexity, we use it as an entry point to demonstrate from three angles how the learning-based methods can help improve the network performance. In the first part, we leverage the graphical neural network (GNN) techniques and propose a two-stage topology-aware machine learning framework, which trains a graph embedding unit and a link usage prediction module jointly to discover links that are likely to be used in optimal scheduling. The second part of the thesis is an attempt to find a learning method that has a closer algorithmic affinity to the traditional DCG method. We make use of reinforcement learning to incrementally generate a better partial solution such that a high quality solution may be found in a more efficient manner. As the third part of the research, we revisit the MCF problem from a novel viewpoint: instead of leaning on the neural networks to directly generate the good solutions, we use them to associate the current problem instance with historical ones that are similar in structure. These matched instances’ solutions offer a highly useful starting point to allow efficient discovery of the new instance’s solution.
Show less
- Title
- Exploiting contextual information for deep learning based object detection
- Creator
- Zhang, Chen
- Date
- 2020
- Description
-
Object detection has long been an important research topic in computer vision area. It forms the basis of many applications. Despite the great...
Show moreObject detection has long been an important research topic in computer vision area. It forms the basis of many applications. Despite the great progress made in recent years, object detection is still a challenging task. One of the keys to improving the performance of object detection is to utilize the contextual information from the image itself or from a video sequence. Contextual information is defined as the interrelated condition in which something exists or occurs. In object detection, such interrelated condition can be related background/surroundings, support from image segmentation task, and the existence of the object in the temporal domain for video-based object detection. In this thesis, we propose multiple methods to exploit contextual information to improve the performance of object detection from images and videos.First, we focus on exploiting spatial contextual information in still-image based object detection, where each image is treated independently. Our research focuses on extracting contextual information using different approaches, which includes recurrent convolutional layer with feature concatenation (RCL-FC), 3-D recurrent neural networks (3-D RNN), and location-aware deformable convolution. Second, we focus on exploiting pixel-level contextual information from a related computer vision task, namely image segmentation. Our research focuses on applying a weakly-supervised auxiliary multi-label segmentation network to improve the performance of object detection without increasing the inference time. Finally, we focus on video object detection, where the temporal contextual information between video frames are exploited. Our first research involves modeling short-term temporal contextual information using optical flow and modeling long-term temporal contextual information using convLSTM. Another research focuses on creating a two-path convLSTM pyramid to handle multi-scale temporal contextual information for dealing with the change in object's scale. Our last work is the event-aware convLSTM that forces convLSTM to learn about the event that causes the performance to drop in a video sequence.
Show less