Georgia Institute of Technology

Displaying 1-40 of 335 results

  • White Papers // Oct 2014

    CAMEO:A Two-Level Memory Organization with Capacity of Main Memory and Flexibility of Hardware-Managed Cache

    In this paper, the authors analyze the trade-offs in architecting stacked DRAM either as part of main memory or as a hardware-managed cache. Using stacked DRAM as part of main memory increases the effective capacity, but obtaining high performance from such a system requires Operating System (OS) support to migrate...

    Provided By Georgia Institute of Technology

  • White Papers // Apr 2014

    Improving MapReduce Performance in a Heterogeneous Cloud: A Measurement Study

    Hybrid clouds, geo-distributed cloud and continuous upgrades of computing, storage and networking resources in the cloud have driven datacenters evolving towards heterogeneous clusters. Unfortunately, most of MapReduce implementations are designed for homogeneous computing environments and perform poorly in heterogeneous clusters. Although a fair of research efforts have dedicated to improve...

    Provided By Georgia Institute of Technology

  • White Papers // Mar 2014

    Software-Based Techniques for Reducing the Vulnerability of GPU Applications

    As highly-parallel accelerators such as graphics processing units become more important in high-performance computing, so does the need to ensure their reliable operation. In response, research has been directed at several efforts to characterize and understand the hardware vulnerability of GPU micro-architecture structures, as well as to detecting and correcting...

    Provided By Georgia Institute of Technology

  • White Papers // Mar 2014

    A Measure of Heterogeneity in Multi-Agent Systems

    Heterogeneous multi-agent systems have previously been studied and deployed to solve a number of different tasks. Despite this, the users still lack a basic understanding of just what \"Heterogeneity\" really is. For example, what makes one team of agents more heterogeneous than another? In this paper, the authors address this...

    Provided By Georgia Institute of Technology

  • White Papers // Mar 2014

    Road-Network Aware Trajectory Clustering: Integrating Locality, Flow and Density

    Mining trajectory data has been gaining significant interest in recent years. However, existing approaches to trajectory clustering are mainly based on density and Euclidean distance measures. The authors argue that when the utility of spatial clustering of mobile object trajectories is targeted at road-network aware location-based applications, density and Euclidean...

    Provided By Georgia Institute of Technology

  • White Papers // Feb 2014

    Algorithmic Time, Energy, and Power on Candidate HPC Compute Building Blocks

    The authors conducted a micro-benchmarking study of the time, energy, and power of computation and memory access on several existing platforms. These platforms represent candidate compute-node building blocks of future high-performance computing systems. Their analysis uses the \"Energy roofline\" model, developed in prior work, which they extend in two ways....

    Provided By Georgia Institute of Technology

  • White Papers // Feb 2014

    Manifold: A Parallel Simulation Framework for Multicore Systems

    In this paper the authors present manifold, an open-source parallel simulation framework for multi-core architectures. It consists of a parallel simulation kernel, a set of micro-architecture components, and an integrated library of power, thermal, reliability, and energy models. Using the components as building blocks, users can assemble multi-core architecture simulation...

    Provided By Georgia Institute of Technology

  • White Papers // Jan 2014

    Methodical Approximate Hardware Design and Reuse

    Design and reuse of approximate hardware components - digital circuits that may produce inaccurate results - can potentially lead to significant performance and energy improvements. Many emerging error-resilient applications can exploit such designs provided approximation is applied in a controlled manner. This paper provides the design abstractions and semantics for...

    Provided By Georgia Institute of Technology

  • White Papers // Jan 2014

    An Optimized 3D-Stacked Memory Architecture by Exploiting Excessive, High-Density TSV Bandwidth

    Memory bandwidth has become a major performance bottleneck as more and more cores are integrated onto a single die, demanding more and more data from the system memory. Several prior studies have demonstrated that this memory bandwidth problem can be addressed by employing a 3D-stacked memory architecture, which provides a...

    Provided By Georgia Institute of Technology

  • White Papers // Jan 2014

    Designing 3D Test Wrappers for Prebond and Postbond Test of 3D Embedded Cores

    3D integration is a promising new technology for tightly integrating multiple active silicon layers into a single chip stack. Both the integration of heterogeneous tiers and the partitioning of functional units across tiers leads to significant improvements in functionality, area, performance, and power consumption. Managing the complexity of 3D design...

    Provided By Georgia Institute of Technology

  • White Papers // Jan 2014

    Smart Refresh: An Enhanced Memory Controller Design for Reducing Energy in Conventional and 3D DieStacked DRAMs

    DRAMs require periodic refresh for preserving data stored in them. The refresh interval for DRAMs depends on the vendor and the design technology they use. For each refresh in a DRAM row, the stored information in each cell is read out and then written back to itself as each DRAM...

    Provided By Georgia Institute of Technology

  • White Papers // Jan 2014

    Noise-Direct: A Technique for Power Supply Noise Aware Floorplanning Using Microarchitecture Profiling

    In this paper, the authors propose Noise-Direct, a design methodology for power integrity aware floorplanning, using microarchitectural feedback to guide module placement. Stringent power constraints have led microprocessor designers to incorporate aggressive power saving techniques such as clock-gating that place a significant burden on the power delivery network. While the...

    Provided By Georgia Institute of Technology

  • White Papers // Jan 2014

    DRAM Decay: Using Decay Counters to Reduce Energy Consumption in DRAMs

    Dynamic Random Access Memories (DRAMs) require periodic refresh for preserving data stored in them. The refresh interval for DRAMs depends on the vendor and the de-sign technology they use. For each refresh in a DRAM row, the stored information in each cell is read out and then writ-ten back to...

    Provided By Georgia Institute of Technology

  • White Papers // Jan 2014

    Asymmetry Aware Scheduling Algorithms for Asymmetric Multiprocessors

    Multiprocessor architecture is becoming popular in both desktop processors and mobile processors. Especially asymmetric architecture shows promise in saving energy and power. However, how to design applications and how to schedule applications in asymmetric multiprocessors are still challenging problems. In this paper, the authors evaluate the performance of applications in...

    Provided By Georgia Institute of Technology

  • White Papers // Jan 2014

    Hardware Support for Real-Time Embedded Multiprocessor System-on-a-Chip Memory Management

    The aggressive evolution of the semiconductor industry smaller process geometries, higher densities, and greater chip complexity has provided design engineers the means to create complex, high-performance Systems-on-Chip (SoC) designs. Such, SoC designs typically have more than one processor and huge memory, all on the same chip. Dealing with the global...

    Provided By Georgia Institute of Technology

  • White Papers // Sep 2013

    ClusterWatch: Flexible, Lightweight Monitoring for High-end GPGPU Clusters

    The ClusterWatch middleware provides runtime flexibility in what system-level metrics are monitored, how frequently such monitoring is done, and how metrics are combined to obtain reliable information about the current behavior of GPGPU clusters. Interesting attributes of ClusterWatch are the ease with which different metrics can be added to the...

    Provided By Georgia Institute of Technology

  • White Papers // Aug 2013

    An Infrastructure for Automating Large-scale Performance Studies and Data Processing

    The cloud has enabled the computing model to shift from traditional data centers to publicly shared computing infrastructure; yet, applications leveraging this new computing model can experience performance and scalability issues, which arise from the hidden complexities of the cloud. The most reliable path for better understanding these complexities is...

    Provided By Georgia Institute of Technology

  • White Papers // Aug 2013

    Oncilla: A GAS Runtime for Efficient Resource Allocation and Data Movement in Accelerated Clusters

    Accelerated and in-core implementations of big data applications typically require large amounts of host and accelerator memory as well as efficient mechanisms for transferring data to and from accelerators in heterogeneous clusters. Scheduling for heterogeneous CPU and GPU clusters has been investigated in depth in the High-Performance Computing (HPC) and...

    Provided By Georgia Institute of Technology

  • White Papers // Aug 2013

    Oncilla: A GAS Runtime for Efficient Resource Allocation and Data Movement in Accelerated Cluster

    Accelerated and in-core implementations of big data applications typically require large amounts of host and accelerator memory as well as efficient mechanisms for transferring data to and from accelerators in heterogeneous clusters. Scheduling for heterogeneous CPU and GPU clusters has been investigated in depth in the High-Performance Computing (HPC) and...

    Provided By Georgia Institute of Technology

  • White Papers // Aug 2013

    Secure Outsourced Garbled Circuit Evaluation for Mobile Devices

    Garbled circuits provide a powerful tool for jointly evaluating functions while preserving the privacy of each user's inputs. While recent research has made the use of this primitive more practical, such solutions generally assume that participants are symmetrically provisioned with massive computing resources. In reality, most people on the planet...

    Provided By Georgia Institute of Technology

  • White Papers // Jul 2013

    Personal Clouds: Sharing and Integrating Networked Resources to Enhance End User Experiences

    End user experiences on mobile devices with their rich sets of sensors are constrained by limited device battery lives and restricted form factors, as well as by the 'scope' of the data available locally. The 'Personal Cloud' distributed software abstractions address these issues by enhancing the capabilities of a mobile...

    Provided By Georgia Institute of Technology

  • White Papers // Jul 2013

    On Symmetric Encryption with Distinguishable Decryption Failures

    The authors propose to relax the assumption that decryption failures are indistinguishable in security models for symmetric encryption. Their main purpose is to build models that better reflect the reality of cryptographic implementations, and to surface the security issues that arise from doing so. They systematically explore the consequences of...

    Provided By Georgia Institute of Technology

  • White Papers // Jun 2013

    Take This Personally: Pollution Attacks on Personalized Services

    Modern web services routinely personalize content to appeal to the specific interests, viewpoints, and contexts of individual users. Ideally, personalization allows sites to highlight information uniquely relevant to each of their users, thereby increasing user satisfaction - and, eventually, the service's bottom line. Unfortunately, as the authors demonstrate in this...

    Provided By Georgia Institute of Technology

  • White Papers // Jun 2013

    An Automated Approach to Create, Store, and Analyze Large-scale Experimental Data in Clouds

    The flexibility and scalability of computing clouds make them an attractive application migration target; yet, the cloud remains a black-box for the most part. In particular, their opacity impedes the efficient but necessary testing and tuning prior to moving new applications into the cloud. A natural and presumably unbiased approach...

    Provided By Georgia Institute of Technology

  • White Papers // Jun 2013

    FlexIO: Location-Flexible Execution of in Situ Data Analytics for Large Scale Scientific Applications

    Increasingly severe I/O bottlenecks on high-end computing machines are prompting scientists to process simulation output data while simulations are running and before placing data on disk - \"In situ\" and/or \"In-transit\". There are several options in placing in-situ data analytics along the I/O path: on compute nodes, on staging nodes...

    Provided By Georgia Institute of Technology

  • White Papers // Jun 2013

    I/O Containers: Managing the Data Analytics and Visualization Pipelines of High End Codes

    Lack of I/O scalability is known to cause measurable slowdowns for large-scale scientific applications running on high end machines. This is prompting researchers to devise 'I/O staging' methods in which outputs are processed via online analysis and visualization methods to support desired science outcomes. Organized as online workflows and carried...

    Provided By Georgia Institute of Technology

  • White Papers // Jun 2013

    A Study of the Effect of Partitioning on Parallel Simulation of Multicore Systems

    There has been little research that studies the effect of partitioning on parallel simulation of multicore systems. This paper presents their paper of this important problem in the context of null-message-based synchronization algorithm for parallel multicore simulation. This paper focuses on coarse grain parallel simulation where each core and its...

    Provided By Georgia Institute of Technology

  • White Papers // Jun 2013

    Cloud Manufacturing: Drivers, Current Status, and Future Trends

    Cloud Manufacturing (CM) refers to a customer-centric manufacturing model that exploits on-demand access to a shared collection of diversified and distributed manufacturing resources to form temporary, reconfigurable production lines which enhance efficiency, reduce product lifecycle costs, and allow for optimal resource loading in response to variable-demand customer generated tasking. The...

    Provided By Georgia Institute of Technology

  • White Papers // May 2013

    Resolution-Aware Network Coded Storage

    In this paper, the authors show that coding can be used in Storage Area Networks (SANs) to improve various Quality of Service metrics under normal SAN operating conditions, without requiring additional storage space. For their analysis, they develop a model which captures modern characteristics such as constrained I/O access bandwidth...

    Provided By Georgia Institute of Technology

  • White Papers // May 2013

    Variations in Performance Measurements of Multi-Core Processors: A Study of n-Tier Applications

    The prevalence of multi-core processors has raised the question of whether applications can use the increasing number of cores efficiently in order to provide predictable Quality of Service (QoS). In this paper, the authors study the horizontal scalability of n-tier application performance within a Multi-Core Processor (MCP). Through extensive measurements...

    Provided By Georgia Institute of Technology

  • White Papers // May 2013

    Secure Cloud Storage Service with An Efficient DOKS Protocol

    Storage services based on public clouds provide customers with elastic storage and on-demand accessibility. However, moving data to remote cloud storage also raises privacy concerns. Cryptographic cloud storage and search over encrypted data have attracted attentions from both industry and academics. In this paper, the authors present a new approach...

    Provided By Georgia Institute of Technology

  • White Papers // May 2013

    Software-Controlled Transparent Management of Heterogeneous Memory Resources in Virtualized Systems

    This paper presents a software-controlled technique for managing the heterogeneous memory resources of next generation multicore platforms with fast 3D die-stacked memory and additional slow off-chip memory. Implemented for virtualized server systems, the technique detects the 'Hot' pages critical to program performance in order to then maintain them in the...

    Provided By Georgia Institute of Technology

  • White Papers // May 2013

    Efficient Trajectory Cover Search for Moving Object Trajectories

    Given a set of query locations and a set of query keywords, a Trajectory Cover (CT) query over a repository of mobile trajectories returns a minimal set of trajectories that maximally covers the query keywords and are also spatially close to the query locations. Processing CT queries over mobile trajectories...

    Provided By Georgia Institute of Technology

  • White Papers // May 2013

    Residency-Aware Virtual Machine Communication Optimization: Design Choices and Techniques

    Network I/O workloads are dominating in many data centers and cloud computing environments today. One way to improve inter Virtual Machine (VM) communication efficiency is to support co-resident VM communication by using shared memory based approaches and to resort to the traditional TCP/IP for inter-VM communications between VMs that are...

    Provided By Georgia Institute of Technology

  • White Papers // May 2013

    Performance Overhead Among Three Hypervisors: An Experimental Study using Hadoop Benchmarks

    Hypervisors are widely used in cloud environments and their impact on application performance has been a topic of significant research and practical interest. The authors conduct experimental measurements of several benchmarks using Hadoop MapReduce to evaluate and compare the performance impact of three popular hypervisors: a commercial hypervisor, Xen, and...

    Provided By Georgia Institute of Technology

  • White Papers // Apr 2013

    Residency Aware Inter-VM Communication in Virtualized Cloud: Performance Measurement and Analysis

    A known problem for virtualized cloud data centers is the inter-VM communication inefficiency for data transfer between co-resident VMs. Several engineering efforts have been made on building a shared memory based channel between co-resident VMs. The implementations differ in terms of whether user/program transparency, OS kernel transparency or VMM transparency...

    Provided By Georgia Institute of Technology

  • White Papers // Apr 2013

    SLIM: A Scalable Location-Sensitive Information Monitoring Service

    Location-sensitive information monitoring services are a centerpiece of the technology for disseminating content-rich information from massive data streams to mobile users. The key challenges for such monitoring services are characterized by the combination of spatial and non-spatial attributes being monitored and the wide spectrum of update rates. A typical example...

    Provided By Georgia Institute of Technology

  • White Papers // Apr 2013

    Measuring SSL Indicators on Mobile Browsers: Extended Life, or End of the Road?

    Mobile browsers are increasingly being relied upon to perform security sensitive operations. Like their desktop counterparts, these applications can enable SSL/TLS to provide strong security guarantees for communications over the web. However, the drastic reduction in screen size and the accompanying reorganization of screen real estate significantly changes the use...

    Provided By Georgia Institute of Technology

  • White Papers // Apr 2013

    Cloud-Based Information Technology Framework for Data Driven Intelligent Transportation Systems

    The authors present a novel cloud based IT framework, CloudTrack, for data driven intelligent transportation systems. They describe how the proposed framework can be leveraged for real-time fresh food supply tracking and monitoring. Cloud-Track allows efficient storage, processing and analysis of real-time location and sensor data collected from fresh food...

    Provided By Georgia Institute of Technology

  • White Papers // Mar 2013

    Optimizing Parallel Simulation of Multicore Systems Using Domain-Specific Knowledge

    In this paper the authors present two optimization techniques for the basic null-message algorithm in the context of parallel simulation of multicore computer architectures. Unlike the general, application-independent optimization methods, these are application-specific optimizations that make use of system properties of the simulation application. They demonstrate in two aspects that...

    Provided By Georgia Institute of Technology

  • White Papers // May 2008

    Double-DIP: Augmenting DIP with Adaptive Promotion Policies to Manage Shared L2 Caches

    In this paper, the authors study how the Dynamic Insert Policy (DIP) cache mechanism behaves in a multi-core shared-cache environment. Based on their observations, they explore a new direction in the design space of caches called the promotion policy. In a conventional LRU-based cache, a hit causes the line to...

    Provided By Georgia Institute of Technology

  • White Papers // Mar 2013

    Optimizing Parallel Simulation of Multicore Systems Using Domain-Specific Knowledge

    In this paper the authors present two optimization techniques for the basic null-message algorithm in the context of parallel simulation of multicore computer architectures. Unlike the general, application-independent optimization methods, these are application-specific optimizations that make use of system properties of the simulation application. They demonstrate in two aspects that...

    Provided By Georgia Institute of Technology

  • White Papers // Feb 2013

    Accelerating Simulation of Agent-Based Models on Heterogeneous Architectures

    The wide usage of GPGPU programming models and compiler techniques enables the optimization of data-parallel programs on commodity GPUs. However, mapping GPGPU applications running on discrete parts to emerging integrated heterogeneous architectures such as the AMD Fusion APU and Intel Sandy/Ivy Bridge with the CPU and the GPU on the...

    Provided By Georgia Institute of Technology

  • White Papers // Dec 2012

    Software Reliability Enhancements for GPU Applications

    As the role of highly-parallel accelerators becomes more important in high performance computing, so does the need to ensure their reliable operation. In applications where precision and correctness is a necessity, bit-level reliable operation is required. While there exist mechanisms for error detection and correction, the cost-effective implementation in massively...

    Provided By Georgia Institute of Technology

  • White Papers // Aug 2012

    Throughput Regulation in Multicore Processors via IPA

    In this paper the authors present an online controller for regulating the throughput of instruction-sequences in multicore processors using dynamic voltage-frequency scaling. The proposed control law comprises an integral controller whose gain is adjusted online based on the derivative of the frequency throughput relationship. This relationship is modeled as a...

    Provided By Georgia Institute of Technology

  • White Papers // Dec 2012

    Eiger: A Framework for the Automated Synthesis of Statistical Performance Models

    As processor architectures continue to evolve to increasingly heterogeneous and asymmetric designs, the construction of accurate performance models of execution time and energy consumption has become increasingly more challenging. Models that are constructed, are quickly invalidated by new features in the next generation of processors while many interactions between application...

    Provided By Georgia Institute of Technology

  • White Papers // Nov 2012

    Kernel Weaver: Automatically Fusing Database Primitives for Efficient GPU Computation

    Data warehousing applications represent an emerging application arena that requires the processing of relational queries and computations over massive amounts of data. Modern general purpose GPUs are high bandwidth architectures that potentially offer substantial improvements in throughput for these applications. However, there are significant challenges that arise due to the...

    Provided By Georgia Institute of Technology

  • White Papers // Feb 2013

    Instruction-Based Energy Estimation Methodology for Asymmetric Manycore Processor Simulations

    Processor power is a complex function of device, packaging, micro-architecture, and application. Typical approaches to power simulation require detailed micro-architecture models to collect the statistical switching activity counts of processor components. In many-core simulations, the detailed core models are the main simulation speed bottleneck. In this paper, the authors propose...

    Provided By Georgia Institute of Technology

  • White Papers // Jul 2010

    Dynamic Partitioned Global Address Spaces for Power Efficient DRAM Virtualization

    Dynamic Partitioned Global Address Spaces (DPGAS) is an abstraction that allows for quick and efficient remapping of physical memory addresses within a global address space, enabling more efficient sharing of remote DRAM. While past paper has proposed several uses for DPGAS, the most pressing issue in today's data centers is...

    Provided By Georgia Institute of Technology

  • White Papers // Jan 2010

    Exploring The Latency and Bandwidth Tolerance of CUDA Applications

    CUDA applications represent a new body of parallel programs. Although several paradigms exist for programming distributed systems and many-core processors, many users struggle to achieve a program that is scalable across systems with different hardware characteristics. This paper explores the scalability of CUDA applications on systems with varying interconnect latencies,...

    Provided By Georgia Institute of Technology

  • White Papers // Jan 2014

    Hardware Support for Real-Time Embedded Multiprocessor System-on-a-Chip Memory Management

    The aggressive evolution of the semiconductor industry smaller process geometries, higher densities, and greater chip complexity has provided design engineers the means to create complex, high-performance Systems-on-Chip (SoC) designs. Such, SoC designs typically have more than one processor and huge memory, all on the same chip. Dealing with the global...

    Provided By Georgia Institute of Technology

  • White Papers // Jan 2014

    Methodical Approximate Hardware Design and Reuse

    Design and reuse of approximate hardware components - digital circuits that may produce inaccurate results - can potentially lead to significant performance and energy improvements. Many emerging error-resilient applications can exploit such designs provided approximation is applied in a controlled manner. This paper provides the design abstractions and semantics for...

    Provided By Georgia Institute of Technology

  • White Papers // Mar 2014

    Software-Based Techniques for Reducing the Vulnerability of GPU Applications

    As highly-parallel accelerators such as graphics processing units become more important in high-performance computing, so does the need to ensure their reliable operation. In response, research has been directed at several efforts to characterize and understand the hardware vulnerability of GPU micro-architecture structures, as well as to detecting and correcting...

    Provided By Georgia Institute of Technology

  • White Papers // Feb 2014

    Manifold: A Parallel Simulation Framework for Multicore Systems

    In this paper the authors present manifold, an open-source parallel simulation framework for multi-core architectures. It consists of a parallel simulation kernel, a set of micro-architecture components, and an integrated library of power, thermal, reliability, and energy models. Using the components as building blocks, users can assemble multi-core architecture simulation...

    Provided By Georgia Institute of Technology

  • White Papers // Aug 2013

    Oncilla: A GAS Runtime for Efficient Resource Allocation and Data Movement in Accelerated Cluster

    Accelerated and in-core implementations of big data applications typically require large amounts of host and accelerator memory as well as efficient mechanisms for transferring data to and from accelerators in heterogeneous clusters. Scheduling for heterogeneous CPU and GPU clusters has been investigated in depth in the High-Performance Computing (HPC) and...

    Provided By Georgia Institute of Technology

  • White Papers // Jun 2013

    A Study of the Effect of Partitioning on Parallel Simulation of Multicore Systems

    There has been little research that studies the effect of partitioning on parallel simulation of multicore systems. This paper presents their paper of this important problem in the context of null-message-based synchronization algorithm for parallel multicore simulation. This paper focuses on coarse grain parallel simulation where each core and its...

    Provided By Georgia Institute of Technology

  • White Papers // Mar 2013

    CHiP: A Profiler to Measure the Effect of Cache Contention on Scalability

    Programmers are looking for ways to exploit the multi-core processors which have become commonplace today. One of the options available is to parallelize the existing serial programs using frameworks like OpenMP etc. However, such parallelization does not always yield the speedup expected by the programmer. This is due to various...

    Provided By Georgia Institute of Technology

  • White Papers // Mar 2013

    OpenCL Performance Evaluation on Modern Multi Core CPUs

    Utilizing heterogeneous platforms for computation has become a general trend making the portability issue important. OpenCL (Open Computing Language) serves the purpose by enabling portable execution on heterogeneous architectures. However, unpredictable performance variation on different platforms has become a burden for programmers who write OpenCL programs. This is especially true...

    Provided By Georgia Institute of Technology

  • White Papers // Jun 2012

    FLEXclusion: Balancing Cache Capacity and On-Chip Bandwidth Via Flexible Exclusion

    Exclusive Last-Level Caches (LLCs) reduce memory accesses by effectively utilizing cache capacity. However, they require excessive on-chip bandwidth to support frequent insertions of cache lines on eviction from upper-level caches. Non-inclusive caches, on the other hand, have the advantage of using the on-chip bandwidth more effectively but suffer from a...

    Provided By Georgia Institute of Technology

  • White Papers // Feb 2012

    Predicting Potential Speedup of Serial Code via Lightweight Profiling and Emulations with Memory Performance Model

    The authors present parallel prophet, which papers potential parallel speedup from an annotated serial program before actual parallelization. Programmers want to see how much speedup could be obtained prior to investing time and effort to write parallel code. With parallel prophet, programmers simply insert annotations that describe the parallel behavior...

    Provided By Georgia Institute of Technology

  • White Papers // Feb 2012

    TAP: A TLP-Aware Cache Management Policy for a CPU-GPU Heterogeneous Architecture

    Combining CPUs and GPUs on the same chip has become a popular architectural trend. However, these heterogeneous architectures put more pressure on shared resource management. In particular, managing the Last-Level Cache (LLC) is very critical to performance. Lately, many researchers have proposed several shared cache management mechanisms, including dynamic cache...

    Provided By Georgia Institute of Technology

  • White Papers // May 2013

    Variations in Performance Measurements of Multi-Core Processors: A Study of n-Tier Applications

    The prevalence of multi-core processors has raised the question of whether applications can use the increasing number of cores efficiently in order to provide predictable Quality of Service (QoS). In this paper, the authors study the horizontal scalability of n-tier application performance within a Multi-Core Processor (MCP). Through extensive measurements...

    Provided By Georgia Institute of Technology

  • White Papers // Feb 2013

    Brawny vs. Wimpy: Evaluation and Analysis of Modern Workloads on Heterogeneous Processors

    Heterogeneous processors, consisting of a mix of high-performance 'Brawny' processors and low-power 'Wimpy' processors, have been proposed to achieve higher energy-efficiency by making it possible for different applications within a diverse mix of workloads to be run on the 'Most appropriate' cores. This paper performs a comparative analysis of such...

    Provided By Georgia Institute of Technology

  • White Papers // Oct 2014

    CAMEO:A Two-Level Memory Organization with Capacity of Main Memory and Flexibility of Hardware-Managed Cache

    In this paper, the authors analyze the trade-offs in architecting stacked DRAM either as part of main memory or as a hardware-managed cache. Using stacked DRAM as part of main memory increases the effective capacity, but obtaining high performance from such a system requires Operating System (OS) support to migrate...

    Provided By Georgia Institute of Technology

  • White Papers // Dec 2009

    A Fill-Rate Service Level Model for Integrated Network Design and Inventory Allocation Problem

    Today's global economy relies heavily on transportation and warehousing to deliver goods. Outsourcing has made the distribution networks complex with several layers of National (import) Distribution Centers (NDCs) and Regional Distribution Centers (RDCs) between suppliers and retailers. More distribution nodes in the supply chain also means more inventory stockpiles. The...

    Provided By Georgia Institute of Technology

  • White Papers // May 2009

    Performance Aware Regeneration in Virtualized Multitier Applications

    Virtual machine technology enables highly agile system deployments in which components can be cheaply moved, cloned, and allocated controlled hardware resources. This paper examines in the context of multitier Enterprise applications, how these facilities can be used to provide enhanced solutions to the classic problem of ensuring high availability without...

    Provided By Georgia Institute of Technology

  • White Papers // Dec 2010

    A Stochastic Approach in Modeling Cooperative Line Networks

    The authors consider a quasi-stationary Markov chain as a model for a decode and forward wireless multi-hop cooperative transmission system that forms successive Opportunistic Large Arrays (OLAs). This paper treats a linear network topology, where the nodes form a one-dimensional horizontal grid with equal spacing. In this OLA approach, all...

    Provided By Georgia Institute of Technology

  • White Papers // Dec 2010

    Cooperative Transmission Range Extension for Duty Cycle-Limited Wireless Sensor Networks

    In this paper, the authors discuss how Cooperative Transmission (CT) can provide better services and/or lower the initial cost compared to non-CT in wireless sensor networks. To see the performance of CT and non-CT in both batteries operated and energy harvesting networks, they look at the duty cycle instead of...

    Provided By Georgia Institute of Technology

  • White Papers // Sep 2009

    SNR Estimation for a Non-Coherent M-FSK Receiver in a Rayleigh Fading Environment

    This paper deals with the problem of estimating average Signal-to-Noise Ratio (SNR) for a communication system employing Non-Coherent M-ARY Frequency Shift Keying (NCMFSK) over fading channels and White Gaussian Noise (AWGN). The authors derive two estimators; one using Maximum Likelihood (ML) approach and other using the data statistics. Various scenarios...

    Provided By Georgia Institute of Technology

  • White Papers // Dec 2009

    Cluster Transmission Time Synchronization for Cooperative Transmission Using Software Defined Radio

    The authors consider the time synchronization of concurrent transmissions from a cluster of cooperating Software Defined Radios (SDRs). In this type of Cooperative Transmission (CT), each cooperating node derives its timing autonomously from a signal received from a source or another cluster. Timing errors, caused by SDR processing delays, cause...

    Provided By Georgia Institute of Technology

  • White Papers // Jan 2010

    Residual-Energy-Activated Cooperative Transmission (REACT) to Avoid the Energy Hole

    In a multi-hop Wireless Sensor Network (WSN) with a constant node density, the nodes that are one hop away from the Sink die first and cause an "Energy hole," because they must forward the traffic from the rest of the network. When this hole forms, a large amount of "Excess"...

    Provided By Georgia Institute of Technology

  • White Papers // Jul 2009

    Multiple Repeater Placement for Assisting Long- Range LOS MIMO Links

    A wireless MIMO link is considered in a LOS configuration with multiple SISO wireless repeaters assisting. A theoretical analysis is presented in which repeater positions are considered and sufficient conditions proposed and evaluated for maximizing the MIMO multiplexing gain. The assisting repeaters would be useful when space constraints, for example,...

    Provided By Georgia Institute of Technology

  • White Papers // Jan 2011

    Energy-Aware Media Transcoding in Wireless Systems

    In distributed systems, transcoding techniques have been used to customize multimedia objects, utilizing trade-offs between the quality and sizes of these objects to provide differentiated services to clients. The authors' research uses transcoding techniques in wireless systems to customize video streams to the requirements of users, while minimizing the energy...

    Provided By Georgia Institute of Technology

  • White Papers // Sep 2009

    Evaluating Bluetooth as a Medium for Botnet Command and Control

    Malware targeting mobile phones is being studied with increasing interest by the research community. While such attention has previously focused on viruses and worms, many of which use near-field communications in order to propagate, none have investigated whether more complex malware such as botnets can effectively operate in this environment....

    Provided By Georgia Institute of Technology

  • White Papers // Aug 2009

    State Of The Art Survey Of Commercial Software For Supply Chain Design

    The excellence of supply chain management is vital not only to obtain high operational efficiency but also to survive in the global competition. In supporting decision makers to make rational and sophisticated plan with huge number of parameters and their complicated trade-off relations, supply chain design software has played a...

    Provided By Georgia Institute of Technology

  • White Papers // May 2010

    Look Who's Talking: Discovering Dependencies Between Virtual Machines Using CPU Utilization

    A common problem experienced in datacenters and utility clouds is the lack of knowledge about the mappings of the services being offered to or run by external users to the sets of Virtual Machines (VMs) realizing them. This makes it difficult to manage VM ensembles to attain provider goals like...

    Provided By Georgia Institute of Technology

  • White Papers // Aug 2010

    Subverting Linux On-the-Fly Using Hardware Virtualization Technology

    In this paper, the authors address the problem faced by modern operating systems due to the exploitation of Hardware-Assisted Full-Virtualization technology by attackers. Virtualization technology has been of growing importance these days. With the help of such a technology, multiple operating systems can be run on a single piece of...

    Provided By Georgia Institute of Technology

  • White Papers // Dec 2009

    Online Detection of Utility Cloud Anomalies Using Metric Distributions

    The online detection of anomalies is a vital element of operations in data centers and in utility clouds like Amazon EC2. Given ever-increasing data center sizes coupled with the complexities of systems software, applications, and workload patterns, such anomaly detection must operate automatically, at runtime, and without the need for...

    Provided By Georgia Institute of Technology

  • White Papers // Apr 2010

    Performance and Availability Aware Regeneration for Cloud Based Multitier Applications

    Virtual machine technology enables agile system deployments in which software components can be cheaply moved, replicated, and allocated hardware resources in a controlled fashion. This paper examines how these facilities can be used to provide enhanced solutions to the classic problem of ensuring high availability while maintaining performance. By regenerating...

    Provided By Georgia Institute of Technology

  • White Papers // Apr 2009

    Automated Staging Testing Framework for Amazon EC2 (Enhancing Elba in to EC2 With MySQL Cluster)

    Increasing complexity of large enterprise and distributed application systems makes the management of the system is an increasingly important and increasing expensive technical challenge. Staging comes as a handy tool when solving those important challenges. The key advantage in staging is, it provides system developers and system administrators to monitor...

    Provided By Georgia Institute of Technology