- Home
- Conference Proceedings
- Qatar Foundation Annual Research Forum Proceedings
- Conference Proceeding
Qatar Foundation Annual Research Forum Volume 2011 Issue 1
- Conference date: 20-22 Nov 2011
- Location: Qatar National Convention Center (QNCC), Doha, Qatar
- Volume number: 2011
- Published: 20 November 2011
161 - 180 of 281 results
-
-
Interference-Aware Spectrum Sharing Techniques for Next Generation Wireless Networks
Authors: Marwa Khalid Qaraqe, Ziad Bouida, Mohamed Abdallah and Mohamed-Slim AlouiniAbstractBackground: Reliable high-speed data communication that supports multimedia application for both indoor and outdoor mobile users is a fundamental requirement for next generation wireless networks and requires a dense deployment of physically coexisting network architectures. Due to the limited spectrum availability, a novel interference-aware spectrum-sharing concept is introduced where networks that suffer from congested spectrums (secondary-networks) are allowed to share the spectrum with other networks with available spectrum (primary-networks) under the condition that limited interference occurs to primary networks.
Objective: Multiple-antenna and adaptive rate can be utilized as a power-efficient technique for improving the data rate of the secondary link while satisfying the interference constraint of the primary link by allowing the secondary user to adapt its transmitting antenna, power, and rate according to the channel state information.
Methods: Two adaptive schemes are proposed using multiple-antenna transmit diversity and adaptive modulation in order to increase the spectral-efficiency of the secondary link while maintaining minimum interference with the primary. Both the switching efficient scheme (SES) and bandwidth efficient scheme (BES) use the scan-and-wait combining antenna technique (SWC) where there is a secondary transmission only when a branch with an acceptable performance is found; else the data is buffered.
Results: In both these schemes the constellation size and selected transmit branch are determined to minimized the average number of switches and achieve the highest spectral efficiency given a minimum bit-error-rate (BER), fading conditions, and peak interference constraint. For delayed sensitive applications, two schemes using power control are used: SES-PC and BES-PC. In these schemes the secondary transmitter sends data using a nominal power level, which is optimized to minimize the average delay. Several numerical examples show that the BES scheme increases the capacity of the secondary link.
Conclusion: The SES and BES schemes reach high spectral efficiency and BER performance at the expense of an increased delay. The SES-PC and BESPC minimize the average delay, satisfy the BER, and maintain a high spectral efficiency. The proposed power optimization and power control processes minimize the delay and the dropping probability especially if we extend the presented work to a multiuser scenario.
-
-
-
Conceptual Weighted Feature Extraction and Support Vector Model: A Good Combination for Text Categorization
Authors: Ali Mohamed Jaoua, Sheikha Ali Karam, Samir Elloumi and Fethi FerjaniAbstractWhile weighted features are known in information retrieval (IR) systems to be used for increasing recall during the document selection step, conceptual methods helped for finding good features. Starting from the features of a sample of Arabic news belonging to k different financial categories, and using the support vector model (SVM), k(k-1) classifiers are generated using one-against-one classification. A new document is submitted to k(k-1) different classifiers then by using the voting heuristic, is assigned to the most selected category. Categorization results obtained for two different methods for feature extraction: one based on the optimal concepts and the other based on isolated labels, proved that isolated labels generate better feature, because of the specificity of the selected features. Therefore, we can say that the quality of the feature, added to weighting methods, using SVM is an important factor for a more accurate classification. The proposed method based on isolated labels gives a good classification rate of Arabic news greater than 80% in the financial domain for five categories. Generalized to English Texts and to more categories, it becomes a good preprocessing filtering preceding automatic annotation step, and therefore helps for more accurate event structuring. Here attached a figure showing the different steps of the new categorization method.
-
-
-
Record Linkage and Fusion over Web Databases
Authors: Mourad Ouzzani, Eduard Dragut, El Kindi and Amgad MadkourAbstractMany data-intensive applications on the Web require integrating data from multiple sources (Web databases) at query time. Online sources may refer to the same real world entity in different ways and some may provide outdated or erroneous data. An important task is to recognize and merge the various references that refer to the same entity at query time. Almost all existing duplicate detection and fusion techniques work in the offline setting and, thus, do not meet the online constraint. There are at least two aspects that differentiate online duplicate detection and fusion from its offline counterpart. First, the latter assumes that the entire data is available, while the former cannot make such a hard assumption. Second, several iterations (query submissions) may be required to compute the “ideal” representation of an entity in the online setting.
We propose a general framework to address this problem: an interactive caching solution. A set of frequently requested records is cleaned off-line and cached for future references. Newly arriving records in response to a stream of queries are cleaned jointly with the records in the cache, presented to users and appended to the cache.
We introduce two online record linkage and fusion approaches: (i) a record-based and (ii) a graph-based. They chiefly differ in the way they organize data in the cache as well as computationally. We conduct a comprehensive empirical study of the two techniques with real data from the Web. We couple their analysis with commonly used cache settings: static/dynamic, cache size and eviction policies.
-
-
-
Call Admission Control with Resource Reservation for OFDM Networks
Authors: Mehdi Khabazian, Osama Kubbar and Hossam HassaneinAbstractThe scarcity of the radio resources and variable channel quality cause many challenges to the resource management for future all-IP wireless communications. One technique to guarantee a certain level of quality of service (QoS) is call admission control (CAC). Briefly, CAC is a mechanism which decides whether a new call could be admitted or rejected depending on its impacts on the current calls’ QoS. Conventional CACs such as guard channel, channel borrowing and queuing priority techniques only consider the instantaneous radio resource availabilities to make a decision on admission problem, thus they are neither able to prevent the network congestion problem nor meet the QoS requirements of different users with multi-service requirements.
In this work, we propose a new CAC technique with a future look into the needed extra resources through a reservation technique to offset the changes of the channel condition due to mobility. We show that during a call session, the needed radio resources may be increased compared with the negotiated resources during call setup. Although such fluctuations are fairly low for a single call, it is not negligible when the network is congested. As a result, some ongoing calls may experience QoS degradation. We show that such a consideration is critical in orthogonal frequency division multiplexing (OFDM) wireless networks such as 3GPP LTE where the radio resources are assigned to the users depending on the channel quality. The study assumes two types of applications denoted by wide-band and narrow-band and the performance of the proposed algorithm is modeled through queuing theory and event-driven simulation approaches. The results show that such a reservation technique improves the call admission performance significantly in terms of call blocking, call drop and call QoS degradation probabilities, and it outperforms the conventional CACs with insignificant loss in network capacity.
-
-
-
A Distributed Reconfigurable Active Solid State Drive Platform for Data Intensive Applications
Authors: Mazen Saghir, Hassan Artail, Haitham Akkary, Hazem Hajj and Mariette AwadAbstractThe ability to efficiently extract useful information from volumes of data distributed across myriad networks is hindered by the latencies inherent to magnetic storage devices and computer networks. We propose overcoming these limitations by leveraging solid-state drive (SSD) and field-programmable gate array (FPGA) technologies to process large streams of data directly at the storage sites.
Our proposed reconfigurable, active, solid-state drive (RASSD) platform consists of distributed nodes that couple SSDs with FPGAs. While SSDs store data, FPGAs implement processing elements that couple soft-core RISC processors with dynamically reconfigurable logic resources. The processors execute data-processing software drivelets, and the logic resources implement hardware for accelerating performance-critical operations. Executing appropriate drivelets and using matching hardware accelerators enables us to efficiently process streams of data stored across SSDs.
To manage the configuration of RASSD nodes and provide a transparent interface to applications, our platform also consists of distributed middleware software. Client local middleware (CLM) resides on client host machines to interpret application data processing requests, locate storage sites, and exchange data-processing requests and results with middleware servers (MWS). MWS connect to clusters of RASSD nodes and contain libraries of drivelets and accelerator configuration bit streams. An MWS loads appropriate drivelet and accelerator bit streams onto a RASSD node's FPGA, aggregates processed data, and returns it to a CLM.
To evaluate our platform, we implemented a simple system consisting of a host computer connected to a RASSD node over a peer-to-peer network. We ran a keyword search application on the host computer, which also provided middleware functionality. We then evaluated this platform under three configurations. In configuration C1, the RASSD node was only used to store data while all data was processed by the MWS running on the host computer. In configuration C2, the data was processed by a drivelet running on the RASSD node. Finally, in configuration C3, the data was processed both by a drivelet and a hardware accelerator.
Our experimental results show that C3 is 2x faster than C2, and 6x faster than C1. This demonstrates our platform's potential for enhancing the performance of data-intensive applications over current systems.
-
-
-
A Dynamic Physical Rate Adaptation for Multimedia Quality-Based Communications in IEEE_802.11 Wireless Networks
By Mariam FlissAbstractAssuming that the IEEE802.11 Wireless Local Area Networks (WLANs) are based on a radio/infrared link, they are more sensitive to the channel variations and connection ruptures. Therefore the support for multimedia applications over WLANs becomes non-convenient due to the compliance failure in term of link rate and transmission delay performance. We studied link adaptation facets and the Quality of Service (QoS) requirements essential for successful multimedia transmissions. In fact, the efficiency of rate control diagrams is linked to the fast response for channel variation. The 802.11 physical layers provide multiple transmission rates (different modulation and coding schemes). The last 802.11g-version maintains 12 physical rates up to 54 Mbps at the 2.4 GHz band. As a result, Mobile Stations (MSs) are able to select the appropriate link rate depending on the required QoS and instantaneous channel conditions to enhance the overall system performance. Hence, the implemented link adaptation algorithm symbolizes a vital fraction to achieve highest transmission capability in WLANs. “When to decrease and when to increase the transmission rate?” Are the two fundamental matters that we will be faced when designing a new physical-rate control mechanism? Many research works focus on tuning channel estimation schemes to better detect when the channel condition was improved enough to accommodate a higher rate, and then adapts its transmission rate accordingly. However, those techniques usually entail modifications on the current 802.11 standard. Another way to perform link control is based on local Acknowledgment (Ack) information for the transmitter station. Consequently, two techniques where accepted by the standard due to their efficiency and implementation simplicity.
We propose a new dynamic time-based link adaptation mechanism, called MAARF (Modified Adaptive Auto Rate Fallback). Beside the transmission frame results, the new model implements a Round Trip Time (RTT) technique to select adequately an instantaneous link rate. This proposed model is evaluated with most recent techniques adopted by the IEEE 802.11 standard: ARF (Auto Rate Fallback) and AARF (Adaptive ARF) schemes. Simulation results will be given to illustrate the link quality improvement of multimedia transmissions over Wi-Fi networks and to compare its performance with previous published results.
-
-
-
Interference Cancellation of Hop-By-Hop Beamforing for Dual-Hop MIMO Relay Networks
Authors: Fawaz AL-Qahtani and Hussein AlnuweiriAbstractCooperative communication relaying systems are gaining much interest because they can improve average link signal to nose ratio by replacing longer hops with multiple shorter hops. The method of relaying has been introduced to enable a source (i.e. mobile terminal) communicate with a target destination via a relaying (i.e. mobile terminal). Furthermore, multiple-input multiple output (MIMO) communication systems have been considered as powerful candidates for the fourth generation of wireless communication standards because they can achieve further performance improvements including an increase in the achievable spectral efficiency and the peak data rates (Multiplexing), and robustness against severe effects of fading (transmit beamforming). In this work, we consider hop-by-hop beamforming relaying system over Rayleigh fading channels. In wireless communication environments, it is well known understood that the performance of wireless networks can be limited by both fading and co-channel interference (CCI). In this work, the multiple antennas at each node of relaying systems can be used to adaptively modify the radiation pattern of the array to reduce the interference by placing nulls in the direction of the dominant interferers. In this paper, we investigate the effect of CCI on the performance of hop-by-hop beamforming relaying system. First, we derive exact closed-form expressions for the outage probability and average symbol error rates. Moreover, we look into the high signal to noise ratio (SNR) regime and study the diversity order and coding gain achieved by the system.
-
-
-
Novel Reduced-Feedback Wireless Communication Systems
Authors: Mohammad Obaidah Shaqfeh, Hussein Alnuweiri and Mohamed-Slim AlouiniAbstractModern communication systems apply channel-aware adaptive transmission techniques and dynamic resource allocation in order to exploit the peak conditions of the fading wireless links and to enable significant performance gains. However, conveying the channel state information among the users’ mobile terminals into the access points of the network consumes a significant portion of the scarce air-link resources and depletes the battery resources of the mobile terminals rapidly. Despite its evident drawbacks, the channel information feedback cannot be eliminated in modern wireless networks because blind communication technologies cannot support the ever-increasing transmission rates and high quality of experience demands of current ubiquitous services.
Developing new transmission technologies with reduced-feedback requirements is sought. Network operators will benefit from releasing the bandwidth resources reserved for the feedback communications and the clients will enjoy the extended battery life of their mobile devices. The main technical challenge is to preserve the prospected transmission rates over the network despite decreasing the channel information feedback significantly. This is a noteworthy research theme especially that there is no mature theory for feedback communication in the existing literature despite the growing number of publications about the topic in the last few years. More research efforts are needed to characterize the trade-off between the achievable rate and the required channel information and to design new reduced-feedback schemes that can be flexibly controlled based on the operator preferences. Such schemes can be then introduced into the standardization bodies for consideration in next generation broadband systems.
We have recently contributed to this field and published several journal and conference papers. We are the pioneers to propose a novel reduced-feedback opportunistic scheduling scheme that combines many desired features including fairness in resources distribution across the active terminals and distributed processing at the MAC layer level. In addition our scheme operates close to the upper capacity limits of achievable transmission rates over wireless links. We have also proposed another hybrid scheme that enables adjusting the feedback load flexibly based on rates requirements. We are currently investigating other novel ideas to design reduced-feedback communication systems.
-
-
-
Learning to Recognize Speech from a Small Number of Labeled Examples
AbstractMachine learning methods can be used to train automatic speech recognizers (ASR). When porting ASR to a new language, however, or to a new dialect of spoken Arabic, we often have too few labeled training data to allow learning of a high-precision ASR. It seems reasonable to think that unlabeled data, e.g., untranscribed television broadcasts, should be useful to train the ASR; human infants, for example, are able to learn the distinction between phonologically similar words after just one labeled training utterance. Unlabeled data tell us the marginal distribution of speech sounds, p(x), but do not tell us the association between labels and sounds, p(y|x). We propose that knowing the marginal is sufficient to rank-order all possible phoneme classification functions, before the learner has heard any labeled training examples at all. Knowing the marginal, the learner is able to compute the expected complexity (e.g., derivative of the expected log covering number) of every possible classifier function, and based on measures of complexity, it is possible to compute the expected mean-squared probable difference between training-corpus error and test-corpus error. Upon presentation of the first few labeled training examples, then, the learner simply chooses, from the rank-ordered list of possible phoneme classifiers, the first one that is reasonably compatible with the labeled examples. This talk will present formal proofs, experimental tests using stripped-down toy problems, and experimental results from English-language ASR; future work will test larger-scale implementations for ASR in the spoken dialects of Arabic.
-
-
-
Demonstration Prototype of a High-Fidelity Robotic-Assisted Suturing Simulator
Authors: Georges Younes, George Turkiyyah and Jullien Abi NahedAbstractRapid advances in robotic surgical devices have put significant pressure on physicians to learn new procedures using newer and sophisticated instruments. This in turn has increased the demand for effective and practical training methods using these technologies, and has motivated the development of surgical simulators that promise to provide practical, safe, and cost-effective environments for practicing demanding robotic-assisted procedures.
However, despite the significant interest and effort in the development of such simulators, the current state-of-art surgical simulators are lacking. They introduce significant simplifications to obtain real-time performance, and these simplifications often come at the expense of realism and fidelity. There is a need to develop and build the next-generation of surgical simulators that improve haptic and visual realism. The primary challenges for building such high-fidelity simulations for soft-tissue organ simulations come from two computationally demanding tasks that must execute in real time: managing the complexity of the geometric environment that is being dynamically modified during the procedure; and modeling the stresses and deformations of the soft tissue interacting with surgical instruments and subjected to cutting and suturing. The mechanics of soft-tissue behavior are complicated by their anisotropic and nonlinear behavior.
In this presentation, we describe an initial prototype of a robotic-assisted simulator applied to a simplified task in a prostatectomy procedure (anastomosis). The simulator demonstrates new methodologies for modeling nonlinear tissue models, integrated high-resolution geometric contact detection for handling inter- and intra-organ collisions in the dynamically changing geometric environment of the simulation, and suturing with threads. The prototype is deployed on a bi-manual haptic feedback frame and serves as a building block for simulations operating in more complex anatomical structures.
-
-
-
Investigating the Dynamics of Densely Crowded Environments at the Hajj Using Image Processing Techniques
Authors: Khurom Hussain Kiyani and Maria PetrouAbstractBackground: With the world's population projected to grow from the current 6.8 billion to around 9 billion by 2050, the resultant increase of megacities and the associated demands on public transport, there is an urgent imperative to understand the dynamics of crowded environments. Very dense crowds that exceed 3 people per square metre present many challenges for efficiently measuring quantities such as density and pedestrian trajectories. The Hajj and the associated minor Muslim pilgrimage of Umrah, present some of the most densely crowded human environments in the world, and thus present an excellent observational laboratory for the study of dense crowd dynamics. An accurate characterisation of such dense crowds cannot only improve existing models, but can also help to develop better intervention strategies for mitigating crowd disasters such as the Hajj 2006 Jamarat stampede that killed over 300 pilgrims. With Qatar set to be one of the cultural centres in the region, e.g. FIFA World Cup 2022, the proper control and management of large singular events is important for not only our safety but also our standing in the international stage.
Objectives: To use the data gathered from the Hajj to assess the dynamics of large dense crowds with a particular focus on crowd instabilities and pattern formation.
Methods: We will make use of advanced image processing and pattern recognition techniques (mathematical morphology, feature selection etc.) in assessing some of the bulk properties of crowds such as density and flow, as well as the finer details such as the ensemble of pedestrian trajectories. We are currently in the process of taking multiple wide-angle stereo videos at this year's Hajj, with our collaborators in Umm Al-Qurra University in Mecca. Multiple video capture of the same scene from different angles allows one to overcome the problem of occlusion in dense crowds.
Results: We will present our field study in the Hajj this year, where we took extensive high quality multiple camera video data. We will also present some of the techniques, which we will be using over the coming year in analyzing this large data set that we have now successfully collated.
-
-
-
Software for Biomechanical Performance Analysis of Force Plate Data
Authors: Manaf Kamil and Dino PalazziAbstractBackground: Force plates have been used in sports biomechanics since the 1960's. However, extracting useful information related to performance from curve analysis is a complicated process. It requires a combined knowledge in signal processing (filtering), mechanical equations, testing protocols, biomechanics, and discrete mathematical analysis to properly process the data.
Objectives: The aim is to provide a practical and accurate tool to analyze force curves from select standard biomechanical performance tests (e.g., counter movement jump, drop jump).
Methods The software is a tool built using Microsoft.Net framework. Key features of the software include:
* Real-time data acquisition module able to acquire data from third-party 8 channel 3D force plates with real-time results for immediate feedback during tests.
* Digital filtering module where the signal is treated for best fit for the analysis.
* Analysis module able to calculate Force, Power, Velocities, Trajectories, Mechanical Impulse and Timing during the different phases of the tests using discrete analysis algorithms.
* Reporting module for plotting and exporting selected variables.
Results The software has been used by ASPIRE Academy Sport Scientists in performance assessments of professional and semi-professional athletes from Qatar and other countries.
Currently, the software can analyze Counter Movement Jump, Drop Jump, Isometric Pulls and Squat Jump.
It contains automatic algorithms to detect specific points for each test type, but allows the user to change these suggestions when needed. Feedback is immediate, in both graphical and numerical form.
Conclusions: This novel software has proven to be a useful tool for immediate and accurate analysis and reporting of select field and lab based biomechanical tests. Going forward, further feedback from the applied users can lead to more features added. Considering the architecture of the software, adding more analysis modules is relatively simple. For example work is currently underway on a sprint running analysis module.
-
-
-
Autonomous Coverage Management in Smart Camera Networks
Authors: Vinay Kolar, Vikram Munishwar, Nael Abu-Ghazaleh and Khaled HarrasAbstractRecent advances in imaging and communication have lead to the development of smart cameras that can operate autonomously and collaboratively to meet various application requirements. Networks of such cameras, called Smart Camera Networks (SCNs), have a range of applications in areas such as monitoring and surveillance, traffic management and health care. The self-configuring nature of cameras, by adjusting their pan, tilt and zoom (PTZ) settings, coupled with wireless connectivity differentiate them substantially from classical multi-camera surveillance networks.
One of the important problems in SCNs is: “How to configure cameras to obtain the best possible coverage of events happening within the area of interest?” As the scale of cameras grows from tens to hundreds of cameras, it is impractical to rely on humans to configure cameras to best track areas of interest. Thus, supporting autonomous configuration of cameras to maximize their collective coverage is a critical requirement in SCNs.
Our research first focuses on a simplified version of the problem, where the field-of-view (FoV) of a camera can be adjusted only by adjusting its pan in discrete manner. Even with such simplifications, solving the problem optimally is NP-hard. Thus, we propose centralized, distributed and semi-centralized heuristics that outperform the state-of-the-art approaches. Furthermore, the semi-centralized approach provides coverage accuracy close to the optimal, while reducing the communication latency by 97% and 74% compared to the centralized and distributed approaches, respectively.
Next, we consider the problem without FoV constraints; we allow FoVs to be adjusted in PTZ dimensions in continuous manner. While, PTZ configurations significantly increase the coverable area, the continuous adjustment nature eliminates any sub-optimality resulting from the discrete settings. However, supporting these features typically results in generating extremely large number of feasible FoVs per camera, out of which only one optimal FoV will be selected. We show that the problem of finding minimum feasible FoVs per camera is NP-hard. However, due to the geometric constraints introduced by the camera's FoV, the problem can be solved in polynomial time. Our proposed algorithm has polynomial-time worst-case complexity of O(n3).
-
-
-
Query Processing in Private Data Outsourcing Using Anonymization
Authors: Ahmet Erhan Nergiz and Chris CliftonAbstractData outsourcing is a growing business. Cloud computing developments such as Amazon Relational Database Service promise further reduced cost. However, use of such a service can be constrained by privacy laws, requiring specialized service agreements and data protection that could reduce economies of scale and dramatically increase costs.
We propose a private data outsourcing approach where the link between identifying information and sensitive (protected) information is encrypted, with the ability to decrypt this link residing only with the client. As the server no longer has access to individually identifiable protected information, it is not subject to privacy laws, and can offer a service that does not need to be customized to the needs of each country- or sector-specific requirements; any risk of violating privacy through releasing sensitive information tied to an individual remains with the client. The data model used in this work is shown with an example in Figure 1 .
This work presents a relational query processor operating within this model. The goal is to minimize communication and client-side computation, while ensuring that the privacy constraints captured in the anatomization are maintained. At first glance, this is straightforward: standard relational query processing at the server, except that any joins involving the encrypted key must be done at the client; an appropriate distributed query optimizer should do a reasonably good job of this. However, two issues arise that confound this simple approach:
1. By making use of the anatomy groups, and the knowledge that there is a one-to-one mapping (unknown to the server) between tuples in such groups, we can perform portions of the join between identifying and sensitive information at the server without violating privacy constraints, and
2. Performing joins at the client and sending results back to the server for further processing can violate privacy constraints.
-
-
-
GreenLoc: Energy Efficient Wifi-Based Indoor Localization
Authors: Mohamed Abdellatif, Khaled Harras and Abderrahmen MtibaaAbstractUser-localization and positioning systems have been a core challenge in the domain of context-aware pervasive systems and applications. GPS has been the de-facto standard for outdoor localization; however, geo-satellite signals upon which GPS rely, are inaccurate in indoor environments. Therefore, various indoor localization techniques based on triangulation, scene analysis, or proximity, have been introduced. The most prominent technologies over which these techniques are applied include WiFi, Bluetooth, RFID, Infrared, and UWB. Due to the ubiquitous deployment of access points, WiFi-based localization via triangulation has emerged to be among the most prominent indoor positioning solutions. A major deployment obstacle for such systems, however, is the high-energy consumption rates of Wifi adapters in mobile devices where energy is the most valuable resource.
We propose GreenLoc, an indoor green localization system that exploits sensors prevalent in today's smart-phones in order to dynamically adapt the frequency of location updates required. Significant energy gains can, therefore, be acquired when users are not mobile. For example, accelerometers can aid in detecting different user states such as walking, running or stopping. Based on these states, mobile devices can dynamically decide upon the appropriate update frequency. We accommodate various motion speeds by estimating the velocity of the device using the latest two location coordinates, and the time interval between these two-recorded locations. We have taken the first steps towards implementing GreenLoc, based on the infamous Ekahau system. We have also conducted preliminary tests utilizing the accelerometer, gravity, gyroscope, and light sensors residing on the HTC Nexus One and IPhone4 smart-phones.
To further save energy in typical indoor environments, such as malls, schools, and airports, GreenLoc exploits people's proximity when moving in groups. Devices within short-range of each other do not necessarily require that they each be individually tracked. Therefore, GreenLoc detects and clusters users moving together and elects a reference node (RN) based on device energy levels and needs. The elected RN will then be tracked via triangulation while other nodes in the group will be tracked based on the RN's location using Bluetooth. Our initial analysis demonstrates very promising results with this system.
-
-
-
A Data Locality and Skew Aware Task Scheduler for MapReduce in Cloud Computing
Authors: Mohammad Hammoud, Suhail Rehman and Majd SakrAbstractInspired by the success and the increasing prevalence of MapReduce, this work proposes a novel MapReduce task scheduler. MapReduce is by far one of the most successful realizations of large-scale, data-intensive, cloud computing platforms. As compared to traditional programming models, MapReduce automatically and efficiently parallelizes computation by running multiple Map and/or Reduce tasks over distributed data across multiple machines. Hadoop, an open source implementation of MapReduce, schedules Map tasks in the vicinity of their input splits seeking diminished network traffic. However, when Hadoop schedules Reduce tasks, it neither exploits data locality nor addresses data partitioning skew inherent in many MapReduce applications. Consequently, MapReduce experiences a performance penalty and network congestion as observed in our experimental results.
Recently there has been some work concerned with leveraging data locality in Reduce task scheduling. For instance, one study suggests a locality-aware mechanism that inspects Map inputs and predicts corresponding consuming reducers. The input splits are subsequently assigned to Map tasks near the future reducers. While such a scheme addresses the problem, it targets mainly public-resource grids and doesn't fully substantiate the accuracy of the suggested prediction process. In this study, we propose Locality-Aware Skew-Aware Reduce Task Scheduler (LASAR), a practical strategy for improving MapReduce performance in clouds. LASAR attempts to schedule each reducer at its center-of-gravity node. It controllably avoids scheduling skew, a situation where some nodes receive more reducers than others, and promotes effective pseudo-asynchronous Map and Reduce phases resulting in earlier completion of submitted jobs, diminished network traffic, and better cluster utilization.
vWe implemented LASAR in Hadoop-0.20.2 and conducted extensive experimentations to evaluate its potential. We found that it outperforms current Hadoop by 11%, and by up to 26% for the utilized benchmarks. We believe LASAR is applicable to several cloud computing environments and multiple essential applications, including but not limited to shared environments and scientific applications. In fact, a large body of work observed partitioning skew in many of critical scientific applications. LASAR paves the way for these applications, and others, to get effectively ported to various clouds.
-
-
-
Multi-Layered Performance Monitoring for Cloud Computing Applications
Authors: Suhail Rehman, Mohammad Hammoud and Majd SakrAbstractCloud computing revolutionizes the way large amounts of data are processed and offers a compelling paradigm to organizations. An increasing number of data-intensive scientific applications are being ported to cloud environments such as virtualized clusters, in order to take advantage of increased cost-efficiency, flexibility, scalability, improved hardware utilization and reduced carbon footprint, among others.
However, due to the complexity of the application execution environment, routine tasks such as monitoring, performance analysis and debugging of applications deployed on the cloud become cumbersome and complex. These tasks often require close interaction and inspection of multiple layers in the application and system software stack. For example, when analyzing a distributed application that has been provisioned on a cluster of virtual machines, a researcher might need to monitor the execution of his program on the VMs, or the availability of physical resources to the VMs. This would require the researcher to use different sets of tools to collect and analyze performance data from each level.
Otus is a tool that enables resource attribution in clusters and currently reports only the virtual resource utilization and not the physical resource utilization on virtualized clusters. This is insufficient to fully understand application behavior on a cloud platform; it would fail to account for the state of the physical infrastructure, its availability or the variation in load by other VMs on the same physical host, for example.
We are extending Otus to collect metrics from multiple layers, starting with the Hypervisor. Otus can now collect information from both the VM level as well as the Hypervisor level; and this information is collected in an OpenTSDB database, which is scalable to large clusters. A web-based application allows the researcher to selectively visualize these metrics in real-time or for a particular time range in the past.
We have tested our multi-layered monitoring technique on several Hadoop Mapreduce applications and clearly identified the causes of several performance problems that would otherwise not be clear using existing methods. Apart from helping researchers understand application needs, our technique could also help accelerate the development and testing of new platforms for cloud researchers.
-
-
-
Towards a New Termination Checker for the Coq Proof Assistant
More LessAbstractModern societies rely on software applications for performing many critical tasks. As this trend is increasing, so it is the necessity to develop cost-effective methods of writing software that ensure that essential safety and security requirements are met. In this context, dependent type theories are recently gaining adoption as a valid tool for performing formal verification of software.
The focus of this work is Coq, a proof assistant based on a dependent type theory called the Calculus of Inductive Constructions. Developed at INRIA (France) for over 20 years, it is arguably one of the most successful proof assistant to this date. It has been used in several real-world large-scale projects such as the formalization of a verification framework for the Java Virtual Machine, a proof of the Four Color Theorem, and a formally verified compiler for the C programming language (project Compcert).
Coq is both a proof assistant and a programming language. To ensure soundness of the formal verification approach, Coq imposes several conditions on the source programs. In particular, all programs written in Coq must be terminating. The current implementation of the termination checker uses syntactic criteria that are too restrictive and limiting in practice, hindering the usability of the system.
In previous work we have proposed an extension of Coq using a termination checker based on the theory of sized types and we have shown that the soundness of the approach. Furthermore, compared to syntactic criteria currently used, our approach is more powerful, easier to understand, and easier to implement, as evidenced by a prototype implementation we developed.
Our objective is to turn our prototype into an implementation of a new core theory and termination checker for Coq. We expect that the resulting system will be more efficient and easier to understand for users. Furthermore it will increase the expressive power and usability of Coq, permitting the use of formal verification on a wider range of applications.
-
-
-
A Natural Language Processing-Based Active and Interactive Platform for Accessing English Language Content and Advanced Language Learning
Authors: Kemal Oflazer, Teruko Mitamura, Tomas By, Hideki Shima and Eric RieblingAbstractSmartReader is a general-purpose “reading appliance” being implemented at Carnegie Mellon University (Qatar and Pittsburgh) - building upon an earlier prototype version. It is an artificial intelligence system that employs advanced language processing technologies and can interact with the reader and respond to queries about the content, words and sentences in a text. We expect it to be used by students in Qatar and elsewhere to help improve their comprehension of English text. SmartReader is motivated by the observation that text is still the predominant medium for learning especially at the advanced level and that text, being ``bland’’, is hardly a conducive and motivating medium for learning, especially when one does not have access to tools that enable one get over language roadblocks, ranging from unknown words to unrecognized and forgotten names, to hard-to-understand sentences. SmartReader strives to make reading (English) textual material, an “active” and an “interactive” process with the user interacting with the text using anytime-anywhere contextually-guided query mechanism based-on contextual user intent recognition. With SmartReader, a user can -inquire about the contextually correct meaning or synonyms of a word or idiomatic and multi-word constructions, -select a person's name, and then get an immediate ``flashback’’ to the first (or the last) time the person was encountered in text to remind herself the details of the person, -extract a summary of a section to remember important aspects of the content at the point she left off, and continue reading with a significantly refreshed context, -select a sentence that she may not be able to understand fully and ask SmartReader to break it down, simplify or paraphrase to comprehend it better. -test her comprehension of the text in a page or a chapter, by asking SmartReader to dynamically generate quizzes and answering them. -ask questions about the content of the text and get answers in addition to many other functions. SmartReader is being implemented as a multi-platform (tablet/PC) client-server system using HTML5 technology, with Unstructured Information Management Architecture -UIMA technology (used recently in IBM's Watson Q/A system in the Jeopardy Challenge) as the underlying language processing framework.
-
-
-
NEXCEL, A Deductive Spreadsheet
More LessAbstractUsability and usefulness have made the spreadsheet one of the most successful computing applications of all times: millions rely on it every day for anything from typing grocery lists to developing multimillion dollar budgets. One thing spreadsheets are not very good at is manipulating symbolic data and helping users make decisions based on them. By tapping into recent research in logic programming, databases and cognitive psychology, we propose a deductive extension to the spreadsheet paradigm which addresses precisely this issue. The accompanying tool, which we call NEXCEL, is intended as an automated assistant for the daily reasoning and decision-making needs of computer users, in the same way as a spreadsheet application such as Microsoft Excel assists them every day with calculations simple and complex. Users without formal training in logic or computer science can interactively define logical rules in the same simple way as they define formulas in Excel. NEXCEL immediately evaluates these rules thereby returning lists of values that satisfy them, again just like with numerical formulas. The deductive component is seamlessly integrated into the traditional spreadsheet so that a user not only still has access to the usual functionalities, but is able to use them as part of the logical inference and, additionally, to embed deductive steps in a numerical calculation.
Under the hood, NEXCEL uses a small logic programming language inspired by Datalog to define derived relations: the target region of the spreadsheet contains a set of logical clauses in the same way that calculated cells contain a numerical formula in a traditional spreadsheet. Therefore, logical reasoning reduces to computing tables of data by evaluating Datalog-like definitions, a process that parallels the calculation of numerical formulas. Each row in the calculated relation is a tuple of values satisfying the definition for this relation, so that the evaluated table lists all such solutions, without repetitions. This linguistic extension significantly enriches the expressive power of the spreadsheet paradigm. Yet, it is provided to the user through a natural extension of the mostly excellent interfaces of modern spreadsheet applications.
-