- Home
- Conference Proceedings
- Qatar Foundation Annual Research Conference Proceedings
- Conference Proceeding
Qatar Foundation Annual Research Conference Proceedings Volume 2014 Issue 1
- Conference date: 18-19 Nov 2014
- Location: Qatar National Convention Center (QNCC), Doha, Qatar
- Volume number: 2014
- Published: 18 November 2014
351 - 400 of 480 results
-
-
Modelling The Power Produced By Photovoltaic Systems
Authors: Fotis Mavromatakis, Yannis Franghiadakis and Frank VignolaThe development and improvement of a model that can provide accurate estimates of the power produced by a photovoltaic system is useful for several reasons. A reliable model contributes to the proper operation of a photovoltaic power system since any deviations between modeled and experimental power can be flagged and studied for possible problems that can be identified and addressed. It is also useful to grid operators to know hours or a day ahead the contribution from different PV systems or renewable energy systems in general. In this way, they will be able to manage and balance production and demand. The model was designed to use the smallest number of free parameters. Apart from the incoming irradiance and module temperature, the model takes into account the effects introduced by the instantaneous angle of incidence and the air mass. The air mass is related to the position of the sun during its apparent motion across the sky since light travels through an increasing amount of atmosphere as the sun gets lower in the sky. In addition, the model takes into account the reduction in efficiency at low solar irradiance conditions. The model is versatile and can incorporate a fixed or variable percentage for the losses due to the deviation of MPPT tracking from ideal, the losses due to the mismatch of the modules, soiling, aging, wiring losses and the deviation from the nameplate rating. Angle of incidence effects were studied experimentally around solar noon by rotating the PV module at predetermined positions and recording all necessary variables (beam & global irradiances, module temperature and short circuit current, sun and module coordinates). Air mass effects were studied from sunrise to solar noon with the PV module always normal to the solar rays (global irradiance, temperature and short circuit current were recorded). Stainless steel meshes were used to artificially reduce the level of the incoming solar irradiance. A pyranometer and a reference cell were placed behind the mesh, while the unobstructed solar irradiance was monitored with a second reference cell. The different mesh combinations allowed us to reach quite low levels of irradiance (10%) with respect to the unobstructed irradiance (100%). Seasonal dust effects were studied by comparing the transmittance of glass samples exposed to outdoor conditions, at weekly time intervals, against a cleaned one. Data from several different US sites as well as from PV systems located in Crete, Greece are currently used to validate the model. Instantaneous values as well as daily integrals are compared to check the performance of the model. At this stage of analysis, it turns out that the typical accuracy of the model is better than 10% for angles of incidence less than sixty degrees. In addition, the performance of the model as a function of the various parameters is being studied and how these affect the calculations. In addition to the functions that have been determined from our measurements, functions available in the literature are also being tested.
-
-
-
A New Structural View Of The Holy Book Based On Specific Words: Towards Unique Chapters (surat) And Sentences (ayat) Characterization In The Quran
Authors: Meshaal Al-saffar, Ali Mohamed Jaoua, Abdelaali Hassaine and Samir ElloumiIn the context of web Islamic data analysis and authentication an important task is to be able to authenticate the holy book if published in the net. For that purpose, in order to detect texts contained in the holy book, it seems obvious to first characterize words which are specific to existing chapters (i.e. "Sourat") and words characterizing each sentence in any chapter (i.e. "Aya"). In this current research, we have first mapped the text of the Quran to a binary context R linking each chapter to all words contained in it, and by calculating the fringe relation F of R, we have been able to discover in a very short time all specific words in each chapter of the holy book. By applying the same approach we have found all specific words of each sentence (i.e. "Aya") in the same chapter whenever it is possible. We have found that almost all sentences in the same chapter have one or many specific words. Only sentences repeated in the same chapter or those sentences included in each other might not have specific words. Observation of words simultaneously specific to a chapter in the holy book and to the sentence in the same chapter gave us the idea for characterizing all specific sentences in each chapter with respect to the whole Quran. We found that for 42 chapters all specific words of a chapter are also specific of some sentence in the same chapter. Such specific words might be used to detect in a shorter time website containing some part of the Quran and therefore should help for checking their authenticity. As a matter of fact by goggling only two or three specific words of a chapter, we observed that search results are directly related to the corresponding chapter in the Quran. Al results have been obtained for Arabic texts with or without vowels. Utilization of adequate data structures and threads enabled us to have efficient software written in Java language. The present tool is directly useful for the recognition of different texts in any domain. In the context of our current project, we project to use the same methods to characterize Islamic books in general. ACKNOWLEDGMENT: This publication was made possible by a grant from the Qatar National Research Fund through National Priority Research Program (NPRP) No. 06-1220-1-233. Its contents are solely the responsibility of the authors and do not necessarily represent the official views of the Qatar National Research Fund or Qatar University.
-
-
-
A Novel Approach To Detection Of Glandular Structures In Colorectal Cancer Histology Images
Authors: Korsuk Sirinukunwattana, David Snead and Nasir RajpootBackground: Gland is a prevalent organ in a human body, synthesizing hormones and other vital substances. Gland morphology is an important feature in diagnosing malignancy and assessing the tumor grade in colorectal adenocarcinomas. However, a good detection and segmentation of glands is required prior to the extraction of any morphological features. Objectives: The aim of this work is to generate a glandular map for a histopathological image containing glandular structures. The map indicates the likelihood of different image regions belonging to glandular structures. This information can then be used as a clue for initial detection of glands. Methods: A pipeline to generate the probability map consists of the following steps. First, a statistical region merging algorithm is employed to generate superpixels. Second, texture and color features are extracted from each superpixel. For texture features, we calculate the coefficients of scattering trans- form. This transformation produces features at different scale-spaces which are translation-invariant and Lipschitz stable to deformation. To summarize the relationship across different scale-spaces, a region-covariance descriptor, which is a symmetric positive definite (SPD) matrix, is calculated. We call this image descriptor, scattering SPD. For color features, we quantize colors in all training images to reduce the number of features and to reduce the effect of stain variation between different images. Color information is encoded by a normalized histogram. Finally, we train a decision tree classifier to recognize superpixels belonging to glandular and nonglandular structures, and assign the probability of a superpixel belonging to the glandular class. Results: We tested our algorithm on a benchmark dataset consisting of 72 images of Hematoxylin & Eosin (H&E) stained colon biopsy from 36 patients. The images were captured at 20× magnification and the expert annotation is provided. One third of the images were used for training and the remaining for testing. Pixels with a probability value greater than 0.5 were considered as the detected glands. Table 1 shows that, in terms of the Dice index, the proposed method performs 5% better than local binary patterns and the combination between scattering SPD and color histogram results in 25% better accuracy than the baseline. Table 1: Average Segmentation Performance ApproachesSensitivitySpecificityAccuracyDice?Farjam et al. (baseline)0.50 ± 0.130.80 ± 0.150.62 ± 0.090.59 ± 0.14 superpixels + local binary pattern0.77 ± 0.060.67 ± 0.100.73 ± 0.040.77 ± 0.05 superpixels + scattering SPD0.77 ± 0.070.85 ± 0.090.81 ± 0.060.82± 0.06 superpixels + color histogram0.74 ± 0.220.82 ± 0.170.77 ± 0.10 0.79 ± 0.10 superpixels + scattering SPD + color histogram0.78 ± 0.07 0.88 ± 0.070.82± 0.060.84 ± 0.06 Conclusions: We present a superpixel-based approach for glandular structure detection in colorectal cancer histology images. We also present a novel texture descriptor derived from the region covariance matrix of scattering coefficients. Our approach generates highly promising results for initial detection of glandular structures in colorectal cancer histology images.
-
-
-
City-wide Traffic Congestion Prediction In Road Networks
Authors: Iman Elghandour and Mohamed KhalefaTraffic congestion is a major problem in many big cities around the world. According to a study performed by the world bank in Egypt in 2010 and concluded in 2012, the traffic congestion was estimated to 14 Billion EGP in the Cairo metropolitan area and to 50 Billion EGP (4\% of the GDP) in the entire Egypt. Few of the reasons of the high monetary cost of the traffic congestion are: (1) travel time delay, (2) travel time unreliability, and (3) excess fuel consumption. Smart traffic management addresses some of the causes and consequences of traffic congestion. It can predict congested routes, take preventive decisions to reduce congestion, disseminate information about accidents and work zones, and identify the alternate routes that can be taken. In this project, we develop a real-time and scalable data storage and analysis framework for traffic prediction and management. The input to this system is a stream of GPS and/or cellular data that has been cleaned and mapped to the road network. Our proposed framework allows us to (1) predict the roads that will suffer from traffic congestion in the near future, and traffic management decisions that can relieve this congestion; and (2) a what-if traffic system that is used to simulate what will happen if a traffic management or planning decision is taken. For example, it answers questions, such as: "What will happen if an additional ring road is built to surround Cairo?" or "What will happen if point of interest X is moved away from the downtown to the outskirts of the city. This framework has the following three characteristics. First, it predicts the flow of the vehicles in the road based on historical data. This is done by tracking vehicles every day trajectories and using them in a statistical model to predict the vehicles movement on the road. It then predicts the congested traffic zones based on the current vehicles in the road and their predicted paths. Second, historical traffic data are heavily exploited in the approach we use to predict traffic flow and traffic congestion. Therefore, we develop new techniques to efficiently store traffic data in the form of graphs for fast retrieval. Third, it is required to update the traffic flow of vehicles and predict congested areas in real-time, therefore we deploy our framework in the cloud and employ optimization technique to speedup the execution of our algorithms.
-
-
-
Integration Of Solar Generated Electricity Into Interconnected Microgrids Modeled As Partitioning Of Graphs With Supply And Demand In A Stochastic Environment
Authors: Raka Jovanovic and Abdelkader BousselhamA significant research effort has been dedicated in developing smartgrids in the form of interconnected microgrids. Their use is especially suitable for integration of solar generated electricity, due to the fact that by separating the electrical grid into smaller subsections, the fluctuations in the voltage and frequency that occur, can be to, a certain extent, isolated from the main grid. For the new topology, it is essential to optimize several important properties like the self-adequacy, reliability, supply-security and the potential for self-healing. These problems are frequently hard to solve, in the sense that they are hard combinatorial ones for which no polynomial time algorithm exists that can find the desired optimal solutions. Due to this fact research has been directed in finding approximate solutions, using different heuristic and metaheuristic methods. Another issue is that such systems are generally of a gigantic size. This resulted in two types of models, detailed ones that are applied to small systems and simplified ones for large ones. In the case of the former, graph models have shown to be very suitable especially ones that are based on graph partitioning problems[4]. One of the questions with the majority of previously developed graph models for large scales systems, is that they are deterministic. They are used for modeling an electrical grid which is in essence a stochastic system. In this work we focus on developing a stochastic graph model for including solar generated electricity to a system of interconnected microgrids. More precisely we focus on maximizing the self-adequacy of the individual microgrids, while trying to maximize the level of included solar generated energy, with a minimal amount of necessary energy storage. In our model we include the unpredictability of the generated electricity, and under such circumstances maintain a high probability that all demands in the system are satisfied. In practice we adapt and extend the concept of partitioning graphs with supply and demand for the problem of interest. This is done by having multiple values corresponding to the demand for one node in the graph. These values are used to represent energy usage in different time periods in one day. In a similar fashion we introduce a probability for the amount of electrical energy that will be produced by the generating nodes, and the maximal amount of storage in such nodes. Finally, we also include a heuristic approach to optimize this multi-objective optimization problem.
-
-
-
Empower The Vle With Social Computing Tools: System Prototype
Authors: Khaled Hussein, Jassim Al-jaber and Yusuf ArayiciSome lecturers gave reports and showed instances of moving part or all of their electronic course support from the Virtual Learning Environment (VLE) to social networking systems like Youtube, MySpace and Facebook because of greater student engagement with these kinds of social networking tools. Recent student interviews in Aspire Academy in Qatar have revealed that students are not concerned with what they are taught (e.g. through lectures, seminars, distance learning sessions, or through a blended learning approach) so long as the instruction was good. The latter reason opens great opportunities for delivering courses through SC media over the VLE, but also raises the question: To what extent can VLE and Social Media be leveraged as a good practice in learning and teaching in different modalities? In this research, the new experience of enriching the VLE with SC tools in Aspire Academy is presented through developing a new system prototype as a more effective solution. The prototyping process included usability testing with Aspire student-athletes and lecturers, plus heuristic evaluation by Human Computer Interaction (HCI) experts. Implementing the prototype system in academic institutions is expected to develop better learning levels and consequently better educational outcomes.
-
-
-
Gate Simulation Of A Clinical Pet Using The Computing Grid
Authors: Yassine Toufique and Othmane BouhaliNowadays, Nuclear Medicine becomes a potential research field of a growing importance at the universities. This fact can be explained, on one hand, by the increasing number of purchases in medical imaging devices by the hospitals, and, in other hand, by the number of PhD students and researchers becoming interested to medical studies. A Positron Emission Tomography (PET) system is a functional medical imaging technique which provides 3D images of the living processes inside the body relying on radioisotopes usage. The physics of PET systems is based on the detection in coincidence of the two 511 keV ?-rays, produced by an electron-positron annihilation, and emitted in opposite directions, as dictated by the conservation of energy and momentum physics laws. The radioactive nuclei used as sources of emission of positrons for PET systems are mainly 11C, 13N, 15O, and 18F, which are produced in cyclotrons, and decay with half-lives of 20.3 min, 9.97 min, 124 sec, and 110 min, respectively. These radioisotopes can be incorporated in a wide variety of radiopharmaceuticals that are inhaled or injected, leading to a medical diagnosis based on images obtained from a PET system. The PET scanners consists mainly of a large number of detector crystals arranged in a ring which surround the patient organ (or phantom in simulations) where the radioisotope tracer (e.g.: 18F-FDG) is inoculated. The final 3D image, representing the distribution of the radiotracer in the organ (or the phantom), is obtained by processing the signals delivered by the detectors of the scanner (when the ?-rays emitted from the source interact with the crystals) and using image reconstruction algorithms. This allows measuring important body functions, such as blood flow, oxygen use, and glucose metabolism, to help doctors evaluate how well organs and tissues are functioning and to diagnose and determine the severity of or treat a variety of diseases. The simulation of a real experiment using a GATE-modeled clinical Positron Emission Tomography (PET) scanner, namely PHILIPS Allegro, has been carried out using a computing Grid infrastructure. In order to reduce the computing time, the PET simulation tasks are split into several jobs submitted to the Grid to run simultaneously. The splitting technique and merging the outputs are discussed. Results of the simulation are presented and good agreements are observed with experimental data. Keywords—Grid Computing; Monte Carlo simulation; GATE; Positron Emission Tomography; splitting
-
-
-
A Cross-platform Benchmark Framework For Mobile Semantic Web Reasoning Engines In Clinical Decision Support Systems
Authors: William Van Woensel, Newres Al Haider and Syed Sr AbidiBackground & Objectives Semantic Web technologies are used extensively in the health domain to enable expressive, standards-based reasoning. Deploying Semantic Web reasoning processes directly on mobile devices has a number of advantages, including robustness to connectivity loss and more timely results. By leveraging local reasoning processes, Clinical Decision Support Systems (CDSS) can thus present timely alerts given dangerous health issues, even when connectivity is lacking. However, a number of challenges arise as well, related to mobile platform heterogeneity and limited computing resources. To tackle these challenges, developers should be empowered to benchmark mobile reasoning performance across different mobile platforms, with rule- and datasets of varying scale and complexity, and under typical CDSS reasoning process flows. To deal with the current heterogeneity of rule formats, a uniform interface on top of mobile reasoning engines also needs to be provided. System We present a mobile, cross-platform benchmark framework, comprising two main components: 1) a generic Semantic Web layer, supplying a uniform, standards-based rule- and dataset interface to mobile reasoning engines; and 2) a Benchmark Engine, to investigate mobile reasoning performance. This framework was implemented using the PhoneGap cross-platform development tool, allowing it to be deployed on a range of mobile platforms. During benchmark execution, the benchmark rule- and dataset (encoded using the SPARQL Inferencing Notation (SPIN) and Resource Description Framework (RDF)) are first passed to the generic Semantic Web layer. In this layer, the local Proxy component contacts an external Conversion Web Service, where converters perform conversion into the different rule engine formats. Developers may develop new converters to support other engines. The results are then communicated back to the Proxy and passed on to the local Benchmark Engine. In the Benchmark Engine, reasoning can be conducted using different process flows, to better align the benchmarks with real-world CDSS. To plugin new reasoning engines (JavaScript or native), developers need to implement a plugin realizing a uniform interface (e.g., load data, execute rules). New process flows can also be supplied. In the benchmarks, data and rule loading times, as well as reasoning times, are measured. From our work in clinical decision support, we identified two useful reasoning process flows: * Frequent Reasoning: To infer new facts, the reasoning engine is loaded with the entire datastore each time a certain timespan has elapsed, and the relevant ruleset is executed. * Incremental Reasoning: In this case, the datastore is kept in-memory, whereby reasoning is applied each time a new fact has been added. Currently, 4 reasoning engines (and their custom formats) are supported, including RDFQuery (https://code.google.com/p/rdfquery/wiki/RdfPlugin), RDFStore-JS (http://github.com/antoniogarrote/rdfstore-js), Nools (https://github.com/C2FO/nools) and AndroJena (http://code.google.com/p/androjena/). Conclusion In this paper, we introduced a mobile, cross-platform and extensible benchmark framework for comparing mobile Semantic Web reasoning performance. Future work consists of investigating techniques to optimize mobile reasoning processes.
-
-
-
Automatic Category Detection Of Islamic Content On The Internet Using Hyper Concept Keyword Extraction And Random Forest Classification
Authors: Abdelaali Hassaine and Ali JaouaThe classification of Islamic content on the Internet is a very important step towards authenticity verification. Many Muslims complain that the information they get from the Internet is either inaccurate or simply wrong. With the content growing in an exponential way, its manual labeling and verification is simply an impossible task. To the extent of our knowledge, no previous work has been carried out regarding his task. In this study, we propose a new method for automatic classification of Islamic content on the Internet. A dataset of four Islamic groups has been created containing texts from four different Islamic groups, namely: Sunni (Content representing Sunni Islam), Shia (Content representing Shia Islam), Madkhali (Content forbidding politics and warning against all scholars with different views) and Jihadi (Content promoting Jihad). We collected a dataset containing 20 different texts for each of those groups, totalizing 80 texts, out of which 56 are used for training and 24 for testing. In order to classify those contents automatically, we first preprocessed the texts using normalization, stop words removal, stemming and segmentation into words. Then, we used the hyper-concepts method which makes it possible to represent any corpus through a relation and to decompose it into non-overlapping rectangular relations and to highlight the most representative attributes or keywords in a hierarchical way. The hyper concept keywords extracted from the training set are subsequently used as predictors (containing either 1 when the text contains the keyword and 0 otherwise). Those predictors are fed to a random forest classifier of 5000 random trees. The number of extracted keywords varies according to the depth of the hyper concept tree, ranging from 47 keywords (depth 1) to 296 keywords (depth 15). The average classification accuracy starts at 45.79% for depth 1 and remains roughly stable at 68.33% from depth 10. This result is very interesting as there four different classes (a random predictor would therefore score around 25%). This study is a great step towards the automatic classification of Islamic content on the Internet. The results show that the hyper concept method successfully extracts relevant keywords for each group and helps in categorizing them automatically. The method needs to be combined with some semantic method in order to reach even higher classification rates. The results of the method are also to be compared with manual classification in order to foresee the improvement one can expect as some texts might indifferently belong to more than one category.
-
-
-
Optimal Communication For Sources And Channels With Memory And Delay-sensitive Applications
More LessShannon's theory of information was developed to address the fundamental problems of communication, such as the reliable data transmission over a noisy channel and the optimal data compression. During the years, it has expanded to find wide range of applications in many areas ranging from cryptography and cyber security to economics and genetics. Recent technological advances designate information theory as a promising and elegant tool to analyze and model information structures within living organisms. The key characteristics of data transmission within organisms are that they consider sources and channels with memory and feedback, they handle their information in a fascinating Shannon-optimum way, while the transmission of the data is delayless. Despite the extensive literature on memoryless sources and channels, the literature regarding sources and channels with memory is limited. Moreover, the optimality of communication schemes for these general sources and channels is completely unexplored. Optimality is often addressed via Joint Source Channel Coding (JSCC) and it is achieved if there exists an encoder-decoder scheme such that the Rate Distortion Function (RDF) of the source is equal to the capacity of the channel. This work is motivated by neurobiological data transmission and aims to design and analyze optimal communication systems consisting of channels and sources with memory, within a delay-sensitive environment. To this aim, we calculate the capacity of the given channel with memory and match it to a Markovian source via an encoder-decoder scheme, utilizing concepts from information theory and stochastic control theory. The most striking result to emerge from this research is that optimal and delayless communication for sources and channels with memory is not only feasible, but also it is achieved with the minimum complexity and computational cost. Though the current research is stimulated by a neurobiological application, the proposed approach and methodology as well as the provided results deliver several noteworthy contributions to a plethora of applications. These, among others, include delay sensitive and real time communication systems, control-communication applications and sensor networks. It addresses issues such as causality, power efficiency, complexity and security, extends the current knowledge of channels and source with memory, while it contributes to the inconclusive debates of real time communication and uncoded data transmission.
-
-
-
On And Off-body Path Loss Model Using Planar Inverted F Antenna
Authors: Mohammad Monirujjaman Khan and Qammer Hussain AbbasiThe rapid development of biosensors and wireless communication devices brings new opportunities for Body-Centric Wireless Networks (BCWN) which has recently received increasing attention due to their promising applications in medical sensor systems and personal entertainment technologies. Body-centric wireless communications (BCWCs) is a central point in the development of fourth generation mobile communications. In body-centric wireless networks, various units/sensors are scattered on/around the human body to measure specified physiological data, as in patient monitoring for healthcare applications [1-3]. A body-worn base station will receive the medical data measured by the sensors located on/around the human body. In BCWCs, communications among on-body devices are required, as well as communications with external base stations. Antennas are the essential component for wearable devices in body-centric wireless networks and they play a vital role in optimizing the radio system performance. The human body is considered an uninviting and even hostile environment for a wireless signal. The diffraction and scattering from the body parts, in addition to the tissue losses, lead to strong attenuation and distortion of the signal [1]. In order to design power-efficient on-body and off-body communication systems, accurate understanding of the wave propagation, the radio channel characteristics and attenuation around the human body is extremely important. In the past few years, researchers have been thoroughly investigating narrow band and ultra wideband on-body radio channels. In [4], on-body radio channel characterisation was presented at ultra wideband frequencies. In body-centric wireless communications, there is a need of communications among the devices mounted on the body as well as off-body devices. In previous study, researchers have designed the antennas for on-body communications and investigated the on-body radio channel performance both in narrowband and Ultra wideband technologies. This paper presents the results of on-body and off-body path loss model using Planar Inverted F Antenna (PIFA). The antenna used in this study works at two different frequency bands as 2.45 GHz (ISM band) and 1.9 GHz (PCS band). The 2.45 GHz is used for the communication over human body surface (on-body) and 1.9 GHz is used for the communication from body mounted devices to off-body units (off-body communications). Measurement campaigns were performed in the indoor environment and anechoic chamber. A frequency-domain measurement set-up was applied. Antenna design and on and off-body path loss model results will be presented.
-
-
-
Performance Analysis Of Heat Pipe-based Photovoltaic-thermoelectric Generator (hp-pv/teg) Hybrid System
Authors: Adham Makki and Siddig OmerPhotovoltaic (PV) cells can absorb up to 80% of the incident solar radiation of the solar spectrum, however, only certain percentage of the absorbed incident energy is converted into electricity depending on the conversion efficiency of the PV cell technology used, while the remainder energy is dissipated as heat accumulating on the surface of the cells causing elevated temperatures. Temperature rise at the PV cell level is addressed as one of the most critical issues influencing the performance of the cells causing serious degradations and shortens the life-time of the PV cells, hence cooling of the PV module during operation is essential. Hybrid PV designs which are able to simultaneously generate electrical energy and utilize the waste heat have been proven to be the most promising solution. In this study, analytical investigation of a hybrid system comprising of a Heat Pipe-based Photovoltaic-Thermoelectric Generator (HP-PV/TEG) for further enhanced performance is presented. The system presented incorporates a PV panel for direct electricity generation, a heat pipe to absorb excessive heat from the PV cells and assist uniform temperature distribution on the surface of the panel, and a thermoelectric generator (TEG) to perform direct heat-to-electricity conversion. A mathematical model based on the heat transfer process within the system is developed to evaluate the cooling capability and predict the overall thermal and electrical performances of the hybrid system. Results are presented in terms electrical efficiencies of the system. It was observed that the integration of TEG modules with PV cells aid improving the performance of the PV cells through utilizing the waste-heat available, leading to higher output power. The system presented can be applied in regions with hot desert climates where electricity demand is higher than thermal energy.
-
-
-
Effective Recommendation Of Reviewers For Research Proposals
Authors: Nassma Salim Mohandes, Qutaibah Malluhi and Tamer ElsayedIn this project, we address the problem that a research funding agency may face when matching potential reviewers with submitted research proposals. A list of potential reviewers for a given proposal is typically selected manually by a small technical group of individuals in the agency. However, the manual approach can be an exhausting and challenging task, and (more importantly) might lead to ineffective selections that affect the subsequent funding decisions. This research work presents an effective automated system that recommends reviewers for proposals and helps program managers in the assignment process. This system views the CVs of the reviewers and rank them by assigning weights for each CV against the list of all the proposals. We propose an automatic method to effectively recommend (for a given research proposal) a short list of potential reviewers who demonstrate expertise in the given research field/topic. To accomplish this task, our system extracts information from the full-text of proposals and the CVs of reviewers. We discuss the proposed solution, and the experience in using the solution within the workflow of the Qatar National Research Fund (QNRF). We evaluate our system on a QNRF/NPRP dataset that includes the submitted proposals and approved list of reviewers from the first 5 cycles of the NPRP funding program. Experimental results on this dataset validate the effectiveness of the proposed approach, and show that the best performance of our system demonstrated for proposals in three research areas: natural science, engineering, and medical. The system does not perform as well for proposals in the other two domains, i.e., humanities and social sciences. Our approach performs very well in overall evaluation with 68% of relevant results, i.e., from each 10 recommendations 7 are matching perfectly. Our proposed approach is general and flexible. Variations of the approach can be used in other applications such as conference paper assignment to reviewers and teacher-course assignment. Our research demonstrates also that there are significant advantages to applying recommender system concepts to the proposal-to-reviewer assignment problem. In summary, the problem of automatic assignment of proposals to reviewers is challenging and time-consuming when it is conducted manually by the program managers. Software systems can offer automated tools that significantly facilitate the role of program managers. We follow previous approaches in treating reviewers finding system as an information retrieval task. We use the same basic tools but the goal is to find relevant people rather than searching for relevant documents. For a specific user query (proposal), the system returns a list of qualified reviewers, ranked by their relevance to the query.
-
-
-
Intelligent Active Management Of Distribution Network To Facilitate Integration Of Renewable Energy Resources
More LessHarvesting electric energy from renewable energy resources is seen as one of the solutions to secure the energy sustainability due to depleting resources of fossil fuel, the conventional resources of electric energy. Renewable energy is typically connected to conventional distribution network which were not designed to accomodate any sources of electricity reduce the security of the energy supply system. Moreover the variency of renewable resources create many operational challenges to the distribution network operator. Higher shares of distributed energy sources lead to unpredictable network flows, greater variations in voltage, and different network reactive power characteristics as already evidence in many distribution networks. Local network constraints occurs more frequently, adversely affecting the quality of supply. Yet distribution network operators are nevertheless expected to continue to operate their networks in a secure way and to provide high-quality service to their customers. Active management of distribution network may provide some answers to these problems. Indeed, distribution management will allow grids to integrate renewable energu resources efficiently by leveraging the inherent characteristics of this type of generation. The growth of renewable energy resources requires changes to how distribution networks are planned and operated. Bi-directional flows need to be taken into account: they must be monitored, simulated and managed. This paper will describe features of smart grid concept that can be employed in distribution network for active management to facilitate the integration of renewable energy resources. The concepts include coordinated voltage control, microgrid operation and intelligent reactive power management to name a few. The development of physical testbed to test these new strategies in managing distribution network will also be described. The heart of these strategies is intelligent controller which acting as energy management system. The development of this controller will also be described and its operationality will be eplained.
-
-
-
Dynamic Team Theory With Nonclassical Information Structures Of Discrete-time Stochastic Dynamic Systems
Static Team Theory is a mathematical formalism of decision problems with multiple Decision Makers (DMs) that have access to different information and aim at optimizing a common pay-off or reward functional. It is often used to formulate decentralized decision problems, in which the decision-making authority is distributed through a collection of agents or players, and the information available to the DMs to implement their actions is non-classical. Static team theory and decentralized decision making originated from the fields of management, organization behavior and government by Marschak and Radner. However, it has far reaching implications in all human activity, including science and engineering systems, that comprise of multiple components, in which information available to the decision making components is either partially communicated to each other or not communicated at all. Team theory and decentralized decision making can be used in large scale distributed systems, such as transportation systems, smart grid energy systems, social network systems, surveillance systems, communication networks, financial markets, etc. As such, these concepts are bound to play key roles in emerging cyber-physical systems and align well with ARC'14 themes on Computing and Information Technology and Energy and Environment. Since the late 1960's several attempts have been made to generalize static team theory to dynamic team theory, in order to account for decentralized decision-making taken sequentially over time. However, to this date, no mathematical framework has been introduced to deal with non-classical information structures of stochastic dynamical decision systems, much as it is successfully done over the last several decades for stochastic optimal control problems, which presuppose centralized information structures. In this presentation, we put forward and analyze two methods, which generalize static team theory to dynamic team theory, in the context of discrete-time stochastic nonlinear dynamical problems, with team strategies, based on non-classical information structures. Both approaches are based on transforming the original discrete-time stochastic dynamical decentralized decision problem to an equivalent one in which the observations and/or the unobserved state processes are independent processes, and hence the information structures available for decisions are not affected by any of the team decisions. The first method is based on deriving team optimality conditions by direct application of static team theory to the equivalent transformed team problem. The second method is based on discrete-time stochastic Pontryagin's maximum principle. The team optimality conditions are captured by a "Hamiltonian System" consisting of forward and backward discrete-time stochastic dynamical equations, and a conditional variational Hamiltonian with respect to the information structure of each team member, while all other team members hold the optimal values.
-
-
-
Cunws/rgo Based Transparent Conducting Electrodes As A Replacement Of Ito In Opto-electric Devices
Transparent electrodes that conduct electrical current and allow light to pass through are widely used as the essential component in various opto-electric devices such as light emitting diodes, solar cells, photodectectors and touch screens. Currently, Indium Tin oxide (ITO) is the best, commercially available transparent conducting electrode (TCE). However, ITO is too expensive owing high cost on indium. Furthermore ITO thin films are too brittle to be used in flexible devices. To fulfill the demand of TCEs for wide range of applications, high performance ITO alternatives are required. Herein we demonstrate an approach for the successful, solution based synthesis of high aspect ratio copper nanowires, which were later combined with reduced graphene oxide (rGO), in order to produce smooth thin film TCEs on both glass and flexible substrate. Structure and component characterization for these electrodes was carried out through Four Probe, Spectrophotometer, Scanning electron Microscope (SEM), Transmission Electron Microscope (TEM) and Atomic Field Microscopy (AFM). In addition to the morphological and electrical characterization, these samples were also tested for their durability by carrying out experiments that involved exposure to various environmental conditions and electrode bending. Our fabricated transparent electrodes exhibited high performance with a transmittance of 91.6% and a sheet resistance of 9 O/sq. Furthermore, the electrodes showed no notable loss in performance during the durability testing experiments. Such results make them as replacement for indium tin oxide as a transparent electrode and presents a great opportunity to accelerate the mass development of devices like high efficiency hybrid silicon photovoltaics via simple and rapid soluble processes.
-
-
-
Development Of A Remote Sma Experiment - A Case Study
Authors: Ning Wang, Jun Weng, Michael Ho, Xuemin Chen, Gangbing Song and Hamid ParsaeiA remote laboratory containing a specially designed experiment was built to demonstrate and visualize the characteristics of wire-shape shape memory alloys (SMAs). In particular, the unit helps the study of the hysteretic behavior of SMAs as well as how the electrical driving frequency changes the hysteresis loop. One such SMA remote experiment with a novel unified framework was constructed at the Texas A&M University at Qatar. In this project, we developed a new experiment data transaction protocol and software package used in the remote lab. It is clear that the new platform makes some improvements in traversing network firewall function and software plug-in free. In order to provide a more realistic experience to the user in conducting the remote SMA experiment, the new solution also implements a Real-Time experiment video function. Compared to the traditional remote SMA experiment that uses the LabVIEW remote panel, the new SMA remote experiment solution has three advantages. The user interface of the new remote SMA experiment is plug-in free and can run in different web browsers. The new remote lab also resolves the issue of traversing a network firewall. End users only need to access the Internet and use a web browser to operate the SMA experiment. The experiment control webpage is developed by JavaScript which is a universally used computer language. Meanwhile, any regular web browsers are able to use all the features of the remote panel without requiring any extra software plug-ins. An additional function of the new remote lab is the real-time delivery of the video transmission from the experiment, thus providing a more realistic experience for the user. This new remote SMA experiment user interface can also be used through smart phones and tablet computers. Compared to LabVIEW based experiments, the experiment data collected from the use of the novel unified framework are similar except for the amplitude of the reference signal. The amplitude can be different because they are defined by the users. The data recorded from the new remote SMA experiment GUI has fewer samples per second comparing to that in remote SMA experiment with LabVIEW. The data transmission in the GUI is limited to 140 samples per second to minimize the memory and increase the connection speed. In the remote SMA experiment with LabVIEW, the sampling rate is 1000 samples per second; however, the hysteresis of SMA has been successfully demonstrated by the data recorded in the new remote SMA experiment with the novel unified framework, which matches the original results collected locally. The study compares two different implementation approaches for the remote SMA experiment; one is the traditional approach with the LabVIEW remote panel and the other is the new approach with the novel unified framework. The difference of these two solutions is listed, and the advantage of the new SMA remote experiment based on the novel unified framework is presented. The capability of running remote experiments on portable devices allows users to learn by observing and interacting with the real experiment in an efficient way.
-
-
-
Semantic Web Based Execution-time Merging Of Processes
Authors: Borna Jafarpour and Syed Sibte Raza AbidiA process is a series of actions executed in a particular environment in order to achieve a goal. It is often the case that several concurrent processes coexist in an environment in order to achieve several goals simultaneously. However, executing multiple processes is not always a possibility in an environment due to the following reasons: (1) All processes might be needed to be executed by a single agent that is not capable of executing more than one process at a time; (2) Multiple processes may have interactions between them that hamper their concurrent executions by multiple agents. As an example, there might be conflicting actions between several processes that their concurrent execution will stop those processes from achieving their goals. The existing solution to address the abovementioned complications is to merge several processes into a unified conflict-free and improved process before execution. This unified merged process is then executed by a single agent in order to achieve goals of all processes. However, we believe this is not the optimal solution because (a) in some environments, it is unrealistic to assume execution of all processes merged into one single process can be delegated to a single agent; (b) since merging is performed before actual execution of the unified process, some of the assumptions made regarding execution flow in individual processes may not be true during actual execution which will render the merged process irrelevant. In this paper, we propose a semantic web based solution to merge multiple processes during their concurrent execution in several agents in order to address the above-mentioned limitations. Our semantic web Process Merging Framework features a Web Ontology Language (OWL) based ontology called Process Merging Ontology (PMO) capable of representing a wide range of workflow and institutional Process Merging Constraints, mutual exclusivity relations between those constraints and their conditions. Process Merging Constraints should be respected during concurrent execution of processes in several agents in order to achieve execution-time process merging. We use OWL axioms and Semantic Web Rule Language (SWRL) rules in the PMO to define the formal semantics of the merging constraints. A Process Merging Engine has also been developed to coordinate several agents, each executing a process pertaining to a goal, to perform process merging during execution. This engine runs the Process Merging Algorithm that utilizes Process Merging Execution Semantics and an OWL reasoner to infer the necessary modifications in actions of each of the processes so that Process Merging Constraints are respected. In order to evaluate our framework we have merged several clinical workflows each pertaining to a disease represented as processes so that they can be used for decision support for comorbid patients. Technical evaluations show efficiency of our framework and evaluations with the help of domain expert shows expressivity of PMO in representation of merging constraints and capability of Process Merging Engine in successful interpretation of the merging constraints. We plan to extend our work to solve problems in business process model merging and AI plan merging research areas.
-
-
-
Performance Of Hybrid-access Overlaid Cellular Mimo Networks With Transmit Selection And Receive Mrc In Poisson Field Interference
Authors: Amro Hussen, Fawaz Al Qahtani, Mohamed Shaqfeh, Redha M. Radaydeh and Hussein AlnuweiriThis paper analyzes the performance of a hybrid control access scheme for small cells in the context of two-tier cellular networks. The analysis considers MIMO transmit/receive arrays configuration that implements transmit antenna selection (TAS) and maximal ratio combining (MRC) under Rayleigh fading channels when the interfering sources are described using Poisson field processes. The adopted models of aggregate interference at each receive station is modeled as a shot noise that follows a Stable distribution. Furthermore, based on the interference awareness at the receive station, two TAS approaches are considered through the analysis, which are the signal-to-noise (SNR)-based selection and signal-to-interference-plus-noise ratio (SINR)-based selection. In addition, the effect of delayed TAS due to imperfect feedback channel on the performance measures is investigated. New analytical results the hybrid-access scheme's downlink outage probability and error rate are obtained. To gain further insight on the system's behavior at limiting cases, asymptotic results for the outage probability and error rate at high signal-to-noise (SNR) are also obtained, which can be useful to describe diversity orders and coding gains. The derived analytical results are validated via Monte Carlo simulations.
-
-
-
Oryx Gtl Data Integration And Automation System For 21st Century Environmental Reporting
Authors: Sue Sung, Pon Saravanan Neerkathalingam, Ismail Al-khabani, Kan Zhang and Arun KanchanORYX GTL is an environmentally responsible company committed to creating an efficient, diversified energy business, developing its employees, and adding value to Qatar's natural resources. ORYX GTL considers the monitoring and reporting consistent environmental data and setting accurate targets is a critical component to increase the operational efficiency on sustainable manner. Monitoring key metrics such as air emissions, criteria pollutants, flaring and energy can provide an opportunity to reduce the environmental impacts and cost savings. ORYX GTL has adopted a state-of-art information technology (IT) solution to enhance the data handling process in support of the company's environmental performance reports such as the greenhouse gas (GHG) accounting and reporting (A&R) program required by Qatar Petroleum (QP). The automated system to report environmental data is proven to be more efficient and accurate which also increases consistency & requires fewer resources to report the data in a reliable manner. The system selected by ORYX GTL is the Data Integration and Automation (DIA) system designed developed by Trinity Consultants on the Microsoft® .net platform. The objective of this paper is to share the challenges and experience during the design, develop and implement this advanced DIA system for critical environmental reporting functions at ORYX GTL as a part of the company's commitment to improve environmental performance. The DIA application can be used as the central data storage/handling system for all environmental data reporting. The DIA software includes several functions built on a state-of-art IT platform to achieve near real-time environmental monitoring, performance tracking, and reporting. The key functions include: -Hourly data retrieval, aggregation, validation, and reconciliation from the plant process historian on a pre-defined schedule. The data retrieved from the process historian may include data such as hourly fuel usage and continuous emission monitoring data, and sampling data collected on routine basis. -Powerful calculation engine allows user to build complex emission calculation equations. Calculated results are stored in the database for use in reporting. In addition to user specified equations, the system also includes a complete calculation module to handle complex calculations for tank emissions. -Through the web interface of the DIA, users can manage system reporting entity hierarchy and user security, set up tags/sources, create manual data entries, create/modify equations, and execute emission reports. The DIA application sends email notifications of errors of tag data and calculation results at user specified intervals. Email recipients can provide timely response to the system with proper root causes and corrective actions. -Custom reports can be designed to generate regulatory reports in the format required by QP or the Qatar Ministry of Environment. The DIA system has significantly enhanced the quality of ORYX GTL's environmental reporting by reducing human interactions required for process data extraction, validation, reconciliation, calculations, and reporting. ORYX GTL's proactive approach to implement and integrate DIA system provided the opportunity to improve reporting functions and stakeholder & regulator satisfaction as well as it ensures the principles of environmental data reporting such as complete, consistent, transparent and accurate.
-
-
-
An Integrated Framework For Verified And Fault Tolerant Software
Authors: Samir Elloumi, Ishraf Tounsi, Bilel Boulifa, Sharmeen Kakil, Ali Jaoua and Mohammad SalehFault tolerance techniques should let the program continue servicing in spite of the presence of errors. They are of primary importance mainly in case of mission-critical systems. Their eventual failure may produce important human and economic casualties. For these reasons, researchers have assigned the software reliability as an important research area in terms of checking its design and functionality. As a matter of fact, software testing aims to increase the software correctness by verifying the program outputs w.r.t an input space generated in a bounded domain. Also, the fault tolerance approach has many effective error detection mechanisms as per as the Backward recovery, Forward recovery or redundancy algorithm. Our work consists of developing an integrated approach for software testing in a bounded domain. It tolerates transient faults to solve deficiencies and to obtain a robust and well-designed program. The developed framework comprises two types of tests: i) Semi-automatic test that enables the user to check the software by manually entering the values of the method and testing with specified values, ii) Automatic test that computerizes the test with the prepared instances of the program and generated values of a chosen method that exists inside the software. For generating the input values of a program, we have involved “Korat” that requires a class invariant, a bounded domain and Java Predicates (or preconditions). The framework uses the reflection technique in order to verify the correctness of the method under test. Based on the pre-post conditions, or Java predicates, previously fixed by the user, the backward recovery and the Forward recovery algorithm are applied to tolerate the transient faults. In case of Forward recovery, an efficient original solution has been developed based on reducing the number of re-executing a bloc of instructions. In fact, the re-execution is started from the current state instead of the initial state under the hypothesis of no loss of critical information. A plugin Java library has been implemented for fault tolerant version. The Framework was experimented for several java programs and was applied for improving the robustness of the Gas purification software. ACKNOWLEDGMENT: This publication was made possible by a grant from the Qatar National Research Fund through National Priority Research Program (NPRP) No. 04-1109-1-174. Its contents are solely the responsibility of the authors and do not necessarily represent the official views of the Qatar National Research Fund or Qatar University
-
-
-
Car Make And Model Detection System
Authors: Somaya Al-maadeed, Rayana Boubezari, Suchithra Kunhoth and Ahmed BouridaneThe deployment of highly intelligent and efficient machine vision systems accomplished to achieve new heights in multiple fields of human activity. A successful replacement of manual intervention with their automated systems assured safety, security and alertness in the transportation field. Automatic number plate recognition (ANPR) has become a common aspect of the intelligent transportation systems. In addition to the license plate information, identifying the exact make and model of the car is suitable to provide many additional cues in certain applications. Authentication systems may find it useful with an extra confirmation based on the model of the vehicle also. Different car models are characterized by the uniqueness in the overall car shape, position and structure of headlights etc. Majority of the research works rely on frontal/rear view of car for the recognition while some others are also there based on an arbitrary viewpoint. A template matching strategy is usually employed to find an exact match for the query image from a database of known car models. It is also possible to select and extract certain discriminative features from the region of interest (ROI) in the car image. And with the help of a suitable similarity measure such as euclidean distance it is able to demarcate between the various classes/models. The main objective of the paper is to understand the significance of certain detectors and descriptors in the field of car make and model recognition. The performance evaluation of SIFT, SURF, ORB feature descriptors for implementing a car recognition system was already available in literature. In this paper, we have studied the effectiveness of various combinations of feature detectors and descriptors on car model detection. The combination of the 6 detectors DoG, Hessian, Harris Laplace, Hessian Laplace, Multiscale Harris, Multiscale Hessian with the 3 descriptors SIFT, liop and patch was tested on three car databases. Scale Invariant Feature Transform (SIFT), a popular object detection algorithm allows the user to match different images and spot the similarities between them. The algorithm based on keypoints selection and description offers feature independent of illumination, scale, noise and rotation variations. Matching between images has been executed using Euclidian distance between descriptors. For the given keypoints in the test image, the smallest Euclidian distance between corresponding descriptor and all the descriptors of the training image indicates the best match. Our experiments were carried out in MATLAB using the VLFeat ToolBox. It was found to achieve a maximum accuracy of 91.67% with DoG-SIFT approach in database 1 comprising cropped ROI of toy car images. For the database 2 consisting of cropped ROI of real car images, the Multiscale Hessian-SIFT yielded the maximum accuracy of 96.88%. The database 3 comprised of high resolution real car images with background. The testing was conducted on the cropped and resized ROI's of these images. A maximum accuracy of 93.78% was obtained when the Multiscale Harris-SIFT feature descriptor was employed. As a whole these feature detectors and descriptors succeeded in recognizing the car models with an overall accuracy above 90%.
-
-
-
Visual Simultaneous Localization And Mapping With Stereo And Wide-angle Imagery
Authors: Peter Hansen, Muhammad Emaduddin, Sidra Alam and Brett BrowningMobile robots provide automated solutions for a range of tasks in industrial settings including but not limited to inspection. Our interest is automated inspection tasks including gas leakage detection in natural gas processing facilities such as those in Qatar. Using autonomous mobile robot solutions remove humans from potentially hazardous environments, eliminate potential human errors from fatigue, and provide data logging solutions for visualization and off-line post-processing. A core requirement for a mobile robot to perform any meaningful inspection task is to localize itself within the operating environment. We are developing a visual Simultaneous Localization And Mapping (SLAM) system for this purpose. Visual SLAM systems enable a robot to localize within an environment while simultaneously building a metric 3D map using only imagery from an on-board camera head. Vision has many advantages over alternate sensors used for localization and mapping. It requires minimal power compared to Lidar sensors, is relatively inexpensive compared to Inertial Navigation Systems (INS), and can operate in GPS denied environments. There is extensive work related to visual SLAM with most systems using either a perspective stereo camera head or a wide-angle of view monocular camera. Stereo cameras enable Euclidean 3D reconstruction from a single stereo pair and provide metric pose estimates. However, the narrow angle of view can limit pose estimation accuracy as visual features can typically be 'tracked' only across a small number of frames. Moreover, the limited angle of view presents challenges for place recognition whereby previously visited locations can be detected and loop closure performed to correct for long-range integrated position estimate inaccuracies. In contrast, wide-angle of view monocular cameras (e.g. fisheye and catadioptric) trade spatial resolution for an increased angle of view. This increased angle can enables visual scene points to be tracked over many frames and can improve rotational pose estimates. The increased angle of view can also improve visual place recognition performance as the same areas of a scene can be imaged under much larger changes in position and orientation. The primary disadvantage of a monocular wide-angle visual SLAM system is a scale ambiguity in the translational component of pose/position estimates. The visual SLAM system being developed in this work uses a combined stereo and wide-angle fisheye camera system with the aim of exploiting the advantages of each. For this we have combined visual feature tracks from both the stereo and fisheye camera within a single non-linear least-squares Sparse Bundle Adjustment (SBA) framework for localization. Initial experiments using large scale image datasets (approximately 10 kilometers in length) collected within Education City have been used to evaluate improvements in localization accuracy using the combined system. Additionally, we have demonstrated performance improvements in visual place recognition using our existing Hidden Markov Model (HMM) based place recognition algorithm.
-
-
-
Secured Scada System
More LessAbstract: SCADA ( Supervisory control and data acquisition) is a system which allows a control of remote industrial equipments, over a communication channel. These legacy communication channels were designed before the cyber space era,and hence they lack any security measures in their network which makes them vulnerable to any cyber attack. RasGas a joint venture between QP and ExxonMobil, was one victim of such attack, it was hit with an unknown virus. The nuclear facility in Iran was hit with a virus called “Stuxnet”, it particulary targets Siemens industrial control systems. The goal of this project is to design a model of a SCADA system that is secured against network attacks. Lets consider for example a simple SCADA system which consist of a Water tank in a remote location and a local control room. the operator controls the water level and temperature using a control panel. The communication channels uses a TCP/IP , protocols through WIFI. The operator raises the temperature of the water by raising the power of the heater, then reads the real temperature of the heater and the water via installed sensors. We consider a man-In-The middle (Adversary) which has access to the network through WIFI. With basic skills s/he is able to redirects tcp/ip traffic to his machine (tapping) and alter data. He can for instance raise water level to reach overflow, or increase the temperature above the "danger zone", and sends back fake sensors data by modifying their response. We introduce an encryption device that encrypt the data such that without the right security credentials , the adversary wont be able to interpret the data and hence not able to modify it. The device is installed at both the control room and the remote tank, and we assume both places are physically secured. To demonstrate the Model. We design and setup a SCADA Model emulator server, that represents and serves as a Water tank, which consists of actuators and sensors. Which is connected to a work station through a network switch. We also setup an adversary workstation that taps and alters the communication between them. We Design Two hardware encryption/decryption devices using FPGA boards and connect them at the ports of both the server and control workstation which we assume to be in a secured zone. and then we analyze the flow of data stream through both secured and non secured state of the channel.
-
-
-
Measurement Platform Of Mains Zero Crossing Period For Powerline Communication
Authors: Souha Souissi and Chiheb RebaiPower lines (PLC), mainly dedicated and optimized for delivering electricity, are, nowadays, used to transfer data. Its low cost and wide coverage makes it one of the pivotal technologies in building up smart grid. The actual role of PLC technology is controversial while some preconize that PLC systems are very good candidates for some applications others discard it and look at wireless as a more elaborated alternative. It is obvious that Smart Grid will include multiple types of communications technologies ranging from optics to wireless and wireline. Among wireline solutions, PLC appears to be the only technology that has deployment cost comparable to wireless as power installation and line are already existent. Narrowband PLCs are a key point in smart grid that is elaborated to support several applications such as Automatic Meter Reading (AMR), Advanced Metering Infrastructure (AMI), demand side management, in-home energy management and Vehicle-to-grid communications A critical stage in designing an efficient PLC system remains in getting sufficient knowledge about channel behavior characteristics such as attenuation, access impedance, multiple noise scenarios and synchronization. That's why; characterizing power line network has been the interest of several research works aiming to compromise between robustness of powerline communication and higher data rate. This interest in narrowband PLC systems to find the adequate one is inciting to deeply focus on channel characterization and modeling methods. It represents a first step to simulate channel behavior then propose a stand-alone hardware for emulation. Authors are investigating the building blocks of a narrowband PLC channel emulator that helps designers to evaluate and verify their systems design. It allows a reproduction of real conditions for any narrowband PLC equipment (single carrier or multicarrier) by providing three major functionalities: noise scenarios, signal attenuation and zero crossing reference mainly used for single carriers systems. For this purpose, authors deploy a bottom up approach to identify a real channel transfer function (TF) based on a prior knowledge of used power cables characteristics and connected loads. A simulator is, then, defined based on Matlab that generates a given TF according to defined parameters. The AC mains zero crossing variation is also studied. In field exhaustive measurements of this reference have shown a perpetual fluctuation presented as a jitter error. It is the reflection of variant AC mains characteristics (frequency, amplitude) which could be related to non-linearity of connected loads at network and used detection circuit in PLC systems. Authors propose a ZC variation model according to system environment (home/ lab, rural). This model will be embedded on channel emulator to reproduce ZC reference variation. Regarding noise, few models are found in literature, specific to narrowband PLCs. An implementation of some models is done and tested on a DSP platform which will include the two previous elements: TF and ZC variation.
-
-
-
Social Media As A Source Of Unbiased News
By Walid MagdyNews media are usually biased toward some political views. Also, the coverage of news is limited to news reported by news agencies. Social media is currently a hub for users to report and discuss news. This includes news reported or missed by news media. Developing a system that can generate news reports from social media can give a global unbiased view on what is hot in a given region. In this talk, we present the research work performed in QCRI for two years, which tackle the problem of using social media to track and follow posts on ongoing news in different regions and for different topics. Initially, we show examples of the presence of bias in reporting news by different news media. We then explore the nature of social media platforms and list the research questions that motivated this work. The challenges for tracking topics related to news are discussed. An automatically adapting information filtering approach is presented that allows tracking broad and dynamic topics in social media. This technique enables automatically tracking posts on news in social media while coping with the high changes occurring in news stories. Our developed system, TweetMogaz, is the demoed, which is an Arabic news portal platform that generated news from Twitter. TweetMogaz reports in real-time what is happening in hot regions in the Middle East, such as Syria and Egypt, in the form of comprehensive reports that include top tweets, images, videos, and news article shared by users on Twitter. It also reports news on different topics such as sports. Moreover, Search is enabled to allow users to get news reports on any topic of interest. The demo would be showing www.tweetmogaz.com live, where emerging topics in news would appear live in front of the audience. By the end of the talk, we would show some of the interesting examples that were noticed on the website in the past year. In addition, a quick overview would be presented on one of the social studies, which was carried out based on the news trend changes on TweetMogaz. The study shows the changes of people behavior when reporting and discussing news during major political changes such as the one happened in Egypt in July 2013. This work is an outcome of two years of research in the Arabic Language Technology group in Qatar Computing Research Institute. The work is published in the form of six research and demo papers in tier 1 conferences such as SIGIR, CSCW, CIKM, and ICWSM. The TweetMogaz system is protected by two patent applications filed in 2012 and 2014. Currently the website serves around 10,000 users, and the number is expected to significantly increase when officially advertised. Please feel free to visit TweetMogaz website for checking the system live: www.tweetmogaz.com Note: A new release with a better design to the website is expected by the time of the conference
-
-
-
Semantic Model Representation For Human's Pre-conceived Notions In Arabic Text With Applications To Sentiment Mining
Authors: Ramy Georges Baly, Gilbert Badaro, Hazem Hajj, Nizar Habash, Wassim El Hajj and Khaled ShabanOpinion mining is becoming of high importance with the availability of opinionated data on the Internet and the different applications it can be used for. Intensive efforts have been made to develop opinion mining systems, and in particular for the English language. However, models for opinion mining in Arabic remain challenging due to the complexity and rich morphology of the language. Previous approaches can be categorized into supervised approaches that use linguistic features to train machine learning classifiers, and unsupervised approaches that make use of sentiment lexicons. Different features have been exploited such as surface-based, syntactic, morphological, and semantic features. However, the semantic extraction remains shallow. In this paper, we propose to go deeper into the semantics of the text when considered for opinion mining. We propose a model that is inspired by the cognitive process that humans follow to infer sentiment, where humans rely on a database of preconceived notions developed throughout their life experiences. A key aspect for the proposed approach is to develop a semantic representation of the notions. This model consists of a combination of a set of textual representations for the notion (Ti), and a corresponding sentiment indicator (Si). Thus
denotes the representation of a notion. However, notions can be constructed at different levels of text granularity ranging from ideas covered by words to ideas covered in full documents. The range also includes clauses, phrases, sentences, and paragraphs. To demonstrate the use of this new semantic model of preconceived notions, we develop the full representation of one-word notions by including the following set of syntactic features for Ti: word surfaces, stems, and lemmas represented by binary presence and TFIDF. We also include morphological features such as part of speech tags, aspect, person, gender, mood, and number. As for the notion sentiment indicator Si, we create a new set of features that indicate the words' sentiment scores based on an internally-developed Arabic sentiment lexicon called ArSenL, and using a third-party lexicon called Sifaat. The aforementioned features are extracted at the word-level, and are considered as raw features. We also investigate the use of additional "engineered" features that reflect the aggregated semantics of a sentence. Such features are derived from word-level information, and include count of subjective words, average of sentiment scores per sentence. Experiments are conducted on a benchmark dataset collected from the Penn Arabic TreeBank (PATB) already annotated with sentiment labels. Results reveal that raw word-level features do not achieve satisfactory performance in sentiment classification. Feature reduction was also explored to evaluate the relative importance of the raw features, where the results showed low correlations between individual raw features and sentiment labels. On the other hand, the inclusion of engineered features had a significant impact on classification accuracy. The outcome of these experiments is a comprehensive set of features that reflect the one-word notion or idea representation in a human mind. The results from one-word also show promises towards higher level context with multi-word notions.
-
-
-
Intelligent M-Health Technology For Enhanced Smoking Cessation Management
Authors: Abdullah Alsharif and Nada PhilipAbstract Smoking-related illnesses are costly to the NHS and a leading cause of morbidity and mortality. Pharmacological treatments including nicotine replacement, some antidepressants, and nicotine receptor partial agonists, as well as individual- and group-based behavioural approaches, can help stop people from smoking. Circa 40% of smokers attempt to quit smoking each year, yet most have rapid relapses. The development of new tools acceptable by a wide range of smokers should be of particular interest. Smartphone interventions such as text messaging have shown some promise in helping people stop smoking. However most of these studies were based on text-messaging interventions with no interactive functionality that can provide better feedback to the smoker. In addition there is increasing evidence that smart mobile phones act as a conduit to behavioural change in other forms of healthcare. A study of currently available iPhone apps for smoking cessation have shown a low level of adherence to key guidelines for smoking cessation; few, if any, recommended or linked the user to proven treatments such as pharmacotherapy, counselling or a “quit line” and smoking cessation program. Hence there is a need for clinical validation of the feasibility of app-based intervention in supporting smoking cessation programmes in community pharmacy settings. The goal of this study is to design and develop an m-health programme platform to support smoking cessation in a community setting. The primary objectives are ascertaining what users require from a mobile app-based smoking cessation system targeting and supporting smokers, and looking into the literature for similar solutions. The study also involves the design and development of an end-to-end smoking cessation management system based on these identified needs; this includes the Patients Hub, Cloud Hub, and Physician/Social Worker Hub, as well as the design and development of a decision support system based on data mining and an artificial intelligent algorithm. Finally, it will implement the system and evaluate it in a community setting.
-
-
-
MobiBots: Risk Assessment Of Collaborative Mobile-to-Mobile Malicious Communication
Authors: Abderrahmen Mtibaa, Hussein Alnuweiri and Khaled HarrasCyber security is moving from traditional infrastructure to sophisticated mobile infrastreless threats. We believe that such imminent transition is happening at a rate exceeding by far the evolution of security solutions. In fact, the transformation of mobile devices into highly capable computing platforms makes the possibility of security attacks originating from within the mobile network a reality. All recent security report emphasize on the steadily increase of malicious mobile applications. Trend Micro, in their last security report, shows that the number of malicious application doubled in just six months to reach more than 700000 malwares in June 2013. This represents a major issue for today's cyber security in the world and particularly in the middle east. The last Trend Micro report shows that the United Arab Emirates has “by far” the highest malicious Android application download volume worldwide. Moreover, Saudi Arabia, another middle eastern country, register the highest downloads of high-risk applications. We believe that today mobile devices are capable of initiating sophisticated cyberattacks especially when they coordinate together forming what we call a mobile distributed botnet (MobiBot). MobiBots leverage the absence of basic mobile operating system security mechanism and the advantages of classical botnets which make them a serious security threat to any machine and/or network. In addition, MobiBot's distributed architecture (see attached figure), its communication model, and its mobility make it very hard to track, identify and isolate. While there has been many android security studies, we find that the proposed solutions can not be adopted in the challenging MobiBot environment due to its de-centralized architecture (figure). MoBiBots bring significant challenges to network security. Thus, securing mobile devices by vetting malicious tasks can be considered as one important first step towards MobiBot security. Motivated by the trends mentioned above, in our project we first investigate the potential for and impact of the large scale infection and coordination of mobile devices. We highlight how mobile devices can leverage short range wireless technologies in attacks against other mobile devices that come within proximity. We quantitatively measure the infection and propagation rates within MobiBots using short range wireless technology such as Bluetooth. We adopt an experimental approach based on a Mobile Device Cloud platform we have build as well as three real world data traces. We show that Mobibot infection can be really fast by infecting all nodes in a network in only few minutes. Stealing data however requires longer period of time and can be done more efficiently if the botnet utilizes additional sinks. We also show that while MobiBots are difficult to detect and isolate compared to common botnet networks, traditional prevention techniques costs at least 40% of the network capacity. We also study the scalability of MobiBots in order to understand the strengths and weaknesses of these malicious networks. We based our analysis on a dataset that consists of multiple disjoint communities, each one is a real world mobility trace. We show that MobiBots succeed on infecting up to 10K bots in less than 80 minutes.
-
-
-
The Infrastructure Of Critical Infrastructure: Vulnerability And Reliability Of Complex Networks
By Martin Saint* Background & Objectives All critical infrastructure can be modeled as networks, or systems of nodes and connections, and many systems such as the electric grid, water supply, or telecommunications exist explicitly as networks. Infrastructures are interdependent, for instance, telecommunications depend on electric power, and control of the electric grid depends increasingly upon telecommunications, creating the possibility for a negative feedback loop following a disturbance. The performance of these systems under disturbance are related to their inherent network characteristics, and network architecture plays a fundamental role in reliability. What characteristics of networks affect their robustness? Could, for instance, the vulnerability of the electric grid to cascading failure be reduced? * Methods We create a failure model of the network where each node and connection is initially in the operative state. At the first discrete time step a network element is changed to the failed state. At subsequent time steps a rule is applied which determines the state of random network elements based upon the state of their neighbors. Depending upon the rule and the distribution of the degree of connectedness of the network element, failures may be contained to a few nodes or connections, or may cascade until the entire network fails. * Results Quantitative measures from the model are the probability of network failure based upon the loss of a network element, and the expected size distribution of failure cascades. Additionally, there is a critical threshold below which infrastructure networks fail catastrophically. The electrical grid is especially vulnerable as it operates close to the stability limit, and there is a low critical threshold after which the network displays a sharp transition to a fragmented state. Failures in the electrical grid result not only in the loss of capacity in the network element itself, but load shifting to adjacent network elements, which contributes to further instability. While most failures are small, failure distributions are heavy tailed indicating occasional catastrophic failure. Many critical infrastructure networks are robust to random failure, but the existence of highly connected hubs give them a high clustering coefficient which makes the network vulnerable to targeted attacks. * Conclusions It is possible to design network architectures which are robust to two different conditions: random failure and targeted attack. It is also possible to alter architecture to increase the critical threshold at which failed network elements cause failure of the network as a whole. Surprisingly, adding more connections or capacity sometimes reduces robustness by creating more routes for failure to propagate. Qatar is in an ideal position to analyze and improve critical infrastructure from a systemic perspective. Modeling and simulation as detailed above are readily applicable to analyzing real infrastructure networks.
-
-
-
Annotation Guidelines For Non-native Arabic Text In The Qatar Arabic Language Bank
Authors: Wajdi Zaghouani, Nizar Habash, Behrang Mohit, Abeer Heider, Alla Rozovskaya and Kemal OflazerAnnotation Guidelines for Non-native Arabic Text in the Qatar Arabic Language Bank The Qatar Arabic Language Bank (QALB) is a corpus of naturally written unedited Arabic and its manual edited corrections. QALB has about 1.5 million words of text written and post-edited by native speakers. The corpus was the focus of a shared task on automatic spelling correction in the Arabic Natural Language Processing Workshop that was held in conjunction with 2014 Conference on Empirical Methods for Natural Language Processing (EMNLP) in Doha, with nine research teams from around the world competing. In this poster we discuss some of the challenges of extending QALB to include non-native Arabic text. Our overarching goal is to use QALB data to develop components for automatic detection and correction of language errors that can be used to help Standard Arabic learners (native and non-native) improve the quality of the Arabic text they produce. The QALB annotation guidelines have focused on native speaker text. Learners of Arabic as a second language (L2 speakers) typically have to adapt to a different script and a different vocabulary with new grammatical rules. These factors contribute to the propagation of errors made by L2 speakers that are of different nature than those produced by native speakers (L1 speakers), who are mostly affected by their dialects and levels of education and use of standard Arabic. Our extended L2 guidelines build on our L1 guidelines with a focus on the types of errors usually found in the L2 writing style and how to deal with problematic ambiguous cases. Annotated examples are provided in the guidelines to illustrate the various annotation rules and their exceptions. As with the L1 guidelines, the L2 texts should be corrected with a minimum number of edits that produce semantically coherent (accurate) and grammatically correct (fluent) Arabic. The guidelines also devise a priority order for corrections that prefer less intrusive edits starting with inflection, then cliticization, derivation, preposition correction, word choice correction, and finally word insertion. This project is supported by the National Priority Research Program (NPRP grant 4-1058-1-168) of the Qatar National Research Fund (a member of the Qatar Foundation). The statements made herein are solely the responsibility of the authors.
-
-
-
Truc: Towards Trusted Communication For Emergency Scenarios In Vehicular Adhoc Networks (vanets) Against Illusion Attack
Authors: Maria Elsa Mathew and Arun Raj Kumar P,With data proliferation at an unprecedented rate, data accessibility while in motion has increased the demand of VANETs of late. VANETs use moving cars as nodes to create a mobile network. VANETs are designed with the goals of enhancing driver safety and providing passenger comfort. Providing security to VANETs is important in terms of providing user anonymity, authentication, integrity and privacy of data. Attacks in VANETs include the Sybil attack, DDoS attack, misbehaving and faulty nodes, sinkhole attack, spoofing, traffic analysis attack, position attack and illusion attack. Illusion attack is a recent threat in which the attacker generates fraud traffic messages to mislead the passengers thereby changing the passenger's driving behaviour. Thus, the attacker achieves his goal by moving the target vehicle along a traffic free route and creating traffic jams in areas where he wishes. Illusion attack is devised mainly by thieves and terrorists who require a clean get away path. The existing method used to prevent Illusion attack is the Plausibility Validation Network (PVN). In PVN the message validation is based on a set of rules depending on the message type. This is an overhead as the rule set for all possible message types have to be stored and updated in the database. Hence, an efficient mechanism is required to prevent illusion attacks. In this paper, our proposed system: TRUC verifies the message by a Message Content Validation (MCV) Algorithm thus ensuring the safety of the car drivers and passengers. The possibilities of attacker creating an Illusion attack is explored in all dimensions and the security goals are analyzed by our proposed design, TRUC
-
-
-
Software-hardware Co-design Approach For Gas Identification
Authors: Amine Ait Si Ali, Abbes Amira, Faycal Bensaali, Mohieddine Benammar, Muhammad Hassan and Amine BermakGas detection is one of the major processes that has to be taken into consideration as an important part of a monitoring system for production and distribution of gases. Gas is a critical resource. Therefore, for safety reason, it is imperative to keep monitoring in real time all parameters such as temperature, concentration and mixture. The presented research in this abstract on gas identification is part of an ongoing research project aiming at the development of a low power reconfigurable self-calibrated multi-sensing platform for gas applications. Gas identification system can be described as a pattern recognition problem. Decision tree classifier is a widely used classification technique in data mining due to its low implementation complexity. It is a supervised learning that consists in a succession of splits that leads to the identification of the predefined classes. The decision tree algorithm has been applied and evaluated for the hardware implementation of a gas identification system. The data used for training is collected from a 16-Array SnO2 gas sensor, the sensor array is exposed to three types of gases (CO, Ethanol and H2) at ten different concentrations (20, 40, 60, 80, 100, 120, 140, 160, 180 and 200ppm), the experiment is repeated twice to generate 30 patterns for training and another 30 patterns for testing. Training is performed in MATLAB. It is first done using the raw data, which is the steady states, and then using transformed data by applying principal component analysis. Table 1 shows the decision tree training results. These include the trees obtained from the learning on the original data and on different combinations of principal components. The resulted models are implemented in C and synthesised using Vivado High Level Synthesis (HLS) tool for a quick prototyping on the heterogeneous Zynq platform. Table 2 illustrates the on-chip resources usage, maximum running frequency and latency for the implementation of the trained decision tree models. The use of Vivado HLS helped to optimise the hardware design by applying different directives such as the one that allow loop unrolling for a better parallelism. The best performance is obtained when the first three principal components were used for the training which resulted in an accuracy of 90%. The hardware implementation illustrated that a trade-off is to be found between the accuracy of the identification and the performance in terms of processing time. It is planned to use drift compensation techniques to get more accurate steady states and increase the performance of the system. The system can be easily adapted to other types of gases by exposing the new gas to the sensor array, collecting the data, performing the training and finally implementing the model.
-
-
-
E-government Alerts Correlation Model
Authors: Aadil Salim Al-mahrouqi, Sameh Abdalla and Tahar KechadiBackground & Objectives Qatar's IT infrastructure is rapidly growing to encompass the evolution of businesses and economical growth the country is increasingly witnessing throughout its industries. It is now evident that the country's e-government requirements and associated data management systems are becoming large in number, highly dynamic in nature, and exceptionally attractive for cybercrime activities. Protecting the sensitive data e-government portals are relying on for daily activities is not a trivial task. The techniques used to perform cybercrimes are becoming sophisticated relatively with the firewalls protecting them. Reaching high-level of data protection, in both wired and wireless networks, in order to face recent cybercrime approaches is a challenge that is continuously proven hard to achieve. In a common IT infrastructure, the deployed network devices contain a number of event logs that reside locally within its memory. These logs are in large numbers, and therefore, analyzing them is a time consuming task for network administrators. In addition, a single network event often generates a redundancy of similar event logs that belong to the same class within short time intervals. This makes it difficult to manage them during forensics investigation. In most cybercrime cases, a single alert log does not contain sufficient information about malicious actions background and invisible network attackers. The information for a particular malicious action or attacker is often distributed among multiple alert logs and among multiple network devices. Forensic investigators mission is to reconstruct incident scenarios is now very complex considering the number as well as the quality of these event logs. Methods My research will focus on involving mathematics and algorithm science for each proposed sub models in the alerts correlation model. After collecting alert logs from network sensors; then it will be stored in the alert logs warehouse. The stored alert log contains redundancy data and irrelevant information. The alert correlation model used to filter out all redundancy data and irrelevant information from the alert logs. This model contains two stages; format standardization and redundancy management. The format standardization process aims unified different event logs format into one format, while the redundancy management process aims to reduce the duplication of the single event. Furthermore, this research will try to utilized criminology science to enhance security level of the proposed model and forensics experiments tools to validate the proposal approach. Results In response to attacks and potential of attacks against network infrastructure and assets, my research focuses on how to build an organized legislative e-government environment. The idea of this approach is to forensically utilize the current network security output by collect, analysis and present evidence of network attack in an efficient manner. After data mining process we can utilize our preprocessing results for e-government awareness purpose. Conclusions This research proposed Qatar e-government alerts correlation model. The proposed model used to process and normalize the captured network event logs. The main point of designing the model is to find a way to forensically visualize the evidence and attack scenario in e-government infrastructure.
-
-
-
First Hybrid 1gbps/0.1 Gbps Free-space Optical /rf System Deployment And Testing In The State Of Qatar
Authors: Syed Jawad Hussain, Abir Touati, Mohammad Elamri, Hossein Kazemi, Farid Touati and Murat UysalI.BACKGROUND & OBJECTIVES Owing to its high-bandwidth, robustness to EMI, and operation in unregulated spectrum, free-space optical communication (FSO) is uniquely qualified as a promising alternative or complementary technology to fiber optic and wireless radio-frequency (RF) links. Despite the vibrant advantages of FSO technology and the variety of its applications, its widespread adoption has been hampered by rather disappointing link reliability for long-range links due to atmospheric turbulence-induced fading and sensitivity to detrimental climate conditions. A major challenge of such hybrid systems is to provide a strong backup system with soft-switching capabilities when the FSO link becomes down. The specific objective of this work is to study for the first time in Qatar and the GCC the link capacity, link availability, and link outage of an FSO system with RF back up (i.e. hybrid FSO/RF) under harsh environment. II.METHODS In this work, a practical demonstration of hybrid FSO/RF link system is shown. The system has a capacity of 1 Gbps and 100 Mbps for FSO and RF, respectively. It is installed in Qatar University at two different buildings 600 m away and 20 feet high. This system is basically a point-to-point optical link that uses Infrared laser lights to wirelessly transmit data. Moreover, the proposed system has capability to make parallel transmission between links. In order to analyze the two transport media, we used the tool IPERF. This Java based GUI (jperf) application can either act as a server or client, and is available on a variety of platforms. We have tested end-to-end throughput by running IPERF tool in server mode on one Laptop and in client mode on another. III.RESULTS Figure1 shows a block diagram of the system used. Initial results were obtained for the two links under same climatic and environmental conditions, where the average ambient temperature reached 50°C and RH above 80% (July-August 2014). Both FSO and RF links allowed transfer rates of around 80% of their full capacity. During all experiments while running both links simultaneously, there was no FSO link failure. In case of an FSO failure, the RF is expected to back up within 2 seconds (hard switching), which might cause a loss of data. Detailed results on FSO-to-RF switching and induced packet loss will be reported in the full manuscript and during the presentation. IV.CONCLUSION Tests on FSO/RF link have been carried for the first time in Qatar. Initial results showed that both FSO and RF links operated close to their capacity. During summer, Qatari weather did not induce FSO link outage. The team is focusing on developing a seamless FSO-RF soft switching using NetFPGA boards and raptor coding.
-
-
-
Wigest: A Ubiquitous Wifi-based Gesture Recognition System
Authors: Heba Abdelnasser, Khaled Harras and Moustafa YoussefMotivated by freeing the user from specialized devices and leveraging natural and contextually relevant human movements, gesture recognition systems are becoming popular as a fundamental approach for providing HCI alternatives. Indeed, there is a rising trend in the adoption of gesture recognition systems into various consumer electronics and mobile devices. These systems, along with research enhancing them by exploiting the wide range of sensors available on such devices, generally adopt various techniques for recognizing gestures including computer vision, inertial sensors, ultra-sonic, and infrared. While promising, these techniques experience various limitations such as being tailored for specific applications, sensitivity to lighting, high installation and instrumentation overhead, requiring holding the mobile device, and/or requiring additional sensors to be worn or installed. We present WiGest, a ubiquitous WiFi-based hand gesture recognition system for controlling applications running on off-the-shelf WiFi-equipped devices. WiGest does not require additional sensors, is resilient to changes within the environment, and can operate in non-line-of-sight scenarios. The basic idea is to leverage the effect of the in-air hand motion on the wireless signal strength received by the device from an access point to recognize the performed gesture. As shown in Figure 1, WiGest parses combinations of signal primitives along with other parameters, such as the speed and magnitude of each primitive, to detect various gestures, which can then be mapped to distinguishable application actions. There are several challenges we address WiGest including handling noisy RSSI values due to multipath interference and other electromagnetic noise in the wireless medium; handling gesture variations and their attributes for different humans or the same human at different times; handling interference due to the motion of other people within proximity of the user's device; and finally being energy-efficient to suit mobile devices. To address these challenges, WiGest leverages different signal processing techniques that can preserve signal details while filtering out the noise and variations in the signal. We implement WiGest on off-the-shelf laptops and evaluate frequencies on the RSSI, creating a signal composed of three primitives: rising edge, falling edge, and pause. We evaluate its performance with different users in an apartment and engineering building settings. Various realistic scenarios are tested covering more than 1000 primitive actions and gestures each in the presence of interfering users in the same room as well as other people moving in the same floor during their daily life. Our results show that WiGest can detect the basic primitives with an accuracy of 90% using a single AP for distances reaching 26 ft including through-the-wall non-line-of-sight scenarios. This increases to 96% using three overheard APs, a typical case for many WiFi deployments. When evaluating the system using a multimedia player application case study, we achieve a classification accuracy of 96%.
-
-
-
Physical Layer Security For Communications Through Compound Channels
Authors: Volkan Dedeoglu and Joseph BoutrosSecure communications is one of the key challenges faced in the field of information security as the transmission of information between legitimate users is vulnerable to interception by illegitimate listeners. The state-of-the -art secure communication schemes employ cryptographic encryption methods. However, the use of cryptographic encryption methods requires generation, distribution and management of keys to encrypt the confidential message. Recently, physical layer security schemes that exploit the difference between the channel conditions of the legitimate users and the illegitimate listeners have been proposed for enhanced communication security. We propose novel coding schemes for secure transmission of messages over compound channels that provides another level of security in the physical layer on top of the existing cryptographic security mechanisms in the application layer. Our aim is to provide secrecy against illegitimate listeners while still offering good communication performance for legitimate users. We consider the transmission of messages over compound channels, where there are multiple parallel communication links between the legitimate users and an illegitimate listener intercepts one of the communication links that is unknown to the legitimate users. We propose a special source splitter structure and a new family of low density parity check code ensembles to achieve secure communications against an illegitimate listener and provide error correction capability for the legitimate listener. First, the source bit sequence is split into multiple bit sequences by using a source splitter. The source splitter is required to make sure that the illegitimate listener does not have access to the secret message bits directly. Then, a special error correction code is applied to the bit sequences, which are the outputs of the source splitter. The error correction code is based on a special parity check matrix which is composed of some subblocks with specific degree distributions. We show that the proposed communication schemes can provide algebraic and information theoretic security. Algebraic security means that the illegitimate listener is unable to solve any of the individual binary bits of the secret message. Furthermore, information theoretic security guarantees the highest level of secrecy by revealing no information to the illegitimate listener about the secret message. The error correction encoder produces multiple codewords to be sent on parallel links. Having access to the noisy outputs of the parallel links, the legitimate receiver recovers the secret message. The finite length performance analysis of the proposed secure communications scheme for the legitimate listener shows good results in terms of the bit error rate and the frame error over binary input additive white Gaussian noise channel. The asymptotic performance analysis of our scheme for a sufficiently large block length is found via the density evolution equations. Since the proposed low density parity check code is a multi-edge type code on graphs, there are two densities that characterize the system performance. The thresholds obtained by the density evolution equations of our scheme show comparable or improved results when compared to the fully random low density parity check codes.
-
-
-
Sparsity-aware Multiple Relay Selection In Large Decode-and-forward Relay Networks
Authors: Ala Gouissem, Ridha Hamila, Naofal Al-dhahir and Sebti FoufouCooperative communication is a promising technology that has attracted significant attention recently thanks to its ability to achieve spatial diversity in wireless networks with only single-antenna nodes. The different nodes of a cooperative system can share their resources so that a virtual Multiple Input Multiple Output (MIMO) system is created which leads to spatial diversity gains. To exploit this diversity, a variety of cooperative protocols have been proposed in the literature under different design criteria and channel information availability assumptions. Among these protocols, two of the most-widely used are the amplify-and-forward (AF) and decode-and-forward (DF) protocols. However, in large-scale relay networks, the relay selection process becomes highly complex. In fact, in many applications such as device-to-device (D2D) communication networks and wireless sensor networks, a large number of cooperating nodes are used, which leads to a dramatic increase in the complexity of the relay selection process. To solve this problem, the sparsity of the relay selection vector has been exploited to reduce the multiple relay selection complexity for large AF cooperative networks while also improving the bit error rate performance. In this work, we extend the study from AF to large-scale decode-and-forward (DF) relay networks. Based on exploiting the sparsity of the relay selection vector, we propose and compare two different techniques (referred to as T1 and T2) that aim to improve the performance of multiple relay selection in large-scale decode-and-forward relay networks. In fact, when only few relays are selected from a large number of relays, the relay selection vector becomes sparse. Hence, utilizing recent advances in sparse signal recovery theory, we propose to use different signal recovery algorithms such as the Orthogonal Matching Pursuit (OMP) to solve the relay selection problem. Our theoretical and simulated results demonstrate that our two proposed sparsity-aware relay selection techniques are able to improve the outage performance and reduce the computation complexity at the same time compared with conventional exhaustive search (ES) technique. In fact, compared to ES technique, T1 reduces the selection complexity by O(K^2 N) (where N is the number of relays and K is the number of selected relays) while outperforming it in terms of outage probability irrespective of the relays' positions. Technique T2 provides higher outage probability compared to T1 but reduces the complexity making a compromise between complexity and outage performance. The best selection threshold for T2 is also theoretically calculated and validated by simulations which enabled T2 to also improve the outage probability compared with ES techniques. Acknowledgment This publication was made possible by NPRP grant #6-070-2-024 from the Qatar National Research Fund (a member of Qatar Foundation). The statements made herein are solely the responsibility of the authors.
-
-
-
Inconsistencies Detection In Islamic Texts Of Law Interpretations ["fatawas"]
Authors: Jameela Al-otaibi, Samir Elloumi, Abdelaali Hassaine and Ali Mohamed JaouaIslamic web content offers a very convenient way for people to learn more about Islam religion and the correct practices. For instance, via these web sites they could ask for fatwas (Islamic advisory opinion) with more facilities and serenity. Regarding the sensitivity of the subject, large communities of researchers are working on the evaluation of these web sites according to several criteria. In particular there is a huge effort to check the consistency of the content with respect to the Islamic shariaa (or Islamic law). In this work we are proposing a semiautomatic approach for evaluating the web sites Islamic content, in terms of inconsistency detection, composed of the following steps: (i) Domain selection and definition: It consists of identifying the most relevant named entities related to the selected domain as well as their corresponding values or keywords (NEV). At that stage, we have started building the Fatwas ontology by analyzed around 100 fatwas extracted from the online system. (ii) Formal representation of the Islamic content: It consists of representing the content as formal context relating fatwas to NEV. Here, each named entity is split into different attributes in the database where each attribute is associated to a possible instantiation of the named entity. (iii) Rules extraction: by applying the ConImp tools, we extract a set of implications (or rules) reflecting cause-effect relations between NEV. As an extended option aiming to provide more precise analysis, we have proposed the inclusion of negative attributes. For example for word "licit", we may associate "not licit" or "forbidden", for word "recommended" we associated "not recommended", etc. At that stage by using an extension of Galois Connection we are able to find different logical associations in a minimal way by using the same tool ConImp. (iv) Conceptual reasoning: the objective is to detect a possibly inconsistency between the rules and evaluate their relevance. Each rule is mapped to a binary table in a relational database model. By joining obtained tables we are able to detect inconsistencies. We may also check if a new law is not contradicting existing set of laws by mapping the law into a logical expression. By creating a new table corresponding to its negation we have been able to prove automatically its consistencies as soon as we obtain an empty join of the total set of joins. This preliminary study showed that the logical representation of fatwas gives promising results in detecting inconsistencies within fatwa ontology. Future work includes using automatic named entity extraction and automatic transformation of law into a formatted database; we should be able to build a global system for inconsistencies detection for the domain. ACKNOWLEDGMENT: This publication was made possible by a grant from the Qatar National Research Fund through National Priority Research Program (NPRP) No. 06-1220-1-233. Its contents are solely the responsibility of the authors and do not necessarily represent the official views of the Qatar National Research Fund or Qatar University.
-
-
-
A Low Power Reconfigurable Multi-sensing Platform For Gas Application
Presence of toxic gases and accidental explosions in gas industries have turned the researcher to innovate an electronic nose system which can indicate the nature and the parameters of the gas passing through different vessels. Therefore, in this research we propose a low power Radio Frequency Identification (RFID) based gas sensor tag which can monitor the parameters and indicate the type of gas. The research work is divided in to three main parts. The first two parts cover the design and analysis of low power multi-sensors and processing unit, while the last part focuses on a passive RFID module which can provide communication between the sensor and the processing unit, as shown in Fig. 1. In passive RFID applications, power consumption is one of the most prominent parameter because most of the power is harvested from the coming RF signal. Therefore a ring-oscillator based low power temperature sensor is designed to measure the gas thermodynamic conditions. The oscillator is designed using the Thyristor based delay element [7], in which the current source present for temperature compensation has been displaced to make the delay element as temperature dependent. The proposed temperature sensor consumes 47nW power at 27 °C, which increases linearly with temperature. Moreover, a 4x4 array of tin-oxide gas sensor based on convex Micro hotplates (MHP), is also utilized to identify the type of gas. The array is designed such that each sensor of an array provide different pattern for the same gas. The power consumption caused by the temperature and gas sensor is in the order of few µW's. The prime advantage of MHP can be visualized by the 950 °C annealed MHP, which exhibit the thermal efficiency of 13 °C /mW. Moreover it requires a driving voltage of only 2.8V to reach 300 °C in less than 5ms, which make it compatible with power supplies required by CMOS ICs. The gas sensor will provide 16 feature points at a time, which can results in hardware complexity and throughput degradation of the processing unit. Therefore, a principle component analysis (PCA) algorithm is implemented to reduce the number of feature points. Thereafter, a binary decision tree algorithm is adopted to classify the gases. We implemented both algorithms on heterogeneous Zynq platform. It is observed that the execution of PCA on Zynq programmable SoC is 1.41 times faster than the corresponding software execution, with a resource utilization of only 23% . Finally, a passive ultrahigh-frequency (UHF) RFID transponder is developed for communicating between the sensing block and processing unit. The designed module is responsible to harvest the power from the coming RF signal and accomplish the power requirement of both sensors. The designed transponder IC achieved minimum sensitivity of -17dBm with a minimum operational power of 2.6µW.
-
-
-
Utilizing Monolingual Gulf Arabic Data For Qatari Arabic-english Statistical Machine Translation
Authors: Kamla Al-mannai, Hassan Sajjad, Alaa Khader, Fahad Al Obaidli, Preslav Nakov and Stephan VogelWith the recent rise of social media, Arabic speakers have started increasingly using dialects in writing, which has constituted research in dialectal Arabic (DA) as a field of interest in natural language processing (NLP). DA NLP is still in its infancy, both in terms of its computational resources and in its tools, e.g. lack of dialectal morphological segmentation tools. In this work, we present a 2.7M-token collection of monolingual corpora of Gulf Arabic extracted from the Web. The data is unique since it is genre-specific, i.e. romance genre, in spite of the various sub-dialects of Gulf Arabic that it covers, e.g., Qatari, Emirati, Saudi. In addition to the monolingual Qatari data collected, we use existing parallel corpora of Qatari (0.47M-token), Egyptian (0.3M-token), Levantine (1.2M-token) and Modern Standard Arabic (MSA) (3.5M-token) to English to develop a Qatari Arabic to English statistical machine translation system (QA-EN SMT). We exploit the monolingual data to 1) develop a morphological segmentation tool for Qatari Arabic, 2) generate a uniform segmentation scheme for the various variants of Arabic employed, and 3) build a Qatari language model in the opposite translation direction. Proper morphological segmentation of Arabic plays a vital role in the quality of a SMT system. Using the monolingual Qatari data collected in combination with the QA side of the small QA-EN existing parallel data, we trained an unsupervised morphological segmentation model for Arabic, i.e. Morfessor, to create a word segmenter for Qatari Arabic. We then extrinsically compare the impact of the resulting segmentation (as opposed to using tools for MSA) on the quality of QA-EN machine translation. The results show that this unsupervised segmentation can yield better translation quality. Unsurprisingly, we found that removing the monolingual data from the training set of the segmenter affects the translation quality with a loss of 0.9 BLEU points. Arabic dialect resources, when adapted for the translation of one dialect are generally helpful in achieving better translation quality. We show that a standard segmentation scheme can improve vocabulary overlap between dialects by segmenting words with different morphological forms in different dialects to a common root form. We train a generic segmentation model for Qatari Arabic and the other variants used using the monolingual Qatari data and the Arabic side of the parallel corpora. We train the QA-EN SMT system using the different parallel corpora (one at a time) in addition to the QA-EN parallel corpus segmented using the generic statistical segmenter. We show a consistent improvement of 1.5 BLEU points when compared with their respective baselines with no segmentation. In the reverse translation direction, i.e. EN_QA, we show that adding a small amount of in-domain data to the language model used results in a relatively large improvement compared to the degradation resulted by adding a large amount of out-of-domain data.
-
-
-
Securing The E-infrastructure In Qatar Through Malware Inspired Cloud Self-protection
Authors: Elhadj Benkhelifa and Thomas WelshWhilst the state of security within the Cloud is still a contentious issue, some privacy and security issues are well known or deemed to be a likely threat. When considering the ongoing threat of malicious insiders the promised security expertise might be deemed untrusted. The focus of our research is determining the extent of issues related to the underlying technology, which support Cloud environments, mainly virtualization platforms. It is often argued that virtualization is secure over conventional shared resources due to the inherent isolation. However much literature may be seen which cites examples to the contrary and as such it should be considered that, as with all software, virtualization applications are susceptible to exploitation and subversion. In fact, it might even be argued that the complexity and heterogeneous nature of the environment may even facilitate further security vulnerabilities. To illustrate and investigate this point we consider the security threat of malware within the context of cloud environments. With this evolution of malware combined with the knowledge that Cloud software is susceptible to vulnerabilities, it is argued that complex malware might exist for the Cloud and if it were successful, would shed light on the security of these technologies. Whilst there are many examples of state of the art malware detection and protection for Cloud environments, this work tends to focus on examining virtual machines (VMs) from another layer. The primary flaw identified in all of the current approaches is failing to take into account malware, which is aware of the Cloud environment; thus be in a position to subvert this detection process. Traditional malware security applications tend to take a defensive approach by looking for existing malware through signature analysis or behavior monitoring. Whilst such approaches are acceptable for traditional environments they become less effective for distributed and dynamic ones. We argued that due to this dynamic nature of the Cloud as well as its uncertain security concerns, a malware type application may be a suitable security defense and thus operate as a proactive, self-protecting element. We present an architecture for Multi-Agent Cloud-Aware Self-Propagating Agents for Self-Protection. By adapting this architecture to include constraints (such as a kill switch) the application may be effectively controlled and thus any negative effects minimized. This application will then cross the multiple layers within the network, having high privilege. Its dynamic and distributed architecture will allow it survive removal from malware whilst hunting down malicious agents and patching systems as necessary. In order to survive in the hostile and dynamic cloud environment, the software incorporates a multi-component and multi-agent architecture which has shown success in the past with malware that propagate in heterogeneous environments. The components consist of passive and active sensors to learn about the environment, distributed storage to provide redundancy and controller/constructor agents for localized coordination. The proposed architecture has been implemented with success and desired results were achieved. The research outputs hold significant potentials, particularly for complex and highly dynamic infrastructures such as those aimed for DigitalQatar.
-
-
-
Watch Health (w-health)
More LessSmart watches have been available for quite some time now. However,with the announcement of"Apple Watch"by Apple Inc., a strong buzz about smart watches has been created. A highly anticipated success of Apple watches can also be linked to an expected increase in the smart watch market in the very near future. Apart from Apple, other big companies such as Sony, Motorola, Nike have their own brand of smart watches. Therefore, a strong market competition would arise leading to competitive prices, technologies, and designs which would possibly lead to increased popularity of smart watches. Following the recent announcement of apple watch, several online and newspaper articles have suggested that the most important application of Apple watch would be in the field of healthcare. This is also backed by the applications available in the Apple watch which includes GPS tracking, gyroscope, accelerometer, pulse monitor, calorie counter, activity tracker, Siri voice tracker, and host of various other applications. Further, the Apple watch is backed by powerful operating systems and hardware processors. This buzz about the smart watches arises one question - How effective can these smart watches be used for providing healthcare solutions? The use of smart devices for healthcare services has been a topic of extensive research for the last decade which has resulted in developing several healthcare solutions for various types of disease management, patient monitoring especially for chronic lifetime diseases and old people. With the emergence of smart watches, now it is time to further explore the possibility of using smart watches for healthcare services. Some of the advantages of smart watches for healthcare services are: ?They are easily wearable and portable and can almost be a part of everyday attire similar to regular watches. ?They are relatively cheaper than other smart devices such as smart mobile phones and similar gadgets. ?With the advancements in hardware and software technologies they are now as powerful as a high-end smart phone and can host several types of applications. ?They can be adapted and customised to provide various disease diagnosis and management according to the individual patient needs. ?The smart watches can include several sensors and also provide a platform for software applications for patient health monitoring and reporting. ?With the use of voice based applications such as Siri, patients having difficulty to use modern gadgets or to read and write can also use the device more easily. There is no doubt that iWatch or other smart watches not only provide numerous possibilities of adapting and implementing existing smart healthcare solutions but also a new platform for developing novel solutions for healthcare. The research and development in the current mobile-health solutions should now also focus their attention towards research in developing smart watches based healthcare solutions. Hence, watch health (w-health) is a division of electronic health practice, which is defined as "a medical practice supported with the use of smart watches technology which includes smart watches, smart mobile phones, wireless devices, patent monitoring devices and many others"
-
-
-
An Enhanced Locking Range 10.5-ghz Divide-by-3 Injection-locked Frequency Divider With Even-harmonic Phase Tuning In 0.18-um Bicmos
Authors: Sanghun Lee, Sunhwan Jang and Cam NguyenA frequency divider is one of the key building blocks in phased-lock loops (PLL) and frequency synthesizers, which are essential subsystems in wireless communication and sensing systems. A frequency divider divides the output frequency of a voltage-controlled oscillator (VCO) in order to compare it with a reference clock signal. Among the different types of frequency dividers, the injection-locked frequency divider (ILFD) is becoming more popular due to its low power and high frequency characteristics. However, the ILFD has an inherent problem of narrow locking range, which basically defines a range over which a frequency-division operation is supported. In order to increase the locking range, one of the most obvious ways is to inject higher power. In fully integrated PLLs, however, the injection signal is supplied by an internal VCO, which typically has limited fixed output power and hence enhancing the locking range with large injection signal power poses difficulty. In this work, we present the development of a fully integrated 10.5-GHz divide-by-3 (1/3) injection-locked frequency divider (ILFD) that can provide extra locking range with a small fixed injection power. The ILFD consists of a previously measured on-chip 10.5-GHz VCO functioning as an injection source, a 1/3 ILFD core, and an output inverter buffer. A phase tuner implemented using an asymmetric inductor is proposed to increase the locking range through even-harmonic (the 2rd harmonic for this design) phase tuning. With a fixed internal injection signal power of only -18 dBm (measured output power of the standalone VCO with a 50-? reference), a 25% enhancement in the locking range from 12 to 15 MHz is achieved with the proposed phase tuning technique without consuming an additional DC power. The frequency tuning range of the integrated 1/3 ILFD is from 3.3 GHz to 4.2 GHz. The proposed 1/3 ILFD is realized in a 0.18-µm BiCMOS process, occupies 0.6 mm × 0.7 mm, and consumes 10.6 mA while the ILFD alone consumes 6.15 mA from 1.8-V supply. The main objective of this work is proposing a new technique of phase-tuning of the even-harmonics that can "further" increase the locking range with an "extra" amount beyond what can be achieved by other techniques. Since the developed technique can enhance the locking range further at a fixed injection power, it can be used in conjunction with other techniques for further enhancing the locking range. For instance, the locking range can be increased by using different injection powers and then further enhanced by tuning the phase of the even harmonics at each power level. The "extra" locking range, not achievable without the even-harmonic phase tuning, amounts to 25%, which is very attractive for PLL applications. Furthermore, additional tuning mechanisms such as use of a capacitor bank can be employed to achieve even wider tuning range for applications such as PLL.
-
-
-
Exprimental Study Of Mppt Algorithms For Pv Solar Water Pumping Applications
By Badii GmatiThe energy utilization efficiency of commercial photovoltaic (PV) pumping systems can be significantly improved by employing many MPPT methods available in the literature such as the constant voltage, short-Current Pulse, Open Voltage, Perturb and Observe, Incremental-Conductance and non-linear methods (Fuzzy Logic and Neural Networks). This paper presents a detailed experimental study of the two DSP implementation techniques: Constant Voltage and Perturb and Observe "P&O" used for standalone PV pumping systems. The influence of algorithm parameters on system behavior is investigated and the various advantages and disadvantages of each technique are identified for different weather conditions. Practical results obtained using dSpace DS1104 show excellent performance and optimal operating system is attained regardless of climate change.
-
-
-
Geometrical Modeling And Kinematic Analysis Of Articulated Tooltips Of A Surgical Robot
INTRODUCTION: The advent of da Vinci surgical robot (Intuitive Surgical, California, USA) has allowed complex surgical procedures in urology, gynecology, cardiothoracic, and pediatric to be performed with better clinical outcomes. The end effectors of these robots exhibits enhanced dexterity with improved range of motion leading to better access and precise control during the surgery. Understanding the design and kinematics of these end effectors (imitating surgical instruments' tooltips) would assist in replication of their complex motion in a computer-generated environment. This would further support the development of computer-aided robotic surgical applications. The aim of this work is to develop a software framework comprising of the geometric three-dimensional models of the surgical robot tool-tips along with their kinematic analysis. METHODS: The geometric models of the surgical tooltips were designed based on the EndoWristTM instruments of the da Vinci surgical robot. Shapes of the link and inter-link distances of the EndoWristTM instrument were measured in detail. A three-dimensional virtual model was then recreated using CAD software (Solidworks, Dassault Systems, Massachusetts, USA). The kinematic analysis was performed considering trocar as the base-frame for actuation. The actuation mechanism of the tool composed of a prismatic joint (T1) followed by four revolute joints (Ri ; i = 1 to 4) in tandem (Figure 1). The relationship between the consequent joints was expressed in form of transformation matrices using Denavit-Hartenberg (D-H) convention. Equations corresponding to the forward and the inverse kinematics were then computed using D-H parameters and applying geometrical approach. The kinematics equations of the designed tooltips were implemented through a modular cross-platform software framework developed using C/C++. In the software, the graphical rendering was performed using openGL and a multi-threaded environment was implemented using Boost libraries. RESULTS AND DISCUSSION: Five geometric models simulating the articulated motion of the EndoWristTM instruments were designed (Figure 2). These models were selected based on the five basic interactions of the surgical tooltip with the anatomical structures, which included: Cauterization of the tissue, Stitching using needles, Applying clips on vascular structures, Cutting using scissors, and Grasping of the tissues. The developed software framework, which includes kinematics computation and graphical rendering of the designed components, was evaluated for applicability in two scenarios (Figure 3). The first scenario demonstrates the integration of the software with a patient-specific simulator for pre-operative surgical rehearsal and planning (Figure 3a). The second scenario shows the applicability of the software in generation of virtual overlays of the tooltips superimposed with the stereoscopic video stream and rendered on the surgeon's console of the surgical robot (Figure 3b). This would further assist in development of vision-based guidance for the tooltips. CONCLUSION: The geometrical modeling and kinematic analysis allowed the generation of the motion of the tooltips in a virtual space that could be used for both pre-operatively and intra-operatively, before and during the surgery, respectively. The resulting framework can also be used to simulate and test new tooltip designs.
-
-
-
Relate-me: Making News More Relevant
Authors: Tamim Jabban and Ingmar WeberTo get readers of international news stories interested and engaged, it is important to show how a piece of far-away news relates to them and how it might affect their own country. As a step in this direction, we have developed a tool to automatically create textual relations between news articles and readers. To produce such connections, the first step is to detect the countries mentioned in the article. Many news sites, including Al Jazeera (http://www.aljazeera.com), use automated tools such as OpenCalais (http://opencalais.com) to detect place references in a news article and list those as a list of countries in a dedicated section at the bottom (say, List A: [Syria, Germany]). If not already included, relevant countries could be detected using existing tools and dictionaries. The second step is to use the reader's IP address to infer the country they are currently located in (say, Country B: Qatar). Knowing this country gives us a "bridge" to the reader as now we can try relate the countries from List A, to the reader's country, Country B. Finally, we have to decide which type of contextual bridges to build between the pairs of countries. Currently, we are focusing on four aspects: 1) Imports & Exports: this section displays imports and exports between the Country B and List A, if any. For instance: "Qatar exports products worth $272m every year to Germany, 0.27% of total exports." Upon clicking on this information, it will redirect the user to another website, showing a breakdown of these imports and exports. 2) Distances: this simply states the direct distance in kilometers from Country B to every country in List A. For instance: "Syria is 2,110km away from Qatar." Upon clicking on this information, it will navigate to a Google Maps display, showing this distance. 3) Relations: this provides a link to the designated Wikipedia page between Country B and every country in List A. For instance, "Germany - Qatar Relations (Wikipedia)." It also shows a link relating the countries using Google News: "More on Qatar - Germany (Google News)" 4) Currencies: this shows the currency conversions between Country B's currency and every other country in List A, for instance: "1 QAR = 0.21 EUR (Germany's currency)." Our current tool, which will be demonstrated live during the poster presentation, was built using JavaScript. With the use of tools such as Greasemonkey, this allowed us to test and display the results of the project on Al Jazeera (http://www.aljazeera.com), without having site ownership. We believe that the problem of making connections between countries explicit is of particular relevance to small countries such as Qatar. Whereas usually a user from, say, Switzerland, might not be interested in events in Qatar, showing information regarding trade between the two could change their mind.
-
-
-
Energy And Spectrally Efficient Solutions For Cognitive Wireless Networks
Authors: Zied Bouida, Ali Ghrayeb and Khalid QaraqeAlthough different spectrum bands are allocated to specific services, it has been identified that these bands are unoccupied or partially used most of the time. Indeed, recent studies show that 70% of the allocated spectrum is not utilized. As wireless communication systems evolve, an efficient spectrum management solution is required in order to satisfy the need of current spectrum-greedy applications. In this context, cognitive radio (CR) has been proposed as a promising solution to optimize the spectrum utilization. Under the umbrella of cognitive radio, spectrum-sharing systems allow different wireless communication systems to coexist and cooperate in order to increase their spectral efficiency. In these spectrum-sharing systems, primary (licensed) users and secondary (unlicensed) users are allowed to coexist in the same frequency spectrum and transmit simultaneously as long as the interference of the secondary user to the primary user stays below a predetermined threshold. Several techniques have been proposed in order to meet the required quality of service of the secondary user while respecting the primary user's constraints. While these techniques, including multiple-input multiple-output (MIMO) solutions, are optimized from a spectrally-efficiency perspective, they are generally not well designed to address the related complexity and power consumption issues. Thus, the achievement of high data rates with these techniques comes at the expense of high-energy consumption and increased system complexity. Due to these challenges, a trade-off between spectral and energy efficiencies has to be considered in the design of future transmission technologies. In this context, we have recently introduced adaptive spatial modulation (ASM), which comprises both adaptive modulation (AM) and spatial modulation (SM), with the aim of enhancing the average spectral efficiency (ASE) of multiple antenna systems. This technique was shown to offer high energy efficiency and low system complexity thanks to the use of SM while achieving high data rates thanks to the use of AM. Motivated by this technique and the need of such performance in a CR scenario, we study in this abstract the concept of ASM in spectrum sharing systems. In this work, we propose the ASM-CR scheme as an energy-efficient, spectrally-efficient, and low-complexity scheme for spectrum sharing systems. The performance of the proposed scheme is analyzed in terms of ASE and average bit error rate and confirmed with selected numerical results using Monte-Carlo simulations. These results confirm that the use of such techniques comes with an improvement in terms of spectral efficiency, energy efficiency, and overall system complexity.
-
-
-
Spca: Scalable Principle Component Analysis For Big Data
More LesssPCA: Scalable Principle Component Analysis for Big Data Web sites, social networks, sensors, and scientific experiments today generate massive amounts of data i.e Big Data. Owners of this data strive to obtain insights from it, often by applying machine learning algorithms.Thus, designing scalable machine learning algorithms that run on a cloud computing infrastructure is an important area of research. Many of these algorithms use the MapReduce programming model. In this poster presentation, we show that MapReduce machine learning algorithms often face scalability bottlenecks, commonly because the distributed MapReduce algorithms for linear algebra do not scale well. We identify several optimizations that are crucial for scaling various machine learning algorithms in distributed settings. We apply these optimizations to the popular Principal Component Analysis (PCA) algorithm. PCA is an important tool in many areas including image processing, data visualization, information retrieval, and dimensionality reduction. We refer to the proposed optimized PCA algorithm as sPCA. sPCA is implemented in the MapReduce framework. It achieves scalability via employing efficient large matrix operations, effectively leveraging matrix sparsity, and minimizing intermediate data.Experiments show that sPCA outperforms the PCA implementation in the popular Mahout machine learning library by wide margins in terms of accuracy, running time, and volume of intermediate data generated during the computation. For example, on a 94 GB dataset of tweets from Twitter, sPCA achieves almost 100% accuracy and it terminates in less than 10,000 s (about 2.8 hours), whereas the accuracy of Mahout PCA can only reach up to 70% after running for more than 259,000 s (about 3 days). In addition,both sPCA and Mahout PCA are iterative algorithms, where the accuracy improves by running more iterations until a target accuracy is achieved. In our experiments, when we fix the target accuracy at 95%, Mahout PCA takes at least two orders of magnitudes longer than sPCA to achieve that target accuracy. Furthermore, Mahout PCA generates about 961 GB of intermediate data, whereas sPCA produces about 131 MB of such data, which is a factor of 3,511x less data. This means that, compared to Mahout PCA, sPCA can achieve more than three orders of magnitudes saving in network and I/O operations, which enables sPCA to scale well.
-
-
-
How To Improve The Health Care System By Predicting The Next Year Hospitalization
By Dhoha AbidIt is very common to study the patient's data hospitalization to get useful information to improve the health care system. According to the American Hospital Association, in 2006, over $30 billion was spent on unnecessary hospital admissions. If patients that are likely to be hospitalized can be identified, the admission will be avoided as they will get the necessary treatments earlier. In this context, in 2013, the Heritage Provider Network (HPN) launched the $3 million Heritage Health Prize in order to develop a system that uses the available patient data (health records and claims)to predict and avoid unnecessary hospitalizations. In this work we take this competition data, and we try to predict the patient's hospitalization number. The data encompasses more than 2,000,000 of patient admission records over three years. The aim is to use the data of the ?rst and second year to predict the number of hospitalization of the third year. In this context, a set of operations mainly: data transformation, outlier detection, clustering, and regression algorithms are applied. Data transformation operations are mainly: (1) As the data is big enough to be processed, dividing the data into chunks is mandatory. (2) Missing values are either replaced or removed. (3) As the data is raw and cannot be labeled, different operations of aggregation are applied. After transforming the data, outlier detection, clustering, and regression algorithms are applied in order to predict the third year hospitalization number for each patient. Results show, by applying directly regression algorithms, the relative error is only 79%. However, by applying the DBSCAN clustering algorithm followed by the regression algorithm, the relative error decreased to be 67%. This is because the attribute that has been generated by the pre-processing clustering step helped the regression algorithm to predict more accurately the number of hospitalization; and this is why the relative error has dropped. The relative error can be decreased more if we apply the clustering pre-processing step twice. That means, the clusters generated in the first clustering step are re-clustered to generated sub-clusters. Then, the regression algorithm is applied to these sub-clusters. The relative error dropped significantly from 67% to 32%. Patients share common hospitalization history are grouped into clusters. This clustering information is used to enhance the regression results.
-