- Home
- Conference Proceedings
- Qatar Foundation Annual Research Conference Proceedings
- Conference Proceeding
Qatar Foundation Annual Research Conference Proceedings Volume 2018 Issue 3
- Conference date: 19-20 Mar 2018
- Location: Qatar National Convention Center (QNCC), Doha, Qatar
- Volume number: 2018
- Published: 15 March 2018
61 - 77 of 77 results
-
-
Applied Internet of Things IoT: Car monitoring system for Modeling of Road Safety and Traffic System in the State of Qatar
Authors: Rateb Jabbar, Khalifa Al-Khalifa, Mohamed Kharbeche, Wael Alhajyaseen, Mohsen Jafari and Shan JiangOne of the most interesting new approaches in the transportation research field is the Naturalistic Driver Behavior which is intended to provide insight into driver behavior during everyday trips by recording details about the driver, the vehicle and the surroundings through an unobtrusive data gathering equipment and without experimental control. In this paper, an Internet of Things solution that collects and analyzes data based on Naturalistic Driver Behavior approach is proposed. The analyzed and collected data will be used as a comprehensive review, and analysis of the existing Qatar traffic system, including traffic data infrastructure, safety planning, engineering practices and standards. Moreover, data analytics for crash prediction and the use of these predictions for the purpose of systemic and systematic network hotspot analysis, risk-based characterization of roadways, intersections, and roundabouts are developed. Finally, an integrated safety risk solution was proposed. This latter, enables decision makers and stakeholders (road users, state agencies, and law enforcement) to identify both high-risk locations and behaviors by measuring a set of dynamic variables including event-based data, roadway conditions, and driving maneuvers. More specifically, the solution consists of a driver behaviors detector system that uses mobile technologies. The system can detect and analyze several behaviors like drowsiness and yawning. Previous works are based on detecting and extracting facial landmarks from images. However, the new suggested system is based on a hybrid approach to detect driver behavior utilizing a deep learning technique using a multilayer perception classifier. In addition, this solution can also collect data about every day trips like start time, end time, average speed, maximum speed, distance and minimum speed. Furthermore, it detects for every fifteen seconds measurements like GPS position, distance, acceleration and rotational velocity along the Roll, Pitch and Yaw axes. The main advantage of the solution is to reduce safety risks on the roads while optimizing safety mitigation costs to a society. The proposed solution has three-layer architecture, namely, the perception, network, and application layers as detailed below. I. The perception layer is the physical layer, composed from several Internet of Thing devices that uses mainly use the smart phones equipped with cameras and sensors (Magnetometer, Accelerometers Gyroscope and Thermometer, GPS sensor and Orientation sensor) for sensing and gathering information about the driver behavior roads and environment as shown in Fig. 1. II. The network layer is responsible for establishing the connection with the servers. Its features are also used for transmitting and processing sensor data. In this solution, hybrid system that collect data and store them locally before sending them to the server is used. This technique proves its efficiency in case of Poor Internet coverage and unstable Internet connection. III. The application layer is responsible for delivering application specific services to end user. It consists in sending the data collected to web server in order to be treat and analyzed before displaying it to the final end user. The web service which part of the application layer is the component responsible for collecting data not only from devices but also from other sources such General Traffic Directorate at Minister of Interior to gather the crash details. This web service stocks all stored data in database server and analyses them. Then, the stored data and analysis will be available for end user via website that has direct access to the web services. Figure 1: Architecture of Car monitoring system Keywords: Driver Monitoring System, DrowsinessDetection, Deep Learning, Real-time Deep Neural Network, Fig. 1: Architecture of IoT solution Keywords: Driver Monitoring System, Drowsiness Detection, Deep Learning, Real-time Deep Neural Network,
-
-
-
Importance of CapabilityDriven Requirements for Smart City Operations
Capability oriented requirements engineering is an emerging research area where designers are faced with the challenge of analyzing changes in the business domain, capturing user requirements, and developing adequate IT solutions taking into account these changes and answering user needs. In this context, researching the interplay between design-time and run-time requirements with a focus on adaptability is of great importance. Approaches to adaptation in the requirements engineering area consider issues underpinning the awareness of requirements and the evolution of requirements. We are focusing on researching the influence of capability-driven requirements on architectures for adaptable systems to be utilized in smart city operations. We investigate requirements specification, algorithms, and prototypes for smart city operations with a focus on intelligent management of transportation and on validating the proposed approaches. In this framework, we conducted a systematic literature review (SLR) of requirements engineering approaches for adaptive system (REAS). We investigated the modeling methods used, the requirements engineering activities performed, the application domains involved, and the deficiencies that need be tackled (in REAS in general, and in SCOs in particular). We aimed at providing an updated review of the state of the art in order to support researchers in understanding trends in REAS in general, and in SCOs in particular. We also focused on the study of Requirement Traceability Recovery (RTR). RTR is the process of constructing traceability links between requirements and other artifacts. It plays an important role in many parts of the software life-cycle. RTR becomes more important and exigent in the case of systems that change frequently and continually, especially adaptive systems where we need to manage the requirement changes in such systems and analyze their impact. We formulated RTR as a mono and a multi-objective search problem using a classic Genetic Algorithm (GA) and a Non-dominated Sorting-based Genetic Algorithm (NSGA-II) respectively. The mono-objective approach takes as input the software system, a set of requirements and generates as output a set of traces between the artifacts of the system and the requirements introduced in the input. This is done based on the textual similarity between the description of the requirements and the artifacts (name of code elements, documentation, comments, etc.). The multi-objective approach takes into account three objectives, namely, the recency of change, the frequency of change, and the semantic similarity between the description of the requirement and the artifact. To validate the two approaches, we used three different open source projects. The reported results confirmed the effectiveness of the two approaches in correctly generating the traces between the requirements and artifacts with high precision and a recall. A comparison between the two approaches shows that the multi-objective approach is more effective than the mono-objective one. We also proposed an approach aiming at optimizing service composition in service-oriented architectures in terms of security goals and cost using NSGA-II in order to help software engineers to map the optimized service composition to the business process model based on security and cost. To do this, we adapted the DREAD model for security risk assessment by suggesting new categorizations for calculating DREAD factors based on a proposed service structure and service attributes. To validate the proposal, we implemented the YAFA-SOA Optimizer. The evaluation of this optimizer shows that risk severity for the generated service composition is less than 0.5, which matches the validation results obtained from a security expert. We also investigated requirements modeling for an event with a large crowd using the capability-oriented paradigm. The motivation was the need for the design of services that meet the challenges of alignment, agility, and sustainability in relation to dynamically changing enterprise requirements especially in large-scale events such as sports events. We introduced the challenges to stakeholders involved in this process and advocated a capability-oriented approach for successfully addressing these challenges. We also investigated a multi-type, proactive and context-aware recommender system in the environment of smart cities. The recommender system recommends gas stations, restaurants, and attractions, proactively, in an internet of things environment. We used a neural network to do the reasoning and validated the system on 7000 random contexts. The results are promising. We also conducted a user's acceptance survey (on 50 users) that showed satisfaction with the application. We also investigated capturing uncertainty in adaptive intelligent transportation systems, which need to monitor their environment at run-time and adapt their behavior in response to changes in this environment. We modelled an intelligent transportation case study using the KAOS goal model and modelled uncertainty by extending our case study using variability points, and hence having different alternatives to choose from depending on the context at run-time. We handled uncertainty by molding our alternatives using ontologies and reasoning to select the optimal alternative at run-time when uncertainty occurs. We also devised a framework, called Vehicell that exploits 5G mobile communication infrastructures to increase the effectiveness of vehicular communications and enhance the relevant services and applications offered in urban environments. This may help in solving some of the mobility problems, and smooth the way for innovative services to citizens and visitors and improve the overall quality of life.
-
-
-
Secure RF Energy Harvesting Scheme for Future Wireless Networks
More LessWireless communication is shaping the future of seamless and reliable connectivity of billions of devices. The communication sector in Qatar is mainly driven by rising demand of higher data rates and uninterrupted connectivity of wireless devices. Wireless fidelity (Wi-Fi), cellular telephone and computer interface devices (e.g. Bluetooth) are a few of the commonly used applications for wireless distribution of information in Qatar. According to analysts, strong growth of Islamic banking, increase in IT consolidation and increased adoption of mobility solutions are some of the key contributors to the growth of digital infrastructure in Qatar. Modernization of legacy infrastructure is another focal point of government of Qatar to enable e-government initiative in rural areas and Tier II/ III cities of Qatar which has long term effects in various domains such as health, electricity, water, heat, communication and trade. Considering this exponential rise of wireless communication in Qatar, a great deal of research is being done from the perspective of secure deployment of new wireless networks. There is also a growing demand to develop more energy-efficient communication techniques to reduce the consumption of fossil fuel without compromising the quality of experience of users. This is also beneficial for solving economic issues that cellular operators are faced with the ever growing number of users. Moreover, with the upcoming FIFA world cup in 2022, millions of dollars are being spent to enhance the capacity and security of existing and upcoming communication networks. However, with greater connectivity and ultimate functionality come several important challenges. The first challenge is the security of data or more specifically who has access to the data. The broadcast nature of wireless channels implies that the transmitted information signals are also received by nodes other than the intended receiver, which results in the leakage of information. Encryption techniques at higher layers are used to secure transmitted information. However, the high computational complexity of these cryptographic techniques consumes significant amount of energy. Moreover, secure secret key management and distribution via an authenticated third party is typically required for these techniques, which may not be realizable in a dense wireless networks. Therefore, a considerable amount of work has recently been devoted to information-theoretic physical layer security (PLS) as a secure communication technique which exploits the characteristics of wireless channels, such as fading, noise, and interferences. The varying nature of these factors causes randomness in the wireless channel which can be exploited to achieve security. The transmission of secret messages takes place when the receiver»s channel experiences less fading than the eavesdropper»s channel, otherwise transmission remains suspended. The second concern is regarding the limited lifetime of wireless devices, especially when massive amount of data needs to be collected and transferred across the network. This challenge can be addressed by innovative means of powering, and for small energy limited devices, this implies the use of energy harvesting (EH) techniques. In this context, the transfer of data and power over a common electromagnetic (EM) wave has gained significant research interest over the past decade. The technique which merges wireless information transfer (WIT) with wireless power transmission (WPT) is commonly termed as simultaneous wireless information and power transfer (SWIPT). However, SWIPT systems cannot be supported using conventional design of transmitter and receiver. To address this issue, two broad categories of receiver architectures have been proposed in SWIPT literature i.e. separated and integrated architecture. In separated receiver architecture, the information decoder and energy harvester act as dedicated and separate units. This although not only increases the cost of receiver but also increases the complexity of the hardware. In contrast, the integrated receiver architecture jointly processes the information and energy using a unified circuitry for both. This architecture reduces the cost and hardware complexity Our work attempts to address the aforementioned issues by evaluating secrecy performance and proposing practical secrecy enhancement scheme in EH wireless devices. In particular, we investigate PLS in SWIPT systems in the presence of multiple eavesdroppers. The secrecy performance of the SWIPT system is analyzed for Rician faded communication links. The security performance is analyzed for imperfect channel estimation, and both separated and integrated receiver architectures for the SWIPT system. We derive closed-form expressions of the secrecy outage probability and the ergodic secrecy rate for the considered scenario and validate the derived analytical expressions through extensive simulations. Our results reveal that an error floor appears due to channel estimation errors at high values of signal to noise ratio (SNR); such that outage probability cannot be further minimized despite an increase in the SNR of the main link. Moreover, the results show that largest secrecy rate can be achieved when the legitimate receiver is equipped with separated SWIPT receiver architecture and the eavesdroppers have an integrated SWIPT receiver architecture. It is also demonstrated that the power splitting factor at both legitimate receiver and at eavesdroppers play a prominent role in determining the secrecy performance of SWIPT. We prove that a larger power splitting factor is required to ensure link security for poor channel estimation. Finally, our work discusses transmit antenna selection and baseline antenna selection schemes to improve security. Therein, it is shown that transmit antenna selection outperforms baseline antenna selection. The results provided in this work can be readily used to evaluate the secrecy performance of SWIPT systems operating in the presence of multiple eavesdroppers.
-
-
-
Implementing and Analyzing a Recursive Technique for Building Path Oblivious RAM
Authors: Maan Haj Rachid, Ryan Riley and Qutaibah MalluhiIt has been demonstrated that encrypting confidential data before storing it is not sufficient because data access patterns can leak significant information about the data itself (Goldreich & Ostrovsky, 1996). Oblivious RAM (ORAM) schemes exist in order to protect the access pattern of data in a data-store. Under an ORAM algorithm, a client accesses a data store in such a way that does not reveal which item it is interested in. This is typically accomplished by accessing multiple items each access and periodically reshuffling some, or all, of the data on the data-store. One critical limitation of ORAM techniques is the need to have large storage capacity on the client, which is typically a weak device. In this work, we utilize an ORAM technique that adapts itself to working with clients having very limited storage. A trivial implementation for an oblivious RAM scans the entire memory for each actual memory access. This scheme is called linear ORAM. Goldreich and Ostrovsky (Goldreich & Ostrovsky, 1996) presented two ORAM constructions with a hierarchical layered structure: the first, Square-root ORAM, provides square root access complexity and constant space requirement; the second, Hierarchical ORAM, requires logarithmic space and polylogarithmic access complexity. Square-root ORAM was revisited by (Zahur, et al.) to improve its performance in a multi-party secure computation setting. The work of Shi et al. (Shi, Chan, Stefanov, & Li, 2011) adopted a sequence of binary trees as the underlying structure. (Stefanov, et al., 2013) utilized this concept to build a simple ORAM called path ORAM. In path ORAM, every block (item) of data in the input array A is mapped to a (uniformly) random leaf in a tree (typically a binary tree) on the server. This is done using a position map stored in the client memory. Each node in the tree has exactly Z blocks which are initially dummy blocks. Each data item is stored in a node on the path extending from the leaf to which the data item was mapped, to the root. When a specific item is requested, a position map is used to point out the leaf to which the block is mapped. Then the whole path is read starting from the block's mapped leaf up to the root into a stash. We call this procedure a read path method. The stash is a space which also exists on the client. The block is mapped to a new leaf (uniformly random). The client gets the required block from the stash. We then try to evict the contents of the stash to the same path which we read from starting from the leaf to the root. We call this procedure a write path method. To transfer a block into a node in the tree: - The node that is tested should have enough space. - The node should be on the path to which the tested block in the stash is mapped. If both conditions are met, a block is evicted to the tested node. Clearly, the tree is encrypted and whenever the client reads a path into the stash, all read blocks are decrypted and the requested block is sent to the client. The client encrypts the blocks before writing them back to the path. The security proof of this ORAM type is explained in (Stefanov, et al., 2013). We assume that a position map can be fit in the client's memory. Since it requires O(N) space, that could be a problem. (Stefanov, et al., 2013) mentioned a general idea for a solution that uses another smaller ORAM O1 on the server to store the position map and stores the position map for O1 in the client's memory. We employ a recursive generalized version of this approach. If the position map is still larger than the client's capacity, a smaller ORAM O2 is built to store the position map for O1. We call these additional trees auxiliary trees. We implemented the recursive technique for Path ORAM and studied the effect of the threshold size of position map (and consequently, the number of auxiliary trees) on the performance of path ORAM. We tested our implementation on 1 million items using several threshold sizes of position map. The number of accesses is 10,000 in all tests. Our results show expected negative correlation between the time consumption and the threshold size of the position map. However, the results suggest that unless the increase in the threshold size of position map decreases the number of trees, no significant improvement in performance will be noticed. It is also clear that the initialization process which is the process of building the items» tree and the auxiliary trees and filling the initial values comprises more than 98% of the consumed time. Accordingly, this type of ORAM suits the case of large number of accesses since the server can fulfil client»s request very fast after finishing the initialization process. References Goldreich, O., & Ostrovsky, R. (1996). Software protection and simulation on oblivious rams. Journal of the ACM (JACM), vol. 43, no. 3, pp. 431–473. Shi, E., Chan, T.-H., Stefanov, E., & Li, M. (2011). Oblivious ram with o ((logn) 3) worst-case cost. International Conference on The Theory and Application of Cryptology and Information Security. Springer, pp. 197–214. Stefanov, E., Dijk, M. V., Shi, E., Fletcher, C., Ren, L., Yu, X., et al. (2013). Path oram: an extremely simple oblivious ram protocol. Proceedings of the 2013 ACM SIGSAC conference on Computer & communications security. ACM, pp. 299–310.. Zahur, S., Wang, X., Raykova, M., Gascon, A., Doerner, J., Evans, D., et al. (n.d.). Revisiting square-root oram: Efficient random access in multi-party computation. 2016: Security and Privacy (SP), IEEE Symposium on. IEEE, pp. 218–234.
-
-
-
CONNECT: CONtextual NamE disCovery for blockchainbased services in the IoT
More LessThe Internet of Things is gaining momentum thanks to the provided vision of seamlessly interconnected devices. However, a unified way to discover and to interact with the surrounding smart environment is missing. As an outcome, we have been assisting to the development of heterogeneous ecosystems, where each service provider adopts its own protocol—thus preventing IoT devices from interacting when belonging to different providers. And, the same is happening again for the blockchain technology which provides a robust and trusted way to accomplish tasks—unfortunately not providing interoperability thus creating the same heterogeneous ecosystems above highlighted.In this context, the fundamental research question we address is how do we find things or services in the Internet of Things. In this paper, we propose the firstIoT discovery approach which provides an answer to the above question by exploiting hierarchical and universal multi-layered blockchains. Our approach does neither define new standards nor force service providers to change their own protocol. On the contrary, it leverages the existing and publicly available information obtained from each single blockchain to have a better knowledge of the surrounding environment. The proposed approach is detailed and discussed with the support of relevant use cases.
-
-
-
Multiple Input Multiple Output InVivo Communication for Nano Sensor at Terahertz Frequencies
Authors: Aya Fekry Abdelaziz, Ke Yang, Khalid Qaraqe, Joseph Boutros, Qammer Abbasi and Akram AlomainyThis study presents a preliminary feasibility investigation of signal propagation and antenna diversity techniques inside the human skin tissues in the frequency range (0.8-1.2)Terahertz (THz) by applying multiple input-single output (MISO) technique. THz application in in-vivo communication has received a great attention for its unique properties as non ionizing characteristics, strong interaction with water content inside the human tissues, and molecular sensitivity [1]. This study helps to evaluate the usage and the performance of MISO system for nanoscale network. The human skin tissue is represented by three main layers: stratum corneum, epidermis, and dermis. The path loss model and the channel characterization inside the human skin was investigated in [2]. The diversity gain (DG) for two different in-vivo channels resulting from the signal propagation between two transmitting antennas, located at the dermis layer, and one receiving antenna, located at epidermis layer, is calculated to evaluate the system performance. Different diversity combining techniques are applied in this study: selection combining (SC), equal-gain combing (EGC), and maximum-ratio combining (MRC). In the simulation setting in CST microwave studio, the distance between the transmitting antennas is fixed; while the effect of the distance between the receiver and the transmitters is analyzed at different frequencies. Although MIMO antenna systems are used in wireless communication to enhance data throughput, from the initial study it is predicted that might be it is not useful in in-vivo nano communication. Results demonstrates that there is a high cross correlation between the two channels. Figure 1 shows the CDF plot for the two channel with different diversity combining techniques used in the study.[1] J. M. Jornet and I. F. Akyildiz, “Channel modeling and capacity analysis for electromagnetic wireless nanonetworks in theterahertz band,” IEEE Transactions on Wireless Communications, vol. 10, no. 10, pp. 3211–3221, 2011.[2] Q. H. Abbasi, H. El Sallabi, N. Chopra, K. Yang, K. A. Qaraqe, and A. Alomainy, “Terahertz channel characterizationinside the human skin for nano-scale body-centric networks,” IEEE Transactions on Terahertz Science and Technology, vol.6, no. 3, pp. 427–434, 2016.
-
-
-
A Cyberphysical Testbed for Smart Water Networks Research Education and Development in Qatar
Smart water networks integrate sensor data, computation, control, and communication technologies to enhance system performance, reliability and consumer satisfaction. The cyber-physical systems are built from, and rely upon, the tight integration of physical elements of real-time sensors and data integration and cyber algorithmic, control computational and communication layers. A cyber-physical testbed has been developed at Texas A&M University at Qatar to simulate a real smart water network for research, education, and technology development purposes. The physical components include pipes, an automated pump-storage system, programmable logic controllers, controllable valves, a disinfectant injector, sensors, and data acquisition devices. Flow, pressure, temperature, and specific water quality parameters, such as pH and conductivity are continuously monitored by sensors, providing an operator with an up-to-date performance of the current state of the system. The pump-storage system is controlled by programmable logic controllers, and are designed to enable evaluation and enhancement of feedback and model-predictive control algorithms physically. The water tank is equipped with heating apparatus to conduct experimental studies for understanding the effect of water temperature on the fate and transport of chlorine and disinfection byproducts in the drinking water distribution of Qatar. The physical facility is integrated with a cyber-data acquisition and communications layer, and a cloud-based data storage, analytics, and visualization platform. Acquired data is stored and maintained in the format of a non-relational database on a cloud storage service. Additionally, a MongoDB server is set up to query and write data records. The analytics backend engine performs a variety of data transforms including, but not limited to, data cleansing and time series imputation and forecast. The visualization frontend provides a graphical interface to allow the operators interact with the backend engine by doing queries, plotting time series, running data analytics tasks, and generating reports. Together, this integrated physical and cyber layers unleash opportunities for education, research, development, evaluation, and commercialization of a variety of smart water networks technologies. It provides an environment that can predict leaks and pipe bursts based on real-time analytics of high-frequency pressure readings data on the cloud. This also enables developing smart pump-storage control technologies that help reducing non-revenue water loss, energy costs, and carbon emissions. The research team is also investigating harnessing the profound solar power resources available in Qatar for powering treatment plants and pumps by innovating control strategies that can handle the intermittency of such renewable power sources. Two asset management models are also developed and implemented on the testbed. (1) Performance assessment model of water distribution systems, which comprises four assessment modules for water pipelines, water accessories, water segments, and water networks. The model identifies critical factors affecting the performance of a water network and schedules pipe maintenance and replacement plans. (2) Risk assessment model for water pipelines failure, which evaluates the risks of performance and mechanical failures and applies a hierarchical fuzzy model to determine the risk using four risk factors as inputs (i.e., environmental, physical, operational, and post-failure). In addition to the research and technology purposes, this testbed has also provided a valuable learning resource for both operators and students. There are already several undergraduate students who are involved in design and construction of this facility. This has created an opportunity to train, educate, and empower undergraduate students, the future engineers and industry leaders of Qatar.
-
-
-
Realtime Object Detection on Android using Tensorflow
More LessDetection of images or moving objects have been highly worked upon, and has been integrated and used in commercial, residential and industrial environments. But, most of the strategies and techniques have heavy limitations. One of the limitations is due to low computational resources at user level. Other important limitations that need to be tackled are lack of proper data analysis of the measured trained data, dependency on the motion of the objects, inability to differentiate one object from other, and also concern over speed of the object under detection and Illuminacy. Hence, there is a need to draft, apply and recognize new techniques of detection that tackle the existing limitations. In our project we have worked upon a model based on Scalable Object Detection, using Deep Neural Networks to localize and track people, cars, potted plants and 16 others categories in the camera preview in real-time. The large Visual Recognition ImageNet package ‘inception5h’ from google is used. This is a trained model, with images of the respective categories, which is then converted to a graph file using neural networks. The graph nodes are usually huge in number and these are optimized for the use in android. The use of already available trained model is just for the purpose of ease and convenience, nevertheless any set of images can be trained and used in the android application. An important point to note is that training the images will need huge computational speeds and more than one computer supporting GPU. Also a.jar file built with the help of bazel is added to android studio to support the integration of Java and tensorflow. This jar file is the key to getting tensorflow in a mobile device. The jar file is built with the help of openCV, which is a library of programming functions mainly aimed at real-time computer vision. Once this has been integrated, any input to the android application inputed at real-time, is predicted with the help of tiny-yolo (you look only once – a darknet reference network). This application supports multi object detection, which is very useful. All the steps occur simultaneously with great speeds giving remarkable results, detecting all the categories of the trained model within a good illuminacy. The real time detection reference network used also works with an acceptable acceleration of moving objects but is not quite effective with low illumination. The objects are limited to identify only 20 categories but the scope can be broadened with a revised trained model. The 20 categories include «aeroplane»,»bicycle»,»bird»,»boat», «bottle», «bus», «car», «cat», «chair», «cow»,»diningtable»,»dog»,»horse»,»motorbike»,»person», «pottedplant», «sheep»,»sofa»,»train» and «tvmonitor». The application can be used handy in a mobile phone or any other smart device with minimal computational resources ie.., no connectivity to the internet. The application challenges speed and illuminacy. Effective results will help in real-time detection of traffic signs and pedestrians from a moving vehicle. This goes hand in hand with the similar intelligence in cameras which can be used as an artificial eye, and can be used in many areas such as surveillance, Robotics, Traffic, facial recognition, etc.
-
-
-
An integrated multiparametric system for infrastructure monitoring and earlywarning based on internet of things
Authors: Farid Toauti, Damiano Crescini, Alessio Galli and Adel Ben MnaouerOur daily life strictly depends on distributed civil and industrial infrastructures in which Qatar and other heavily industrialized countries have large investments. Recently, fails in such infrastructures have incurred enormous economic losses and development disruptions as well as human lives. Infrastructures are strategic assets for a sustainable development requiring correct management. To this end, their health levels and serviceability should be continuously assessed. Geophysical and mechanical quantities that determine such serviceability are for instance tilt angles, vibration levels, applied forces, stress, and existence of previous structural defects. It follows that for a feasible serviceability assessment, appropriate sensing and data processing of those parameters have to be achieved. For example, bridges are monitored for structure movements and stress level while earthquake early warning systems detect primary seismic waves before arrival of strong waves. In case of riverbank conservation, water level must be monitored together with the associated mass flow for load estimation. In addition, precipitation rate and groundwater level are paramount indicators to anticipate slope fault. Finally, strain/temperature measurement can be used to sense the health of concrete gravity or arch dams. End-users, engineers and owners can take the most appropriate decisions based on the sensed parameters. The Structural Health Assessment (SHA) is not straightforward. The structural condition is generally complex in terms of architectural parameters like damage existence, distributed masses, damping factors, stiffness matrices, and/or applied distributed forces. The above factors make such SHA extremely difficult and/or exceptionally expensive. With the aim to alleviate this difficulty, possible approaches in SHA are based on vibration measurements. The analysis of such measurements reveals the structure dynamic behaviour, which in turn reflects the characteristics and distributed forces on structures. Also, structural soundness is obtained by the estimation of the inverse analyses of the dynamic performance. However, this dynamic behaviour that is inherently complex in both time and/or spatial scale, is more complicated by the fact that for example deterioration/damage/erosion is essentially a local phenomenon. Commonly, technicians with specific domain knowledge achieve SHAs manually. Obviously, this incurs high costs and inadequate monitoring frequency. Also, there is a high probability of making errors due to improper positioning of the instrumentation or to mere mistakes during data collection. Nevertheless, for commonly large buildings (e.g. towers, general buildings, bridges and tunnels), data from just few distributed sensors cannot accurately fulfil the SHA. Consequently, the use of dense distributed sensors working at a sufficiently high sampling frequency becomes a must. Physical wiring of the site under observation is impractical due to cost and architectural constraints. Thus, for Structural Health Monitoring (SHM), networks of dense distributed sensors, which are wirelessly connected, become imperative. When a copious number of transducers are adopted, wireless communication appears to be attractive. Also, the high cost needed for the installation of wired sensors can be strongly reduced by employing wireless sensors. The present research the authors implemented a WSN-based approach for widespread monitoring without forcing intolerable boundary conditions, i.e., requiring wiring the measuring nodes, triggering manual data collection or imposing strong modifications to the site before the deployment of the sensory hardware (less intrusive). In view of the above discussion, the investigators explored some key issues on the above challenges by referring to several SHM engineering paradigms. The author designed a novel multi-parametric system dedicated to stability monitoring and control of soils, engineering works (e.g. bridges, stadium, tunnels), underground rail tunnels, offshore platform in order to continuously evaluate danger levels of potentially instable areas. The proposed system can be assembled ‘in situ’ structuring an underground-instrumented column, where different modules are joined together on a digital bus (e.g. via RS485 or CANBUS communication). Each module contains up to ten different sensors (e.g. accelerometers, magnetometers, inclinometers, extensometers, temperature sensors, and piezometers) and an electronic board for data collection, conversion, filtering and data transmission. Special flexible joints that permit strong, continuous adaptability to bends and twists of the drilling hole, link the modules. A control unit, installed outside the ground provides the readings at regular time intervals and it is connected to other columns via wireless communication forming a wide network. In particular, the proposed approach allows both analysing the response of the infrastructure to vibrations on the fly, so an early warning signal can be triggered, and saving the corresponding measurements for further analysis. Authors believe that this proposal is original and unique in three aspects. First, as most of the earlier studies on SHM were carried out by adapting existing hardwired solution for snap shot measurements rather than representative long-term monitoring, our proposal presents the first initiative to develop green WSN technologies applied to sustainable SHM applications. Second, it will develop tailored sensor technology and new techniques for SHM taking into account metrological and physical parameters such as resolution, cost, accuracy, size, and power consumption. Third, the project will commission a novel multi-parametric SHM system, which can be customized to other areas (e.g. environmental monitoring, traffic monitoring, etc.). The research is to support innovations at system and component levels leading to out-of-the-box know-how. The proposed solution is based on novel/customized sensors and data processing, environmentally powered communication platform, and communication networks and algorithms embracing the visionary nature of the IoT with out-of-the-box solutions. Specific outcomes have been experimental proof-of-concept, through testing and prototyping, of a tailored SHM sensor technology and smart techniques that uniquely provide self-calibration and self-diagnostics of faults, a multi-sensor viable instrumented column for SHM with advanced techniques, and environmentally powered wireless platform with innovative MAC protocols (power-aware, context-aware, cognitive and polymorphic). This work employs tools and techniques of modern sensing, processing, and networking in order to generate novel SHM solutions that uniquely provide precision measurement, green IoT-based communication approach, viability, and cost-effectiveness.
-
-
-
Learning Spatiotemporal Latent Factors of Traffic via a Regularized Tensor Factorization: Imputing Missing Values and Forecasting
Authors: Abdelkader Baggag, Tahar Zanouda, Sofiane Abbar and Fethi FilaliSpatiotemporal data related to traffic has become common place due to the wide availability of cheap sensors and the rapid deployment of IoT platforms. Yet, this data suffer several challenges related to sparsity, incompleteness, and noise, which makes traffic analytics difficult. In this paper, we investigate the problem of missing data or noisy information in the context of real-time monitoring and forecasting of traffic congestion for road networks. The road network is represented as a directed graph in which nodes are junctions and edges are road segments. We assume that the city has deployed high-fidelity sensors for speed reading in a subset of edges. Our objective is to infer speed readings for the remaining edges in the network as well as missing values to malfunctioning sensors. We propose a tensor representation for the series of road network snapshots, and develop a regularized factorization method to estimate the missing values, while learning the latent factors of the network. The regularizer, which incorporates spatial properties of the road network, improves the quality of the results. The learned factors along with a graph-based temporal dependency are used in an autoregressive algorithm to predict the future state of the road network with long horizon. Extensive numerical experiments with real traffic data from the cities of Doha(Qatar) and Aarhus (Denmark) demonstrate that the proposed approach is appropriate for imputing missing data and predicting traffic state.Main contributions. The main contributions are:We propose a novel temporal regularized tensor factorization framework (TRTF) for high-dimensional traffic data. TRTF provides a principled approach to account for both the spatial structure and the temporal dependencies.We introduce a novel data-driven graph-based autoregressive model, where the weights are learned from the data. Hence, the regularizer can account for both positive and negative correlations.We show that incorporating temporal embeddings into CP-WOPT leads to accurate multi-step forecasting, compared to state of the art matrix factorization based methods.We conduct extensive experiments on real traffic congestion datasets from two different cities and show the superiority of TRTF for both tasks of missing value completion and multi-step forecasting under different experimental settings. For instance,TRTF outperforms LSM-RN by 24% and TRMF by 29%.Conclusion. We present in this paper TRTF, an algorithm for temporal regularized tensor decomposition. We show how the algorithm can be used for several traffic related tasks such as missing value completion and forecasting. The proposed algorithm incorporates both spa-tial and temporal properties into the tensor decomposition procedures such as CP-WOPT, yielding to learning better factors. We also, extend TRTF with an auto-regressive procedure to allow for multi step-ahead forecasting of future values. We compare our method to recently developed algorithms that deal with the same type of problems using regularized matrix factorization,and show that under many circumstances, TRTF does provide better results. This is particularly true in cases where the data suffers from high proportions of missing values, which is common in the traffic context. For instance, TRTF achieves a 20% gain in MAPE score compared to the second best algorithm (CP-WOPT) in completing missing values in the case of extreme sparsity observed in Doha. As future work, we will first focus on adding non-negativity constraints to TRTF, although the highest fraction of negative values generated by our method throughout all the experiments did not exceed 0.7%. Our second focus will be to optimize TRTF training phase in order to increase its scalability to handle large dense tensors, and to implement it on a parallel environment.
-
-
-
A Deep Learning Approach for Detection of Electricity Theft Cyber Attacks in Smart Grids
Authors: Muhammad Ismail, Mostafa Shahin, Erchin Serpedin and Khalid QaraqeFuture smart grids rely on advanced metering infrastructure (AMI) networks for monitoring and billing purposes. However, several research works have revealed that such AMI networks are vulnerable to different kinds of cyber attacks. In this research work, we consider one type of such cyber attacks that targets electricity theft and we propose a novel detection mechanism based on a deep machine learning approach. While existing research papers focus on shallow machine learning architectures to detect these cyber attacks, we propose a deep feedforward neural network (D-FF-NN) detector that can thwart such cyber attacks efficiently. To optimize the D-FF-NN hyper-parameters, we apply a sequential grid search technique that significantly improves the detector»s performance while reducing the associated complexity in the learning process. We carry out extensive studies to test the proposed detector based on a publicly available real load profile data of 5000 customers. The detector»s performance is investigated against a mixture of different attacks including partial reduction attacks, selective by-pass attacks, and price-based load control attacks. Our study reveals that the proposed D-FF-NN detector presents a superior performance compared with state-of-the-art detector»s that are based on shallow machine learning architectures
-
-
-
A simple and secure framework for protecting sensitive data stored on the cloud
Authors: Elias Yaacoub, Ali Sakr and Hassan NouraIn the past decade, Cloud-Computing emerged as a new computing concept with a distributed nature using virtual network and systems. Many businesses rely on this technology to keep their systems running but concerns are rising about security breaches in cloud computing. This work presents a secure approach for storing data on the cloud. the proposed methodology is described as follows: 1) The client who wants to store data on the cloud subscribes with n cloud providers (CPs). 2) A file F to be stored on the cloud is subdivided into n parts, or subfiles: F1, F2,.... Fn. 3) Each part is encrypted with an encryption key Kf. The encrypted parts are denoted by F1*, F2*,…, Fn*. 4) A random permutation vector P(f) is generated. 5) The encrypted parts are stored on the n clouds according to P(F); in other words, due to P(F), F1* could be stored on CP3 for example F2* on CPn, etc. 6) In order to be able to retrieve his files, the client needs to maintain some information related to the distribution of the various parts and to the encryption of the file. Thus, he maintains two tables. 7) The first table contains a hash of the file name H(F_name), and the key Kf. 8) The second table contains a hash of the file name H(F_name), a hash of the file content itself (unencrypted) H(F), a hash of the encrypted file content H(F*), and the permutation vector P(F), encrypted with Kf. 9) The two tables are stored on different servers protected by advanced security measures, and preferably located at different locations. 10) In order to obtain the file, the client enters the file name. Then the system computes the hash value of the name, finds the corresponding entry in Table 1, and obtains the key. Then, the corresponding entry in Table 2 is found. The key, obtained from Table 1, is used to decrypt the permutation vector P(f). Then, the encrypted parts are downloaded from the different cloud providers. Afterwards, they are assembled in the correct order and decrypted. The hash values of the encrypted and unencrypted versions are then computed and compared to their corresponding values stored in Table2 in order to check for the integrity of the file downloaded from the cloud. This approach allows the client to use the same storage space on the cloud: Instead of using a single cloud provider to store a file of size S bits, the client is using n cloud providers, where he stores S/n bits with each cloud provider. Thus, the storage costs of the two methods are comparable. If the client wishes to introduce redundancy in the file, such that he can recover the whole file from j parts instead of n parts, with j< = n, then redundancy can be added to the original file as appropriate. In this case, the storage costs will increase accordingly, but this is an added enhancement that can be used with or without the proposed approach. On the other hand, the overhead due to the proposed approach consists of maintaining two tables containing the information relevant to the file. The storage required to maintain the entry corresponding to each file in these two tables is small compared to a typical file size: in fact, we only need to store a few hash values (of fixed size), along with the encryption key. This seems a reasonable price to pay for a client that has sensitive data that he cannot post unencrypted on the cloud, or even posting it encrypted with a single provider is risky in case a security breach occurs at the provider»s premises.
-
-
-
Substring search over encrypted data
More LessOur data, be it personal or professional, is increasingly outsourced. This results from the development of cloud computing in the past ten years, a paradigm that shifts computing to a utility. Even without realizing it, cloud computing has entered our lives inexorably: every owner of a smartphone, every user of a social network is using cloud computing, as most IT companies and tech giants in particular are using infrastructure as a service to offer services in the model of software as a service. These services (dropbox, google, facebook, twitter…) are simple to use, flexible…and free! Users just send their data and they get all services without paying. Actually, these companies are making most of their revenues by profiling the users thanks to the data that the users willingly provide. The data is the indirect payment to benefit from these services. This raises privacy concerns at the personal level, as well as confidentiality issues for sensitive documents in a professional environment. The classical way of dealing with confidentiality is to conceal the data through encryption. However, cloud providers need access to data in order to provide useful services, not only to profile users. Take a cloud email service as example, where the emails are stored and archived in the cloud and only downloaded to the user's phone or computer when the user wants to read them. If the emails are encrypted in the cloud, the cloud cannot access them and confidentiality is enforced. However, the cloud can also not provide any useful service to the user such as a search functionality over emails. To meet these conflicting requirements (hiding the data and accessing the data) a solution is to develop mechanisms that allow computation on encrypted data. While generic protocols for computation on encrypted data have been researched developed, such as Gentry's breakthrough fully homomorphic encryption, their performance remains unsatisfactory. On the contrary, tailoring solutions to specific needs result in more practical and efficient solution. In the case of searching over encrypted data, searchable encryptions algorithms have been developed for over decade and achieve now satisfactory performance (linear in the size of the dictionary). Most of the work in this field focus on single keyword search in the symmetric setting. To overcome this limitation, we first proposed a scheme based on letter orthogonalization that allows testing of string membership by performing efficient inner products (AsiaCCS 2013). Going further, we now propose a general solution to the problem of efficient substring search over encrypted data. The solution enhances existing “keyword” searchable encryption schemes by allowing searching for any part of encrypted keywords without requiring one to store all possible combinations of substrings from a given dictionary. The proposed technique is based on the previous idea of letter orthogonalization. We first propose SED-1, the base protocol for substring search. We then identify some attacks on SED-1 that demonstrate the complexity of the substring search problem under different threat scenarios. This leads us to propose our second and main protocol SED-2. The protocol is also efficient in that the search complexity is linear in the size of the keyword dictionary. We run several experiments on a sizeable real world dataset to evaluate the performance of our protocol. This final work has been accepted for publication in the IOS journal of computer security https://content.iospress.com/articles/journal-of-computer-security/jcs14652.
-
-
-
Almost BPXOR Coding Technique for Tolerating Three Disk Failures in RAID7 Architectures
Authors: Naram Mhaisen, Mayur Punkar, Yongge Wang, Yvo Desmedt and Qutaibah MalluhiRedundant Array of Independent Disks (RAID) storage architectures provide protection of digital infrastructure against potential disks failures. For example, RAID-5 and RAID-6 architectures provide protection against one and two disk failures, respectively. Recently, the data generation has significantly increased due to the emergence of new technologies. Thus, the size of storage systems is also growing rapidly to accommodate such large data sizes, which increases the probability for disks failures. This necessitates a new RAID architecture that can tolerate up to three disk failures. RAID architectures implement coding techniques. The code specifies how data is stored among multiple disks and how lost data can be recovered from surviving disks. This abstract introduces a novel coding scheme for new RAID-7 architectures that can tolerate up to three disks failures. The code is an improved version of the existing BP-XOR code and is called “Almost BP-XOR”.There are multiple codes that can be used for RAID-7 architectures. However, [5,2] BP-XOR codes have significantly lower encoding and decoding complexities than most common codes [1]. Regardless of this fact, this code does not achieve the fastest data decoding and reconstruction speeds due to its relatively low efficiency of 0.4. Furthermore, the existence of MDS [6,3] bx6 BP-XOR codes, b>2 (which achieves efficiency of 0.5) is still an open research question. This work proposes [6,3] 2 x 6 Almost BP-XOR codes. These codes largely utilize the simple and fast BP-XOR decoder while achieving an efficiency of 0.5, leading to the fastest recovery from disk failures among other state-of-the-art codes. An algorithm to generate a [6, 3] 2 x 6 Almost BP-XOR code has been developed and an example code is provided in Table 1. The [6, 3] 2 x 6 Almost BP-XOR codes are constructed in a way that any three-column-erasure pattern will result in one of the following two main scenarios. First: At least one of the surviving degree-three encoding symbols contains two known information symbols. This scenario occurs in 70% of three-column erasure cases (i.e. 14 out of the 20 possible cases). The recovery process in such scenario is identical to that of the BP-XOR codes; Knowing any two information symbols in a degree-three encoding symbol is sufficient to know the third information symbol through performing a simple XOR operation. Second: None of the surviving degree-three encoding symbols contains two known information symbols. This scenario occurs in the remaining 30% of three-column erasure cases (i.e., 6 out if the possible 20). The BP-XOR decoder fails in such a scenario. However, due to the construction of the codes, at least one surviving degree-three encoding symbol contains a known information symbol. Thus, knowing one of the reaming two information symbols in such a degree-three encoding symbol will initiate the BP-XOR decoder again.Table 2 shows these erasure patterns along with an expression for one of the missing information symbols. these expressions can be stored in buffers and used whenever the corresponding erasure pattern occurs. Solutions in Table 2 are derived from the inverse of a 6x6 submatrix that results from a generator matrix G by deleting columns from G corresponding to erased code columns. The read complexity of almost BP-XOR codes is 1. On the other hand. The decoding of almost BP-XOR codes require just 6 XOR operations when for a given three-column-erasure pattern BP-XOR decoding succeeds. However, when the BP- XOR decoder fails, it will require up to 15 XOR operations in total. The normalized repairing complexity is 15/6 = 2.5.Experimentally, Fig.1 shows that the proposed Almost BP-XOR codes require the least amount of time to decode and reconstruct erased columns. Thus, it is concluded that the [6, 3] 2x6 almost BP-XOR codes are best suited for RAID-7 system that requires storage efficiency of 0.5. References [1] Y. Wang. “Array BP-XOR codes for reliable cloud storage systems,” in Proc. of the 2013 IEEE International Symposium on Information Theory (ISIT), pp. 326–330, Istanbul, Turkey, July 2013.NoteFigures, Tables, and more details are provided in the complete attached file titled Abstract-ARC18.pdf, (respecting the same word count restriction).
-
-
-
Leveraging Online Social Media Data for Persona Profiling
Authors: Bernard J. Jansen, Soon-gyo Jung, Joni Salminen, Jisun An and Haewoon KwakThe availability of large quantities of online data affords the isolation of key user segments based on demographics and behaviors for many online systems. However, there is an open question of how organizations can best leverage this user information in communication and decision-making. The automatic generation of personas to represent customer segments is an interactive design technique with considerable potential for product development, policy decision, and content creation. A persona is an imaginary but characteristic person that is the representation of a customer, audience, or user segment. The representative segment shares common characteristics in terms of behavioral attributes or demographics. Representing a user segment, a persona is generally developed in the form of a detailed profile narrative, typically expressed in one or two pages, about a representative but an imaginary individual that embodies the collection of users with similar behaviors or demographics. In order to make the fictitious individual appear as a real person to system developers and other decision-makers, the persona profile usually comprises a variety of demographic and behavioral details, such as socioeconomic status, gender, hobbies, family members, friends, possessions, among other data and information. Along with this data, the persona profiles typically address the goals, needs, wants, frustrations and other attitudinal aspects of the fictitious individual that are relevant to the product being developed and designed. Personas have typically been fairly static once created by using manual, qualitative methods. In this research, we demonstrate a data-driven approach for creating and validating personas in real time, based on automated analysis of actual user data. Using a variety of data collection sites and research partners from various verticals (digital content, non-profits, retail, service, etc., we are specifically interested in understanding the users of these organizations by identifying (1) whom the organizations are reaching (i.e., user segment) and (2) what content are associated with each user segment. Focusing on one aspect of user behavior, we collect tens of millions of instances of interaction by users to online content, specifically examining the topics of content interaction. We then decompose the interaction patterns, discover related impactful demographics, and add personal properties; this approach creates personas based on these behavioral and demographic aspects that represent the core user segments for each organization. We conduct analysis to remove outliers and use non-negative matrix factorization to identify first the meaningful behavioral patterns and then the impactful demographic groupings. We then demonstrate how these findings can be leveraged to generate real-time personas based on actual user data to facilitate organizational communication and decision-making. Demonstrating that these insights can be used to develop personas in near real-time, the research results provide insights into user segmentation, competitive marketing, topical interests, and preferred system features for the users. Overall, research implications are that personas can be generated in near real-time representing the core users groups of online products.
-
-
-
A Fast and Secure Approach for the Transmission of Monitoring Data over MultiRATs
Authors: Elias Yaacoub, Rida Diba and Hassan NouraIn an mHealth remote patient monitoring scenario, usually control units/data aggregators receive data from the body area network (BAN) sensors then send it to the network or “cloud”. The control unit would have to transmit the measurement data to the home access point (AP) using WiFi for example, or directly to a cellular base station (BS), e.g. using the long-term evolution (LTE) technology, or both (e.g. using multi-homing to transmit over multiple radio access technologies (Multi-RATs). Fast encryption or physical layer security techniques are needed to secure the data. In fact, during normal conditions, monitoring data can be transmitted using best effort transmission. However, when real-time processing detects an emergency situation, the current monitoring data should be transmitted real-time to the appropriate medical personnel in emergency response teams. The proposed approach consists of benefiting of the presence of multi-RATs in order to exchange the secrecy information more efficiently while optimizing the transmission time. It can be summarized as follows (assuming there are two RATs):1) The first step is to determine the proportion of data bits to be transmitted over each RAT in order to minimize the transmission time, given the data rates achievable on each RAT. Denoting the data rates by R1, and R2, and the total number of bits to be transmitted by D = SUM(D1,D2), where D1 and D2 are the number of bits to be transmitted over RAT1 and RAT2 respectively, then they should be selected such that D1/R1 = D2/R22) Then, the exchange of the secrecy parameters between sender and receiver is done over the two RATs in order to maintain the security of the transmission. To avoid the complexity of public key cryptography, a three way handshake can be used: 2-1) The sender decides to divide the data into n parts, with a fraction n1 sent on RAT1 and a fraction n2 sent on RAT2, according to the ratios determined in 1) above (i.e. the sum of the bits in the n1 parts should be close to D1 bits, and the sum of the n2 parts should be close to D2 bits) 2-2) The sender generates a scrambling vector P(D,n) to scramble the n data parts and transmit them out of order. 2-3) The sender groups the secret information consisting of S = {n, n1, n2, P(D,n)}, and could add additional information to protect against replay attacks, e.g. timestamp, nonce, etc., and sends this information on the two RATs, encrypted by a different key: K11 on RAT1 and K12 on RAT2. Thus, {S}_K11 is sent on RAT1 and {S}_K12 is sent on RAT2. 2-4) The receiver does not know K11 and K12. Thus, it encrypts the received information with two other keys K21 (over RAT1) and K22 (over RAT2) and sends them back: {{S}_K11,K21} is sent on RAT1 and {{S}_K12,K22} is sent on RAT2. 2-5) The sender decodes the received encrypted vectors using his keys and sends back {S}_K21 on RAT1 and {S}_K22 on RAT2. The secret information is still securely encoded by the receiver's secret keys K21 and K22. 2-6) The receiver can now decrypt the information and obtain S. 3. The two parties can now communicate using the secret scrambling approach provided by S. This information can be changed periodically as needed. For example, if the data is subdivided over 10 parts, with 40% to be sent over LTE and 60% to be sent over WiFi, according to the scrambling vector P(D,n) = {4,1,10,7,3,9,5,2,6}, then parts {4,1,10,7} are sent over LTE and parts {7,3,9,5,2,6} are sent over WiFi. The receiver will sort them out in the correct order.
-
-
-
Crowdsourced MultiView Live Video Streaming using Cloud Computing
Authors: Aiman Erbad and Kashif BilalMulti-view videos are composed of multiple video streams captured simultaneously using multiple cameras from various angles (different viewpoints) of a scene. Multi-view videos offer more appealing and realistic view of the scene leading to higher user satisfaction and enjoyment. However, displaying realistic and live multiview scenes captured from a limited view-points faces multiple challenges, including excessive number of precise synchronization of many cameras, color differences among cameras, large bandwidth, computation and storage requirements, and complex encoding. current multi-view video setups are very limited and based in studios. We propose a novel system to collect individual video streams (views) captured for the same event by multiple attendees, and combine them into multi-view videos, where viewers can watch the event from various angles, taking crowdsourced media streaming to a new immersive level. The proposed system is called Cloud based Multi-View Crowdsourced Streaming (CMVCS), and it delivers multiple views of an event to viewers at the best possible video representation based on each viewer's available bandwidth. CMVCS is a complex system having many research challenges. In this study, we focus on resource allocation of the CMVCS system. The objective of the study is to maximize the overall viewer satisfaction by allocating available resources to transcode views in an optimal set of representations, subject to computational and bandwidth constraints. We choose the video representation set to maximize QoE using Mixed Integer Programming (MIP). Moreover, we propose a Fairness Based Representation Selection (FBRS) heuristic algorithm to solve the resource allocation problem efficiently. We compare our results with optimal and Top-N strategies. The simulation results demonstrate that FBRS generates near optimal results and outperforms the state-of-the-art Top-N policy, which is used by a large scale system (Twitch). Moreover, we consider region based distributed datacenters to minimize the overall end-to-end latency. To further enhance the viewers’ satisfaction level and Quality of Experience (QoE), we propose an edge based cooperative caching and online transcoding strategy to minimize the delay and backhaul bandwidth consumption. Our main research contributions are: We present the design and architecture of a Cloud based Multi-View Crowdsourced Streaming (CMVCS) system that allows viewers to experience the captured events from various angles. We propose a QoE metric to determine the overall user satisfaction based on the received view representation, viewers’ bandwidth capability, and end-to-end latency between viewer and transcoding site. We formulate a Mixed Integer Programming (MIP) optimization problem for multi-region distributed resource allocation to choose the optimal set of views and representations to maximize QoE in constrained settings. We propose a fairness based heuristic algorithm to find near optimal resource allocation efficiently. We propose an edge computing based video caching and online transcoding strategy to minimize delay and backhaul network consumption. We use multiple real-world traces to simulate various scenarios and show the efficiency of the proposed solution.
-