11,635 research outputs found

    The Emerging Internet of Things Marketplace From an Industrial Perspective: A Survey

    Get PDF
    The Internet of Things (IoT) is a dynamic global information network consisting of internet-connected objects, such as Radio-frequency identification (RFIDs), sensors, actuators, as well as other instruments and smart appliances that are becoming an integral component of the future internet. Over the last decade, we have seen a large number of the IoT solutions developed by start-ups, small and medium enterprises, large corporations, academic research institutes (such as universities), and private and public research organisations making their way into the market. In this paper, we survey over one hundred IoT smart solutions in the marketplace and examine them closely in order to identify the technologies used, functionalities, and applications. More importantly, we identify the trends, opportunities and open challenges in the industry-based the IoT solutions. Based on the application domain, we classify and discuss these solutions under five different categories: smart wearable, smart home, smart, city, smart environment, and smart enterprise. This survey is intended to serve as a guideline and conceptual framework for future research in the IoT and to motivate and inspire further developments. It also provides a systematic exploration of existing research and suggests a number of potentially significant research directions.Comment: IEEE Transactions on Emerging Topics in Computing 201

    App-based feedback on safety to novice drivers: learning and monetary incentives

    Get PDF
    An over-proportionally large number of car crashes is caused by novice drivers. In a field experiment, we investigated whether and how car drivers who had recently obtained their driving license reacted to app-based feedback on their safety-relevant driving behavior (speeding, phone usage, cornering, acceleration and braking). Participants went through a pre-measurement phase during which they did not receive app-based feedback but driving behavior was recorded, a treatment phase during which they received app-based feedback, and a post-measurement phase during which they did not receive app-based feedback but driving behavior was recorded. Before the start of the treatment phase, we randomly assigned participants to two possible treatment groups. In addition to receiving app-based feedback, the participants of one group received monetary incentives to improve their safety-relevant driving behavior, while the participants of the other group did not. At the beginning and at the end of experiment, each participant had to fill out a questionnaire to elicit socio-economic and attitudinal information. We conducted regression analyses to identify socio-economic, attitudinal, and driving-behavior-related variables that explain safety-relevant driving behavior during the pre-measurement phase and the self-chosen intensity of app usage during the treatment phase. For the main objective of our study, we applied regression analyses to identify those variables that explain the potential effect of providing app-based feedback during the treatment phase on safety-relevant driving behavior. Last, we applied statistical tests of differences to identify self-selection and attrition biases in our field experiment. For a sample of 130 novice Austrian drivers, we found moderate improvements in safety-relevant driving skills due to app-based feedback. The improvements were more pronounced under the treatment with monetary incentives, and for participants choosing higher feedback intensities. Moreover, drivers who drove relatively safer before receiving app-based feedback used the app more intensely and, ceteris paribus, higher app use intensity led to improvements in safety-related driving skills. Last, we provide empirical evidence for both self-selection and attrition biases

    Efficient Opportunistic Sensing using Mobile Collaborative Platform MOSDEN

    Get PDF
    Mobile devices are rapidly becoming the primary computing device in people's lives. Application delivery platforms like Google Play, Apple App Store have transformed mobile phones into intelligent computing devices by the means of applications that can be downloaded and installed instantly. Many of these applications take advantage of the plethora of sensors installed on the mobile device to deliver enhanced user experience. The sensors on the smartphone provide the opportunity to develop innovative mobile opportunistic sensing applications in many sectors including healthcare, environmental monitoring and transportation. In this paper, we present a collaborative mobile sensing framework namely Mobile Sensor Data EngiNe (MOSDEN) that can operate on smartphones capturing and sharing sensed data between multiple distributed applications and users. MOSDEN follows a component-based design philosophy promoting reuse for easy and quick opportunistic sensing application deployments. MOSDEN separates the application-specific processing from the sensing, storing and sharing. MOSDEN is scalable and requires minimal development effort from the application developer. We have implemented our framework on Android-based mobile platforms and evaluate its performance to validate the feasibility and efficiency of MOSDEN to operate collaboratively in mobile opportunistic sensing applications. Experimental outcomes and lessons learnt conclude the paper

    MLPerf Inference Benchmark

    Full text link
    Machine-learning (ML) hardware and software system demand is burgeoning. Driven by ML applications, the number of different ML inference systems has exploded. Over 100 organizations are building ML inference chips, and the systems that incorporate existing models span at least three orders of magnitude in power consumption and five orders of magnitude in performance; they range from embedded devices to data-center solutions. Fueling the hardware are a dozen or more software frameworks and libraries. The myriad combinations of ML hardware and ML software make assessing ML-system performance in an architecture-neutral, representative, and reproducible manner challenging. There is a clear need for industry-wide standard ML benchmarking and evaluation criteria. MLPerf Inference answers that call. In this paper, we present our benchmarking method for evaluating ML inference systems. Driven by more than 30 organizations as well as more than 200 ML engineers and practitioners, MLPerf prescribes a set of rules and best practices to ensure comparability across systems with wildly differing architectures. The first call for submissions garnered more than 600 reproducible inference-performance measurements from 14 organizations, representing over 30 systems that showcase a wide range of capabilities. The submissions attest to the benchmark's flexibility and adaptability.Comment: ISCA 202
    corecore