14 research outputs found
Load Balancing in Cloud Computing
Cloud computing is one of the top trending technologies which primarily focuses on the end user’s use cases. The service provider needs to provide services to many clients. These increasing number of requests from the clients are giving rise to the new inventions in the load scheduling algorithms. There are different scheduling algorithms which are already present in the cloud computing, and some of them includes the Shortest Job First (SJF), First Come First Serve (FCFS), Round Robin (RR) etc. Though there are different parameters to consider when load balancing in cloud computing, makespan (time difference between start time of first task and finish of last task on the same machine) and response time are the most important parameters. This research surveys different load balancing algorithms and aims to improve the SJF load balancing algorithm in cloud computing. In this project, a Modified Shortest Job First (MSJF) and Generalized Priority (GP) load scheduling algorithms are combined to reduce the makespan and optimize the resource utilization. Together, MSJF and GP sends the longest task having high MIPS (million instructions per second) requirements to the machine with a high processing power and the shortest task having low MIPS requirements to the machine with a low processing power. Hence, neither the task with the lowest MIPS requirements nor the task with the highest MIPS requirements needs to wait for a very long time for resource allocation. Every task gets fair priority. Results are shown for SJF, MSJF, and GP in order to compare the different number of tasks using cloud simulator
Distributed Information Systems and Data Mining in Self-Organizing Networks
The diffusion of sensors and devices to generate and collect data is capillary. The infrastructure that envelops the smart city has to react to the contingent situations and to changes in the operating environment. At the same time, the complexity of a distributed system, consisting of huge amounts of components fixed and mobile, can generate unsustainable costs and latencies to ensure robustness, scalability, and reliability, with type architectures middleware. The distributed system must be able to self-organize and self-restore adapting its operating strategies to optimize the use of resources and overall efficiency. Peer-to-peer systems (P2P) can offer solutions to face the requirements of managing, indexing, searching and analyzing data in scalable and self-organizing fashions, such as in cloud services and big data applications, just to mention two of the most strategic technologies for the next years.
In this thesis we present G-Grid, a multi-dimensional distributed data indexing able to efficiently execute arbitrary multi-attribute exact and range queries in decentralized P2P environments. G-Grid is a foundational structure and can be effectively used in a wide range of application environments, including grid computing, cloud and big data domains.
Nevertheless we proposed some improvements on the basic structure introducing a bit of randomness by using Small World networks, whereas are structures derived from social networks and show an almost uniform traffic distribution. This produced huge advantages in efficiency, cutting maintenance costs, without losing efficacy. Experiments show how this new hybrid structure obtains the best performance in traffic distribution and it a good settlement for the overall performance on the requirements desired in the modern data systems
Cross-Platform Text Mining and Natural Language Processing Interoperability - Proceedings of the LREC2016 conference
No abstract available
Cross-Platform Text Mining and Natural Language Processing Interoperability - Proceedings of the LREC2016 conference
No abstract available
From Data to Software to Science with the Rubin Observatory LSST
The Vera C. Rubin Observatory Legacy Survey of Space and Time (LSST) dataset
will dramatically alter our understanding of the Universe, from the origins of
the Solar System to the nature of dark matter and dark energy. Much of this
research will depend on the existence of robust, tested, and scalable
algorithms, software, and services. Identifying and developing such tools ahead
of time has the potential to significantly accelerate the delivery of early
science from LSST. Developing these collaboratively, and making them broadly
available, can enable more inclusive and equitable collaboration on LSST
science.
To facilitate such opportunities, a community workshop entitled "From Data to
Software to Science with the Rubin Observatory LSST" was organized by the LSST
Interdisciplinary Network for Collaboration and Computing (LINCC) and partners,
and held at the Flatiron Institute in New York, March 28-30th 2022. The
workshop included over 50 in-person attendees invited from over 300
applications. It identified seven key software areas of need: (i) scalable
cross-matching and distributed joining of catalogs, (ii) robust photometric
redshift determination, (iii) software for determination of selection
functions, (iv) frameworks for scalable time-series analyses, (v) services for
image access and reprocessing at scale, (vi) object image access (cutouts) and
analysis at scale, and (vii) scalable job execution systems.
This white paper summarizes the discussions of this workshop. It considers
the motivating science use cases, identified cross-cutting algorithms,
software, and services, their high-level technical specifications, and the
principles of inclusive collaborations needed to develop them. We provide it as
a useful roadmap of needs, as well as to spur action and collaboration between
groups and individuals looking to develop reusable software for early LSST
science.Comment: White paper from "From Data to Software to Science with the Rubin
Observatory LSST" worksho