Your search
Results 87 resources
-
Scheduling periodic real-time tasks on multiple periodic resources is an emerging research issue in the real-time scheduling community and has drawn increased attention over the last few years. This paper studies a sub-category of the scheduling problem which focuses on scheduling a periodic task on multiple periodic resources where none of these resources have sufficient capacity to support the task. Instead of splitting the task into sub-tasks, which is not always practical in real systems, we integrate resources together to jointly support the task. First, we develop a method to integrate two fixed but arbitrary pattern periodic resources into an equivalent periodic resource. Second, for two periodic resources with unknown but fixed resource occurrence patterns, we give the lower and upper bounds of the available time provided by an integrated periodic resource within a period. Third, we present theoretical and empirical analysis on the schedulability of a non-splittable periodic task on two periodic resources and their integrated periodic resource.
-
In this paper, we address the resource and virtual machine instance hour minimization problem for directed-acyclic-graph based deadline constrained applications deployed on computer clouds. The allocated resources and instance hours on computer clouds must: (1) guarantee the satisfaction of a deadline constrained application's end-to-end deadline; (2) ensure that the number of virtual machine (VM) instances allocated to the application is minimized; (3) under the allocated number of VM instances, determine application execution schedule that minimizes the application's makespan; and (4) under the decided application execution schedule, determine a VM operation schedule, i.e., when a VM should be turned on or off, that minimizes total VM instance hours needed to execute the application. We first give lower and upper bounds for the number of VM instances needed to guarantee the satisfaction of a deadline constrained application's end-to-end deadline. Based on the bounds, we develop a heuristic algorithm called minimal slack time and minimal distance (MSMD) algorithm that finds the minimum number of VM instances needed to guarantee the application's deadline and schedules tasks on the allocated VM instances so that the application's makespan is minimized. Once the application execution schedule and the number of VM instances needed are determined, the proposed VM instance hour minimization (IHM) algorithm is applied to further reduce the instance hours needed by VMs to complete the application's execution. Our experimental results show that the MSMD algorithm can guarantee applications' end-to-end deadlines with less resources than the HEFT [32], MOHEFT [16], DBUS [9], QoS-base [40] and Auto-Scaling [25] heuristic scheduling algorithms in the literature. Furthermore, under allocated resources, the MSMD algorithm can, on average, reduce an application's makespan by 3.4 percent of its deadline. In addition, with the IHM algorithm we can effectively reduce the application's execution instance hours compared with when IHM is not applied.
-
The periodic task set assignment problem in the context of multiple processors has been studied for decades. Different heuristic approaches have been proposed, such as the Best-Fit (BF), the First-Fit (FF), and the Worst-Fit (WF) task assignment algorithms. However, when processors are not dedicated but only periodically available to the task set, whether existing approaches still provide good performance or if there is a better task assignment approach in the new context are research problems which, to our best knowledge, have not been studied by the real-time research community. In this paper, we present the Best-Harmonically-Fit (BHF) task assignment algorithm to assign periodic tasks on multiple periodic resources. By periodic resource we mean that for every fixed time interval, i.e., the period, the resource always provides the same amount of processing capacity to a given task set. Our formal analysis indicates that if a harmonic task set is also harmonic with a resource's period, the resource capacity can be fully utilized by the task set. Based on this analysis, we present the Best-Harmonically-Fit task assignment algorithm. The experimental results show that, on average, the BHF algorithm results in 53.26 , 42.54 , and 27.79 percent higher resource utilization rate than the Best-Fit Decreasing (BFD), the First-Fit Decreasing (FFD), and the Worst-Fit Decreasing (WFD) task assignment algorithms, respectively; but comparing to the optimal resource utilization rate found by exhaustive search, it is about 11.63 percent lower.
-
Cloud bursting is one of the key research topics in the cloud computing communities. A well designed cloud bursting module enables private clouds to automatically launch virtual machines (VMs) to public clouds when more resources are needed. One of the main challenges in developing a cloud bursting module is to decide when and where to launch a VM so that all resources are most effectively and efficiently utilized and the system performance is optimized. However, based on system operational data obtained from FermiCloud, a private cloud developed by the Fermi National Accelerator Laboratory for scientific workflows, the VM launching overhead is not a constant. It varies with physical resource utilization, such as CPU and I/O device utilizations, at the time when a VM is launched. Hence, to make judicious decisions as to when and where a VM should be launched, a VM launching overhead reference model is needed. In this paper, we first develop a VM launching overhead reference model based on operational data we have obtained on FermiCloud. Second, we apply the developed reference model on FermiCloud and compare calculated VM launching overhead values based on the model with measured overhead values on FermiCloud. Our empirical results on FermiCloud indicate that the developed reference model is accurate. We believe, with the guidance of the developed reference model, efficient resource allocation algorithms can be developed for cloud bursting process to minimize the operational cost and resource waste.
-
In this age of technology, building quality software is essential to competing in the business market. One of the major principles required for any quality and business software product for value fulfillment is reliability. Estimating software reliability early during the software development life cycle saves time and money as it prevents spending larger sums fixing a defective software product after deployment. The Software Reliability Growth Model (SRGM) can be used to predict the number of failures that may be encountered during the software testing process. In this paper we explore the advantages of the Grey Wolf Optimization (GWO) algorithm in estimating the SRGM’s parameters with the objective of minimizing the difference between the estimated and the actual number of failures of the software system. We evaluated three different software reliability growth models: the Exponential Model (EXPM), the Power Model (POWM) and the Delayed S-Shaped Model (DSSM). In addition, we used three different datasets to conduct an experimental study in order to show the effectiveness of our approach.
-
Design of the Proportional-Integral-Derivative (PID) controller for an industrial process represents a challenge due to process complexity and non-linearity. Traditional methods such as Ziegler-Nichols (ZN) for PID controller tuning do not provide an optimal gain; thus, might leave the system with potential instability condition and cause significant losses and damages to the system. This paper investigates the merits of evolutionary and swarm-based optimization algorithms in fine-tuning the parameters of a PID controller. Here, Genetic Algorithms (GAs) and Particle Swarm Optimization (PSO) algorithm were utilized to optimize the PID controller for a DC motor system. Various fitness functions were provided for the presented algorithms to compute the performance of the controller. A new fitness function was proposed to achieve an outstanding control response for the DC motor system. Results demonstrate the efficacy of the proposed methods in improving closed loop system response.
-
SESSION TITLE: Clinical Prediction and Diagnosis of OSA
-
Robotic systems have been evolving since decades and touching almost all aspects of life, either for leisure or critical applications. Most of traditional robotic systems operate in well-defined environments utilizing pre-configured on-board processing units. However, modern and foreseen robotic applications ask for complex processing requirements that exceed the limits of on-board computing power. Cloud computing and the related technologies have high potential to overcome on-board hardware restrictions and can improve the performance efficiency. This research highlights the advancements in robotic systems with focus on cloud robotics as an emerging trend. There exists an extensive amount of effort to leverage the potentials of robotic systems and to handle arising shortcomings. Moreover, there are promising insights for future breed of intelligent, flexible, and autonomous robotic systems in the Internet of Things era.
-
This work proposes a new evolutionary multilayer perceptron neural networks using the recently proposed Bird Swarm Algorithm. The problem of finding the optimal connection weights and neuron biases is first formulated as a minimization problem with mean square error as the objective function. The BSA is then used to estimate the global optimum for this problem. A comprehensive comparative study is conducted using 13 classification datasets, three function approximation datasets, and one real-world case study (Tennessee Eastman chemical reactor problem) to benchmark the performance of the proposed evolutionary neural network. The results are compared with well-regarded conventional and evolutionary trainers and show that the proposed method provides very competitive results. The paper also considers a deep analysis of the results, revealing the flexibility, robustness, and reliability of the proposed trainer when applied to different datasets.
-
SESSION TITLE: Clinical Prediction and Diagnosis of OSA
-
AbstractThe autonomous navigation of robots in unknown environments is a challenge since it needs the integration of a several subsystems to implement different functionality. It needs drawing a map of the environment, robot map localization, motion planning or path following, implementing the path in real-world, and many others; all have to be implemented simultaneously. Thus, the development of autonomous robot navigation (ARN) problem is essential for the growth of the robotics field of research. In this paper, we present a simulation of a swarm intelligence method is known as Particle Swarm Optimization (PSO) to develop an ARN system that can navigate in an unknown environment, reaching a pre-defined goal and become collision-free. The proposed system is built such that each subsystem manipulates a specific task which integrated to achieve the robot mission. PSO is used to optimize the robot path by providing several waypoints that minimize the robot traveling distance. The Gazebo simulator was used to test the response of the system under various envirvector representing a solution to the optimization problem.onmental conditions. The proposed ARN system maintained robust navigation and avoided the obstacles in different unknown environments. vector representing a solution to the optimization problem.
-
SESSION TITLE: Clinical Prediction and Diagnosis of OSA
-
There is still an urgent need of finding a mathematical model which can provide an accurate relationship between the software project effort/cost and the cost drivers. A powerful algorithm which can optimize such a relationship via developing a mathematical relationship between model variables is urgently needed. In this paper, we explore the use of GP to develop a software cost estimation model utilizing the effect of both the developed line of code and the used methodology during the development. An application of estimating the effort for some NASA software projects is introduced. The performance of the developed Genetic Programming (GP) based model was tested and compared to known models in the literature. The developed GP model was able to provide good estimation capabilities compared to other models.
-
The objective of my sabbatical leave project was to propose a new scheduling algorithm that extends the current MapReduce model to improve system performance. MapReduce, which has been popularized by Google, is a scalable tool that enables the processing of massive volumes of data.
Explore
Department
Resource type
- Book (4)
- Book Section (6)
- Conference Paper (34)
- Journal Article (39)
- Report (4)
Publication year
Resource language
- English (52)