Modern human life is heavily dependent on computing systems and one of the core components affecting the performance of these systems is underlying operating system.Operating systems need to be upgraded to match the n...Modern human life is heavily dependent on computing systems and one of the core components affecting the performance of these systems is underlying operating system.Operating systems need to be upgraded to match the needs of modern-day systems relying on Internet of Things,Fog computing and Mobile based applications.The scheduling algorithm of the operating system dictates that how the resources will be allocated to the processes and the Round Robin algorithm(RR)has been widely used for it.The intent of this study is to ameliorate RR scheduling algorithm to optimize task scheduling.We have carried out an experimental study where we have developed four variations of RR,each algorithm considers three-time quanta and the performance of these variations was compared with the RR algorithm,and results highlighted that these variations performed better than conventional RR algorithm.In the future,we intend to develop an automated scheduler that can determine optimal algorithm based on the current set of processes and will allocate time quantum to the processes intelligently at the run time.This way the task performance of modern-day systems can be improved to make them more efficient.展开更多
Minimizing time cost in time-shared operating systems is considered basic and essential task,and it is the most significant goal for the researchers who interested in CPU scheduling algorithms.Waiting time,turnaround ...Minimizing time cost in time-shared operating systems is considered basic and essential task,and it is the most significant goal for the researchers who interested in CPU scheduling algorithms.Waiting time,turnaround time,and number of context switches are themost time cost criteria used to compare between CPU scheduling algorithms.CPU scheduling algorithms are divided into non-preemptive and preemptive.RoundRobin(RR)algorithm is the most famous as it is the basis for all the algorithms used in time-sharing.In this paper,the authors proposed a novel CPU scheduling algorithm based on RR.The proposed algorithm is called Adjustable Time Slice(ATS).It reduces the time cost by taking the advantage of the low overhead of RR algorithm.In addition,ATS favors short processes allowing them to run longer time than given to long processes.The specific characteristics of each process are;its CPU execution time,weight,time slice,and number of context switches.ATS clusters the processes in groups depending on these characteristics.The traditionalRRassigns fixed time slice for each process.On the other hand,dynamic variants of RR assign time slice for each process differs from other processes.The essential difference between ATS and the other methods is that it gives a set of processes a specific time based on their similarities within the same cluster.The authors compared between ATS with five popular scheduling algorithms on nine datasets of processes.The datasets used in the comparison vary in their features.The evaluation was measured in term of time cost and the experiments showed that the proposed algorithm reduces the time cost.展开更多
CPU scheduling is the basic task within any time-shared operating system.One of the main goals of the researchers interested in CPU scheduling is minimizing time cost.Comparing between CPU scheduling algorithms is sub...CPU scheduling is the basic task within any time-shared operating system.One of the main goals of the researchers interested in CPU scheduling is minimizing time cost.Comparing between CPU scheduling algorithms is subject to some scheduling criteria(e.g.,turnaround time,waiting time and number of context switches(NCS)).Scheduling policy is divided into preemptive and non-preemptive.Round Robin(RR)algorithm is the most common preemptive scheduling algorithm used in the time-shared operating systems.In this paper,the authors proposed a modified version of the RR algorithm,called dynamic time slice(DTS),to combine the advantageous of the low scheduling overhead of the RR and favor short process for the sake of minimizing time cost.Each process has a weight proportional to the weights of all processes.The process’s weight determines its time slice within the current period.The authors benefit from the clustering technique in grouping the processes that are similar in their attributes(e.g.,CPU service time,weight,allowed time slice(ATS),proportional burst time(PBT)and NCS).Each process in a cluster is assigned the average of the processes’time slices in this cluster.A comparative study of six popular scheduling algorithms and the proposed approach on nine groups of processes vary in their attributes was performed and the evaluation was measured in terms of waiting and turnaround times,and NCS.The experiments showed that the proposed algorithm gives better results.展开更多
An effective method to detect stepping-stone intrusion(SSI)is to estimate the length of a connection chain.This type of detection method is referred to as a network-based detection approach.Existing network-based SSI ...An effective method to detect stepping-stone intrusion(SSI)is to estimate the length of a connection chain.This type of detection method is referred to as a network-based detection approach.Existing network-based SSI detection methods are either ineffective in the context of the Internet because of the presence of outliers in the packet round-trip times(RTTs)or inefficient,as many packets must be captured and processed.Because of the high fluctuation caused by the intermediate routers on the Internet,it is unavoidable that the RTTs of the captured packets contain outlier values.In this paper,we first propose an efficient algorithm to eliminate most of the possible RTT outliers of the packets captured in the Internet environment.We then develop an efficient SSI detection algorithm by mining network traffic using an improved version of k-Means clustering.Our proposed detection algorithm for SSI is accurate,effective,and efficient in the context of the Internet.Well-designed network experiments are conducted in the Internet environment to verify the effectiveness,correctness,and efficiency of our proposed algorithms.Our experiments show that the effective rate of our proposed SSI detection algorithm is higher than 85.7%in the context of the Internet.展开更多
Active queue management(AQM)is essential to prevent the degradation of quality of service in TCP/AQM systems with round-trip time(RTT)delay.RTT delays are primarily caused by packet-propagation delays,but they can als...Active queue management(AQM)is essential to prevent the degradation of quality of service in TCP/AQM systems with round-trip time(RTT)delay.RTT delays are primarily caused by packet-propagation delays,but they can also be caused by the processing time of queuing operations and dynamically changing network situations.This study focuses on the design and analysis of an AQM digital controller under time-delay uncertainty.The controller is based on the Smith predictor algorithm and is called the SMITHPI controller.This study also demonstrates the stability of the controller and its robustness against network parameter variations such as the number of TCP connections,time delays,and user datagram protocol flows.The performance,robustness,and effectiveness of the proposed SMITHPI controller are evaluated using the NS-2 simulator.Finally,the performance of the SMITHPI controller is compared with that of a well-known queue-based AQM,called the proportional-integral controller.展开更多
In a majority of cases of long-time numerical integration for initial-value problems, roundoff error has received little attention. Using twenty-nine numerical methods, the influence of round-off error on numerical so...In a majority of cases of long-time numerical integration for initial-value problems, roundoff error has received little attention. Using twenty-nine numerical methods, the influence of round-off error on numerical solutions is generally studied through a large number of numerical experiments. Here we find that there exists a strong dependence on machine precision (which is a new kind of dependence different from the sensitive dependence on initial conditions), maximally effective computation time (MECT) and optimal stepsize (OS) in solving nonlinear ordinary differential equations (ODEs) in finite machine precision. And an optimal searching method for evaluating MECT and OS under finite machine precision is presented. The relationships between MECT, OS, the order of numerical method and machine precision are found. Numerical results show that round-off error plays a significant role in the above phenomena. Moreover, we find two universal relations which are independent of the types of ODEs, initial values and numerical schemes. Based on the results of numerical experiments, we present a computational uncertainty principle, which is a great challenge to the reliability of long-time numerical integration for nonlinear ODEs.展开更多
The error propagation for general numerical method in ordinarydifferential equations ODEs is studied. Three kinds of convergence, theoretical, numerical and actual convergences, are presented. The various components o...The error propagation for general numerical method in ordinarydifferential equations ODEs is studied. Three kinds of convergence, theoretical, numerical and actual convergences, are presented. The various components of round-off error occurring in floating-point computation are fully detailed. By introducing a new kind of recurrent inequality, the classical error bounds for linear multistep methods are essentially improved, and joining probabilistic theory the “normal” growth of accumulated round-off error is derived. Moreover, a unified estimate for the total error of general method is given. On the basis of these results, we rationally interpret the various phenomena found in the numerical experiments in part I of this paper and derive two universal relations which are independent of types of ODEs, initial values and numerical schemes and are consistent with the numerical results. Furthermore, we give the explicitly mathematical expression of the computational uncertainty principle and expound the intrinsic relation between two uncertainties which result from the inaccuracies of numerical method and calculating machine.展开更多
文摘Modern human life is heavily dependent on computing systems and one of the core components affecting the performance of these systems is underlying operating system.Operating systems need to be upgraded to match the needs of modern-day systems relying on Internet of Things,Fog computing and Mobile based applications.The scheduling algorithm of the operating system dictates that how the resources will be allocated to the processes and the Round Robin algorithm(RR)has been widely used for it.The intent of this study is to ameliorate RR scheduling algorithm to optimize task scheduling.We have carried out an experimental study where we have developed four variations of RR,each algorithm considers three-time quanta and the performance of these variations was compared with the RR algorithm,and results highlighted that these variations performed better than conventional RR algorithm.In the future,we intend to develop an automated scheduler that can determine optimal algorithm based on the current set of processes and will allocate time quantum to the processes intelligently at the run time.This way the task performance of modern-day systems can be improved to make them more efficient.
基金The authors extend their appreciation to Deanship of Scientific Research at King Khalid University for funding this work through the Research Groups Project under Grant Number RGP.1/95/42.
文摘Minimizing time cost in time-shared operating systems is considered basic and essential task,and it is the most significant goal for the researchers who interested in CPU scheduling algorithms.Waiting time,turnaround time,and number of context switches are themost time cost criteria used to compare between CPU scheduling algorithms.CPU scheduling algorithms are divided into non-preemptive and preemptive.RoundRobin(RR)algorithm is the most famous as it is the basis for all the algorithms used in time-sharing.In this paper,the authors proposed a novel CPU scheduling algorithm based on RR.The proposed algorithm is called Adjustable Time Slice(ATS).It reduces the time cost by taking the advantage of the low overhead of RR algorithm.In addition,ATS favors short processes allowing them to run longer time than given to long processes.The specific characteristics of each process are;its CPU execution time,weight,time slice,and number of context switches.ATS clusters the processes in groups depending on these characteristics.The traditionalRRassigns fixed time slice for each process.On the other hand,dynamic variants of RR assign time slice for each process differs from other processes.The essential difference between ATS and the other methods is that it gives a set of processes a specific time based on their similarities within the same cluster.The authors compared between ATS with five popular scheduling algorithms on nine datasets of processes.The datasets used in the comparison vary in their features.The evaluation was measured in term of time cost and the experiments showed that the proposed algorithm reduces the time cost.
文摘CPU scheduling is the basic task within any time-shared operating system.One of the main goals of the researchers interested in CPU scheduling is minimizing time cost.Comparing between CPU scheduling algorithms is subject to some scheduling criteria(e.g.,turnaround time,waiting time and number of context switches(NCS)).Scheduling policy is divided into preemptive and non-preemptive.Round Robin(RR)algorithm is the most common preemptive scheduling algorithm used in the time-shared operating systems.In this paper,the authors proposed a modified version of the RR algorithm,called dynamic time slice(DTS),to combine the advantageous of the low scheduling overhead of the RR and favor short process for the sake of minimizing time cost.Each process has a weight proportional to the weights of all processes.The process’s weight determines its time slice within the current period.The authors benefit from the clustering technique in grouping the processes that are similar in their attributes(e.g.,CPU service time,weight,allowed time slice(ATS),proportional burst time(PBT)and NCS).Each process in a cluster is assigned the average of the processes’time slices in this cluster.A comparative study of six popular scheduling algorithms and the proposed approach on nine groups of processes vary in their attributes was performed and the evaluation was measured in terms of waiting and turnaround times,and NCS.The experiments showed that the proposed algorithm gives better results.
基金supported by the the National Centers of Academic Excellence in Cybersecurity(NCAE-C)Grant(No.H98230-20-1-0293)at the National Security Agency with Columbus State University,Georgia,USA。
文摘An effective method to detect stepping-stone intrusion(SSI)is to estimate the length of a connection chain.This type of detection method is referred to as a network-based detection approach.Existing network-based SSI detection methods are either ineffective in the context of the Internet because of the presence of outliers in the packet round-trip times(RTTs)or inefficient,as many packets must be captured and processed.Because of the high fluctuation caused by the intermediate routers on the Internet,it is unavoidable that the RTTs of the captured packets contain outlier values.In this paper,we first propose an efficient algorithm to eliminate most of the possible RTT outliers of the packets captured in the Internet environment.We then develop an efficient SSI detection algorithm by mining network traffic using an improved version of k-Means clustering.Our proposed detection algorithm for SSI is accurate,effective,and efficient in the context of the Internet.Well-designed network experiments are conducted in the Internet environment to verify the effectiveness,correctness,and efficiency of our proposed algorithms.Our experiments show that the effective rate of our proposed SSI detection algorithm is higher than 85.7%in the context of the Internet.
文摘Active queue management(AQM)is essential to prevent the degradation of quality of service in TCP/AQM systems with round-trip time(RTT)delay.RTT delays are primarily caused by packet-propagation delays,but they can also be caused by the processing time of queuing operations and dynamically changing network situations.This study focuses on the design and analysis of an AQM digital controller under time-delay uncertainty.The controller is based on the Smith predictor algorithm and is called the SMITHPI controller.This study also demonstrates the stability of the controller and its robustness against network parameter variations such as the number of TCP connections,time delays,and user datagram protocol flows.The performance,robustness,and effectiveness of the proposed SMITHPI controller are evaluated using the NS-2 simulator.Finally,the performance of the SMITHPI controller is compared with that of a well-known queue-based AQM,called the proportional-integral controller.
文摘In a majority of cases of long-time numerical integration for initial-value problems, roundoff error has received little attention. Using twenty-nine numerical methods, the influence of round-off error on numerical solutions is generally studied through a large number of numerical experiments. Here we find that there exists a strong dependence on machine precision (which is a new kind of dependence different from the sensitive dependence on initial conditions), maximally effective computation time (MECT) and optimal stepsize (OS) in solving nonlinear ordinary differential equations (ODEs) in finite machine precision. And an optimal searching method for evaluating MECT and OS under finite machine precision is presented. The relationships between MECT, OS, the order of numerical method and machine precision are found. Numerical results show that round-off error plays a significant role in the above phenomena. Moreover, we find two universal relations which are independent of the types of ODEs, initial values and numerical schemes. Based on the results of numerical experiments, we present a computational uncertainty principle, which is a great challenge to the reliability of long-time numerical integration for nonlinear ODEs.
基金This work was supported by the Knowledge Innovation Key Project of Chinese Academy of Sciences inthe Resource Environment Field (KZCX1-203) Outstanding State Key Laboratory Project (Grant No. 49823002) the National Natural Science Foundation of C
文摘The error propagation for general numerical method in ordinarydifferential equations ODEs is studied. Three kinds of convergence, theoretical, numerical and actual convergences, are presented. The various components of round-off error occurring in floating-point computation are fully detailed. By introducing a new kind of recurrent inequality, the classical error bounds for linear multistep methods are essentially improved, and joining probabilistic theory the “normal” growth of accumulated round-off error is derived. Moreover, a unified estimate for the total error of general method is given. On the basis of these results, we rationally interpret the various phenomena found in the numerical experiments in part I of this paper and derive two universal relations which are independent of types of ODEs, initial values and numerical schemes and are consistent with the numerical results. Furthermore, we give the explicitly mathematical expression of the computational uncertainty principle and expound the intrinsic relation between two uncertainties which result from the inaccuracies of numerical method and calculating machine.