This paper introduces the integration of the Social Group Optimization(SGO)algorithm to enhance the accuracy of software cost estimation using the Constructive Cost Model(COCOMO).COCOMO’s fixed coefficients often lim...This paper introduces the integration of the Social Group Optimization(SGO)algorithm to enhance the accuracy of software cost estimation using the Constructive Cost Model(COCOMO).COCOMO’s fixed coefficients often limit its adaptability,as they don’t account for variations across organizations.By fine-tuning these parameters with SGO,we aim to improve estimation accuracy.We train and validate our SGO-enhanced model using historical project data,evaluating its performance with metrics like the mean magnitude of relative error(MMRE)and Manhattan distance(MD).Experimental results show that SGO optimization significantly improves the predictive accuracy of software cost models,offering valuable insights for project managers and practitioners in the field.However,the approach’s effectiveness may vary depending on the quality and quantity of available historical data,and its scalability across diverse project types and sizes remains a key consideration for future research.展开更多
This article is an addendum to the 2001 paper [1] which investigated an approach to hierarchical clustering based on the level sets of a density function induced on data points in a d-dimensional feature space. We ref...This article is an addendum to the 2001 paper [1] which investigated an approach to hierarchical clustering based on the level sets of a density function induced on data points in a d-dimensional feature space. We refer to this as the “level-sets approach” to hierarchical clustering. The density functions considered in [1] were those formed as the sum of identical radial basis functions centered at the data points, each radial basis function assumed to be continuous, monotone decreasing, convex on every ray, and rising to positive infinity at its center point. Such a framework can be investigated with respect to both the Euclidean (L2) and Manhattan (L1) metrics. The addendum here puts forth some observations and questions about the level-sets approach that go beyond those in [1]. In particular, we detail and ask the following questions. How does the level-sets approach compare with other related approaches? How is the resulting hierarchical clustering affected by the choice of radial basis function? What are the structural properties of a function formed as the sum of radial basis functions? Can the levels-sets approach be theoretically validated? Is there an efficient algorithm to implement the level-sets approach?展开更多
In several fields like financial dealing,industry,business,medicine,et cetera,Big Data(BD)has been utilized extensively,which is nothing but a collection of a huge amount of data.However,it is highly complicated alon...In several fields like financial dealing,industry,business,medicine,et cetera,Big Data(BD)has been utilized extensively,which is nothing but a collection of a huge amount of data.However,it is highly complicated along with time-consuming to process a massive amount of data.Thus,to design the Distribution Preserving Framework for BD,a novel methodology has been proposed utilizing Manhattan Distance(MD)-centered Partition Around Medoid(MD–PAM)along with Conjugate Gradient Artificial Neural Network(CG-ANN),which undergoes various steps to reduce the complications of BD.Firstly,the data are processed in the pre-processing phase by mitigating the data repetition utilizing the map-reduce function;subsequently,the missing data are handled by substituting or by ignoring the missed values.After that,the data are transmuted into a normalized form.Next,to enhance the classification performance,the data’s dimensionalities are minimized by employing Gaussian Kernel(GK)-Fisher Discriminant Analysis(GK-FDA).Afterwards,the processed data is submitted to the partitioning phase after transmuting it into a structured format.In the partition phase,by utilizing the MD-PAM,the data are partitioned along with grouped into a cluster.Lastly,by employing CG-ANN,the data are classified in the classification phase so that the needed data can be effortlessly retrieved by the user.To analogize the outcomes of the CG-ANN with the prevailing methodologies,the NSL-KDD openly accessible datasets are utilized.The experiential outcomes displayed that an efficient result along with a reduced computation cost was shown by the proposed CG-ANN.The proposed work outperforms well in terms of accuracy,sensitivity and specificity than the existing systems.展开更多
During the pre-design stage of buildings,reliable long-term prediction of thermal loads is significant for cool-ing/heating system configuration and efficient operation.This paper proposes a surrogate modeling method ...During the pre-design stage of buildings,reliable long-term prediction of thermal loads is significant for cool-ing/heating system configuration and efficient operation.This paper proposes a surrogate modeling method to predict all-year hourly cooling/heating loads in high resolution for retail,hotel,and office buildings.16384 surrogate models are simulated in EnergyPlus to generate the load database,which contains 7 crucial building features as inputs and hourly loads as outputs.K-nearest-neighbors(KNN)is chosen as the data-driven algorithm to approximate the surrogates for load prediction.With test samples from the database,performances of five different spatial metrics for KNN are evaluated and optimized.Results show that the Manhattan distance is the optimal metric with the highest efficient hour rates of 93.57%and 97.14%for cooling and heating loads in office buildings.The method is verified by predicting the thermal loads of a given district in Shanghai,China.The mean absolute percentage errors(MAPE)are 5.26%and 6.88%for cooling/heating loads,respectively,and 5.63%for the annual thermal loads.The proposed surrogate modeling method meets the precision requirement of engineering in the building pre-design stage and achieves the fast prediction of all-year hourly thermal loads at the district level.As a data-driven approximation,it does not require as much detailed building information as the commonly used physics-based methods.And by pre-simulation of sufficient prototypical models,the method overcomes the gaps of data missing in current data-driven methods.展开更多
Particle swarm optimization(PSO)is a type of swarm intelligence algorithm that is frequently used to resolve specific global optimization problems due to its rapid convergence and ease of operation.However,PSO still h...Particle swarm optimization(PSO)is a type of swarm intelligence algorithm that is frequently used to resolve specific global optimization problems due to its rapid convergence and ease of operation.However,PSO still has certain deficiencies,such as a poor trade-off between exploration and exploitation and premature convergence.Hence,this paper proposes a dual-stage hybrid learning particle swarm optimization(DHLPSO).In the algorithm,the iterative process is partitioned into two stages.The learning strategy used at each stage emphasizes exploration and exploitation,respectively.In the first stage,to increase population variety,a Manhattan distance based learning strategy is proposed.In this strategy,each particle chooses the furthest Manhattan distance particle and a better particle for learning.In the second stage,an excellent example learning strategy is adopted to perform local optimization operations on the population,in which each particle learns from the global optimal particle and a better particle.Utilizing the Gaussian mutation strategy,the algorithm’s searchability in particular multimodal functions is significantly enhanced.On benchmark functions from CEC 2013,DHLPSO is evaluated alongside other PSO variants already in existence.The comparison results clearly demonstrate that,compared to other cutting-edge PSO variations,DHLPSO implements highly competitive performance in handling global optimization problems.展开更多
Few-shot Learning algorithms can be effectively applied to fields where certain categories have only a small amount of data or a small amount of labeled data,such as medical images,terrorist surveillance,and so on.The...Few-shot Learning algorithms can be effectively applied to fields where certain categories have only a small amount of data or a small amount of labeled data,such as medical images,terrorist surveillance,and so on.The Metric Learning in the Few-shot Learning algorithmis classified by measuring the similarity between the classified samples and the unclassified samples.This paper improves the Prototypical Network in the Metric Learning,and changes its core metric function to Manhattan distance.The Convolutional Neural Network of the embedded module is changed,and mechanisms such as average pooling and Dropout are added.Through comparative experiments,it is found that thismodel can converge in a small number of iterations(below 15,000 episodes),and its performance exceeds algorithms such asMAML.Research shows that replacingManhattan distance with Euclidean distance can effectively improve the classification effect of the Prototypical Network,and mechanisms such as average pooling and Dropout can also effectively improve the model.展开更多
The grey wolf optimizer(GWO)is a swarm-based intelligence optimization algorithm by simulating the steps of searching,encircling,and attacking prey in the process of wolf hunting.Along with its advantages of simple pr...The grey wolf optimizer(GWO)is a swarm-based intelligence optimization algorithm by simulating the steps of searching,encircling,and attacking prey in the process of wolf hunting.Along with its advantages of simple principle and few parameters setting,GWO bears drawbacks such as low solution accuracy and slow convergence speed.A few recent advanced GWOs are proposed to try to overcome these disadvantages.However,they are either difficult to apply to large-scale problems due to high time complexity or easily lead to early convergence.To solve the abovementioned issues,a high-accuracy variable grey wolf optimizer(VGWO)with low time complexity is proposed in this study.VGWO first uses the symmetrical wolf strategy to generate an initial population of individuals to lay the foundation for the global seek of the algorithm,and then inspired by the simulated annealing algorithm and the differential evolution algorithm,a mutation operation for generating a new mutant individual is performed on three wolves which are randomly selected in the current wolf individuals while after each iteration.A vectorized Manhattan distance calculation method is specifically designed to evaluate the probability of selecting the mutant individual based on its status in the current wolf population for the purpose of dynamically balancing global search and fast convergence capability of VGWO.A series of experiments are conducted on 19 benchmark functions from CEC2014 and CEC2020 and three real-world engineering cases.For 19 benchmark functions,VGWO’s optimization results place first in 80%of comparisons to the state-of-art GWOs and the CEC2020 competition winner.A further evaluation based on the Friedman test,VGWO also outperforms all other algorithms statistically in terms of robustness with a better average ranking value.展开更多
In laying routes on printed circuit boards and VLSI chips, wires are usually allowed to run rectilinearly. This makes people consider the number of intersections between two rectangular paths on a plane. A Manhattan p...In laying routes on printed circuit boards and VLSI chips, wires are usually allowed to run rectilinearly. This makes people consider the number of intersections between two rectangular paths on a plane. A Manhattan path on the plane consists of a horizontal line segment and a vertical line segment sharing a common endpoint. A rectangular path P展开更多
文摘This paper introduces the integration of the Social Group Optimization(SGO)algorithm to enhance the accuracy of software cost estimation using the Constructive Cost Model(COCOMO).COCOMO’s fixed coefficients often limit its adaptability,as they don’t account for variations across organizations.By fine-tuning these parameters with SGO,we aim to improve estimation accuracy.We train and validate our SGO-enhanced model using historical project data,evaluating its performance with metrics like the mean magnitude of relative error(MMRE)and Manhattan distance(MD).Experimental results show that SGO optimization significantly improves the predictive accuracy of software cost models,offering valuable insights for project managers and practitioners in the field.However,the approach’s effectiveness may vary depending on the quality and quantity of available historical data,and its scalability across diverse project types and sizes remains a key consideration for future research.
文摘This article is an addendum to the 2001 paper [1] which investigated an approach to hierarchical clustering based on the level sets of a density function induced on data points in a d-dimensional feature space. We refer to this as the “level-sets approach” to hierarchical clustering. The density functions considered in [1] were those formed as the sum of identical radial basis functions centered at the data points, each radial basis function assumed to be continuous, monotone decreasing, convex on every ray, and rising to positive infinity at its center point. Such a framework can be investigated with respect to both the Euclidean (L2) and Manhattan (L1) metrics. The addendum here puts forth some observations and questions about the level-sets approach that go beyond those in [1]. In particular, we detail and ask the following questions. How does the level-sets approach compare with other related approaches? How is the resulting hierarchical clustering affected by the choice of radial basis function? What are the structural properties of a function formed as the sum of radial basis functions? Can the levels-sets approach be theoretically validated? Is there an efficient algorithm to implement the level-sets approach?
文摘In several fields like financial dealing,industry,business,medicine,et cetera,Big Data(BD)has been utilized extensively,which is nothing but a collection of a huge amount of data.However,it is highly complicated along with time-consuming to process a massive amount of data.Thus,to design the Distribution Preserving Framework for BD,a novel methodology has been proposed utilizing Manhattan Distance(MD)-centered Partition Around Medoid(MD–PAM)along with Conjugate Gradient Artificial Neural Network(CG-ANN),which undergoes various steps to reduce the complications of BD.Firstly,the data are processed in the pre-processing phase by mitigating the data repetition utilizing the map-reduce function;subsequently,the missing data are handled by substituting or by ignoring the missed values.After that,the data are transmuted into a normalized form.Next,to enhance the classification performance,the data’s dimensionalities are minimized by employing Gaussian Kernel(GK)-Fisher Discriminant Analysis(GK-FDA).Afterwards,the processed data is submitted to the partitioning phase after transmuting it into a structured format.In the partition phase,by utilizing the MD-PAM,the data are partitioned along with grouped into a cluster.Lastly,by employing CG-ANN,the data are classified in the classification phase so that the needed data can be effortlessly retrieved by the user.To analogize the outcomes of the CG-ANN with the prevailing methodologies,the NSL-KDD openly accessible datasets are utilized.The experiential outcomes displayed that an efficient result along with a reduced computation cost was shown by the proposed CG-ANN.The proposed work outperforms well in terms of accuracy,sensitivity and specificity than the existing systems.
基金This work was supported by the National Natural Science Foundation of China(Grant No.51978481).
文摘During the pre-design stage of buildings,reliable long-term prediction of thermal loads is significant for cool-ing/heating system configuration and efficient operation.This paper proposes a surrogate modeling method to predict all-year hourly cooling/heating loads in high resolution for retail,hotel,and office buildings.16384 surrogate models are simulated in EnergyPlus to generate the load database,which contains 7 crucial building features as inputs and hourly loads as outputs.K-nearest-neighbors(KNN)is chosen as the data-driven algorithm to approximate the surrogates for load prediction.With test samples from the database,performances of five different spatial metrics for KNN are evaluated and optimized.Results show that the Manhattan distance is the optimal metric with the highest efficient hour rates of 93.57%and 97.14%for cooling and heating loads in office buildings.The method is verified by predicting the thermal loads of a given district in Shanghai,China.The mean absolute percentage errors(MAPE)are 5.26%and 6.88%for cooling/heating loads,respectively,and 5.63%for the annual thermal loads.The proposed surrogate modeling method meets the precision requirement of engineering in the building pre-design stage and achieves the fast prediction of all-year hourly thermal loads at the district level.As a data-driven approximation,it does not require as much detailed building information as the commonly used physics-based methods.And by pre-simulation of sufficient prototypical models,the method overcomes the gaps of data missing in current data-driven methods.
基金the National Natural Science Foundation of China(Nos.62066019 and 61903089)the Natural Science Foundation of Jiangxi Province(Nos.20202BABL202020 and 20202BAB202014)the Graduate Innovation Foundation of Jiangxi University of Science and Technology(Nos.XY2021-S092 and YC2022-S641).
文摘Particle swarm optimization(PSO)is a type of swarm intelligence algorithm that is frequently used to resolve specific global optimization problems due to its rapid convergence and ease of operation.However,PSO still has certain deficiencies,such as a poor trade-off between exploration and exploitation and premature convergence.Hence,this paper proposes a dual-stage hybrid learning particle swarm optimization(DHLPSO).In the algorithm,the iterative process is partitioned into two stages.The learning strategy used at each stage emphasizes exploration and exploitation,respectively.In the first stage,to increase population variety,a Manhattan distance based learning strategy is proposed.In this strategy,each particle chooses the furthest Manhattan distance particle and a better particle for learning.In the second stage,an excellent example learning strategy is adopted to perform local optimization operations on the population,in which each particle learns from the global optimal particle and a better particle.Utilizing the Gaussian mutation strategy,the algorithm’s searchability in particular multimodal functions is significantly enhanced.On benchmark functions from CEC 2013,DHLPSO is evaluated alongside other PSO variants already in existence.The comparison results clearly demonstrate that,compared to other cutting-edge PSO variations,DHLPSO implements highly competitive performance in handling global optimization problems.
文摘Few-shot Learning algorithms can be effectively applied to fields where certain categories have only a small amount of data or a small amount of labeled data,such as medical images,terrorist surveillance,and so on.The Metric Learning in the Few-shot Learning algorithmis classified by measuring the similarity between the classified samples and the unclassified samples.This paper improves the Prototypical Network in the Metric Learning,and changes its core metric function to Manhattan distance.The Convolutional Neural Network of the embedded module is changed,and mechanisms such as average pooling and Dropout are added.Through comparative experiments,it is found that thismodel can converge in a small number of iterations(below 15,000 episodes),and its performance exceeds algorithms such asMAML.Research shows that replacingManhattan distance with Euclidean distance can effectively improve the classification effect of the Prototypical Network,and mechanisms such as average pooling and Dropout can also effectively improve the model.
文摘The grey wolf optimizer(GWO)is a swarm-based intelligence optimization algorithm by simulating the steps of searching,encircling,and attacking prey in the process of wolf hunting.Along with its advantages of simple principle and few parameters setting,GWO bears drawbacks such as low solution accuracy and slow convergence speed.A few recent advanced GWOs are proposed to try to overcome these disadvantages.However,they are either difficult to apply to large-scale problems due to high time complexity or easily lead to early convergence.To solve the abovementioned issues,a high-accuracy variable grey wolf optimizer(VGWO)with low time complexity is proposed in this study.VGWO first uses the symmetrical wolf strategy to generate an initial population of individuals to lay the foundation for the global seek of the algorithm,and then inspired by the simulated annealing algorithm and the differential evolution algorithm,a mutation operation for generating a new mutant individual is performed on three wolves which are randomly selected in the current wolf individuals while after each iteration.A vectorized Manhattan distance calculation method is specifically designed to evaluate the probability of selecting the mutant individual based on its status in the current wolf population for the purpose of dynamically balancing global search and fast convergence capability of VGWO.A series of experiments are conducted on 19 benchmark functions from CEC2014 and CEC2020 and three real-world engineering cases.For 19 benchmark functions,VGWO’s optimization results place first in 80%of comparisons to the state-of-art GWOs and the CEC2020 competition winner.A further evaluation based on the Friedman test,VGWO also outperforms all other algorithms statistically in terms of robustness with a better average ranking value.
基金Project supported by the National Natural Science Foundation of China. A preliminary version of this work was presented at the First Chinese SIAM Conference, Beijing. Nov. 1-3, 1990
文摘In laying routes on printed circuit boards and VLSI chips, wires are usually allowed to run rectilinearly. This makes people consider the number of intersections between two rectangular paths on a plane. A Manhattan path on the plane consists of a horizontal line segment and a vertical line segment sharing a common endpoint. A rectangular path P