Table 1 - uploaded by Bilal Abed-alguni
Content may be subject to copyright.
Example of random selection scheme. 

Example of random selection scheme. 

Source publication
Article
Full-text available
The selection process is an important part of any optimization algorithm. Usually, an efficient selection process should balance between exploration of the search space and exploitation of the current knowledge about the best solutions. Cuckoo search (CS) is a simple yet powerful optimization algorithm inspired by the parasitic reproduction behavio...

Contexts in source publication

Context 1
... n is the number of solutions. Table 1 shows an example of random selection scheme for 4 solutions ranked from best to worst. Each solution has an equal opportunity to be selected as a cuckoo that can be enhanced using Lévy flight (p i = 1/4 = 0.25). ...
Context 2
... all comparative tables, the best result for each benchmark function is highlighted with bold. Tables 8 to 11 (10D, 30D, 100D and 1000D, receptively) show the experimental results for each of the 20 benchmark functions described in Table 6. The results are in the format: average of best solutions over 100 independent runs (first row), standard deviation of the best (second row) and error function value EFV (third row) for the average of the best solutions. ...
Context 3
... EFV measures the distance from the global optimal solution to the average of the best solutions [32]. Table 7 shows a summary of Tables 8 to 11. The overall results suggest that all the proposed variations of the CS algorithm outperform the original CS algorithm (RCS variation). ...
Context 4
... in Tables 8 and 9, the obtained results in Table 10 (D=100 test functions) suggest that all the variations of CS perform better than RCS. Apparently, BCS achieves five best results for F2, F5,F10, F16 and F18 functions and EGCS achieves five best results for F4, F6, F7, F11 and F20 functions. ...
Context 5
... BCS achieves five best results for F2, F5,F10, F16 and F18 functions and EGCS achieves five best results for F4, F6, F7, F11 and F20 functions. No significant observation was noted when the dimensionality of the problems was increased to 1,000 as shown in Table 11. The only notable observation is that EGCS performs better than the other variations for six functions. ...
Context 6
... EGCS outperforms the other variations for complex test functions (selected functions from CEC 2005, F11-F20 in the current paper). Table 16 summarizes the experimental results of the variations of CS using the 20 test functions described in Table 6. The results are in the format: average number of iterations ±, standard deviation of iterations and the percentage of enhancement compared to RCS is given between two parentheses. ...
Context 7
... results are in the format: average number of iterations ±, standard deviation of iterations and the percentage of enhancement compared to RCS is given between two parentheses. In Table 16, the number of iterations for each algorithm was recorded when one of the following conditions is satisfied: ...
Context 8
... can be observed from Table 16 that the GCS and PCS variations converge faster to solutions for 5 test functions each (unimodal or basic multimodal functions). This may be because GCS follows a greedy approach that is suitable for reproducing solutions for simple functions. ...
Context 9
... µ 7 H 1 : at least one mean is different from the others. Tables 12 to 15 provide the ANOVA results for all the test functions with different dimensions (Table 12: D=10, Table 13: D=30, Table 14: D=100 and Table 15: D=1,000). As shown in the tables, the P -value for all the problems except for F3 (Step Function) is less than 5% and the value of F calculated (F -cal) is larger than the value of F critical (F -crit). ...
Context 10
... µ 7 H 1 : at least one mean is different from the others. Tables 12 to 15 provide the ANOVA results for all the test functions with different dimensions (Table 12: D=10, Table 13: D=30, Table 14: D=100 and Table 15: D=1,000). As shown in the tables, the P -value for all the problems except for F3 (Step Function) is less than 5% and the value of F calculated (F -cal) is larger than the value of F critical (F -crit). ...
Context 11
... 6.40E-05 5.37E-05 3.60E-05 Table 10: Experimental results of seven variations of cuckoo search for twenty benchmark functions, D=100, runs=100, iterations=10,000. RCS GCS PCS BCS ECS EGCS RLCS F1 1.79E-06 3.21E-07 4.49E-07 4.68E-07 2.84E-07 3.26E-07 2.25E-07 4.04E-06 4.27E-07 7.27E-07 8.56E-07 4.05E-07 7.14E-07 3.60E-07 1.79E-06 ...
Context 12
... 6.40E-05 5.37E-05 3.60E-05 Table 10: Experimental results of seven variations of cuckoo search for twenty benchmark functions, D=100, runs=100, iterations=10,000. RCS GCS PCS BCS ECS EGCS RLCS F1 1.79E-06 3.21E-07 4.49E-07 4.68E-07 2.84E-07 3.26E-07 2.25E-07 4.04E-06 4.27E-07 7.27E-07 8.56E-07 4.05E-07 7.14E-07 3.60E-07 1.79E-06 ...
Context 13
... 0.00E+00 5.10E-04 9.03E-04 0.00E+00 6.91E-04 Continued on next page Table 16 -continued from previous page Function Name RCS GCS PCS BCS ECS EGCS RLCS 0% -3% 12% 6% 0% 22% 30% Continued on next page ...

Similar publications

Preprint
Full-text available
Policy learning in multi-agent reinforcement learning (MARL) is challenging due to the exponential growth of joint state-action space with respect to the number of agents. To achieve higher scalability, the paradigm of centralized training with decentralized execution (CTDE) is broadly adopted with factorized structure in MARL. However, we observe...

Citations

... Nonetheless, CS has one important limitation, which is the possibility of premature convergence and slow convergence rates. So, Abed-alguni and Alkhateeb suggested six modifications to the CS algorithm, which significantly exceeded the benchmark functions results that were achieved by the original CS approach [4]. ...
Article
The scheduling optimization subject is indeed one of the most effective solutions in cloud system area to ensure the high systems performance and resource usage. The novelty presented in this study is a mathematical model that draws on an Improved Cuckoo Search Algorithm as a scheduling non-linear optimizer for the cloud computing environment. The model under consideration takes into account the underlying complications as well as the varying nature of the use of cloud resources, primarily addressing the latency issue with priority given to the increase of the throughput. Improved Cuckoo Search (ICS) Algorithm, developed on 'cuckoo search' algorithm, uses adaptive step size and probabilistic switching, which, in turn, ensure that the solution space is both being properly examined and exploited. The applications’ utilization of the scheduling mathematical model is based on an elaborately designed theoretical background representing the task as a non-linear optimization problem that meets all the specific constraints and objectives important in cloud computing. Statistical reassessment involves the comparison of the proposed model with standard benchmarks. The proposed model superiority is evident in terms of convergence speed, solution quality and robustness against diverse workloads, as shown in varied tests. The findings of this research contributes to the field by deriving a specialized tool commonly used by the cloud service provider in designing resource scheduling, enhancing the service delivery and operational efficiency.
... In the modern landscape of optimization, the dynamic management of solution populations has transitioned from being a mere enhancement to an absolute necessity [11,12]. Addressing this critical need, our study sets forth with a meticulously crafted approach that marries the strengths of both metaheuristics and clustering techniques. ...
... Their selection is predicated on their consistent performance and resilience across diverse optimization challenges. Their legacy of success, coupled with their inherent capabilities of exploration and exploitation, underscores their suitability for our ambitious undertaking [11]. To elevate the potential of these bio-inspired solvers, we incorporate the ability of the Density-Based Spatial Clustering of Applications with Noise (DBSCAN) algorithm, a wide-known clustering methodology by the scientific community [13,14]. ...
... where α and γ are constants. Based on these approximations and idealization, the pseudo-code of BA is shown in Algorithm 2. if acceptance condition satisfied then 9 Accept new solution and update loudness and pulse rate; 10 end 11 Update the best global solution; 12 end 13 Post-process results; ...
Article
Full-text available
In the optimization field, the ability to efficiently tackle complex and high-dimensional problems remains a persistent challenge. Metaheuristic algorithms, with a particular emphasis on their autonomous variants, are emerging as promising tools to overcome this challenge. The term “autonomous” refers to these variants’ ability to dynamically adjust certain parameters based on their own outcomes, without external intervention. The objective is to leverage the advantages and characteristics of an unsupervised machine learning clustering technique to configure the population parameter with autonomous behavior, and emphasize how we incorporate the characteristics of search space clustering to enhance the intensification and diversification of the metaheuristic. This allows dynamic adjustments based on its own outcomes, whether by increasing or decreasing the population in response to the need for diversification or intensification of solutions. In this manner, it aims to imbue the metaheuristic with features for a broader search of solutions that can yield superior results. This study provides an in-depth examination of autonomous metaheuristic algorithms, including Autonomous Particle Swarm Optimization, Autonomous Cuckoo Search Algorithm, and Autonomous Bat Algorithm. We submit these algorithms to a thorough evaluation against their original counterparts using high-density functions from the well-known CEC LSGO benchmark suite. Quantitative results revealed performance enhancements in the autonomous versions, with Autonomous Particle Swarm Optimization consistently outperforming its peers in achieving optimal minimum values. Autonomous Cuckoo Search Algorithm and Autonomous Bat Algorithm also demonstrated noteworthy advancements over their traditional counterparts. A salient feature of these algorithms is the continuous nature of their population, which significantly bolsters their capability to navigate complex and high-dimensional search spaces. However, like all methodologies, there were challenges in ensuring consistent performance across all test scenarios. The intrinsic adaptability and autonomous decision making embedded within these algorithms herald a new era of optimization tools suited for complex real-world challenges. In sum, this research accentuates the potential of autonomous metaheuristics in the optimization arena, laying the groundwork for their expanded application across diverse challenges and domains. We recommend further explorations and adaptations of these autonomous algorithms to fully harness their potential.
... In Algorithm 3, the positions of k white sharks are modified inside the crossover loop (Lines 36-47). At iteration t of this loop, one of the fittest sharks w r t is selected using the roulette wheel selection mechanism [65]. In this selection mechanism (Eq. ...
Article
Full-text available
Intrusion Detection (ID) is an essential task in the cyberattacks domain built to secure Internet applications and networks from malicious actors. The main shortcoming of the current Intrusion Detection Systems (IDSs) is their attack detection performance because the existing solutions fail to detect the threats efficiently. Therefore, improving the prediction model of the IDS can improve its performance. Several studies suggest improving the Machine Learning (ML)-based IDS prediction models by selecting the most informative features from the security data before utilizing it in the ML-IDS model. Thus, meta-heuristic algorithms such as White Shark Optimizer (WSO) are adapted to deal with such feature selection problems. For the IDS prediction model, in this paper, the WSO algorithm is improved to cope with the binary domain of the feature selection task as follows: Firstly, two transfer functions are used to map the continuous domain into binary. Secondly, the modified K-means algorithm is proposed to assess building the initial population with a high level of diversity. Finally, several crossover operators are utilized to improve the evolution process of the binary WSO. The three improved versions are BIWSO1 which is WSO with transfer functions, BIWSO2, which is BIWSO1 with a modified k-means algorithm, and BIWSO3, which is BIWSO2 with crossover operators. The proposed versions of BIWSO are tested using twelve public real-world IDS and IoT datasets. Comparative evaluations against well-established meta-heuristic algorithms are conducted where the BIWSO3 proves its efficiency in terms of classification accuracy, precision, recall, and F1 measurements. For further validation, statistical evidence using Friedman’s, Wilcoxon’s, and Mann–Whitney U tests has been conducted where BIWSO3 proves its performance at a remarkably significant level.
... In the next section, we present three efficient metaheuristics to solve DAAO. The metaheuristic algorithms are among the powerful approaches to deal with the complex optimization problem when the commercial solvers or classical optimization are not as efficient as expected [39][40][41][42]. ...
Article
Intensity-modulated radiation therapy is a well-known technique for treating cancer patients worldwide. A treatment plan in this technique requires decision-making for three main problems: selection of beam angles, intensity map calculation, and leaf sequencing. Previous works investigated these problems sequentially. We present a new integrated framework for simultaneous decision-making of directions, intensities, and aperture shape, called direct angle and aperture optimization, and develop a mixed-integer nonlinear mathematical model for the problem. Due to the nonlinearity and the dimension of the problem, three efficient metaheuristics based on differential evolution (DE) called classic differential evolution (cDE), discrete differential evolution (dDE), and adaptive hybrid discrete differential evolution-particle swarm optimization (ahdDE-PSO) algorithms are designed to solve the problem. Parameters calibration is performed using the Taguchi design of experiments. The performance of the algorithms is evaluated by solving the problem for ten real cases of liver cancer disease from the TROTS data set. The performed ablation study and statistical analysis of computational results demonstrate that ahdDE-PSO is capable of finding high-quality treatment plans.
... Nature-inspired computing is a crucial discipline that brings to the development of novel optimization algorithms inspired by the natural behavior of flora and fauna. The most successful nature-inspired algorithms are Ant Colony Optimization [3], Particle Swarm Optimization [4], Bee algorithm [5], Grey Wolf Optimizer [6], Cuckoo search algorithm [7], and Bald Eagle Search algorithm [8]. ...
... The randomness of the wolves' movement is determined by the parameter 'a, which linearly drops from 2 to 0. The parameter 'a" is random but is vital in controlling the movement of wolves towards the prey. It depends on the current iteration count (i) and the maximum number of iterations (M), as shown in Eq (7). While searching the prey (Eq (8)), the wolves move to different positions according to the equations defined as ...
... . Over a few decades ago, Goldberg et al. [59] proposed the selection methods. These mechanisms were predominately in evolutionary algorithms such as fuzzy intelligent algorithms [60], bat-inspired algorithm [61], Modified Particle Swarm Optimization for Support Vector Machines [62], Whale Optimization algorithm in combination with simulated Annealing [63] and cuckoo search algorithm [7] and Bat algorithm [64]. ...
Article
Full-text available
In a fast-moving world, transportation consumes most of the time and resources. Traffic prediction has become a thrust application for machine learning algorithms to overcome the hurdles faced by congestion. Its accuracy determines the selection and existence of machine learning algorithms. The accuracy of such an algorithm is improved better by the proper tuning of the parameters. Support Vector Regression (SVR) is a well-known prediction mechanism. This paper exploits the Hybrid Grey Wolf Optimization–Bald Eagle Search (GWO-BES) algorithm for tuning SVR parameters, wherein the GWO selection methods are of natural selection. SVR-GWO-BES with natural selection has error performance increases by 48% in Mean Absolute Percentage Error and Root Mean Square Error, with the help of Caltrans Performance Measurement System (PeMS) open-source data and Chennai city traffic data for traffic forecasting. It is also shown that the increasing population of search agents increases the performance.
... CSA incorporates the optimization operators of the simulated annealing (SA) algorithm [28] into the improvement loop of the cuckoo search (CS) algorithm [29]. The experimental and statistical results in [27] suggest that CSA performs better than the particle swarm optimization (PSO) algorithm [30], bat Algorithm (BA) [31], Boltzman CS algorithm, and ε-greedy CS algorithm [32]. Besides, the results indicate that CSA has a stable reliable performance with various problem sizes (small, medium, and large) compared to the other algorithms. ...
... This algorithm iteratively moves the candidate solutions toward the best-so-far solution in the population of candidate solutions instead of moving them toward a neighboring solution as done in the original ABC algorithm. However, moving the candidate solutions repeatedly toward the best solutions causes imbalance between exploration and exploitation of the search space, which may degrade the performance of the algorithm [32]. ...
Article
Full-text available
The job shop scheduling problem (JSSP) is a popular NP-hard scheduling problem that simulates the scheduling problems of different real-world applications in fields such as manufacturing, engineering, and planning. In general, many optimization-based scheduling algorithms are guaranteed to provide an optimal solution for small instances of the JSSP, but do not perform well with large instances of the JSSP. The hybrid cuckoo search and simulated annealing (CSA) algorithm is a new hybrid optimization algorithm for solving continuous optimization problems. It incorporates the optimization operators of the simulated annealing algorithm into the cuckoo search algorithm. In this paper, we present discrete CSA (DCSA) which is a discrete variation of CSA for solving the JSSP. DCSA incorporates four modifications into CSA. First, it uses the opposition-based learning method in the initialization step to produce diverse candidate solutions. Second, it uses a combination of the variable neighborhood search (VNS) and Lévy-flight methods for better exploration of the search space. Third, it uses the elite opposition-based learning method before the abandon step of CSA to jump out of local optima. Finally, it uses the smallest position value with the JSSP to convert the continuous candidate solutions generated by CSA into discrete ones. DCSA was examined and compared to six popular optimization-based scheduling algorithms (DABC (Yin et al. in Sci Res Essays 6:2578–2596, 2011), PSO-VNS (Tasgetiren et al. in Int J Oper Res 3:120–135, 2006), DCS (Quaarab et al., in: Yang (ed) Recent advances in swarm intelligence and evolutionary computation, Springer, New York, 2015), GWO (Jiang and Zhang in IEEE Access 6:26231–26240, 2018), DWPA (Wang et al. in 2019 5th International Conference on Control, Automation and Robotics ICCAR 2019 581–585, 2019) and DGA (Kalshetty et al. in J Sci Res 64:310–321, 2020) using 34 JSSP instances from the OR-Library: Fisher and Thompson (3 instances), and Lawrence (31 instances). The experimental and statistical results indicate that DCSA converges faster to the Best-Known Solution for 29 instances out of the 34 tested JSSP instances and requires less computational time for the 34 tested instances than the computational times of the other algorithms.
... The needs and benefits of optimization techniques based on metaheuristics have motivated many scholars and scientists in the last few decades to apply these algorithms to solve NP-hard and complicated optimization problems [1][2][3][4]. Today, mathematical optimization techniques have played an important role in the industrial and non-industrial sectors [5]. ...
Article
Full-text available
Nowadays, a great deal of attention is paid to meta-heuristic algorithms to reach the approximate solution in an acceptable computational time. As one of the recently-developed successful meta-heuristics, Social Engineering Optimizer (SEO) algorithm is according to the inspiration of the rules of social engineering to solve approximate optimization problems. In this research, a Modified Social Engineering Optimizer algorithm (MSEO) by using an adjustment operator is proposed in which there are some assessment criteria for defender and attacker to determine and calculate the weight simultaneously for the first time. This enhancement comprises adding adjustment operators to improve the performance of SEO in terms of search accuracy and running time. Most notably, this operator is utilized to make a better new generation and improve the interaction between the search phases. The adjustment operator strategy is also applied to a novel division based on the best person. As an extensive comparison, the suggested algorithm is tested on fourteen standard benchmark functions and compared with ten well-established and recent optimization algorithms as well as the main version of the SEO algorithm. This algorithm is also tested for sensitivities on the parameters. In this regard, a set of engineering applications were provided to prove and validate the MSEO algorithm for the first time. The experimental outcomes show that the suggested algorithm produces very accurate results which are better than the SEO and other compared algorithms. Most notably, the MSEO provides a very competitive output and a high convergence rate.
... These facts motivated many researchers in the optimization field to use the PFSSP as a benchmark to prove the efficiency of their discrete optimization algorithms [1,3,6,8,9]. However, many of the proposed algorithms are either hybrid algorithms that require high number of computations [10][11][12] (e.g., Hybrid Harmony Search (HHS) [13], Hybrid Whale Optimization Algorithm (HWA), Improved Jaya (IJA) [8], New Ant Colony Algorithm (NACA) [13]) or are based on optimization algorithms that may converge earlier than expected to local optimal solutions (e.g., Heuristic based on the local best solution (HLBS) [14], New Cuckoo Search (NCS) [1]). ...
Article
Full-text available
The Permutation Flow Shop Scheduling Problem (PFSSP) is an interesting scheduling problem that has many real-world applications. It has been widely used as a benchmark to prove the efficiency of many discrete optimization algorithms. The DJaya algorithm is a discrete variation of the Jaya algorithm that has been recently proposed for solving discrete real-world problems. However, DJaya may get stuck in a local optima because of some limitations in its optimization operators. In this paper, we propose a new discrete optimization algorithm called Discrete Jaya with Refraction Learning and Three Mutation Methods (DJRL3M) for solving the PFSSP. DJRL3M incorporates five modifications into DJaya. First, it utilizes Refraction Learning (RL), which is a special type of opposition learning, to generate a diverse initial population of solutions. Second, it uses three mutation methods to explore the search space of a problem: DJaya mutation, highly disruptive polynomial mutation and Pitch Adjustment mutation. Third, it employs RL at each iteration to generate the opposite solutions of the best and worst solutions in an attempt to jump out local optima. Fourth, it uses the abandon method at the end of each iteration to discard a predefined percentage of the worst solutions and generate new random solutions. Finally, it uses the smallest position value to determine the correct values of the decision variables in a given candidate solution. The performance of DJRL3M was evaluated and compared with six well-recognized optimization algorithms [(New Cuckoo Search (NCS) (Wang et al. in SC 21:4297–4307, 2017), DJaya (Gao et al. in ITC 49:1944–1955, 2018), Hybrid Harmony Search (HHS) (Zhao et al. in EAAI 65:178-199, 2017), Modified Genetic algorithm (MGA) (Mumtaz et al. in: Advances in Manufacturing Technology XXXII: Proceedings of the 16th International Conference on Manufacturing Research, incorporating the 33rd National Conference on Manufacturing Research, 2018), Generalised Accelerations for Insertion-based Heuristics (GAIbH) (Fernandez-Viagas et al. in EJOR 282:858–872, 2020), Memetic algorithm with novel semi-constructive crossover and mutation operators (MASC) (Kurdi in ASC 94:106548, 2020)] using a set of Taillard’s benchmark instances. The experimental and statistical results show that DJRL3M obtains better performance than the performance of NCS, DJaya, HHS and MGA and exhibits competitive performance compared to the performance of MASC and GAIbH.
... Big Sur. The algorithms were all programmed in Java . 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 The size of population n and the fraction of abandon solutions p a in i CSPM and i CSPM2 were dynamically tuned over multiple simulations as recommended in [55]. The maximum number of iterations of each algorithm was limited to 10,000. ...
Preprint
Full-text available
The Island Cuckoo Search ( i CSPM) algorithm is a new variation of Cuckoo Search (CS) that uses the island model and the Highly Disruptive Polynomial (HDP) mutation for solving a broad range of optimization problems. This article introduces an improved i CSPM algorithm called i CSPM with elite opposition-based learning and multiple mutation methods ( i CSPM2). i CSPM2 has three main characteristics. Firstly, it separates candidate solutions into a number of islands (sub-populations) and then divides the islands equally among four improved versions of CS: CS via Le'vy fights (CS1) [1], CS with HDPM mutation (CS10) [2], CS with Jaya mutation (CSJ) and CS with pitch adjustment mutation (CS11) [2]. Secondly, it uses Elite Opposition-based Learning (EOBL) to improve its convergence rate and exploration ability. Finally, it uses the Smallest Position Value (SPV) with scheduling problems to convert continuous candidate solutions into discrete ones. A set of 15 popular benchmark functions was used to compare the performance of iCSPM2 to the performance of the original i CSPM algorithm based on different experimental scenarios. Results indicate that i CSPM2 exhibits improved performance over i CSPM. However, the sensitivity analysis of i CSPM and i CSPM2 to their parameters indicates that their convergence behavior is sensitive to the island model parameters. Further, the single-objective IEEE CEC 2014 functions were used to evaluate and compare the performance of iCSPM2 to four well-known swarm optimization algorithms: DGWO [3], L-SHADE [4], MHDA [5] and FWA-DM [6]. The overall experimental and statistical results suggest that i CSPM2 has better performance than the four well-known swarm optimization algorithms. i CSPM2's performance was also compared to two powerful discrete optimization algorithms (GAIbH [7] and MASC [8]) using a set of Taillard's benchmark instances for the permutation flow shop scheduling problem. The results indicate that i CSPM2 performs better than GAIbH and MASC. The source code of i CSPM2 is publicly available at https://github.com/bilalh2021/iCSPM2
... The relevant efforts can be clustered into two main streams, namely the problem-independent and problem-dependent improvements. The former one enhances the performance by developing novel search operators such as selection scheme [22], migration strategy [23], and self-adapting control mechanism [24], which expands the application scope of the heuristic algorithms. The problem-dependent improvement incorporates the problem knowledge to construct the systematic rules so that the invalid search is reduced. ...
Article
Reduction of carbon emissions is receiving growing attention in manufacturing industry. In the real-life flexible open-shop manufacturing environments, the workpieces generally need frequent transport between machines so that the limitation of the transport resource is non-negligible. Joint scheduling is a better way to achieve global optimum than sequential scheduling due to the interactive relationship between production and transport. However, owing to the extreme complexity, joint scheduling has received little attention over the past years. In this paper, the low-carbon joint scheduling in flexible open-shop environment with constrained automatic guided vehicle (LCJS-FOSCA) is investigated. A mixed-integer programming (MIP) model is formulated with the objectives to minimize the total carbon emission and makespan. Facing the extreme complexity of LCJS-FOSCA, an enhanced multi-objective particle swarm optimization (EMOPSO) is developed. In EMOPSO, several improvement strategies including the initialization method for high-quality solutions, chaotic position updating and mutation for global exploration, and problem-knowledge-based neighborhood search for local exploitation, are proposed. Finally, the following results are obtained through a comprehensive case study: (1) the proposed strategies make significant promotion on the convergence and comprehensive quality, and EMOPSO is verified to be effective for solving LCJS-FOSCA; (2) compared with the two sequential scheduling methods, the makespan obtained by joint scheduling is reduced by approximately 27.7% and 25.5%, and the total carbon emission is reduced by 3.3% and 4.2%; (3) the optimal numbers of AGV in various scales of transport time are provided based on the analysis of optimization results. This work presents a significance on promoting cleaner production of the modern manufacturing industry.