1. School of Information Science and Engineering, Wuhan University of Science and Technology, Wuhan 430081, China
2. School of Economics and Management, East China Normal University, Shanghai 200062, China
3. HUST Wuxi Research Institute, Wuxi 214174, China; Yuanshi Autonomy Co. Ltd, Nantong 226010, China
jding@fem.ecnu.edu.cn
Show less
History+
Received
Accepted
Published
2024-07-21
2024-10-11
Issue Date
Revised Date
2024-10-25
2024-09-29
PDF
(4740KB)
Abstract
Fiber allocation in optical cable production is critical for optimizing production efficiency, product quality, and inventory management. However, factors like fiber length and storage time complicate this process, making heuristic optimization algorithms inadequate. To tackle these challenges, this paper proposes a new framework: the dueling-double-deep Q-network with twin state-value and action-advantage functions (D3QNTF). First, dual action-advantage and state-value functions are used to prevent overestimation of action values. Second, a method for random initialization of feasible solutions improves sample quality early in the optimization. Finally, a strict penalty for errors is added to the reward mechanism, making the agent more sensitive to and better at avoiding illegal actions, which reduces decision errors. Experimental results show that the proposed method outperforms state-of-the-art algorithms, including greedy algorithms, genetic algorithms, deep Q-networks, double deep Q-networks, and standard dueling-double-deep Q-networks. The findings highlight the potential of the D3QNTF framework for fiber allocation in optical cable production.
Optical cables, as a fundamental component of communication networks, are essential for telephone communication, internet access, and data transmission across data centers (Liu et al., 2023). The allocation of optical fibers, a critical step in deploying these systems, faces challenges, including the labor-intensive and time-consuming nature of manual selection and suboptimal use of cable resources (Tan et al., 2024). Multi-objective optimization algorithms offer a key technological solution for efficient fiber allocation within optical cables.
Multi-objective optimization algorithms are broadly classified into three categories: exact algorithms, heuristic algorithms, and artificial intelligence algorithms. Common exact algorithms include branch and bound, decomposition methods, and Lagrangian relaxation algorithms (Silva et al., 2018). However, many multi-objective optimization problems (MOPs) exhibit nonlinearity, discontinuity, and non-differentiability, making exact algorithms difficult to apply effectively in practice (Zhong et al., 2024). As the number of optimization variables increases, solving these problems becomes more complex, time-consuming, and resource-intensive, limiting their use in large-scale, real-world applications.
To address MOPs, approximation methods that do not aim for exact solutions have become widely adopted. These include simulated annealing (Lee and Perkins, 2021), tabu search (Wang et al., 2021a), genetic algorithms (GA) (Xue et al., 2020), particle swarm optimization (Fang et al., 2024), and ant colony optimization (Martin et al., 2020). Significant advancements have been made to these intelligent optimization techniques. For example, Li et al. (2023) introduce a decomposition-based switching multi-objective whale optimizer, Ma et al. (2023) present a particle swarm optimization-assisted deep domain adaptation method, Li et al. (2024) propose a dynamic multi-objective optimization algorithm based on a hierarchical response system, and Wang et al. (2023b) develop an intelligent scheduling application integrated with MOPs to tackle truck scheduling problems. While these methods generally yield satisfactory solutions in reasonable time frames, they do not guarantee global optimality and their performance is highly sensitive to parameter settings, with different configurations potentially producing varied results.
Reinforcement learning, an artificial intelligence approach, adjusts strategies through continuous interaction with the environment to maximize long-term rewards. It is widely used in areas such as robotics control and optimization. Compared to heuristic algorithms, reinforcement learning offers several advantages: (1) it can learn and adapt strategies in real-time through environmental interactions, enabling effective responses to changes and uncertainties (Gui et al., 2023); (2) it is well-suited for high-dimensional and complex decision-making problems, effectively managing the increasing size of state and action spaces (Wang et al., 2021b), and (3) policy improvement is automated, reducing the need for manual parameter tuning (Zheng et al., 2023). The application of deep reinforcement learning (DRL) to MOPs has emerged as a promising approach (Mazyavkina et al., 2021). For example, an adaptive scheduling algorithm based on a deep Q network (DQN) has been developed to handle complex dynamic job-shop scheduling problems (Zhao et al., 2021). Although these methods have improved the efficiency of solving MOPs, challenges remain when applying them to optical fiber allocation. Specifically, DQN algorithms may encounter convergence issues and tend to overestimate action values, compromising strategy effectiveness.
The problem of optical fiber allocation is a typical MOP. An optimized fiber allocation plan not only minimizes fiber loss and breakage, but also improves production efficiency and product quality. In optical cable production, fiber allocation presents a complex MOP, with key factors including fiber color, length, storage time, and the number of segmented fibers. Excessive fiber segmentation can negatively affect production efficiency and inventory fiber quality. Additionally, factors such as fiber color, length, storage time, and selection sequence significantly influence inventory fiber quality. Thus, optical cable manufacturers must develop optimal fiber allocation plans to make the best use of limited inventory resources.
To address these challenges, this paper proposes an end-to-end optimization algorithm: a dueling double DQN framework with twin state-value and action-advantage functions (D3QNTF), designed to handle complex MOPs in fiber selection for optical cable production. The proposed approach automates the fiber allocation process, reducing reliance on manual operations while significantly improving production efficiency and lowering costs. The main contributions of this paper are summarized as follows:
(1) Dual action-advantage and state-value functions are used to address action value overestimation, improving the stability of the learning algorithm.
(2) A random initialization method for feasible solutions is introduced to enhance the network's learning ability, replacing the traditional approach of populating the experience pool through environmental exploration. This increases the proportion of successful samples in the early stages, significantly improving the model’s decision-making capabilities.
(3) Extreme penalties for illegal actions are incorporated into the reward mechanism to increase the agent’s sensitivity to and avoidance of such actions.
The structure of this paper is organized as follows. Section 2 introduces multi-objective optimization problems and reinforcement learning methods. Section 3 presents the proposed D3QNTF framework for resource allocation strategies. Section 4 provides simulation validation and a comparative discussion of the results. Finally, Section 5 offers conclusions and recommendations for future research.
2 Multi-objective optimization problems and reinforcement learning methods
2.1 Multi-objective optimization problems (MOPs)
MOPs represent a class of mathematical optimization tasks in which multiple objectives must be optimized simultaneously (Ming et al., 2023). These problems are generally formulated as follows:
where represents an -dimensional candidate solution, and denotes an -dimensional objective space, with and being an -dimensional bounded continuous decision space. The number of objective functions is denoted by , with . Since these objectives often conflict with one another, optimizing all objectives simultaneously is challenging. To address this, the concept of Pareto dominance is applied to identify a set of Pareto-optimal solutions rather than a single optimal solution. A solution is said to dominate a solution (denoted as ) if, for every , , and there exists at least one such that . A solution is considered Pareto optimal if no other solution dominates . The Pareto optimal set is defined as , and the Pareto optimal front is defined as . The primary goal of MOPs is to obtain a set of Pareto optimal solutions that closely approximate and are well-distributed along in the objective space.
Due to their inherent complexity, MOPs are regarded as particularly challenging in practical applications. The primary difficulty arises from the necessity to concurrently optimize multiple conflicting objectives, which complicates the effectiveness of any single optimization strategy in addressing all goals. Moreover, many real-world MOPs display characteristics such as nonlinearity, discontinuity, and non-differentiability, rendering traditional optimization methods unsuitable for direct application. Among the prevailing methodologies are both exact algorithms and heuristic algorithms, each possessing distinct limitations. While exact algorithms guarantee a global optimal solution through systematic search techniques, they tend to be computationally intensive, especially in large-scale, high-dimensional optimization scenarios. The substantial demand for computational resources and time makes these algorithms impractical in such contexts. Common exact algorithms, such as branch and bound, decomposition methods, and Lagrangian relaxation, perform adequately for small-scale problems, but encounter significant challenges as problem size escalates. Additionally, exact algorithms often necessitate problem-specific adaptations, resulting in limited generalizability and further constraining their application to complex MOPs.
In contrast, heuristic algorithms utilize flexible search strategies that enable them to identify near-optimal solutions within a restricted timeframe. Examples of this category include simulated annealing, tabu search, GAs, PSO, and ant colony optimization (Yao et al., 2023). Heuristic algorithms are particularly effective in tackling large-scale and complex problems, especially when the characteristics of the problem are unclear or when the solution space is extensive. However, they do not guarantee a global optimal solution, and their effectiveness is highly contingent upon parameter selection; different parameter combinations can result in considerable variations in results. Furthermore, heuristic algorithms often exhibit slower convergence and demonstrate less stable performance when applied to complex, constrained multi-objective problems due to their lack of robust theoretical foundations.
To address these limitations, artificial intelligence algorithms, particularly reinforcement learning, have attracted increasing attention in recent years for their application in MOPs. Reinforcement learning learns optimal strategies through interaction with the environment, demonstrating exceptional performance in dynamic and complex settings (Kiran et al., 2022). Compared to traditional exact and heuristic algorithms, reinforcement learning offers significant adaptability and is better suited for handling complex constraints. It does not depend on prior models and can progressively approach optimal solutions in scenarios characterized by high uncertainty and conflicting objectives. Furthermore, reinforcement learning continuously refines its strategy throughout the learning process based on received feedback, providing high flexibility and generalization capabilities. As a result, reinforcement learning effectively addresses the limitations of traditional methods, offering innovative solutions for complex real-world problems in MOP.
2.2 DQN and D3QN algorithm
Q-learning is a fundamental reinforcement learning algorithm, with its core principle centered on continuous interaction with the environment to learn the expected long-term rewards associated with various actions in different states (Moerland et al., 2023). Q-learning utilizes a Q-function to represent the value of state-action pairs and iteratively updates the Q-values using an update rule, enabling the agent to progressively learn the optimal policy. This update process is informed by the Bellman optimal equation, wherein the Q-table is updated using the current state, action, and the immediate reward given by the environment. The primary objective is to maximize cumulative rewards. Although Q-learning faces challenges in handling large-scale discrete state spaces, it lays the theoretical groundwork for DRL methods such as DQN.
DRL combines the perceptual capabilities of deep learning with the decision-making abilities of reinforcement learning (Wang et al., 2024). Its essence lies in the ability to learn multi-dimensional abstract features through multi-layer deep neural networks, facilitating the perception and understanding of complex situations in the environment while making decisions based on this understanding. Currently, some of the most prominent DRL methods include DQN, double DQN (Van Hasselt et al., 2016), deep deterministic policy gradient (DDPG) (Qiu et al., 2019), and deep recurrent Q-learning (Hausknecht and Stone, 2015).
The DQN architecture consists of a value network, a target value network, an error function, and a replay memory unit. It employs a deep neural network (DNN) to estimate the action-value function, while the experience replay mechanism and target value network are utilized to mitigate the instability and non-convergence issues that arise when approximating the action-value function using DNNs (Luo et al., 2021). DQN follows the same update formula as Q-learning but leverages a neural network to fit the Q-table, effectively transforming the elements of the state from discrete to continuous values. According to the Q-learning algorithm’s update formula, the loss function for the DQN algorithm is defined as the mean squared error between the current Q value and the target Q value, as illustrated in the following formula:
where and represent the weight parameters of the decision network and the target network, respectively. Once the loss function is obtained, the gradient descent method can be directly used to solve for the weight parameters of the neural network’s loss function.
The dueling DQN algorithm serves as an enhancement over the original DQN algorithm by introducing a dueling network architecture (Wang et al., 2016). This architecture bifurcates the network into two streams: one dedicated to representing the state value function and the other to representing the action advantage function. These two streams are then combined through a specialized aggregation layer to produce the estimated state-action value function Q. By integrating these two networks, the dueling DQN significantly improves the efficiency and accuracy of the algorithm, making it well-suited for managing large state and action spaces. The final value function can be expressed as follows:
where and are the specific parameters for the state value function and the action advantage function , respectively. denotes the size of the action space. The term ensures that the average value of the advantage function is zero, thereby stabilizing the computation of .
The dueling-double-deep Q-network (D3QN) algorithm is an advanced DRL technique that integrates the features of double DQN and dueling DQN (Gök, 2024). By addressing the issue of Q-value overestimation, it produces more reliable Q estimates, thus facilitating enhanced decision-making and overall performance. The learning process of the D3QN algorithm is depicted in Fig. 1. The agent inputs the state into the policy network, which calculates and outputs the Q values for each action. Then, using the -greedy strategy, the agent selects an action to execute and interacts with the environment, obtaining a reward and a new state (Tokic, 2010). The tuple {,,} is stored in the experience replay buffer for network training. The -greedy strategy means that, with a probability of , a random action is selected, and with a probability of , the action corresponding to the highest Q value calculated by the current policy network is chosen, as shown in the following equation:
where , is the set of possible actions. is a random number uniformly distributed between 0 and 1, and is the probability of selecting a random action.
The loss function of the D3QN algorithm is calculated as follows:
In the D3QN algorithm, the selection of actions for determining the target Q value relies on the parameters of the policy network. Specifically, the action that corresponds to the maximum Q value for the present state within the policy network is identified, and the Q value for this action is then computed within the target network. This methodology effectively reduces the likelihood of Q-value overestimation.
3 DRL-based resource allocation strategy for optical cable production
3.1 Problem formulation
The challenge of optical fiber allocation in optical cable production can be defined as follows: In the context of production scheduling tasks (Wang et al., 2023a), there are a defined set of bundle numbers, N, and a corresponding set of lengths, L, alongside the total quantity of fibers required for cable manufacturing. Each product is associated with a specific length that corresponds to a particular bundle number, establishing a one-to-one relationship between the bundle and length sets. To fulfill production scheduling requirements, manufacturers are tasked with devising an efficient fiber allocation strategy that considers several factors, including the quantity and length of available fibers, the total scheduled production length, the number of tubes, and the number of cores. The total number of fibers needed for cable production is calculated by multiplying the number of tubes by the number of cores.
To address the complexities of fiber allocation under various constraints, a mathematical model has been developed (Huang et al., 2023). This model ensures that multiple selection principles are integrated into the allocation process: the length of a single optical fiber must not exceed the maximum length available in inventory; the production length limit of the tube is determined by its major, minor, and outer diameters; and priority is given to fibers that are 8 km or shorter, colored fibers, return-to-inventory fibers, and those with extended storage durations. The model’s objective is to maximize the proportion of high-quality fibers in inventory while minimizing the number of segmented fibers and fiber allocation operations.
The definitions of the symbols used in the model are presented in Table 1. The mathematical model for fiber allocation constraints is outlined as follows:
Equation (6) ensures that the length of a single optical fiber does not surpass the maximum available fiber length in inventory; Eq. (7) defines the relationship between the production length limit of the tube and its major, minor, and outer diameters; Eqs. (8) – (11) prioritize the use of fibers based on specific characteristics: fibers shorter than 8 km (Eq. (8)), colored fibers (Eq. (9)), return-to-inventory fibers (Eq. (10)), and fibers with prolonged storage durations (Eq. (11)). Finally, Eq. (12) ensures that these four selection principles are considered concurrently during the fiber allocation process.
To enhance understanding of the number of tubes and fibers, an illustration is provided in the form of a cross-sectional diagram of a specific optical cable product model, as depicted in Fig. 2.
In Fig. 2, small colored rings represent tubes, with the number of rings corresponding to the tube count. Within each tube, different colored dots symbolize fibers, with the number of these dots representing the fiber count. Specifically, the optical cable illustrated consists of 8 tubes, each containing 6 fibers.
3.2 Network design and algorithm implementation
This section presents the D3QNTF framework, which enhances the traditional D3QN methodology. For the fiber resource allocation problem in optical cable production, this algorithm aims to efficiently allocate fiber resources while minimizing the number of segmented fibers and fiber allocation operations. The proposed D3QNTF framework is illustrated in Fig. 3. The subsequent sections will outline the basic components of D3QNTF: the state , the action , and the reward .
3.2.1 State vector
The state variable captures inventory information and the length of the optical cable to be produced. Thus, a state vector composed of the most relevant influencing factors in the environment is defined as follows:
where represents the number of fiber groups meeting the length constraints, and represents the remaining order demand.
3.2.2 Action vector
The goal of optical fiber allocation is to meet customer order demands while efficiently utilizing optical fibers of various lengths from inventory. Therefore, an action vector is defined in the algorithm framework.
where represents the length combinations that meet the length constraints.
3.2.3 Reward function
The reward function is pivotal in DRL as it steers the agent toward making appropriate decisions and avoiding detrimental actions. In this framework, the reward function aims to minimize both the number of segmented fibers and the number of fiber allocations, while simultaneously maximizing the overall inventory score. These metrics are essential for optical cable production. Consequently, the agent incurs a substantial penalty when fiber segmentation is excessive or allocation operations occur frequently. In contrast, when there is a significant improvement in the inventory score for the selected fiber lengths, the agent is rewarded.
The reward function is thus defined as the weighted sum of the number of segmented fibers, the number of fiber allocations, and the inventory score, which can be mathematically expressed as follows:
where represents the number of segmented fiber (i.e., the amount of fiber that needs to be cut during optical cable production), denotes the inventory score after fiber allocation, represents the inventory score before fiber allocation (i.e., a metric used to assess the overall quality of inventory fibers), and ν refers to the iteration number of fiber allocation (i.e., the total number of production runs based on orders). Additionally, , , and are the weight coefficients employed to compute the reward function score, while ℓ indicates the validity of the current action (1 for valid, 0 for invalid).
After defining all elements, both the evaluation network and the target network are constructed with identical architectures. Each network processes the input quadruplet {}, which comprises the current state, action, reward, and next state, and outputs the predicted action value and the target value , respectively. Both networks utilize fully connected layers. This framework draws inspiration from the dual-network concept inherent in the twin delayed deep deterministic policy gradient (Fujimoto et al., 2018), which generates two functions: action-advantage and state-value, thereby mitigating the overestimation error of the predicted action value . To enhance training efficiency, the common layers of the networks are merged, with the final layer employing a softmax function to convert predicted probabilities into the range [0, 1).
The detailed training process of the D3QNTF network is outlined in Algorithm 1. It is important to note that the testing process varies slightly from the training process. The specific steps of the testing process are detailed in steps 17 to 29 of Algorithm 1, with the remaining steps not being executed during the testing phase.
Algorithm 2 provides a comprehensive description of the initialization method for the replay buffer utilized by D3QNTF. Initially, several feasible solutions are generated through a random initialization method. These solutions subsequently interact with the environment, resulting in the generation of a Markov decision chain, which is stored in the replay buffer.
4 Experiments and discussion
4.1 Experimental design
4.1.1 Data description
This study utilizes real inventory data from an optical cable manufacturing company, which includes details such as fiber ID, fiber color, fiber length, a flag indicating return-to-inventory fibers, and the fiber coloring sequence. To accurately simulate real-world application scenarios and thoroughly evaluate the performance and stability of D3QNTF, the study establishes the number of tubes and fibers at 4 and 12, respectively. The outer diameter of each tube is set at 0.255 cm, with a width of 63 m, a major diameter of 100 m, a minor diameter of 50 m, and the maximum fiber length in inventory capped at 66 km.
The specific parameters are defined as follows: the fiber length selection interval is set to 1.02, meaning that fibers within the length range are standardized to the nearest minimum value within this range (i.e., the left boundary length). The redundancy length is set at 0.15, representing the additional length of the fiber relative to the tube length. The length factor is established at 1.025, utilized to estimate the relationship between the final product length and the original length. The experiment concentrates on the three most commonly used optical cable production configurations in the factory: 3*20, 3*20 + 5*20, and 3*20 + 5*20 + 7*20, where 3, 5, and 7 denote the order lengths, and 20 represents the number of bundles.
4.1.2 Evaluation criteria
In line with the specific requirements of the optical cable manufacturing plant, we have established four evaluation metrics: inventory score, number of fiber allocations, number of segmented fibers, and an equally weighted sum of these three metrics. Among these, the number of fiber allocations and the number of segmented fibers serve as critical indicators in optical cable production. A reduction in these values can significantly enhance production efficiency while improving the quality of inventory fibers. Conversely, elevated values in these metrics may lead to decreased production efficiency and a decline in the quality of inventory fibers.
4.1.3 Benchmark models
To comprehensively assess the learning performance of D3QNTF in discrete action spaces, we compared it against four categories of algorithms: greedy algorithms, heuristic algorithms (including GA and PSO), multi-objective optimization algorithms (such as NSGA-II and MOPSO), and fundamental reinforcement learning algorithms (including DQN, DDQN, and D3QN). These algorithms are widely utilized across various problem-solving contexts for specific reasons: greedy algorithms offer quick local optima; heuristic algorithms are proficient in resolving complex problems by identifying near-global optima; multi-objective optimization algorithms adeptly balance multiple conflicting objectives; and foundational reinforcement learning algorithms (like DQN, DDQN, and D3QN) serve as direct performance benchmarks for D3QNTF. This comparative methodology allows for a more comprehensive evaluation of D3QNTF's decision-making performance and its applicability.
The hyperparameters utilized in the DRL experiments are outlined in Table 2. These parameters were chosen based on preliminary experimental findings and best practices from relevant literature, aiming to optimize model performance and ensure the reliability of the experimental results. Key parameters include the learning rate, batch size, and the number of iterations, all of which significantly influence the training effectiveness of the model. For the length combination case of 3*20 + 5*20, the network architecture of D3QNTF is detailed in Table 3. In comparison, D3QN omits the and layers, while DQN and DDQN do not include the and layers. Although DQN and DDQN share the same network architecture, they differ in their strategies for target value estimation.
For different order combinations 3*20, 3*20 + 5*20, and 3*20 + 5*20 + 7*20, the dimensions of the state vector for DQN, DDQN, D3QN, and D3QNTF are set to 14, 64, and 165, respectively, while the dimensions of the action vector are set to 13, 62, and 162, respectively.
4.2 Experimental results
D3QNTF is evaluated using the common length combination case of 3*20 + 5*20. Fig. 4 illustrates the training process of the D3QNTF algorithm for this scenario. In Fig. 4(a), it is evident that during the initial stage of training, the returns and rewards obtained by the agent from interactions with the environment exhibit considerable fluctuations, primarily due to the random initialization of the network parameters. As training progresses, these fluctuations gradually diminish and ultimately converge. In Fig. 4(b), as the number of training steps increases, the estimated Q value stabilizes, and the loss curve becomes steady, converging toward zero. This trend indicates that the model's prediction accuracy and stability improve as training advances. The minor oscillations observed during convergence can primarily be attributed to the low probability of random action selection.
To provide a clearer comparison of fiber allocation effectiveness across various algorithms, we smoothed the learning curves for each algorithm, as illustrated in Fig. 5. The results reveal that DQN exhibits the lowest return and average reward values. Although its return curve shows a tendency to stabilize, it continues to experience oscillations and even declines, primarily due to DQN’s failure to achieve full convergence within a finite number of steps and its substantial Q-value estimation error. In contrast, D3QN and DDQN demonstrate notable improvements, with their loss functions converging to near zero, reduced Q-value estimation errors, and higher rewards and returns. Notably, D3QNTF outperforms both D3QN and DDQN, benefitting from its sample initialization operation, which enhances the model’s decision-making capabilities. As shown in Fig. 5(a), 5(b), and 5(d), although D3QNTF converges more slowly due to the necessity of fitting a larger number of network parameters, its return and average reward curves remain more stable throughout the learning process. This stability can be attributed to D3QNTF's implementation of dual action-advantage and state-value functions, which effectively mitigate the issue of action-value overestimation and enhance the algorithm's overall learning stability. In comparison, D3QN displays significant oscillations in its curve between 110,000 and 200,000 steps. The Q-value estimation curves for the various algorithms are presented in Fig. 5(c).
4.3 Comparison results
The experimental findings indicate that D3QNTF consistently outperforms other algorithms across different test sets. As shown in Table 4, for the 3*20 case, all algorithms successfully identify the optimal fiber allocation combination, increasing the inventory score to 0.6207 while reducing the number of segmented fibers to 0 and fiber allocations to 4. The final return of −9.9958 suggests that the greedy algorithm's search rules, alongside the heuristic and multi-objective optimization algorithms' mathematical models, are well-aligned to this scenario (Zhang et al., 2021). However, as the case expands to 3*20 + 5*20, the performance of the greedy algorithm substantially lags behind that of the other algorithms, highlighting its inefficiency in addressing large-scale solution space problems. In contrast, the GA algorithm identifies high-quality feasible solutions within the integer programming model of fiber allocation, achieving a 100% reduction in segmented fibers and a 16.67% reduction in allocations. Due to the requirement to balance multiple metrics and GA's superior performance in one metric, the multi-objective optimization algorithms NSGA-II and MOPSO achieve slightly lower scores than GA. Nevertheless, in comparison to D3QN and D3QNTF, the solutions provided by GA remain uncompetitive, further underscoring the overall superiority of reinforcement learning algorithms for this particular problem.
In the case of 3*20 + 5*20 + 7*20, all algorithms experience a significant increase in the number of segmented fibers and allocations. The greedy algorithm, while attempting to minimize fiber segments, results in a drastic increase in the number of fiber allocations, leading to its overall poorest performance. As illustrated in Fig. 6 and Fig. 7, a visual comparison of the metrics for the 3*20 + 5*20 and 3*20 + 5*20 + 7*20 cases demonstrates that D3QNTF effectively balances the number of segmented fibers and allocations, maximizes the inventory score, and ultimately achieves the highest return. For instance, D3QNTF increases the number of fiber allocations to reduce fiber segmentation, thereby improving the final score. Furthermore, D3QNTF’s architecture, which incorporates dual action-advantage and state-value functions and populates the replay buffer with randomly initialized feasible solutions, enhances its learning capabilities compared to traditional D3QN. This architecture not only improves the stability of the algorithm but also significantly enhances its optimization performance in complex environments. Overall, D3QNTF demonstrates exceptional capability in addressing large-scale and highly complex fiber allocation problems.
5 Conclusions
In optical cable production, fiber allocation faces challenges like optimizing production efficiency, improving quality, and lowering inventory costs. Traditional heuristic algorithms often struggle, especially in complex and dynamic settings. To address these issues, this paper presents an advanced D3QNTF model, which extends the DRL framework for a more adaptive solution. The D3QNTF model incorporates dual action-advantage and state-value functions and introduces a novel random initialization method. This replaces the traditional method of populating the experience pool through environmental exploration. The approach helps reduce action-value overestimation, improves network learning stability, and enhances decision-making capabilities.
The agent interacts with the environment through continuous decisions, feedback, and adjustments to obtain the optimal strategy. Simulations confirm the model's effectiveness. Compared to traditional reinforcement learning, the D3QNTF model significantly improves stability and learning capacity. It handles high-dimensional challenges more efficiently and uses an innovative initialization strategy to enhance learning. As a result, decision-making quality is greatly improved, boosting performance in fiber allocation tasks. The D3QNTF model achieves better inventory scores, reduces segmented fibers, and optimizes fiber allocations, yielding the highest return values in all tests. These results demonstrate strong decision-making and robustness. However, the challenge of managing diverse customer orders in production scheduling remains unresolved. Future research should explore managing multiple orders within a single model.
Fang J,Wang Z,Liu W,Chen L,Liu X, (2024). A new particle-swarm-optimization- assisted deep transfer learning framework with applications to outlier detection in additive manufacturing. Engineering Applications of Artificial Intelligence, 131: 107700
[2]
FujimotoSHoofHMegerD (2018). Addressing function approximation error in actor-critic methods. In: International Conference on Machine Learning: 1587–1596
[3]
Gök M, (2024). Dynamic path planning via Dueling Double Deep Q-network (D3QN) with prioritized experience replay. Applied Soft Computing, 158: 111503
[4]
Gui Y,Tang D,Zhu H,Zhang Y,Zhang Z, (2023). Dynamic scheduling for flexible job shop using a deep reinforcement learning approach. Computers & Industrial Engineering, 180: 109255
[5]
HausknechtMStoneP (2015). Deep recurrent Q-learning for partially observable MDPs. In: 2015 AAAI Fall Symposium Series
[6]
Huang M,Hao Y,Wang Y,Hu X,Li L, (2023). Split-order consolidation optimization for online supermarkets: Process analysis and optimization models. Frontiers of Engineering Management, 10( 3): 499–516
[7]
Kiran B R,Sobh I,Talpaert V,Mannion P,Sallab A A A,Yogamani S,Perez P, (2022). Deep reinforcement learning for autonomous driving: A survey. IEEE Transactions on Intelligent Transportation Systems, 23( 6): 4909–4926
[8]
Lee J,Perkins D, (2021). A simulated annealing algorithm with a dual perturbation method for clustering. Pattern Recognition, 112: 107713
[9]
Li H,Liu H,Lan C,Yin Y,Wu P,Yan C,Zeng N, (2023). SMWO/D: A decomposition-based switching multi-objective whale optimiser for structural optimisation of turbine disk in aero-engines. International Journal of Systems Science, 54( 8): 1713–1728
[10]
Li H,Wang Z,Lan C,Wu P,Zeng N, (2024). A novel dynamic multiobjective optimization algorithm with hierarchical response system. IEEE Transactions on Computational Social Systems, 11( 2): 2494–2512
[11]
Liu J,Yuan S,Luo B,Biondi B,Noh H Y, (2023). Turning telecommunication fiber-optic cables into distributed acoustic sensors for vibration-based bridge health monitoring. Structural Control and Health Monitoring, 2023: 1–14
[12]
Luo S,Zhang L,Fan Y, (2021). Dynamic multi-objective scheduling for flexible job shop by deep reinforcement learning. Computers & Industrial Engineering, 159: 107489
[13]
Ma G,Wang Z,Liu W,Fang J,Zhang Y,Ding H,Yuan Y, (2023). Estimating the state of health for lithium-ion batteries: A particle swarm optimization-assisted deep domain adaptation approach. IEEE/CAA Journal of Automatica Sinica, 10( 7): 1530–1543
[14]
Martin E,Cervantes A,Saez Y,Isasi P, (2020). IACS-HCSP: Improved ant colony optimization for large-scale home care scheduling problems. Expert Systems with Applications, 142: 112994
[15]
Mazyavkina N,Sviridov S,Ivanov S,Burnaev E, (2021). Reinforcement learning for combinatorial optimization: A survey. Computers & Operations Research, 134: 105400
[16]
Ming F,Gong W,Li D,Wang L,Gao L, (2023). A competitive and cooperative swarm optimizer for constrained multiobjective optimization problems. IEEE Transactions on Evolutionary Computation, 27( 5): 1313–1326
[17]
Moerland T M,Broekens J,Plaat A,Jonker C M, (2023). Model-based reinforcement learning: A survey. Foundations and Trends® in Machine Learning, 16( 1): 1–118
[18]
Qiu C,Hu Y,Chen Y,Zeng B, (2019). Deep deterministic policy gradient (DDPG)-based energy harvesting wireless communications. IEEE Internet of Things Journal, 6( 5): 8577–8588
[19]
Silva Y L T,Subramanian A,Pessoa A A, (2018). Exact and heuristic algorithms for order acceptance and scheduling with sequence-dependent setup times. Computers & Operations Research, 90: 142–160
[20]
Tan F,Yuan Z,Zhang Y,Tang S,Guo F,Zhang S, (2024). Improved genetic algorithm based on rule optimization strategy for fibre allocation. Systems Science & Control Engineering, 12( 1): 2347887
[21]
TokicM (2010). Adaptive ε-greedy exploration in reinforcement learning based on value differences. In: Annual Conference on Artificial Intelligence. Berlin, Heidelberg: Springer, 203–210
[22]
Van Hasselt H,Guez A,Silver D, (2016). Deep reinforcement learning with double Q-learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, 30( 1): 2094–2100
[23]
Wang L,Hu X,Wang Y,Xu S,Ma S,Yang K,Liu Z,Wang W, (2021b). Dynamic job-shop scheduling in smart manufacturing using deep reinforcement learning. Computer Networks, 190: 107969
[24]
Wang X,Wang S,Liang X,Zhao D,Huang J,Xu X,Dai B,Miao Q, (2024). Deep reinforcement learning: A survey. IEEE Transactions on Neural Networks and Learning Systems, 35( 4): 5064–5078
[25]
Wang Y,Han Y,Gong D,Li H, (2023a). A review of intelligent optimization for group scheduling problems in cellular manufacturing. Frontiers of Engineering Management, 10( 3): 406–426
[26]
WangYLiuWWangCFadzilFLauriaSLiuX (2023b). A novel multi-objective optimization approach with flexible operation planning strategy for truck scheduling. International Journal of Network Dynamics and Intelligence, 100002
[27]
Wang Y,Wu Z,Guan G,Li K,Chai S, (2021a). Research on intelligent design method of ship multi-deck compartment layout based on improved taboo search genetic algorithm. Ocean Engineering, 225: 108823
[28]
WangZSchaulTHesselMHasseltHLanctotMFreitasN (2016). Dueling network architectures for deep reinforcement learning. In: International Conference on Machine Learning, 1995–2003
[29]
Xue Z,Zhang Y,Cheng C,Ma G, (2020). Remaining useful life prediction of lithium-ion batteries with adaptive unscented kalman filter and optimized support vector regression. Neurocomputing, 376: 95–102
[30]
Yao F,Du Y,Li L,Xing L,Chen Y, (2023). General modeling and optimization technique for real-world earth observation satellite scheduling. Frontiers of Engineering Management, 10( 4): 695–709
[31]
Zhang Y,Chen L,Li Y,Zheng X,Chen J,Jin J, (2021). A hybrid approach for remaining useful life prediction of lithium-ion battery with adaptive levy flight optimized particle filter and long short-term memory network. Journal of Energy Storage, 44: 103245
[32]
Zhao Y,Wang Y,Tan Y,Zhang J,Yu H, (2021). Dynamic jobshop scheduling algorithm based on deep Q network. IEEE Access: Practical Innovations, Open Solutions, 9: 122995–123011
[33]
ZhengTZhouYHuMZhangJ (2023). Dynamic scheduling for large-scale flexible job shop based on noisy DDQN. International Journal of Network Dynamics and Intelligence, 100015
[34]
ZhongLZengZHuangZShiXBieY (2024). Joint optimization of electric bus charging and energy storage system scheduling. Frontiers of Engineering Management, 11(4): 676–696
RIGHTS & PERMISSIONS
Higher Education Press
AI Summary 中Eng×
Note: Please be aware that the following content is generated by artificial intelligence. This website is not responsible for any consequences arising from the use of this content.