Due to the NP-hardness of the two-sided assembly line balancing (TALB) problem, multiple constraints existing in real applications are less studied, especially when one task is involved with several constraints. In ...Due to the NP-hardness of the two-sided assembly line balancing (TALB) problem, multiple constraints existing in real applications are less studied, especially when one task is involved with several constraints. In this paper, an effective hybrid algorithm is proposed to address the TALB problem with multiple constraints (TALB-MC). Considering the discrete attribute of TALB-MC and the continuous attribute of the standard teaching-learning-based optimization (TLBO) algorithm, the random-keys method is hired in task permutation representation, for the purpose of bridging the gap between them. Subsequently, a special mechanism for handling multiple constraints is developed. In the mechanism, the directions constraint of each task is ensured by the direction check and adjustment. The zoning constraints and the synchronism constraints are satisfied by teasing out the hidden correlations among constraints. The positional constraint is allowed to be violated to some extent in decoding and punished in cost fimction. Finally, with the TLBO seeking for the global optimum, the variable neighborhood search (VNS) is further hybridized to extend the local search space. The experimental results show that the proposed hybrid algorithm outperforms the late acceptance hill-climbing algorithm (LAHC) for TALB-MC in most cases, especially for large-size problems with multiple constraints, and demonstrates well balance between the exploration and the exploitation. This research proposes an effective and efficient algorithm for solving TALB-MC problem by hybridizing the TLBO and VNS.展开更多
Nowadays,succeeding safe communication and protection-sensitive data from unauthorized access above public networks are the main worries in cloud servers.Hence,to secure both data and keys ensuring secured data storag...Nowadays,succeeding safe communication and protection-sensitive data from unauthorized access above public networks are the main worries in cloud servers.Hence,to secure both data and keys ensuring secured data storage and access,our proposed work designs a Novel Quantum Key Distribution(QKD)relying upon a non-commutative encryption framework.It makes use of a Novel Quantum Key Distribution approach,which guarantees high level secured data transmission.Along with this,a shared secret is generated using Diffie Hellman(DH)to certify secured key generation at reduced time complexity.Moreover,a non-commutative approach is used,which effectively allows the users to store and access the encrypted data into the cloud server.Also,to prevent data loss or corruption caused by the insiders in the cloud,Optimized Genetic Algorithm(OGA)is utilized,which effectively recovers the data and retrieve it if the missed data without loss.It is then followed with the decryption process as if requested by the user.Thus our proposed framework ensures authentication and paves way for secure data access,with enhanced performance and reduced complexities experienced with the prior works.展开更多
Background: Popliteal cysts are common and present as asymptomatic lumps in the medial popliteal fossa. Some have complex internal characteristics such as septa and loose-bodies. However, not all are popliteal cysts a...Background: Popliteal cysts are common and present as asymptomatic lumps in the medial popliteal fossa. Some have complex internal characteristics such as septa and loose-bodies. However, not all are popliteal cysts and can be aggressive. These lesions need to be differentiated by the absence of the communicating neck with the joint on ultrasound. Presence of Doppler flow of non-communicating cysts requires further evaluation on MRI, prior to performing a biopsy. Using a case series, we propose an algorithmic approach that is simple and will help identify the malignant lesions and institute appropriate management. Case-Presentation: Popliteal Cyst: On ultrasound: characteristic neck communicating with knee joint. Synovial Sarcoma: Gadolinium enhancement, with areas of low-, iso- and hyper-intense signal to fat on T2. Synovial-Osteochondromatosis: Non-mineralized: T1-low/intermediate intensity;T2-high intensity. Mineralized type: low intensity on T1 & T2. Thrombosed Popliteal Aneurysm: Lamellated appearance-high/low signal intensity on T2. Myxoid-Liposarcomas: Inhomogeneous appearance;homogenous with gadolinium. Usually require a biopsy for diagnosis. Conclusion: The cystic lesions in the medial aspect of the popliteal fossa can be misdiagnosed. Our article reiterates the importance of the communicating neck that separates popliteal cysts from other mimics. We have proposed an algorithm to identify these mimics.展开更多
Steganography is a technique for hiding secret messages while sending and receiving communications through a cover item.From ancient times to the present,the security of secret or vital information has always been a s...Steganography is a technique for hiding secret messages while sending and receiving communications through a cover item.From ancient times to the present,the security of secret or vital information has always been a significant problem.The development of secure communication methods that keep recipient-only data transmissions secret has always been an area of interest.Therefore,several approaches,including steganography,have been developed by researchers over time to enable safe data transit.In this review,we have discussed image steganography based on Discrete Cosine Transform(DCT)algorithm,etc.We have also discussed image steganography based on multiple hashing algorithms like the Rivest–Shamir–Adleman(RSA)method,the Blowfish technique,and the hash-least significant bit(LSB)approach.In this review,a novel method of hiding information in images has been developed with minimal variance in image bits,making our method secure and effective.A cryptography mechanism was also used in this strategy.Before encoding the data and embedding it into a carry image,this review verifies that it has been encrypted.Usually,embedded text in photos conveys crucial signals about the content.This review employs hash table encryption on the message before hiding it within the picture to provide a more secure method of data transport.If the message is ever intercepted by a third party,there are several ways to stop this operation.A second level of security process implementation involves encrypting and decrypting steganography images using different hashing algorithms.展开更多
In recent decades,fog computing has played a vital role in executing parallel computational tasks,specifically,scientific workflow tasks.In cloud data centers,fog computing takes more time to run workflow applications...In recent decades,fog computing has played a vital role in executing parallel computational tasks,specifically,scientific workflow tasks.In cloud data centers,fog computing takes more time to run workflow applications.Therefore,it is essential to develop effective models for Virtual Machine(VM)allocation and task scheduling in fog computing environments.Effective task scheduling,VM migration,and allocation,altogether optimize the use of computational resources across different fog nodes.This process ensures that the tasks are executed with minimal energy consumption,which reduces the chances of resource bottlenecks.In this manuscript,the proposed framework comprises two phases:(i)effective task scheduling using a fractional selectivity approach and(ii)VM allocation by proposing an algorithm by the name of Fitness Sharing Chaotic Particle Swarm Optimization(FSCPSO).The proposed FSCPSO algorithm integrates the concepts of chaos theory and fitness sharing that effectively balance both global exploration and local exploitation.This balance enables the use of a wide range of solutions that leads to minimal total cost and makespan,in comparison to other traditional optimization algorithms.The FSCPSO algorithm’s performance is analyzed using six evaluation measures namely,Load Balancing Level(LBL),Average Resource Utilization(ARU),total cost,makespan,energy consumption,and response time.In relation to the conventional optimization algorithms,the FSCPSO algorithm achieves a higher LBL of 39.12%,ARU of 58.15%,a minimal total cost of 1175,and a makespan of 85.87 ms,particularly when evaluated for 50 tasks.展开更多
The large-scale optimization problem requires some optimization techniques, and the Metaheuristics approach is highly useful for solving difficult optimization problems in practice. The purpose of the research is to o...The large-scale optimization problem requires some optimization techniques, and the Metaheuristics approach is highly useful for solving difficult optimization problems in practice. The purpose of the research is to optimize the transportation system with the help of this approach. We selected forest vehicle routing data as the case study to minimize the total cost and the distance of the forest transportation system. Matlab software helps us find the best solution for this case by applying three algorithms of Metaheuristics: Genetic Algorithm (GA), Ant Colony Optimization (ACO), and Extended Great Deluge (EGD). The results show that GA, compared to ACO and EGD, provides the best solution for the cost and the length of our case study. EGD is the second preferred approach, and ACO offers the last solution.展开更多
The continuously updated database of failures and censored data of numerous products has become large, and on some covariates, information regarding the failure times is missing in the database. As the dataset is larg...The continuously updated database of failures and censored data of numerous products has become large, and on some covariates, information regarding the failure times is missing in the database. As the dataset is large and has missing information, the analysis tasks become complicated and a long time is required to execute the programming codes. In such situations, the divide and recombine (D&R) approach, which has a practical computational performance for big data analysis, can be applied. In this study, the D&R approach was applied to analyze the real field data of an automobile component with incomplete information on covariates using the Weibull regression model. Model parameters were estimated using the expectation maximization algorithm. The results of the data analysis and simulation demonstrated that the D&R approach is applicable for analyzing such datasets. Further, the percentiles and reliability functions of the distribution under different covariate conditions were estimated to evaluate the component performance of these covariates. The findings of this study have managerial implications regarding design decisions, safety, and reliability of automobile components.展开更多
In time series modeling, the residuals are often checked for white noise and normality. In practice, the useful tests are Ljung Box test. Mcleod Li test and Lin Mudholkar test. In this paper, we present a nonparame...In time series modeling, the residuals are often checked for white noise and normality. In practice, the useful tests are Ljung Box test. Mcleod Li test and Lin Mudholkar test. In this paper, we present a nonparametric approach for checking the residuals of time series models. This approach is based on the maximal correlation coefficient ρ 2 * between the residuals and time t . The basic idea is to use the bootstrap to form the null distribution of the statistic ρ 2 * under the null hypothesis H 0:ρ 2 * =0. For calculating ρ 2 * , we proposes a ρ algorithm, analogous to ACE procedure. Power study shows this approach is more powerful than Ljung Box test. Meanwhile, some numerical results and two examples are reported in this paper.展开更多
基金Supported by National Natural Science Foundation of China(Grant Nos.51275366,50875190,51305311)Specialized Research Fund for the Doctoral Program of Higher Education of China(Grant No.20134219110002)
文摘Due to the NP-hardness of the two-sided assembly line balancing (TALB) problem, multiple constraints existing in real applications are less studied, especially when one task is involved with several constraints. In this paper, an effective hybrid algorithm is proposed to address the TALB problem with multiple constraints (TALB-MC). Considering the discrete attribute of TALB-MC and the continuous attribute of the standard teaching-learning-based optimization (TLBO) algorithm, the random-keys method is hired in task permutation representation, for the purpose of bridging the gap between them. Subsequently, a special mechanism for handling multiple constraints is developed. In the mechanism, the directions constraint of each task is ensured by the direction check and adjustment. The zoning constraints and the synchronism constraints are satisfied by teasing out the hidden correlations among constraints. The positional constraint is allowed to be violated to some extent in decoding and punished in cost fimction. Finally, with the TLBO seeking for the global optimum, the variable neighborhood search (VNS) is further hybridized to extend the local search space. The experimental results show that the proposed hybrid algorithm outperforms the late acceptance hill-climbing algorithm (LAHC) for TALB-MC in most cases, especially for large-size problems with multiple constraints, and demonstrates well balance between the exploration and the exploitation. This research proposes an effective and efficient algorithm for solving TALB-MC problem by hybridizing the TLBO and VNS.
文摘Nowadays,succeeding safe communication and protection-sensitive data from unauthorized access above public networks are the main worries in cloud servers.Hence,to secure both data and keys ensuring secured data storage and access,our proposed work designs a Novel Quantum Key Distribution(QKD)relying upon a non-commutative encryption framework.It makes use of a Novel Quantum Key Distribution approach,which guarantees high level secured data transmission.Along with this,a shared secret is generated using Diffie Hellman(DH)to certify secured key generation at reduced time complexity.Moreover,a non-commutative approach is used,which effectively allows the users to store and access the encrypted data into the cloud server.Also,to prevent data loss or corruption caused by the insiders in the cloud,Optimized Genetic Algorithm(OGA)is utilized,which effectively recovers the data and retrieve it if the missed data without loss.It is then followed with the decryption process as if requested by the user.Thus our proposed framework ensures authentication and paves way for secure data access,with enhanced performance and reduced complexities experienced with the prior works.
文摘Background: Popliteal cysts are common and present as asymptomatic lumps in the medial popliteal fossa. Some have complex internal characteristics such as septa and loose-bodies. However, not all are popliteal cysts and can be aggressive. These lesions need to be differentiated by the absence of the communicating neck with the joint on ultrasound. Presence of Doppler flow of non-communicating cysts requires further evaluation on MRI, prior to performing a biopsy. Using a case series, we propose an algorithmic approach that is simple and will help identify the malignant lesions and institute appropriate management. Case-Presentation: Popliteal Cyst: On ultrasound: characteristic neck communicating with knee joint. Synovial Sarcoma: Gadolinium enhancement, with areas of low-, iso- and hyper-intense signal to fat on T2. Synovial-Osteochondromatosis: Non-mineralized: T1-low/intermediate intensity;T2-high intensity. Mineralized type: low intensity on T1 & T2. Thrombosed Popliteal Aneurysm: Lamellated appearance-high/low signal intensity on T2. Myxoid-Liposarcomas: Inhomogeneous appearance;homogenous with gadolinium. Usually require a biopsy for diagnosis. Conclusion: The cystic lesions in the medial aspect of the popliteal fossa can be misdiagnosed. Our article reiterates the importance of the communicating neck that separates popliteal cysts from other mimics. We have proposed an algorithm to identify these mimics.
文摘Steganography is a technique for hiding secret messages while sending and receiving communications through a cover item.From ancient times to the present,the security of secret or vital information has always been a significant problem.The development of secure communication methods that keep recipient-only data transmissions secret has always been an area of interest.Therefore,several approaches,including steganography,have been developed by researchers over time to enable safe data transit.In this review,we have discussed image steganography based on Discrete Cosine Transform(DCT)algorithm,etc.We have also discussed image steganography based on multiple hashing algorithms like the Rivest–Shamir–Adleman(RSA)method,the Blowfish technique,and the hash-least significant bit(LSB)approach.In this review,a novel method of hiding information in images has been developed with minimal variance in image bits,making our method secure and effective.A cryptography mechanism was also used in this strategy.Before encoding the data and embedding it into a carry image,this review verifies that it has been encrypted.Usually,embedded text in photos conveys crucial signals about the content.This review employs hash table encryption on the message before hiding it within the picture to provide a more secure method of data transport.If the message is ever intercepted by a third party,there are several ways to stop this operation.A second level of security process implementation involves encrypting and decrypting steganography images using different hashing algorithms.
基金This work was supported in part by the National Science and Technology Council of Taiwan,under Contract NSTC 112-2410-H-324-001-MY2.
文摘In recent decades,fog computing has played a vital role in executing parallel computational tasks,specifically,scientific workflow tasks.In cloud data centers,fog computing takes more time to run workflow applications.Therefore,it is essential to develop effective models for Virtual Machine(VM)allocation and task scheduling in fog computing environments.Effective task scheduling,VM migration,and allocation,altogether optimize the use of computational resources across different fog nodes.This process ensures that the tasks are executed with minimal energy consumption,which reduces the chances of resource bottlenecks.In this manuscript,the proposed framework comprises two phases:(i)effective task scheduling using a fractional selectivity approach and(ii)VM allocation by proposing an algorithm by the name of Fitness Sharing Chaotic Particle Swarm Optimization(FSCPSO).The proposed FSCPSO algorithm integrates the concepts of chaos theory and fitness sharing that effectively balance both global exploration and local exploitation.This balance enables the use of a wide range of solutions that leads to minimal total cost and makespan,in comparison to other traditional optimization algorithms.The FSCPSO algorithm’s performance is analyzed using six evaluation measures namely,Load Balancing Level(LBL),Average Resource Utilization(ARU),total cost,makespan,energy consumption,and response time.In relation to the conventional optimization algorithms,the FSCPSO algorithm achieves a higher LBL of 39.12%,ARU of 58.15%,a minimal total cost of 1175,and a makespan of 85.87 ms,particularly when evaluated for 50 tasks.
文摘The large-scale optimization problem requires some optimization techniques, and the Metaheuristics approach is highly useful for solving difficult optimization problems in practice. The purpose of the research is to optimize the transportation system with the help of this approach. We selected forest vehicle routing data as the case study to minimize the total cost and the distance of the forest transportation system. Matlab software helps us find the best solution for this case by applying three algorithms of Metaheuristics: Genetic Algorithm (GA), Ant Colony Optimization (ACO), and Extended Great Deluge (EGD). The results show that GA, compared to ACO and EGD, provides the best solution for the cost and the length of our case study. EGD is the second preferred approach, and ACO offers the last solution.
文摘The continuously updated database of failures and censored data of numerous products has become large, and on some covariates, information regarding the failure times is missing in the database. As the dataset is large and has missing information, the analysis tasks become complicated and a long time is required to execute the programming codes. In such situations, the divide and recombine (D&R) approach, which has a practical computational performance for big data analysis, can be applied. In this study, the D&R approach was applied to analyze the real field data of an automobile component with incomplete information on covariates using the Weibull regression model. Model parameters were estimated using the expectation maximization algorithm. The results of the data analysis and simulation demonstrated that the D&R approach is applicable for analyzing such datasets. Further, the percentiles and reliability functions of the distribution under different covariate conditions were estimated to evaluate the component performance of these covariates. The findings of this study have managerial implications regarding design decisions, safety, and reliability of automobile components.
文摘In time series modeling, the residuals are often checked for white noise and normality. In practice, the useful tests are Ljung Box test. Mcleod Li test and Lin Mudholkar test. In this paper, we present a nonparametric approach for checking the residuals of time series models. This approach is based on the maximal correlation coefficient ρ 2 * between the residuals and time t . The basic idea is to use the bootstrap to form the null distribution of the statistic ρ 2 * under the null hypothesis H 0:ρ 2 * =0. For calculating ρ 2 * , we proposes a ρ algorithm, analogous to ACE procedure. Power study shows this approach is more powerful than Ljung Box test. Meanwhile, some numerical results and two examples are reported in this paper.