国内也有不错的开源项目,只是感觉后劲不足,亟须自然科学基金赞助

http://cs.cug.edu.cn/teacherweb/lichanghe/pages/EAlib.html大笑


如果上面的服务器下载EAlib源代码有问题,可以参考这个网盘的

Welcome to Changhe Li's EAlib Page


About Me News Research Teaching Publications EAlib Downloads Funding Links

LIBRARY OF EVOLUTIONARY ALGORITHMS

This is an open sources project with the aim of providing the source codes of state-of-the-art algorithms, problems, and algorithms' performance metrics for researchers in the area of evolutionary computation. Based on this framework, you can easily implement your ideas and add the codes into this library. More important, you will be also able to compare with other state-of-the-art algorithms without implementing them under a fair environment. In addition, this library will be updated as time goes on when new algorithms and problems come to this research area. Anyone, who want to contribute to this project, is welcome, and please contact me without hesitation.

This project is totally for research purpose. As a researcher in the area of evolutionary computation, I know how bad it is if there is no source codes available for the algorithm you want to compare with your own method. It may spend weeks to find out why the results you got are not as good as the results provided in the original paper. The worst case is that you just can not figure out the reason even though you spend weeks. There are many reasons: not enough details available of the algorithm, different implementations, using different languages, the stochastic characteristic of EAs ..., etc. Therefore, here comes the motivation to build up a standard platform for you to carry out experiments. I believe that this project would clear the way for your research and save you a lot of time. Well, in a word, I would like to encourage you to take part in this project and let's together make the research life easy, comfortable, and efficient for you and someone else.


This library was implemented in C++ language, it has three basic components: algorithms, problems, and algorithms' performance evaluation. The class structure is as follows:

Algorithms:

This module provides basic templates for variant EAs. The module includes three basic classes: the Chromosome,Individual, and Population classes. Functions and descriptions of them are as follows.

The Chromosome class: The most fundamental class, represents a solution of a problem and is able to take composite encoding scheme,e.g., A b 1.23 ! & true 59. It was designed based on problems' representation.

The Individual class: A simple template individual class, provides some basic operations for a single individual with variant individuals of different algorithms.

The Population class: A template population class, provides some basic operations for a population


Problems:

This module collects some benchmark problems used for the performance evaluation for EAs in different research topics. Currently, it has two types of problems, function optimization problems (FOPs) and dynamic optimization problems(DOPs).

Problem .................................................... the base class for all problems
|
+BenchmarkFunction ........................................ the base class for all function optimization problems
| |
| +FSphere ................................................ the Sphere function f(x)=x*x
+DynamicProblem ........................................... the base class for dynamic optimization problems(DOPs)
|
+DynamicContinuous ...................................... the base class for DOPs in continuous space
|
+RealDB G.............................................. the base class for GDBG benchmark
| |
| +CompositionDBG ..................................... the Composition problems in GDBG 
| |
| +RotationDBG ......................................... the Rotation problems in GDBG
|
+MovingPeak ............................................ the Moving Peaks problem
|
+DF1 ................................................... the DF1 problem

Performance metrics:

PerformSingleObj ...............................................the performance measurement for single objective optimization problem

Other major files:

Global.h ........................................................ class for global variables 
StructDec.h ..................................................... struct declaration file
newran.h ........................................................ random number generator: From library http://www.robertnz.net/nr02doc.htm


Instruction for new algorithms and new problems

To add a new algorithm:

Take a simple PSO algorithm as an example, you need to develop two classes, which are: PSOIndividual: public Individual and PSOPopulation: public Population<PSOIndividual> , then follow the steps described in the beginning of the main function in main.cpp file to run the new algorithm.

To add a new problem:

If a new problem is to be added into this lib, you need to implement it under a proper position in the Problem class. In addition, encoding and decoding work must be done in the Chromosome.cpp file if a new solution representation is needed to describe the problem.

Each algorithm and problem has a unique ID, please assign an ID to an algorithm or a problem when they are added.

Development:

You can develop your algorithms in both Visual studio 2010 in Windows and CODE::BLOCK in Ubuntu, corresponding project files are EAlib.sln and EAlib.cbp, respectively. A makefile is also available to compile the program. If you have any question about this library, please contact Changhe Li with changhe.lw@gmail.com.


Algorithms collected so far:

This lib has included some algorithms from different research areas, which are listed below:

  • Function optimization:
    • Particle swarm optimization
      1. SPSO07: C. Maurice, Standard PSO 2007, http://www.particleswarm.info/Programs.html, 2007.
      2. SLPSO: Changhe Li, Shengxiang Yang and Trung Thanh Nguyen, A Self-Learning Particle Swarm Optimizer for Global Optimization Problems. accepted by IEEE TSMCB.
    • Differential evolution
      1. Storn, R. and Price, K. (1997), Differential Evolution - A Simple and Efficient Heuristic for Global Optimization over Continuous Spaces, Journal of Global Optimization, 11, pp. 341--359.
      2. Thomsen R (2004) Multimodal optimization using crowding-based differential evolution. In: Proceedings of the 2004 IEEE congress on evolutionary computation, pp 1382--1389. IEEE Press, Portland, OR, USA
    • Genetic algorithms
    • Hybrid algorithms
  • Dynamic optimization
    • Particle swarm optimization
      1. CPSO: S. Yang and C. Li, A clustering particle swarm optimizer for locating and tracking multiple optima in dynamic environments, IEEE Trans. Evol. Comput., vol. 14, no. 6, pp. 959-974, 2010.
      2. CPSOR: C.Li and S.Yang A General Framework of Multi-Population Methods with Clustering in Undetectable Dynamic Environments , accepted by IEEE TEC
    • Differential evolution
    • Genetic algorithms
    • Hybrid algorithms
      1. Lung R, Dumitrescu D (2010) Evolutionary swarm cooperative optimization in dynamic environments. Nat Comput Int J 9(1): 83--94.
  • Multi-Objective optimization

Problems collected so far:

  • Function optimization problems(FOPs) :Sphere, Sphere_Noisy, S_Sphere, R_Sphere, RS_Sphere, S_Sphere_CEC05, Rastrigin, Rastrigin_Noisy, S_Rastrigin, R_Rastrigin, RS_Rastrigin, S_Rastrigin_CEC05, RS_Rastrigin_CEC05, Weierstrass, RS_Weierstrass, R_Weierstrass, RS_Weierstrass_CEC05, Griewank, R_Griewank, RS_Griewank, Ackley, Ackley_Noisy, S_Ackley, R_Ackley, RS_Ackley, RS_Ackley_Bound_CEC05, Step, Quartic_Noisy, Scaffer_F6, Rosenbrock, S_Rosenbrock, S_Rosenbrock_CEC05, Schwefel_2_22, Schwefel_2_22_Noisy, S_Schwefel_2_22, R_Schwefel_2_22, RS_Schwefel_2_22, Schwefel, Schwefel_Noisy, S_Schwefel, R_Schwefel, RS_Schwefel, Schwefel_1_2, Schwefel_1_2_Noisy, S_Schwefel_1_2, S_Schwefel_1_2_Noisy, R_Schwefel_1_2, RS_Schwefel_1_2, S_Schwefel_1_2_CEC05, S_Schwefel_1_2_Noisy_CEC05, Schwefel_2_21, Penalized_1, Penalized_2, Noncont_Rastrigin, RS_Elliptic_CEC05, Elliptic, Com, R_Com, Com_CEC05, H_Com_CEC05, H_Com_Noisy_CEC05, RH_Com_CEC05, RH_Com_NarrowBasin_CEC05, RH_Com_Bound_CEC05, M_global1, M_global2, M_global3, M_global4, M_global5, S_Sphere_CEC08, Schwefel_2_21_CEC08, S_Rosenbrock_CEC08, S_Rastrigin_CEC08, S_Griewank_CEC08, S_Ackley_CEC08, RW_Gear_Train, RW_ParEst_FMSoundWaves.
  • Dynamic optimization problems(DOPs): CompositionDBG_DOP, RotationDBG_DOP, MovingPeak_DOP, DF1_DOP, Binary_DOP.

Please find the description of the above problems in the corresponding files in the library. 

Coming algorithms:

  1. CPSO-Hk: F. van den Bergh and A. P. Engelbrech, "A cooperative approach to particle swarm optimization," IEEE Trans. Evol. Comput., vol. 8, pp. 225-239, 2004.
  2. FIPS: R. Mendes, J. Kennedy, and J. Neves, "The fully informed particle swarm: Simpler, Maybe better," IEEE Trans. Evol. Comput., vol. 8, pp. 204-210, 2004.
  3. CLSPO: J. J. Liang, A. K. Qin, P. N. Suganthan, and S. Baska, "Comprehensive learning particle swarm optimizer for global optimization of multimodal functions," IEEE Trans. Evol. Comput., vol. 10, no. 3, pp. 281-295, 2006.
  4. APSO:Z. Zhan, J. Zhang, Y. Li, and H. S. Chung, "Adaptive particle swarm optimization," IEEE Trans. Syst., Man, Cybern. B: Cybern., vol. 39, pp. 1362-1381, 2009.
  5. FPSO: M. A. M. de Oca, T. Stutzle, M. Birattari, and M. Dorigo, "Frankenstein's pso: A composite particle swarm optimization algorithm," IEEE Trans. Evol. Comput., vol. 13, no. 5, pp. 1120-1132, 2009.
  6. JADE: J. Zhang and A. C. Sanderson, "JADE: adaptive differential evolution with optional external archive," IEEE Trans. Evol. Comput., vol. 13, no. 5, pp. 945-958, 2009.
  7. HRCGA:C. Garcia-Martinez, M. Lozano, F. Herrera, D. Molina, and A. M.Sanchez, "Global and local real-coded genetic algorithms based on parent-centric crossover operators," Europ. J. Oper. Res., vol. 185, pp.1088-1113, 2008.
  8. rSPSO: Bird, S.; Xiaodong Li; , "Using regression to improve local convergence," Evolutionary Computation, 2007. CEC 2007. IEEE Congress on , vol., no., pp.592-599, 25-28 Sept. 2007
  9. SPSO: Parrott, D.; Xiaodong Li; , "Locating and tracking multiple dynamic optima by a particle swarm model using speciation," Evolutionary Computation, IEEE Transactions on , vol.10, no.4, pp.440-458, Aug. 2006.
  10. mQSO,mCPSO: Blackwell, T.; Branke, J.; , "Multiswarms, exclusion, and anti-convergence in dynamic environments," Evolutionary Computation, IEEE Transactions on , vol.10, no.4, pp.459-472, Aug. 2006


To download the EAlib, please click here(updated 20 July 2012).


Contributors:

Ming Yang



last updated: 09/02/2013 00:12:41
1、资源项目源码均已通过严格测试验证,保证能够正常运行; 2、项目问题、技术讨论,可以给博主私信或留言,博主看到后会第一时间与您进行沟通; 3、本项目比较适合计算机领域相关的毕业设计课题、课程作业等使用,尤其对于人工智能、计算机科学与技术等相关专业,更为适合; 4、下载使用后,可先查看README.md或论文文件(如有),本项目仅用作交流学习参考,请切勿用于商业用途。 5、资源来自互联网采集,如有侵权,私聊博主删除。 6、可私信博主看论文后选择购买源代码。 1、资源项目源码均已通过严格测试验证,保证能够正常运行; 2、项目问题、技术讨论,可以给博主私信或留言,博主看到后会第一时间与您进行沟通; 3、本项目比较适合计算机领域相关的毕业设计课题、课程作业等使用,尤其对于人工智能、计算机科学与技术等相关专业,更为适合; 4、下载使用后,可先查看README.md或论文文件(如有),本项目仅用作交流学习参考,请切勿用于商业用途。 5、资源来自互联网采集,如有侵权,私聊博主删除。 6、可私信博主看论文后选择购买源代码。 1、资源项目源码均已通过严格测试验证,保证能够正常运行; 2、项目问题、技术讨论,可以给博主私信或留言,博主看到后会第一时间与您进行沟通; 3、本项目比较适合计算机领域相关的毕业设计课题、课程作业等使用,尤其对于人工智能、计算机科学与技术等相关专业,更为适合; 4、下载使用后,可先查看README.md或论文文件(如有),本项目仅用作交流学习参考,请切勿用于商业用途。 5、资源来自互联网采集,如有侵权,私聊博主删除。 6、可私信博主看论文后选择购买源代码。
应用背景为变电站电力巡检,基于YOLO v4算法模型对常见电力巡检目标进行检测,并充分利用Ascend310提供的DVPP等硬件支持能力来完成流媒体的传输、处理等任务,并对系统性能做出一定的优化。.zip深度学习是机器学习的一个子领域,它基于人工神经网络的研究,特别是利用多层次的神经网络来进行学习和模式识别。深度学习模型能够学习数据的高层次特征,这些特征对于图像和语音识别、自然语言处理、医学图像分析等应用至关重要。以下是深度学习的一些关键概念和组成部分: 1. **神经网络(Neural Networks)**:深度学习的基础是人工神经网络,它是由多个层组成的网络结构,包括输入层、隐藏层和输出层。每个层由多个神经元组成,神经元之间通过权重连接。 2. **前馈神经网络(Feedforward Neural Networks)**:这是最常见的神经网络类型,信息从输入层流向隐藏层,最终到达输出层。 3. **卷积神经网络(Convolutional Neural Networks, CNNs)**:这种网络特别适合处理具有网格结构的数据,如图像。它们使用卷积层来提取图像的特征。 4. **循环神经网络(Recurrent Neural Networks, RNNs)**:这种网络能够处理序列数据,如时间序列或自然语言,因为它们具有记忆功能,能够捕捉数据中的时间依赖性。 5. **长短期记忆网络(Long Short-Term Memory, LSTM)**:LSTM 是一种特殊的 RNN,它能够学习长期依赖关系,非常适合复杂的序列预测任务。 6. **生成对抗网络(Generative Adversarial Networks, GANs)**:由两个网络组成,一个生成器和一个判别器,它们相互竞争,生成器生成数据,判别器评估数据的真实性。 7. **深度学习框架**:如 TensorFlow、Keras、PyTorch 等,这些框架提供了构建、训练和部署深度学习模型的工具和库。 8. **激活函数(Activation Functions)**:如 ReLU、Sigmoid、Tanh 等,它们在神经网络中用于添加非线性,使得网络能够学习复杂的函数。 9. **损失函数(Loss Functions)**:用于评估模型的预测与真实值之间的差异,常见的损失函数包括均方误差(MSE)、交叉熵(Cross-Entropy)等。 10. **优化算法(Optimization Algorithms)**:如梯度下降(Gradient Descent)、随机梯度下降(SGD)、Adam 等,用于更新网络权重,以最小化损失函数。 11. **正则化(Regularization)**:技术如 Dropout、L1/L2 正则化等,用于防止模型过拟合。 12. **迁移学习(Transfer Learning)**:利用在一个任务上训练好的模型来提高另一个相关任务的性能。 深度学习在许多领域都取得了显著的成就,但它也面临着一些挑战,如对大量数据的依赖、模型的解释性差、计算资源消耗大等。研究人员正在不断探索新的方法来解决这些问题。
深度学习是机器学习的一个子领域,它基于人工神经网络的研究,特别是利用多层次的神经网络来进行学习和模式识别。深度学习模型能够学习数据的高层次特征,这些特征对于图像和语音识别、自然语言处理、医学图像分析等应用至关重要。以下是深度学习的一些关键概念和组成部分: 1. **神经网络(Neural Networks)**:深度学习的基础是人工神经网络,它是由多个层组成的网络结构,包括输入层、隐藏层和输出层。每个层由多个神经元组成,神经元之间通过权重连接。 2. **前馈神经网络(Feedforward Neural Networks)**:这是最常见的神经网络类型,信息从输入层流向隐藏层,最终到达输出层。 3. **卷积神经网络(Convolutional Neural Networks, CNNs)**:这种网络特别适合处理具有网格结构的数据,如图像。它们使用卷积层来提取图像的特征。 4. **循环神经网络(Recurrent Neural Networks, RNNs)**:这种网络能够处理序列数据,如时间序列或自然语言,因为它们具有记忆功能,能够捕捉数据中的时间依赖性。 5. **长短期记忆网络(Long Short-Term Memory, LSTM)**:LSTM 是一种特殊的 RNN,它能够学习长期依赖关系,非常适合复杂的序列预测任务。 6. **生成对抗网络(Generative Adversarial Networks, GANs)**:由两个网络组成,一个生成器和一个判别器,它们相互竞争,生成器生成数据,判别器评估数据的真实性。 7. **深度学习框架**:如 TensorFlow、Keras、PyTorch 等,这些框架提供了构建、训练和部署深度学习模型的工具和库。 8. **激活函数(Activation Functions)**:如 ReLU、Sigmoid、Tanh 等,它们在神经网络中用于添加非线性,使得网络能够学习复杂的函数。 9. **损失函数(Loss Functions)**:用于评估模型的预测与真实值之间的差异,常见的损失函数包括均方误差(MSE)、交叉熵(Cross-Entropy)等。 10. **优化算法(Optimization Algorithms)**:如梯度下降(Gradient Descent)、随机梯度下降(SGD)、Adam 等,用于更新网络权重,以最小化损失函数。 11. **正则化(Regularization)**:技术如 Dropout、L1/L2 正则化等,用于防止模型过拟合。 12. **迁移学习(Transfer Learning)**:利用在一个任务上训练好的模型来提高另一个相关任务的性能。 深度学习在许多领域都取得了显著的成就,但它也面临着一些挑战,如对大量数据的依赖、模型的解释性差、计算资源消耗大等。研究人员正在不断探索新的方法来解决这些问题。
深度学习是机器学习的一个子领域,它基于人工神经网络的研究,特别是利用多层次的神经网络来进行学习和模式识别。深度学习模型能够学习数据的高层次特征,这些特征对于图像和语音识别、自然语言处理、医学图像分析等应用至关重要。以下是深度学习的一些关键概念和组成部分: 1. **神经网络(Neural Networks)**:深度学习的基础是人工神经网络,它是由多个层组成的网络结构,包括输入层、隐藏层和输出层。每个层由多个神经元组成,神经元之间通过权重连接。 2. **前馈神经网络(Feedforward Neural Networks)**:这是最常见的神经网络类型,信息从输入层流向隐藏层,最终到达输出层。 3. **卷积神经网络(Convolutional Neural Networks, CNNs)**:这种网络特别适合处理具有网格结构的数据,如图像。它们使用卷积层来提取图像的特征。 4. **循环神经网络(Recurrent Neural Networks, RNNs)**:这种网络能够处理序列数据,如时间序列或自然语言,因为它们具有记忆功能,能够捕捉数据中的时间依赖性。 5. **长短期记忆网络(Long Short-Term Memory, LSTM)**:LSTM 是一种特殊的 RNN,它能够学习长期依赖关系,非常适合复杂的序列预测任务。 6. **生成对抗网络(Generative Adversarial Networks, GANs)**:由两个网络组成,一个生成器和一个判别器,它们相互竞争,生成器生成数据,判别器评估数据的真实性。 7. **深度学习框架**:如 TensorFlow、Keras、PyTorch 等,这些框架提供了构建、训练和部署深度学习模型的工具和库。 8. **激活函数(Activation Functions)**:如 ReLU、Sigmoid、Tanh 等,它们在神经网络中用于添加非线性,使得网络能够学习复杂的函数。 9. **损失函数(Loss Functions)**:用于评估模型的预测与真实值之间的差异,常见的损失函数包括均方误差(MSE)、交叉熵(Cross-Entropy)等。 10. **优化算法(Optimization Algorithms)**:如梯度下降(Gradient Descent)、随机梯度下降(SGD)、Adam 等,用于更新网络权重,以最小化损失函数。 11. **正则化(Regularization)**:技术如 Dropout、L1/L2 正则化等,用于防止模型过拟合。 12. **迁移学习(Transfer Learning)**:利用在一个任务上训练好的模型来提高另一个相关任务的性能。 深度学习在许多领域都取得了显著的成就,但它也面临着一些挑战,如对大量数据的依赖、模型的解释性差、计算资源消耗大等。研究人员正在不断探索新的方法来解决这些问题。
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值