1 |
ZHANG T. An introduction to support vector machines and other kernel-based learning methods[J]. AI Magazine, 2001, 22(2): 103-104.
|
2 |
XIE J H. Kernel optimization of LS-SVM based on damage detection for smart structures[C]// Proceedings of the 2nd IEEE International Conference on Computer Science and Information Technology. Piscataway: IEEE, 2009: 406-409. 10.1109/iccsit.2009.5234791
|
3 |
HUANG C R, CHEN Y T, CHEN W Y, et al. Gastroesophageal reflux disease diagnosis using hierarchical heterogeneous descriptor fusion support vector machine[J]. IEEE Transactions on Biomedical Engineering, 2016, 63(3): 588-599. 10.1109/tbme.2015.2466460
|
4 |
PENG S L, HU Q H, CHEN Y L, et al. Improved support vector machine algorithm for heterogeneous data[J]. Pattern Recognition, 2015, 48(6): 2072-2083. 10.1016/j.patcog.2014.12.015
|
5 |
SONNENBURG S, RÄTSCH G, SCHÄFER C, et al. Large scale multiple kernel learning[J]. Journal of Machine Learning Research, 2006, 7: 1531-1565. 10.1186/1471-2105-7-s1-s9
|
6 |
XIAO Y L, ZHONG S P. An improved online multiple kernel classification algorithm based on double updating online learning[C]// Proceedings of the 2014 International Conference on Cloud Computing and Internet of Things. Piscataway: IEEE, 2014: 109-113. 10.1109/cciot.2014.7062516
|
7 |
BACH F R. Consistency of the group lasso and multiple kernel learning[J]. Journal of Machine Learning Research, 2008, 9: 1179-1225. 10.1145/1390156.1390161
|
8 |
CORTES C, MOHRI M, ROSTAMIZADEH A. Learning sequence kernels[C]// Proceedings of the 2008 IEEE Workshop on Machine Learning for Signal Processing. Piscataway: IEEE, 2008: 2-8. 10.1109/mlsp.2008.4685446
|
9 |
KLOFT M, BREFELD U, LASKOV P, et al. Non-sparse multiple kernel learning[EB/OL]. [2021-03-20]..
|
10 |
ZHENG D N, WANG J X, ZHAO Y N. Non-flat function estimation with a multi-scale support vector regression[J]. Neurocomputing, 2006, 70(1/2/3): 420-429. 10.1016/j.neucom.2005.12.128
|
11 |
LANCKRIET G R G, CRISTIANINI N, BARTLETT P, et al. Learning the kernel matrix with semidefinite programming[J]. Journal of Machine Learning Research, 2002, 5: 27-72.
|
12 |
GHIASI-SHIRAZI K, SAFABAKHSH R, SHAMSI M. Learning translation invariant kernels for classification[J]. Journal of Machine Learning Research, 2014, 11: 1353-1390.
|
13 |
汪洪桥,孙富春,蔡艳宁,等. 多核学习方法[J]. 自动化学报, 2010, 36(8):1037-1050. 10.3724/SP.J.1004.2010.01037
|
|
WANG H Q, SUN F C, CAI Y N, et al. On multiple kernel learning methods[J]. Acta Automatica Sinica, 2010, 36(8): 1037-1050. 10.3724/SP.J.1004.2010.01037
|
14 |
BENNETT K P, MOMMA M, EMBRECHTS M J. MARK: a boosting algorithm for heterogeneous kernel models[C]// Proceedings of 8th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. New York: ACM, 2002: 24-31. 10.1145/775047.775051
|
15 |
SONNENBURG S, RÄTSCH G, SCHÄFER C. A general and efficient multiple kernel learning algorithm[C]// Proceedings of the 18th International Conference on Neural Information Processing Systems. Cambridge: MIT Press, 2005: 1273-1280. 10.1007/11415770_30
|
16 |
RAKOTOMAMONJY A, BACH F R, CANU S, et al. SimpleMKL[J]. Journal of Machine Learning Research, 2008, 9(11): 2491-2521.
|
17 |
ARGYRIOU A, HAUSER R, MICCHELLI C A, et al. A DC-programming algorithm for kernel selection[C]// Proceedings of the 23rd International Conference on Machine Learning. New York: ACM, 2006: 41-48. 10.1145/1143844.1143850
|
18 |
ALIOSCHA-PEREZ M, OVENEKE M C, SAHLI H. SVRG-MKL: a fast and scalable multiple kernel learning solution for features combination in multi-class classification problems[J]. IEEE Transactions on Neural Networks and Learning Systems, 2020, 31(5): 1710-1723. 10.1109/tnnls.2019.2922123
|
19 |
WANG X M, WANG S T, DU Y J, et al. Minimum class variance multiple kernel learning[J]. Knowledge-Based Systems, 2020, 208(5): No.106469. 10.1016/j.knosys.2020.106469
|
20 |
LIU X W, WANG L, ZHU X Z, et al. Absent multiple kernel learning algorithms[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020, 42(6): 1303-1316. 10.1109/tpami.2019.2895608
|
21 |
王梅,薛成龙,张强. 基于秩空间差异的多核组合方法[J]. 山东大学学报(工学版), 2021, 51(1):108-113.
|
|
WANG M, XUE C L, ZHANG Q. Multi-kernel combination method based on rank spatial difference[J]. Journal of Shandong University (Engineering Science), 2021, 51(1): 108-113.
|
22 |
贾涵,连晓峰,潘兵. 基于模糊松弛约束的外观缺陷多核学习技术[J]. 测控技术, 2019, 38(8):43-47,73. 10.19708/j.ckjs.2019.08.009
|
|
JIA H, LIAN X F, PAN B. Appearance defects multiple kernel learning technology based on fuzzy relaxation constraints[J]. Measurement and Control Technology, 2019, 38(8): 43-47, 73. 10.19708/j.ckjs.2019.08.009
|
23 |
HE Q, ZHANG Q S, WANG H Y. Kernel-target alignment based multiple kernel one-class support vector machine[C]// Proceedings of the 2019 IEEE International Conference on Systems, Man and Cybernetics. Piscataway: IEEE, 2019: 2083-2088. 10.1109/smc.2019.8914503
|
24 |
KOLTCHINSKII V, PANCHENKO D. Rademacher processes and bounding the risk of function learning[J]. GINÉ E, MASON D M, WELLNER J A. Progress in Probability II, PRPR 47. Boston: Birkhäuser, 2000: 443-459. 10.1007/978-1-4612-1358-1_29
|
25 |
BARTLETT P L, MENDELSON S. Rademacher and Gaussian complexities: risk bounds and structural results[C]// Proceedings of the 2001 International Conference on Computational Learning Theory, LNCS2111. Berlin: Springer, 2001: 224-240.
|
26 |
KLOFT M, BLANCHARD G. On the convergence rate of ℓp-norm multiple kernel learning[J]. Journal of Machine Learning Research, 2012, 13: 2465-2501.
|
27 |
CORTES C, KLOFT M, MOHRI M. Learning kernels using local Rademacher complexity[C]// Proceedings of the 26th International Conference on Neural Information Processing Systems. Red Hook, NY: Curran Associates Inc., 2013: 2760-2768.
|
28 |
LIU Y, LIAO S Z. Eigenvalues ratio for kernel selection of kernel methods[C]// Proceedings of the 29th AAAI Conference on Artificial Intelligence. Palo Alto, CA: AAAI Press, 2015: 2814-2820. 10.1609/aaai.v33i01.33013462
|
29 |
LIU Y, LIAO S Z, LIN H L, et al. Infinite kernel learning: generalization bounds and algorithms[C]// Proceedings of the 31st AAAI Conference on Artificial Intelligence. Palo Alto, CA: AAAI Press, 2017: 2280-2286. 10.1609/aaai.v34i04.5892
|
30 |
WILLIAMS C K I. Computing with infinite networks[C]// Proceedings of the 9th International Conference on Neural Information Processing Systems. Cambridge: MIT Press, 1996:295-301.
|
31 |
LEE J, BAHRI Y, NOVAK R, et al. Deep neural networks as Gaussian processes[EB/OL]. (2018-03-03) [2020-12-18]..
|
32 |
JACOT A, GABRIEL F, HONGLER C. Neural tangent kernel: convergence and generalization in neural networks[C]// Proceedings of the 32nd International Conference on Neural Information Processing Systems. Red Hook, NY: Curran Associates Inc., 2018: 8580-8589.
|
33 |
LEE J, XIAO L, SCHOENHOLZ S S, et al. Wide neural networks of any depth evolve as linear models under gradient descent[EB/OL]. (2019-12-08) [2020-12-20].. 10.1088/1742-5468/abc62b
|
34 |
ARORA S, DU S S, HU W, et al. On exact computation with an infinitely wide neural net[EB/OL]. (2019-11-04) [2020-12-24]..
|
35 |
HUANG K X, WANG Y Q, TAO M L, et al. Why do deep residual networks generalize better than deep feedforward networks? — a neural tangent kernel perspective[EB/OL]. (2020-12-22) [2020-12-30]..
|
36 |
DU S S, HOU K C, PÓCZOS B, et al. Graph neural tangent kernel: fusing graph neural networks with graph kernels[EB/OL]. (2019-11-03) [2020-11-04]..
|
37 |
LI Z Y, WANG R S, YU D L, et al. Enhanced convolutional neural tangent kernels[EB/OL]. (2019-11-03) [2021-01-05].. 10.1109/icassp.2019.8682265
|
38 |
ARORA S, DU S S, LI Z Y, et al. Harnessing the power of infinitely wide deep nets on small-data tasks[EB/OL]. (2019-10-27) [2021-01-08]..
|
39 |
CRISTIANINI N, SHAWE-TAYLOR J, ELISSEEFF A, et al. On kernel-target alignment[C]// Proceedings of the 14th International Conference on Neural Information Processing Systems: Natural and Synthetic. Cambridge: MIT Press, 2001: 367-373. 10.7551/mitpress/1120.003.0052
|
40 |
BARTLETT P L, BOUSQUET O, MENDELSON S. Local Rademacher complexities[J]. The Annals of Statistics, 2005, 33(4): 1497-1537. 10.1214/009053605000000282
|