Metamaterials Design Method based on Deep learning Database
Tóm tắt
Từ khóa
Tài liệu tham khảo
Wei, 2018, Deep-learning-enabled on-demand design of chiral theory metamaterials, ACS Nano, 12
Itzik, 2018, Plasmonic nanostructure design and characterization via Deep Learning, Light: Science & Application, 7, 60, 10.1038/s41377-018-0060-7
Valley, 2019, Metamaterials How Close Are We to a Klingon Cloaking Device or Harry Potter Invisibility Cloak? (No SAND2019-14245B)
Zhang, 2019, Machine-learning designs of anisotropic digital coding metasurfaces, Adv. Theory Simul., 2, 1800132, 10.1002/adts.201800132
Qiu, 2019, Deep learning: a rapid and efficient route to automatic metasurface design, Advanced Science, 6, 1900128, 10.1002/advs.201900128
Nadell, 2019, Deep learning for accelerated all-dielectric metasurface design, Optics express, 27, 27523, 10.1364/OE.27.027523
An, 2019, A deep learning approach for objective-driven all-dielectric metasurface design, ACS Photonics, 6, 3196, 10.1021/acsphotonics.9b00966
John-Herpin, 2021, Infrared Metasurface augmented by deep learning for monitoring dynamics between all major classes of biomolecules, Advanced Materials, 33, 2006054, 10.1002/adma.202006054
Liu, 2018, Generative model for the inverse design of metasurfaces, Nano letters, 18, 6570, 10.1021/acs.nanolett.8b03171
Sajedian, 2019, Double-deep Q-learning to increase the efficiency of metasurface holograms, Scientific reports, 9, 1, 10.1038/s41598-019-47154-z
Pestourie, 2020, Active learning of deep surrogates for PDEs: application to metasurface design., npj Computational Materials, 6, 1, 10.1038/s41524-020-00431-2
Mall, 2020, Fast design of plasmonic metasurfaces enabled by deep learning, Journal of Physics D: Applied Physics, 53, 49LT01, 10.1088/1361-6463/abb33c
Itzik, 2018, Plasmonic nanostructure design and characterization via deep learning
Scherer, 2010, 92
Li, 2018, Hyperspectral remote sensing image classification based on maximum overlap pooling convolutional neural network, Sensors, 18, 3587, 10.3390/s18103587
Zhang, 2018, Twelve-layer deep convolutional neural network with stochastic pooling for tea category classification on GPU platform, Multimedia Tools and Applications, 77, 22821, 10.1007/s11042-018-5765-3
Pagola, 2017
Sharma, 2019, 1
Boureau, 2010, A theoretical analysis of feature pooling in visual recognition., 111
Cybenko, 1989, Approximation by superpositions of a sigmoidal function, Mathematics Of Control Signals and Systems, 2, 303, 10.1007/BF02551274
Kalman, 1992, Why tanh: choosing a sigmoidal function. In [Proceedings 1992], 578
Barron, 1993, Universal approximation bounds for superpositions of a sigmoidal function, IEEE Transactions on Information Theory, 39, 930, 10.1109/18.256500
Chandra, 2003, Sigmoidal function classes for feedforward artificial neural networks, Neural Processing Letters, 18, 205, 10.1023/B:NEPL.0000011137.04221.96
Kyurkchiev, 2015, Sigmoidal functions: some computational and modelling aspects, Biomath Communications, 1, 10.11145/j.bmc.2015.03.081
Fan, 2000, Extended tanh-function method and its applications to nonlinear equations, Physics Letters A, 277, 212, 10.1016/S0375-9601(00)00725-8
Parkes, 1996, An automated tanh-function method for finding solitary wave solutions to non-linear evolution equations, Computer Physics Communications, 98, 288, 10.1016/0010-4655(96)00104-X
Elwakil, 2002, Modified extended tanh-function method for solving nonlinear partial differential equations, Physics Letters A, 299, 179, 10.1016/S0375-9601(02)00669-2
Abdou, 2006, Modified extended tanh-function method and its application on nonlinear physical equations, Physics Letters A, 353, 487, 10.1016/j.physleta.2006.01.013
Agarap, 2018
Eckle, 2019, A comparison of deep networks with ReLU activation function and linear spline-type methods, Neural Networks, 110, 232, 10.1016/j.neunet.2018.11.005
Schmidt-Hieber, 2020, Nonparametric regression using deep neural networks with ReLU activation function, The Annals of Statistics, 48, 1875
He, 2018, Relu deep neural networks and linear finite elements.
Bottou, 2010, Large-scale machine learning with stochastic gradient descent In, 177
Shamir, 2013, Stochastic gradient descent for non-smooth optimization: Convergence results and optimal averaging schemes, 71
Qian, 2015, Efficient distance metric learning by adaptive sampling and mini-batch stochastic gradient descent (SGD) Machine Learning, 99, 353
Wijnhoven, 2010, Fast training of object detection using stochastic gradient descent, 424
Liu, 2020, An improved analysis of stochastic gradient descent with momentum
Cutkosky, 2019, Momentum-based variance reduction in non-convex sgd.
Wang, 2020
Khan, 2019, Design of momentum fractional stochastic gradient descent for recommender systems, IEEE Access, 7, 179575, 10.1109/ACCESS.2019.2954859
Cutkosky, 2020, Momentum improves normalized sgd, 2260
Zhang, 2018, Improved adam optimizer for deep neural networks, 1
Bock, 2019, 232