Yayın:
Comparing of deep neural networks and extreme learning machines based on growing and pruning approach

dc.contributor.authorAkyol, Kemal
dc.date.accessioned2026-01-04T13:55:54Z
dc.date.issued2020-02-01
dc.description.abstractAbstract Recently, the studies based on Deep Neural Networks and Extreme Learning Machines have become prominent. The models of parameters designed in these studies have been chosen randomly and the models have been designed in this direction. The main focus of this study is to determine the ideal parameters i.e. optimum hidden layer number, optimum hidden neuron number and activation function for Deep Neural Networks and Extreme Learning Machines architectures based on growing and pruning approach and to compare the performances of the models designed. The performances of the models are evaluated on two datasets; Parkinson and Self-Care Activities Dataset. Multi experiments have verified that the Deep Neural Networks architectures present a good prediction performance and this architecture outperforms the Extreme Learning Machines.
dc.description.urihttps://doi.org/10.1016/j.eswa.2019.112875
dc.description.urihttps://dx.doi.org/10.1016/j.eswa.2019.112875
dc.identifier.doi10.1016/j.eswa.2019.112875
dc.identifier.issn0957-4174
dc.identifier.openairedoi_dedup___::b8f5af30d634012597d7a335cfdf3407
dc.identifier.orcid0000-0002-2272-5243
dc.identifier.scopus2-s2.0-85070931802
dc.identifier.startpage112875
dc.identifier.urihttps://hdl.handle.net/20.500.12597/37826
dc.identifier.volume140
dc.identifier.wos000495470700009
dc.language.isoeng
dc.publisherElsevier BV
dc.relation.ispartofExpert Systems with Applications
dc.rightsCLOSED
dc.titleComparing of deep neural networks and extreme learning machines based on growing and pruning approach
dc.typeArticle
dspace.entity.typePublication
local.api.response{"authors":[{"fullName":"Kemal Akyol","name":"Kemal","surname":"Akyol","rank":1,"pid":{"id":{"scheme":"orcid","value":"0000-0002-2272-5243"},"provenance":null}}],"openAccessColor":null,"publiclyFunded":false,"type":"publication","language":{"code":"eng","label":"English"},"countries":null,"subjects":[{"subject":{"scheme":"FOS","value":"0209 industrial biotechnology"},"provenance":null},{"subject":{"scheme":"FOS","value":"0202 electrical engineering, electronic engineering, information engineering"},"provenance":null},{"subject":{"scheme":"FOS","value":"02 engineering and technology"},"provenance":null}],"mainTitle":"Comparing of deep neural networks and extreme learning machines based on growing and pruning approach","subTitle":null,"descriptions":["Abstract Recently, the studies based on Deep Neural Networks and Extreme Learning Machines have become prominent. The models of parameters designed in these studies have been chosen randomly and the models have been designed in this direction. The main focus of this study is to determine the ideal parameters i.e. optimum hidden layer number, optimum hidden neuron number and activation function for Deep Neural Networks and Extreme Learning Machines architectures based on growing and pruning approach and to compare the performances of the models designed. The performances of the models are evaluated on two datasets; Parkinson and Self-Care Activities Dataset. Multi experiments have verified that the Deep Neural Networks architectures present a good prediction performance and this architecture outperforms the Extreme Learning Machines."],"publicationDate":"2020-02-01","publisher":"Elsevier BV","embargoEndDate":null,"sources":["Crossref"],"formats":null,"contributors":null,"coverages":null,"bestAccessRight":{"code":"c_14cb","label":"CLOSED","scheme":"http://vocabularies.coar-repositories.org/documentation/access_rights/"},"container":{"name":"Expert Systems with Applications","issnPrinted":"0957-4174","issnOnline":null,"issnLinking":null,"ep":null,"iss":null,"sp":"112875","vol":"140","edition":null,"conferencePlace":null,"conferenceDate":null},"documentationUrls":null,"codeRepositoryUrl":null,"programmingLanguage":null,"contactPeople":null,"contactGroups":null,"tools":null,"size":null,"version":null,"geoLocations":null,"id":"doi_dedup___::b8f5af30d634012597d7a335cfdf3407","originalIds":["S0957417419305858","10.1016/j.eswa.2019.112875","50|doiboost____|b8f5af30d634012597d7a335cfdf3407","2967914531"],"pids":[{"scheme":"doi","value":"10.1016/j.eswa.2019.112875"}],"dateOfCollection":null,"lastUpdateTimeStamp":null,"indicators":{"citationImpact":{"citationCount":49,"influence":5.670477e-9,"popularity":3.8263693e-8,"impulse":41,"citationClass":"C4","influenceClass":"C4","impulseClass":"C3","popularityClass":"C3"}},"instances":[{"pids":[{"scheme":"doi","value":"10.1016/j.eswa.2019.112875"}],"license":"Elsevier TDM","type":"Article","urls":["https://doi.org/10.1016/j.eswa.2019.112875"],"publicationDate":"2020-02-01","refereed":"peerReviewed"},{"alternateIdentifiers":[{"scheme":"mag_id","value":"2967914531"},{"scheme":"doi","value":"10.1016/j.eswa.2019.112875"}],"type":"Article","urls":["https://dx.doi.org/10.1016/j.eswa.2019.112875"],"refereed":"nonPeerReviewed"}],"isGreen":false,"isInDiamondJournal":false}
local.import.sourceOpenAire
local.indexed.atWOS
local.indexed.atScopus

Dosyalar

Koleksiyonlar