Please use this identifier to cite or link to this item: http://hdl.handle.net/11455/43913
DC FieldValueLanguage
dc.contributor.authorJuang, Chia-Fengen_US
dc.contributor.authorChen, Teng-Changen_US
dc.contributor.authorCheng, Wei-Yuanen_US
dc.contributor.otherNational Chung Hsing University,Department of Electrical Engineeringen_US
dc.contributor.other國立中興大學電機工程學系zh_TW
dc.date2011-8zh_TW
dc.date.accessioned2014-06-06T08:11:41Z-
dc.date.available2014-06-06T08:11:41Z-
dc.identifier.issn1063-6706zh_TW
dc.identifier.urihttp://hdl.handle.net/11455/43913-
dc.description.abstractThis paper proposes the implementation of a zero-order Takagi-Sugeno-Kang (TSK)-type fuzzy neural network (FNN) on graphic processing units (GPUs) to reduce training time. The software platform that this study uses is the compute unified device architecture (CUDA). The implemented FNN uses structure and parameter learning in a self-constructing neural fuzzy inference network because of its admirable learning performance. FNN training is conventionally implemented on a single-threaded CPU, where each input variable and fuzzy rule is serially processed. This type of training is time consuming, especially for a high-dimensional FNN that consists of a large number of rules. The GPU is capable of running a large number of threads in parallel. In a GPU-implemented FNN (GPU-FNN), blocks of threads are partitioned according to parallel and independent properties of fuzzy rules. Large sets of input data are mapped to parallel threads in each block. For memory management, this research suitably divides the datasets in the GPU-FNN into smaller chunks according to fuzzy rule structures to share on-chip memory among multiple thread processors. This study applies the GPU-FNN to different problems to verify its efficiency. The results show that to train an FNN with GPU implementation achieves a speedup of more than 30 times that of CPU implementation for problems with high-dimensional attributes.en_US
dc.language.isoen_USzh_TW
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC.en_US
dc.relationIEEE TRANSACTIONS ON FUZZY SYSTEMS, Volume 19, Issue 4, Page(s) 717-728.en_US
dc.subjectClassificationen_US
dc.subjectcompute unified device architecture (CUDA)en_US
dc.subjectfuzzy neural networks (FNNs)en_US
dc.subjectgraphic processing unit (GPU)en_US
dc.subjectneural fuzzy systemsen_US
dc.titleSpeedup of Implementing Fuzzy Neural Networks With High-Dimensional Inputs Through Parallel Processing on Graphic Processing Unitsen_US
dc.typeJournal Articlezh_TW
dc.identifier.doi10.1109/TFUZZ.2011.2140326zh_TW
dc.contributor.catalogerMiao-zhen Luoen_US
Appears in Collections:電機工程學系所
文件中的檔案:

取得全文請前往華藝線上圖書館



Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.