Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Random Forests for CUDA GPUs
Blekinge Institute of Technology, School of Computing.
Blekinge Institute of Technology, School of Computing.
2010 (English)Independent thesis Advanced level (degree of Master (Two Years))Student thesis
Abstract [en]

Context. Machine Learning is a complex and resource consuming process that requires a lot of computing power. With the constant growth of information, the need for efficient algorithms with high performance is increasing. Today's commodity graphics cards are parallel multi processors with high computing capacity at an attractive price and are usually pre-installed in new PCs. The graphics cards provide an additional resource to be used in machine learning applications. The Random Forest learning algorithm which has been showed competitive within machine learning has a good potential for performance increase through parallelization of the algorithm. Objectives. In this study we implement and review a revised Random Forest algorithm for GPU execution using CUDA. Methods. A review of previous work in the area has been done by studying articles from several sources, including Compendex, Inspec, IEEE Xplore, ACM Digital Library and Springer Link. Additional information regarding GPU architecture and implementation specific details have been obtained mainly from documentation available from Nvidia and the Nvidia developer forums. The implemented algorithm has been benchmarked and compared with two state-of-the-art CPU implementations of the Random Forest algorithm, both regarding consumed time for training and classification and for classification accuracy. Results. Measurements from benchmarks made on the three different algorithms are gathered showing the performance results of the algorithms for two publicly available data sets. Conclusion. We conclude that our implementation under the right conditions is able to outperform its competitors. We also conclude that this is only true for certain data sets depending on the size of the data sets. Moreover we conclude that there is potential for further improvements of the algorithm both regarding performance as well as adaption towards a wider range of real world applications.

Place, publisher, year, edition, pages
2010. , p. 39
Keywords [en]
CUDA, Random forests, Parallel computing, Graphics processing units
National Category
Software Engineering
Identifiers
URN: urn:nbn:se:bth-2953Local ID: oai:bth.se:arkivex92EA884698CFD4ABC12577430040AEA6OAI: oai:DiVA.org:bth-2953DiVA, id: diva2:830248
Uppsok
Technology
Supervisors
Note
Mikael: +46768539263, Daniel: +46703040693Available from: 2015-04-22 Created: 2010-06-15 Last updated: 2018-01-11Bibliographically approved

Open Access in DiVA

fulltext(1295 kB)916 downloads
File information
File name FULLTEXT01.pdfFile size 1295 kBChecksum SHA-512
63223b43203e767b6a99cea9291fea916ee340603c95401d9daac90978ed9b6eb5a9c103ee24c9f0c0d752bd83777c364691c5f8ff91f0b641f543749b4b2243
Type fulltextMimetype application/pdf

By organisation
School of Computing
Software Engineering

Search outside of DiVA

GoogleGoogle Scholar
Total: 921 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

urn-nbn

Altmetric score

urn-nbn
Total: 320 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf