{"id":8827,"date":"2013-01-22T23:17:57","date_gmt":"2013-01-22T21:17:57","guid":{"rendered":"http:\/\/hgpu.org\/?p=8827"},"modified":"2013-01-22T23:17:57","modified_gmt":"2013-01-22T21:17:57","slug":"data-parallel-patterns-on-cpugpu-mix","status":"publish","type":"post","link":"https:\/\/hgpu.org\/?p=8827","title":{"rendered":"Data parallel patterns on CPU\/GPU mix"},"content":{"rendered":"<p>We propose a model that uses a small set of quite simple parameters to devise a proper partitioning{between CPU and GPU cores{of the tasks deriving from structured data parallel patterns\/algorithmic skeletons. The model takes into account both hardware related and application dependent parameters. It eventually computes the percentage of tasks to be executed on CPU and GPU cores such that both kind of cores are exploited and performance figures are optimized. Different experimental results on state-of-the-art CPU\/GPU architectures are shown that assess the model properties.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>We propose a model that uses a small set of quite simple parameters to devise a proper partitioning{between CPU and GPU cores{of the tasks deriving from structured data parallel patterns\/algorithmic skeletons. The model takes into account both hardware related and application dependent parameters. It eventually computes the percentage of tasks to be executed on CPU [&hellip;]<\/p>\n","protected":false},"author":351,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_jetpack_memberships_contains_paid_content":false,"footnotes":"","jetpack_publicize_message":"","jetpack_publicize_feature_enabled":true,"jetpack_social_post_already_shared":false,"jetpack_social_options":{"image_generator_settings":{"template":"highway","default_image_id":0,"font":"","enabled":false},"version":2}},"categories":[11,89,3],"tags":[1782,14,263,20,1401,251,378],"class_list":["post-8827","post","type-post","status-publish","format-standard","hentry","category-computer-science","category-nvidia-cuda","category-paper","tag-computer-science","tag-cuda","tag-data-parallelism","tag-nvidia","tag-nvidia-geforce-gt-540","tag-nvidia-geforce-gtx-285","tag-tesla-c2050"],"views":2470,"jetpack_publicize_connections":[],"jetpack_featured_media_url":"","jetpack_sharing_enabled":true,"_links":{"self":[{"href":"https:\/\/hgpu.org\/index.php?rest_route=\/wp\/v2\/posts\/8827","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/hgpu.org\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/hgpu.org\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/hgpu.org\/index.php?rest_route=\/wp\/v2\/users\/351"}],"replies":[{"embeddable":true,"href":"https:\/\/hgpu.org\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=8827"}],"version-history":[{"count":0,"href":"https:\/\/hgpu.org\/index.php?rest_route=\/wp\/v2\/posts\/8827\/revisions"}],"wp:attachment":[{"href":"https:\/\/hgpu.org\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=8827"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/hgpu.org\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=8827"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/hgpu.org\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=8827"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}