Mostrar el registro sencillo del ítem
dc.contributor.author | Torres de la Sierra, Yuri | |
dc.contributor.author | Llanos Ferraris, Diego Rafael | |
dc.contributor.author | González Escribano, Arturo | |
dc.contributor.author | Llanos, Diego | |
dc.date.accessioned | 2024-10-14T08:16:35Z | |
dc.date.available | 2024-10-14T08:16:35Z | |
dc.date.issued | 2012 | |
dc.identifier.citation | 18th International Conference, Euro-Par 2012, Rhodes Island, Greece, August 27-31, 2012, pp 502-513 | es |
dc.identifier.uri | https://uvadoc.uva.es/handle/10324/70779 | |
dc.description | Producción Científica | es |
dc.description.abstract | Programming models and techniques to exploit parallelism in accelerators, such as GPUs, are different from those used in traditional parallel models for shared- or distributed-memory systems. It is a challenge to blend different programming models to coordinate and exploit devices with very different characteristics and computation powers. This paper presents a new extensible framework model to encapsulate run-time decisions related to data partition, granularity, load balance, synchronization, and communication for systems including assorted GPUs. Thus, the main parallel code becomes independent of them, using internal topology and system information to transparently adapt the computation to the system. The programmer can develop specific functions for each architecture, or use existent specialized library functions for different CPU-core or GPU architectures. The high-level coordination is expressed using a programming model built on top of message-passing, providing portability across distributed- or shared-memory systems. We show with an example how to produce a parallel code that can be used to efficiently run on systems ranging from a Beowulf cluster to a machine with mixed GPUs. Our experimental results show how the run-time system, guided by hints about the computational-power ratios of different devices, can automatically part and distribute large computations across heterogeneous systems, improving the overall performance. | es |
dc.format.extent | 11 p. | es |
dc.format.mimetype | application/pdf | es |
dc.language.iso | eng | es |
dc.publisher | Springer | es |
dc.rights.accessRights | info:eu-repo/semantics/openAccess | es |
dc.subject | Informática | es |
dc.title | Encapsulated synchronization and load-balance in heterogeneous programming | es |
dc.type | info:eu-repo/semantics/conferenceObject | es |
dc.identifier.doi | 10.1007/978-3-642-32820-6_50 | es |
dc.relation.publisherversion | https://link.springer.com/chapter/10.1007/978-3-642-32820-6_50 | es |
dc.title.event | 18th International European Conference on Parallel and Distributed Computing | es |
dc.type.hasVersion | info:eu-repo/semantics/publishedVersion | es |
dc.subject.unesco | 1203 Ciencia de Los Ordenadores | es |
dc.subject.unesco | 3304 Tecnología de Los Ordenadores | es |