Por favor, use este identificador para citar o enlazar este ítem:https://uvadoc.uva.es/handle/10324/70440
Título
Speculative parallelization
Año del Documento
2006
Editorial
IEEE Press
Descripción
Producción Científica
Documento Fuente
Computer, vol. 39, no. 12, pp. 126-128, December 2006.
Abstract
The most promising technique for automatically parallelizing loops when the system cannot determine dependences at compile time is speculative parallelization. Also called thread-level speculation, this technique assumes optimistically that the system can execute all iterations of a given loop in parallel. A hardware or software monitor divides the iterations into blocks and assigns them to different threads, one per processor, with no prior dependence analysis. If the system discovers a dependence violation at runtime, it stops the incorrectly computed work and restarts it with correct values. Of course, the more parallel the loop, the more benefits this technique delivers. To better understand how speculative parallelization works, it is necessary to distinguish between private and shared variables. Informally speaking, private variables are those that the program always modifies in each iteration before using them. On the other hand, values stored in shared variables are used in different iterations.
Materias (normalizadas)
Informática
Materias Unesco
1203 Ciencia de Los Ordenadores
3304 Tecnología de Los Ordenadores
ISSN
0018-9162
Revisión por pares
SI
Version del Editor
Idioma
eng
Tipo de versión
info:eu-repo/semantics/publishedVersion
Derechos
openAccess
Collections
Files in this item