Suchergebnisse - Shared-memory parallel computing

  1. 1

    Computational performance of a smoothed particle hydrodynamics simulation for shared-memory parallel computing von Nishiura, Daisuke, Furuichi, Mikito, Sakaguchi, Hide

    ISSN: 0010-4655, 1879-2944
    Veröffentlicht: Elsevier B.V 01.09.2015
    Veröffentlicht in Computer physics communications (01.09.2015)
    “… The computational performance of a smoothed particle hydrodynamics (SPH) simulation is investigated for three types of current shared-memory parallel computer devices …”
    Volltext
    Journal Article
  2. 2

    On Pervasive Shared Memory Parallel Computing von Rattan, Ishwar

    ISSN: 0026-2005, 2167-8634
    Veröffentlicht: Alma Michigan Academy of Science, Arts & Letters 22.06.2021
    Veröffentlicht in Michigan academician (22.06.2021)
    “… Nowadays one can find parallel computing (PC) capability in laptops, desktops, smart phones and embedded devices in, use today …”
    Volltext
    Journal Article
  3. 3

    Adaptive parallel applications: from shared memory architectures to fog computing (2002–2022) von Galante, Guilherme, da Rosa Righi, Rodrigo

    ISSN: 1386-7857, 1573-7543
    Veröffentlicht: New York Springer US 01.12.2022
    Veröffentlicht in Cluster computing (01.12.2022)
    “… The evolution of parallel architectures points to dynamic environments where the number of available resources or configurations may vary during the execution of applications …”
    Volltext
    Journal Article
  4. 4

    Max-PIM: Fast and Efficient Max/Min Searching in DRAM von Zhang, Fan, Angizi, Shaahin, Fan, Deliang

    Veröffentlicht: IEEE 05.12.2021
    “… Recently, in-DRAM computing is becoming one promising technique to address the notorious 'memory-wall' issue for big data processing …”
    Volltext
    Tagungsbericht
  5. 5

    Chasing Away RAts: Semantics and evaluation for relaxed atomics on heterogeneous systems von Sinclair, Matthew D., Alsop, Johnathan, Adve, Sarita V.

    Veröffentlicht: ACM 01.06.2017
    “… An unambiguous and easy-to-understand memory consistency model is crucial for ensuring correct synchronization and guiding future design of heterogeneous …”
    Volltext
    Tagungsbericht
  6. 6

    Reverse-Mode Automatic Differentiation and Optimization of GPU Kernels via Enzyme von Moses, William S., Churavy, Valentin, Paehler, Ludger, Huckelheim, Jan, Narayanan, Sri Hari Krishna, Schanen, Michel, Doerfert, Johannes

    ISSN: 2167-4337
    Veröffentlicht: ACM 14.11.2021
    “… Computing derivatives is key to many algorithms in scientific computing and machine learning such as optimization, uncertainty quantification, and stability analysis …”
    Volltext
    Tagungsbericht
  7. 7

    SLATE: Design of a Modern Distributed and Accelerated Linear Algebra Library von Gates, Mark, Kurzak, Jakub, Charara, Ali, YarKhan, Asim, Dongarra, Jack

    ISSN: 2167-4337
    Veröffentlicht: ACM 17.11.2019
    “… SLATE will provide coverage of existing ScaLAPACK functionality, including the parallel BLAS …”
    Volltext
    Tagungsbericht
  8. 8

    A Parallel Computing Method for the Computation of the Moore-Penrose Generalized Inverse for Shared-Memory Architectures von Gelvez-Almeida, Elkin, Barrientos, Ricardo J., Vilches-Ponce, Karina, Mora, Marco

    ISSN: 2169-3536, 2169-3536
    Veröffentlicht: Piscataway IEEE 01.01.2023
    Veröffentlicht in IEEE access (01.01.2023)
    “… In this paper, we propose a parallel computing method for the computation of the Moore-Penrose generalized inverse of large-size full-rank rectangular matrices …”
    Volltext
    Journal Article
  9. 9

    Domain decomposition with discrete element simulations using shared-memory parallel computing for railways applications von Hoang, T.M.P., Saussine, G., Dureisseix, D., Alart, P.

    ISSN: 1779-7179, 1958-5829
    Veröffentlicht: Routledge 01.01.2012
    Veröffentlicht in European journal of computational mechanics (01.01.2012)
    “… To reduce computational costs, we study the use of two strategies: domain decomposition methods and shared-memory parallelisation with OpenMP …”
    Volltext
    Journal Article
  10. 10

    DASP: Specific Dense Matrix Multiply-Accumulate Units Accelerated General Sparse Matrix-Vector Multiplication von Lu, Yuechen, Liu, Weifeng

    ISSN: 2167-4337
    Veröffentlicht: ACM 11.11.2023
    “… Sparse matrix-vector multiplication (SpMV) plays a key role in computational science and engineering, graph processing, and machine learning applications. Much …”
    Volltext
    Tagungsbericht
  11. 11

    Bringing Order to Sparsity: A Sparse Matrix Reordering Study on Multicore CPUs von Trotter, James D., Ekmekcibasi, Sinan, Langguth, Johannes, Torun, Tugba, Duzakin, Emre, Ilic, Aleksandar, Unat, Didem

    ISSN: 2167-4337
    Veröffentlicht: ACM 11.11.2023
    “… Many real-world computations involve sparse data structures in the form of sparse matrices. A common strategy for optimizing sparse matrix operations is to …”
    Volltext
    Tagungsbericht
  12. 12

    Understanding Priority-Based Scheduling of Graph Algorithms on a Shared-Memory Platform von Yesil, Serif, Heidarshenas, Azin, Morrison, Adam, Torrellas, Josep

    ISSN: 2167-4337
    Veröffentlicht: ACM 17.11.2019
    “… It performs a detailed empirical performance analysis of several advanced CPS designs in a state-of-the-art graph analytics framework running on a large shared-memory server …”
    Volltext
    Tagungsbericht
  13. 13
  14. 14

    Almost Deterministic Work Stealing von Shiina, Shumpei, Taura, Kenjiro

    ISSN: 2167-4337
    Veröffentlicht: ACM 17.11.2019
    “… With task parallel models, programmers can easily parallelize divide-and-conquer algorithms by using nested fork-join structures …”
    Volltext
    Tagungsbericht
  15. 15

    NestedMP: Enabling cache-aware thread mapping for nested parallel shared memory applications von He, Jiangzhou, Chen, Wenguang, Tang, Zhizhong

    ISSN: 0167-8191, 1872-7336
    Veröffentlicht: Elsevier B.V 01.01.2016
    Veröffentlicht in Parallel computing (01.01.2016)
    “… •Task-core mapping schemas for nested-parallel applications may affect performance …”
    Volltext
    Journal Article
  16. 16

    Novel parallel method for association rule mining on multi-core shared memory systems von Vu, Lan, Alaghband, Gita

    ISSN: 0167-8191, 1872-7336
    Veröffentlicht: Elsevier B.V 01.12.2014
    Veröffentlicht in Parallel computing (01.12.2014)
    “… •ShaFEM: a novel association rule mining method for multi-core shared memory systems …”
    Volltext
    Journal Article
  17. 17

    STM-Multifrontal QR: Streaming Task Mapping Multifrontal QR Factorization Empowered by GCN von Lin, Shengle, Yang, Wangdong, Wang, Haotian, Tsai, Qinyun, Li, Kenli

    ISSN: 2167-4337
    Veröffentlicht: ACM 14.11.2021
    “… % higher than that of prior work. Moreover, for numerical factorization, an optimized tasks stream parallel processing strategy is proposed and a more efficient computing task mapping framework for NUMA architecture is adopted …”
    Volltext
    Tagungsbericht
  18. 18

    TSHMEM: Shared-Memory Parallel Computing on Tilera Many-Core Processors von Lam, Bryant C., George, Alan D., Lam, Herman

    Veröffentlicht: IEEE 01.05.2013
    “… Fortunately, a partitioned global address space (PGAS) programming model has demonstrated realizable performance and productivity potential for large parallel computing systems with distributed-memory architectures …”
    Volltext
    Tagungsbericht
  19. 19

    Implementation of QR and LQ decompositions on shared memory parallel computing systems von Egunov, V., Andreev, A.

    Veröffentlicht: IEEE 2016
    “… The paper presents some results of research of applicability of various computing schemes of implementing QR and LQ matrix decompositions for shared memory parallel systems …”
    Volltext
    Tagungsbericht
  20. 20

    Concurrent computation of topological watershed on shared memory parallel machines von Mahmoudi, Ramzi, Akil, Mohamed, Bedoui, Mohamed Hédi

    ISSN: 0167-8191, 1872-7336
    Veröffentlicht: Elsevier B.V 01.11.2017
    Veröffentlicht in Parallel computing (01.11.2017)
    “… •We set up an adequate parallelization approach that guides a parallel watershed computing …”
    Volltext
    Journal Article