阿根廷vs墨西哥竞猜
 library logo
    • login
    view item 
    •   knowledge commons home
    • electronic theses and dissertations
    • electronic theses and dissertations from 2009
    • view item
    •   knowledge commons home
    • electronic theses and dissertations
    • electronic theses and dissertations from 2009
    • view item
    javascript is disabled for your browser. some features of this site may not work without it.
    quick search

    browse

    all of knowledge commonscommunities & collectionsby issue dateauthorstitlessubjectsdisciplineadvisorcommittee memberthis collectionby issue dateauthorstitlessubjectsdisciplineadvisorcommittee member

    my account

    login

    power-aware caches for gpgpus

    thumbnail
    view/open
    saghira2015m-1a.pdf (1.342mb)
    date
    2015
    author
    saghir, ahsan
    metadata
    show full item record
    abstract
    in this thesis, we propose two optimization techniques to reduce power consumption in l1 caches (data, texture and constant), shared memory and l2 cache. the first optimization technique targets static power. evaluation of gpgpu applications shows that once a cache block is accessed by a thread, it takes several hundreds of clock cycles until the same block is accessed again. the long inter-access cycle can be used to put cache cells into drowsy mode and reduce static power. while drowsy cells reduce static power, they increase access time as voltage of a cache cell in drowsy mode should be raised before the block can be accessed. to mitigate performance impact of drowsy cells, we propose a novel technique called coarse grained drowsy mode. in coarse grained drowsy mode, we partition each cache into regions of consecutive cache blocks and wake up a region upon cache access. due to temporal and spatial locality of cache accesses, this method dramatically reduces performance impact caused by drowsy cells. the second optimization technique relies on branch divergence in gpgpus. the execution model in gpgpus is single instruction multiple thread (simt) which means processing cores execute the same instruction with different data for gpgpu threads. the simt execution model may result in divergence of threads when a control instruction is executed. gpgpus execute branch instructions in two phases. in the first phase, threads in the taken path are active and the rest are idle. in the second phase, threads in the not-taken path are executed and the rest are idle. contemporary gpgpus access all portions of cache blocks, although some threads are idle due to branch divergence. we propose accessing only portions of cache blocks corresponding to active threads. by disabling unnecessary sections of cache blocks, we are able to reduce dynamic power of caches. our results show that on average, the two optimization techniques together reduce power of caches by up to 98% and 15% for static and dynamic power, respectively.
    uri
    http://knowledgecommons.lakeheadu.ca/handle/2453/711
    collections
    • electronic theses and dissertations from 2009 [1612]

    阿根廷vs墨西哥竞猜 library
    contact us | send feedback

     

     


    阿根廷vs墨西哥竞猜 library
    contact us | send feedback