Yonggang Che, Chuanfu Xu, Pingjing Lu, and Zhenghua Wang


  1. [1] K. Skadron, M. Martonosi, D.I. August, M.D. Hill et al., Chal-lenges in computer architecture evaluation, IEEE Computer,36(8), August 2003, 30–36.
  2. [2] H. Lee, L. Jin, K. Lee, S. Demetriades et al., Two-phasetrace-driven simulation (TPTS): A fast multicore processorarchitecture simulation approach, Software – Practice andExperience, 40, 2010, 239–258.
  3. [3] J.E. Miller, H. Kasture, G. Kurian, C. Gruenwald III et al.,Graphite: A distributed parallel simulator for multicores, Proc.16th International Symposium on High-Performance ComputerArchitecture, Bangalore, India, January 9–14, 2010.
  4. [4] G. Zheng, G. Kakulapati, and L.V. Kal´e, Bigsim: A parallelsimulator for performance prediction of extremely large parallelmachines, Proc. 18th International Parallel and DistributedProcessing Symposium, Santa Fe, New Mexico, April 2004.
  5. [5] G. Zheng and R. Mokos, BigSim tutorial, Charm++ Workshop2009, University of Illinois at Urbana-Champaign, Urbana,Illinois, USA.
  6. [6] M. Hermanns and M. Geimer, Verifying causal connections be-tween distant performance Phenomena in large-scale message-passing applications, Technical Report FZJ-JSC-IB-2008-05,J¨ulich Supercomputing Centre.
  7. [7] R. Susukita, H. Ando, and M. Aoyagi, Performance predictionof large-scale parallel system and application using macro-levelsimulation, Proc. SC2008, Austin, Texas, USA, November2008.
  8. [8] W.E. Denzel, J. Li, P. Walker, and Y. Jin, A framework forend-to-end simulation of high performance computing systems,Proc. SIMUTools’08, Marseille, France, March 3–7, 2008.
  9. [9] K. Mohror and K.L. Karavanic, Towards scalable event trac-ing for high-end systems, Proc. 3rd International Conf. onHigh Performance Computing and Communications, Houston,Texas, 2007, 695–706.
  10. [10] K. Mohror and K.L. Karavanic, Evaluating similarity-basedtrace reduction techniques for scalable performance analysis,Proc. SC09, Portland, Oregon, November 14–20, 2009.
  11. [11] M. Noeth, P. Ratn, F. Mueller, M. Schulz et al., ScalaTrace:Scalable compression and replay of communication traces forhigh performance computing, Journal of Parallel and Dis-tributed Computing, 69(8), 2009, 696–710.
  12. [12] D. Bhardwaj, Example 26 (gauss_elimination.c).http://www.cse.iitd.ernet.in/dheerajb/MPI/codes/day-3/c/gauss_elimination.c.
  13. [13] F.B. Hanson, Laplace Equation Numerically Solved by JacobiIteration Method on PSC T3E. http://www.math.uic.edu/hanson/mcs572/lap4mpi.c.
  14. [14] Antoine Petitet, Clint Whaley, Jack Dongarra, Andy Cleary.HPL 2.0 – A Portable Implementation of the High-PerformanceLinpack Benchmark for Distributed-Memory Computers.http://www.netlib.org/benchmark/hpl/hpl-2.0.tar.gz.
  15. [15] R. Clint Whaley, Antoine Petitet and Jack J. Dongarra.ATLAS source code distribution (stable, version 3.6.0).http://www.netlib.org/atlas/atlas3.6.0.tgz.
  16. [16] K. Mohror and K.L. Karavanic, An investigation of tracingoverheads on high end systems, Technical Report TR-06-06,Department of Computer Science, Portland State University.
  17. [17] J. Vetter, Dynamic statistical profiling of communication ac-tivity in distributed applications, Proc. International Conf. onMeasurement and Modeling of Computer Systems, CA, 2002.
  18. [18] T. Mytkowicz, A. Diwan, M. Hauswirth, and P.F. Sweeney,Understanding measurement perturbation in trace-based data,Proc. NSF Next Generation Software Program Workshop ’07,Delft, Netherlands, March, 2007.
  19. [19] M.M. Tikir, M.A. Laurenzano, L. Carrington, and A. Snavely,PSINS: An open source event tracer and execution simulatorfor MPI applications, Proc. Euro-Par 2009, Long Beach,California, USA, 135–148.
  20. [20] P. Ratn, F. Mueller, B.R. de Supinski, and M. Schulz, Pre-serving time in large-scale communication traces, Proc. ICS’08, Island of Kos, Aegean Sea, Greece, June 2008, 46–55.
  21. [21] P.H. Carns, W.B. Ligon III, R.B. Ross, and R. Thakur, PVFS:A Parallel File System For Linux Clusters. Proc. 4th AnnualLinux Showcase and Conference, Atlanta, GA, October 2000,317–327.39

Important Links:

Go Back