Skip to main content
Log in

QCDgrid: A Grid Resource for Quantum Chromodynamics

  • Published:
Journal of Grid Computing Aims and scope Submit manuscript

Abstract

Quantum Chromodynamics (QCD) is an application area that requires access to large supercomputing resources and generates large amounts of raw data. The UK's national lattice QCD collaboration UKQCD currently stores and requires access to around five Tbytes of data, a figure that is growing dramatically as the collaboration's purpose built supercomputing system, QCDOC [P.A. Boyle, D. Chen, N.H. Christ, M. Clark, S.D. Cohen, C. Cristian, Z. Dong, A. Gara, B. Joo, C. Jung, C. Kim, L. Levkova, X. Liao, G. Liu, R.D. Mawhinney, S. Ohta, K. Petrov, T. Wettig and A. Yamaguchi, “Hardware and software status of QCDOC, arXiv: hep-lat/0309096”, Nuclear Physics. B, Proceedings Supplement, Vol. 838, pp. 129–130, 2004. See: http://www.ph.ed.ac.uk/ukqcd/community/qcdoc/; P.A. Boyle, D. Chen, N.H. Christ, M.A. Clark, S.D. Cohen, C. Cristian, Z. Dong, A. Gara, B. Joo, C. Jung, C. Kim, L.A. Levkova, X. Liao, R.D. Mawhinney, S. Ohta, K. Petrov, T. Wettig and A. Yamaguchi, “Overview of the QCDSP and QCDOC computers”, IBM Journal of Research and Development, Vol. 49, No. 2/3, p. 351, 2005] came into full production service towards the end of 2004. This data is stored on QCDgrid, a data Grid currently composed of seven storage elements at five separate UK sites.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. D. Britton, P. Clarke, J. Coles, D. Colling, A. Doyle, S.M. Fisher, A.C. Irving, J. Jensen, A. McNab and D. Newbold, “A Grid for Particle Physics – From testbed to production”, UK e-Science All Hands Conference, Nottingham, September 2004. See: http://www.gridpp.ac.uk.

  2. Particle Physics and Astronomy Research Council. See http://www.pparc.ac.uk/.

  3. W. Hoschek, J. Jaen-Martinez, A. Samar, H. Stockinger and K. Stockinger, “Data Management in an International Data Grid Project”, IEEE/ACM International Workshop on Grid Computing Grid' 2000, 17–20 December 2000. See: http://www.eu-datagrid.org/.

  4. I. Foster and C. Kesselman, “Globus: A Metacomputing Infrastructure Toolkit”, International Journal of Supercomputer Applications, Vol. 11, No. 2, pp. 115–128, 1997. See: http://www.globus.org/.

    Article  Google Scholar 

  5. The Globus Alliance. See http://www.globus.org/alliance/.

  6. Enabling Grids for E-Science in Europe. See http://www.cern.ch/egee.

  7. LHC Compute Grid project. See http://lcg.web.cern.ch/LCG/.

  8. The Virtual Data Toolkit. See http://www.cs.wisc.edu/vdt/index.html.

  9. ALIce ENvironment. See: http://alien.cern.ch/.

  10. Large Hadron Collider, A Large Particle Accelerator due to Open at CERN in 2007. See http://www.cern.ch/.

  11. The Particle Physics Data Grid Collaboratory Pilot. See: http://www.ppdg.net/.

  12. International Virtual Data Grid Laboratory. See: http://www.ivdgl.org/.

  13. Grid Physics Network. See: http://www.griphyn.org/.

  14. T. Tannenbaum, D. Wright, K. Miller and M. Livny, Condor – A Distributed Job Scheduler, Beowulf Cluster Comput-ing with Linux, The MIT Press, 2002. ISBN: 0-262-69274-0.

  15. A Toroidal LHC Apparatus. See: http://www.usatlas.bnl.gov/.

  16. The CompactMuon Solenoid. See: http://www.uscms.org/SoftwareComputing/Grid/.

  17. I. Foster et al., An Application Grid Laboratory for Science, the Grid’2003 Production Grid: Principles and Practice, Thirteenth IEEE International Symposium on High-Performance Distributed Computing (HPDC13), Honolulu, 2004.

  18. Open Science Grid. See: http://www.opensciencegrid.org/index.html.

  19. CrossGrid. See: http://www.gridstart.org/CROSSGRID.shtml.

  20. ILDG Collaboration. See http://www.lqcd.org/ildg, A.C. Irving et al., “arXiv: hep-lat/0309029”, Nuclear Physics. B, Proceedings Supplement, Vol. 159, pp. 129–130, 2004.

  21. Jefferson Lab Storage Resource Manager. See: http://cc.jlab.org/scicomp/srm/.

  22. W. Meier, eXist: An Open Source Native XML Database, Web, Web- Services, and Database Systems. NODe 2002 Web- and Database-Related Workshops, Erfurt, Germany, October 2002. Springer LNCS Series, 2593. See http://exist.sourceforge.net/.

  23. NeSCForge. See http://forge.nesc.ac.uk/projects/qcdgrid.

  24. GridFTP: Universal Data Transfer for the Grid, Globus White Paper. See http://www-fp.globus.org/datagrid/gridftp.html.

  25. Globus Replica Location Service. See http://wwwfp. globus.org/datagrid/replica-management.html.

  26. Open Grid Services Architecture Data Access and Integration project. See http://www.ogsa-dai.org.uk/.

  27. The Metadata Catalog Service. See: http://www.isi.edu/ deelman/MCS/.

  28. The Storage Resource Broker See: http://www.arl.wustl.edu/ kenw/npaci/sms8.html.

  29. ENACTS project. See http://www.enacts.org/.

  30. The UK eScience Certificate Authority. See: https://ca.grid-support.ac.uk.

  31. The NorGrid Certificate Authority. See: http://www.norgrid.no/.

  32. The Grid-Ireland Certificate Authority. See: http://cagraidsvr06.cs.tcd.ie/.

  33. D.J. Byrne, Work Package 1.2: QCDgrid Stress Testing Requirements, November 2004. See: http://www.gridpp.ac.uk/qcdgrid/.

  34. R.L. Rivest, The MD5 Message Digest Algorithm, Internet RFC 1321 (April 1992).

  35. Lightweight Middleware for Grid Computing. See: http://glite.web.cern.ch/glite/.

  36. G. Andronico, R. Barbera and A. Falzone, Nuclear Instruments and Methods in Physics Research, Vol. A534, pp. 76–79, 2004.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to L. Smith.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Perry, J., Smith, L., Jackson, A.N. et al. QCDgrid: A Grid Resource for Quantum Chromodynamics. J Grid Computing 3, 113–130 (2005). https://doi.org/10.1007/s10723-005-9005-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10723-005-9005-5

Key words

Navigation