iTnews
  • Home
  • News
  • Technology
  • Hardware

Q&A: Large Hadron Collider's IT environment revealed

By Martin Courtney on Jan 18, 2010 6:31AM
Q&A: Large Hadron Collider's IT environment revealed

How CERN is processing data in a virtual environment.

The European Organisation for Nuclear Research (CERN), the scientific research facility using the Large Hadron Collider (LHC) to explain the mysteries of the universe, is testing a virtualised server environment spanning multiple shared data centres that could eventually see all of the facility’s grid based number crunching applications run on virtual, rather than physical, machines.

CERN group leader of fabric infrastructure and operations, Tony Cass, tells Computing how this presents a significant management and security challenge, moreover when it comes to convincing 10,000 exacting physicists that virtual servers can run their batch processing jobs properly.

Computing: What is CERN doing with server virtualisation?

Cass: We are working on a virtualisation model for a batch computing environment as a test at the moment, involving 20 machines offered to one of our existing customers, but we plan to roll out virtualisation much more widely next year. We want to convince the researchers accessing our systems that their [batch] jobs will carry on running much as they did in the past. We have 5,000 physical systems each with 16 cores, which can potentially support 80,000 virtual machines (VMs). But we need something to help us manage and understand what is going on in that virtual environment.

Computing: What runs on those CERN servers?

Cass: It is a batch environment consisting of two types: data from the LHC experiments in the form of ones and zeros, and raw images like those from a digital camera. Around 10,000 physicists around the world process and analyse snapshots of this using the grid, submitting jobs to process the data that is held on a mass disk subsystem at CERN. The specific thing that is relevant is how we manage those jobs, routing them to the compute capacity we have.

We also need to provide enough processing capacity to support the production and analysis of more than 15 petabytes of data per year, including all of the data we collect from the LHC experiments, which we expect to add to year on year.

Computing: Do any of the researchers using CERN resources worry that virtualising systems will mean a performance drop off?

Cass: They have expressed concerns, but people are always nervous about that sort of thing if there is any chance of disruption. The LHC started last year, ran for ten days, then stopped and has only just started up again, and people do not want to think anything can stop their data processing. The CPU and network access provides no performance penalty, the biggest challenge is accessing data on local disk servers, because there is input/output (I/0) all over the network.

Computing: What are the management challenges posed by virtualising servers on this scale?

Cass: We have to make sure all of those VMs are up to date with security patches and everything, and provide a policy driven approach to guarantee a minimum level of service for some customers, whilst making sure we can expand [compute] capacity for others as necessary. In any server environment that uses live migration with a hypervisor, you have to know where all the VMs are, where the physical hardware is, and what will be affected.

Computing: How is CERN addressing this management challenge?

Cass: We are using Platform Computing’s ISF adaptive cluster and LSF grid workload management software to manage both virtual and physical machines, both those short lived VMs used for batch processing and long life machines used for classical server processes running for many days or months.

Back in the mid 1990s CERN was one of the first to support running tens of thousands of jobs on Linux, but now that is mainstream. ISF offers a pretty comprehensive range of software but there are other things we might need, like Open Nebula, an open source toolkit for cloud computing, which works in a virtual management machine. The attraction here is that it will inter-operate with other packages as necessary, because we have to have the whole range rather than a single hypervisor management system from one company.

Got a news tip for our journalists? Share it with us anonymously here.
itweek.co.uk @ 2010 Incisive Media
Tags:
andcasscernfabricgrouphardwareinfrastructureleaderofoperationsqasoftwaretony

Partner Content

The Great Resignation has intensified insider security threats
Promoted Content The Great Resignation has intensified insider security threats
How to turn digital complexity into competitive advantage
Promoted Content How to turn digital complexity into competitive advantage
Security: Understanding the fundamentals of governance, risk & compliance
Promoted Content Security: Understanding the fundamentals of governance, risk & compliance
Avoiding CAPEX by making on-premise IT more cloud-like
Promoted Content Avoiding CAPEX by making on-premise IT more cloud-like

Sponsored Whitepapers

Extracting the value of data using Unified Observability
Extracting the value of data using Unified Observability
Planning before the breach: You can’t protect what you can’t see
Planning before the breach: You can’t protect what you can’t see
Beyond FTP: Securing and Managing File Transfers
Beyond FTP: Securing and Managing File Transfers
NextGen Security Operations: A Roadmap for the Future
NextGen Security Operations: A Roadmap for the Future
Video: Watch Juniper talk about its Aston Martin partnership
Video: Watch Juniper talk about its Aston Martin partnership

Events

  • Micro Focus Information Management & Governance (IM&G) Forum 2022
  • CRN Channel Meets: CyberSecurity Live Event
  • IoT Insights: Secure By Design for manufacturing
  • Cyber Security for Government Summit
  • Forrester Technology & Innovation Asia Pacific 2022
By Martin Courtney
Jan 18 2010
6:31AM
0 Comments

Related Articles

  • South32 to pursue new IT operating model
  • Collins Foods puts IT focus on security controls, cloud services
  • Qantas calls time on IBM, Fujitsu in tech modernisation
  • Sydney Trains looks to IoT to cut back on power use
Share on Twitter Share on Facebook Share on LinkedIn Share on Whatsapp Email A Friend

Most Read Articles

Qantas calls time on IBM, Fujitsu in tech modernisation

Qantas calls time on IBM, Fujitsu in tech modernisation

PayTo rollout kicks off

PayTo rollout kicks off

Researchers hacked Oracle servers to demo serious vulnerability

Researchers hacked Oracle servers to demo serious vulnerability

Neobank Volt exits the banking industry

Neobank Volt exits the banking industry

Digital Nation

Integrity, ethics and board decisions in the digital age
Integrity, ethics and board decisions in the digital age
IBM global chief data officer on the rise of the number crunchers
IBM global chief data officer on the rise of the number crunchers
The security threat of quantum computing
The security threat of quantum computing
COVER STORY: Operationalising net zero through the power of IoT
COVER STORY: Operationalising net zero through the power of IoT
Crypto experts optimistic about future of Bitcoin: Block
Crypto experts optimistic about future of Bitcoin: Block
All rights reserved. This material may not be published, broadcast, rewritten or redistributed in any form without prior authorisation.
Your use of this website constitutes acceptance of nextmedia's Privacy Policy and Terms & Conditions.