Last updated:

lxplus Guide

💡 Tip

lxplus is a shared resource. Long-running or memory-intensive processes should be submitted as batch jobs via HTCondor rather than run interactively.

lxplus is CERN's interactive login service, providing a shared Linux environment where thousands of CERN users compile code, run analyses, access data, and manage their work. It is the primary gateway to CERN's computing infrastructure and is used daily by physicists, engineers, and technical staff across the organisation. If you need to connect to lxplus for the first time, see the Kerberos and SSH setup page for connection instructions.

What Is lxplus

lxplus is a cluster of shared login nodes running AlmaLinux (CERN's standard Linux distribution). When you connect via SSH to lxplus.cern.ch, you are assigned to one of the available nodes in a round-robin fashion. Each session gives you access to a standard Linux environment with common development tools, compilers, editors, and access to CERN's storage systems (EOS and AFS). The machines are powerful but shared among many concurrent users, so resource usage per session is subject to limits.

There are several lxplus flavours available for specific needs. The default lxplus.cern.ch runs the current standard OS version. If you need a specific older or newer version for compatibility reasons, variants like lxplus9.cern.ch (AlmaLinux 9) are available. Check the CERN IT documentation for the current list of available flavours.

Available Resources

Each lxplus node typically offers multiple CPU cores and several gigabytes of RAM per user session. However, since nodes are shared, there are soft limits on per-user CPU time and memory consumption to ensure fair usage. Interactive sessions are intended for development, debugging, testing, and short analyses — not for running production workloads that consume many cores or gigabytes of memory for hours at a time.

If your task requires significant compute resources — for example, processing large datasets, running Monte Carlo simulations, or training machine learning models — you should submit it as a batch job rather than running it interactively on lxplus. The batch system distributes your job across dedicated compute nodes where it can run without competing with other interactive users.

Home Directory: AFS vs EOS

When you log into lxplus, your default home directory is on AFS (Andrew File System) at /afs/cern.ch/user/<first-initial>/<username>/. AFS has been the traditional home directory system at CERN for decades and is still used for configuration files, login scripts, and small files. However, AFS has a limited quota (typically 10 GB) and is gradually being superseded by EOS for most use cases.

Your EOS home directory at /eos/user/<first-initial>/<username>/ provides a much larger default quota (1 TB) and is the recommended location for analysis files, large datasets, and active work. Both AFS and EOS are accessible from any lxplus node, so you can easily move files between them. As a general practice, keep your AFS home clean and use EOS for your main work files.

Software Environments

CERN's software environment on lxplus is managed through CVMFS (CernVM File System), a read-only file system that distributes software packages across the network. Through CVMFS, you have access to a vast catalogue of pre-built software including ROOT, Geant4, Python distributions, and experiment-specific frameworks. The LCG releases (LHC Computing Grid software stacks) provide curated collections of compatible tools and libraries for physics analysis.

To set up a specific software environment, you typically source a setup script. For example, source /cvmfs/sft.cern.ch/lcg/views/LCG_105/x86_64-el9-gcc12-opt/setup.sh sets up a complete LCG 105 environment with ROOT, Python, and many analysis tools. Experiment-specific setups (such as ATLAS's Athena or CMS's CMSSW) have their own initialisation procedures documented by each collaboration.

Batch Jobs with HTCondor

For compute-intensive tasks, CERN provides a batch computing system based on HTCondor. From lxplus, you can submit jobs that run on a large pool of dedicated worker nodes without impacting interactive users. A basic HTCondor submission involves writing a short submit file that specifies your executable, input files, resource requirements (CPU, memory, disk), and output locations.

A minimal submit file might look like: define the executable (your script), specify the output, error, and log file paths, request the desired number of CPUs and memory, and call condor_submit to queue the job. You can then monitor its progress with condor_q and check completed job details with condor_history. For large-scale workflows involving hundreds or thousands of jobs, tools like the CERN HTCondor DAGMan and workflow managers streamline the process.

Resource Limits and Fair-Share

CERN's computing resources operate on a fair-share scheduling model. Each user and experiment has a share of the total resources, and the batch system prioritises jobs based on recent usage — if you have used a large amount recently, your new jobs may have lower priority until your share rebalances. This ensures that no single user or group monopolises the system.

On lxplus itself, interactive resource limits are enforced to protect the shared environment. Processes that consume excessive CPU or memory may be automatically terminated. If you need dedicated resources for a specific project or time period, CERN IT offers virtual machines and OpenStack instances that provide dedicated compute capacity under your control — contact the IT department through ServiceNow to discuss options.

Sources

Review and maintenance

Maintained by Efrén Rodríguez Rodríguez. Content and official sources are reviewed quarterly.

  • Verify legal deadlines and authority links
  • Recheck transport timetables and stop names
  • Revalidate tax thresholds and rates
  • Confirm CERN service URLs and portal names