**3. The role of EU-IndiaGrid & EU-IndiaGrid2 projects**

As discussed in the sections above during the period of the EU-IndiaGrid project activity and right at the start of EU-IndiaGrid2 e-Infrastructures in India marked a considerable progress (Masoni, 2011). The leading responsibilities of EU-IndiaGrid Indian partners and the project bridging role between European and Indian e-Infrastructures gave to EU-IndiaGrid project the opportunity to be at the core of this development and to effectively contribute at improving cooperation between Europe and India in this domain.

In particular the activities related to the High Energy Physics applications were fully integrated in the framework of the Worldwide LHC Computing Grid Collaboration with particular concern with the ALICE and CMS experiments active in India.

290 Grid Computing – Technology and Applications, Widespread Coverage and New Horizons

participating also to the EU-IndiaGrid2 project TIFR is successfully exploiting the TEIN3

GARUDA is India's first national grid initiative bringing together academic, scientific and research communities for developing their data and compute intensive applications with guaranteed Quality of Service. The project is coordinated by Centre for Development of Advanced Computing (CDAC) under the funding of Department of Information Technology (DIT) of Government of India. GARUDA involves 36 Partner institutions with 45 research and academic centres. GARUDA ended its Proof of Concept phase in 2008 and

Afterwards GARUDA has moved successfully to the Operational phase, thereby providing sustainable production grade computing infrastructure as a service to its partners. With this endeavour, the GARUDA grid is playing a key role in accelerating the research work by inter-connecting academicians, researchers, policy makers and the masses at large. Applications of national importance are being hosted on the vast infrastructure offered by the GARUDA grid, to solve grand challenge problems of national priority as Disaster Management (DMSAR) and Bio informatics. At present GARUDA network infrastructure is provided by NKN (see section 2.1), consolidating this way the integration between NKN

Within the grid infrastructure various resources such as High Performance Computing systems (HPC) and satellite based communication systems have been committed by different centres of C-DAC and GARUDA partners. GARUDA Grid is composed of various heterogeneous computing resources such as HPC clusters and SMP systems running on AIX, Linux and Solaris. At the moment GARUDA is offering several thousand of cores. The

GARUDA is well equipped with the provisioning of a customized middleware stack that can effectively harness this diverse range of resources being available on the grid. The middleware stack is enabled with CDAC's in-house developed, efficient reservation managers that ensure high availability & reliability of the resources to the applications running on the production grid infrastructure. Basic grid middleware services are provided by GlobusToolkit –4.0.8 on the top of which Gridway metascheduler has been enabled.

As discussed in the sections above during the period of the EU-IndiaGrid project activity and right at the start of EU-IndiaGrid2 e-Infrastructures in India marked a considerable progress (Masoni, 2011). The leading responsibilities of EU-IndiaGrid Indian partners and the project bridging role between European and Indian e-Infrastructures gave to EU-IndiaGrid project the opportunity to be at the core of this development and to effectively

In particular the activities related to the High Energy Physics applications were fully integrated in the framework of the Worldwide LHC Computing Grid Collaboration with

contribute at improving cooperation between Europe and India in this domain.

particular concern with the ALICE and CMS experiments active in India.

total computational power available today on Garuda is approximately 65 Teraflops.

connectivity for LHC data transfer since December 2010.

**2.3.2 GARUDA: The national grid initiative of India** 

concluded the Foundation Phase in August 2009.

and Indian grid infrastructures and related applications.

**3. The role of EU-IndiaGrid & EU-IndiaGrid2 projects** 

Moreover a dedicated Virtual Organization (VO euindia) was made available by the project to the users from the beginning of 2006. Such a VO included several grid resources distributed across Europe and India and it was fundamental to allow users to deploy and then use the grid infrastructure for their research.

Grid resources available to the user communities via the dedicated VO comprises grid sites installed and configured mainly using gLite 3.2 middleware running on 64bit Linux Operating System. Available hardware is relatively recent and includes several multicore (4, 8,12,24 CPU-cores) Worker Nodes. A snapshot taken at end of December 2010 shows that VO members have at disposal , on a best effort basis, over about 7300 CPU-cores (1800 CPUsockets) that represents a computing power of around 20 MSI00; on the storage side, the VO can use up to 44 TB of total space.

The EU-IndiaGrid2 project installed and currently maintains all the gLite central services needed to support the operational status of the user applications belonging to the EUIndia Virtual Organization.

As discussed in the next section on the top of this standard gLite infrastructure some advanced services have been installed and configured. We stress here the importance to have our own dedicated EUIndia services that allow the project to easily experiment and configure additional services on the request of users.

The increase of usage of the EU-IndiaGrid infrastructure, combined with scientific results obtained and presented at relevant international conferences or published on journals represent a clear measure of success of the user communities activity. The project Workshops and Conferences dedicated to the different applications were an important vehicle for the dissemination of results and for fostering the adoption of grid technology toward the scientific community and not only. In addition, supporting and addressing the problem of the interoperability at the application level further contributed to promote the use of advanced grid technology, and the cooperation between different projects and Institutes. Applications and user communities behind can thus be regarded as a key to sustainability, and they can help motivating the investment in e-Infrastructures.

EU-IndiaGrid2 which started on January 2010, leveraged on the EU-IndiaGrid project achievements and the strong cooperation links established with the foremost European and Indian e-Infrastructure initiatives and then paved the way for successful sustainable cooperation across European and Indian e-Infrastructures.

EU-IndiaGrid2 is strongly integrated in the Indian e-Infrastructure scenario. Its partners take leading roles in NKN, WLCG and GARUDA and a solid and fruitful cooperation has been established between these initiatives and the EU-IndiaGrid2 project.

EU-IndiaGrid2 provided specific support to ensure exploiting the progress in connectivity favouring Euro-India cooperation in e-Science. The project supports the interoperation and interoperability between the European and the Indian grid infrastructures as well as four main application areas in the domain of Biology, Climate Change, High Energy Physics and Material Science. The main landmarks during the time life of the projects includes:


Applications Exploiting e-Infrastructures Across

done generally through the MPI paradigm.

distributed memory parallelism over capable GRID CEs.

WholeNodes: to ask for all the cores on a specified node

attributes are still to be implemented on gLite middleware.

HostNumber: to specify the total number of nodes you wish to run on

**4.1 Parallel support on GRID** 

resources. These are

it to the job requirements.

training).

**4.2 GRIDSEED training tool** 

node

Europe and India Within the EU-IndiaGrid Project 293

Many scientific applications, like for instance climate modelling simulations, require a parallel computing approach and many tasks are also of the tightly coupled type. The question of how to run in parallel on the grid is therefore of great importance and we want to address here. Nowadays multicore architectures are widely available, even on the European GRID, but they are only suitable for small and medium size jobs. Distributed memory, multi-node clusters are still the only viable tool for serious scientific computing

Our aim was thus to provide a simple, transparent and efficient mechanism to exploit MPI

As of today, the gLite middleware does not yet provide proper MPI support. gLite is now integrating the MPI-start mechanism, a set of scripts, which should make it easy to detect and use site specific MPI-related configuration. In fact, it can select the proper MPI distribution, the proper batch scheduler and it can distribute the files if there is no shared disk space. The MPI-Start scripts will also handle user's pre/post execution actions. However from the point of view of users the JDL attributes that could characterize MPI enabled CEs in job descriptor files are misleading and they describe a wrong level of abstraction. The EGEE- MPI working group proposed (more than one year ago) three new attributes added to CPUnumber in order to request explicitly MPI-type distributed

SMPGranularity: to determine how many cores you would like to use on every single

Even if it is still an open question, whether WholeNodes has priority over SMPGranularity and whether SMPGranularity has priority over CPUnumber, they could provide however a great improvement to submit parallel jobs on the gLite infrastructure. Unfortunately these

There are however some patches available to enable the WMS and the CREAM CE to recognize these new attribute. The EUIndia WMS and some computing elements CEs have been therefore patched and the patches are now available and distributed within our Virtual Organization. The new attributes allow the GRID users to submit transparently their MPI parallel jobs to the MPI capable resources and furthermore fine-tune their request matching

The GRIDSEED tool (Gregori, 2011) , a set of virtual VM machines preconfigured with several grid services, is by far the most successful training tool developed within EuindiaGRID projects. GRIDSEED is not just a software package, but a complete training tool containing a software package and all the information to use it as training platforms, targeting different user community and different kind of aspects (users vs. sys administrator


With the transition from ERNET to NKN for the network layer an interoperation problem occurred since all the nodes within the GARUDA grid became not visible to the external world. Thanks to the effort, coordinated by the EU-IndiaGrid2 project, ERNET, NIC and CDAC it was possible to solve this issue in the context of the EU-IndiaGrid2 Workshop in December in Delhi and since end 2010 all the GARUDA infrastructure is visible to worldwide grids. In addition the project supported the interoperability between the European Grid Initiative, EGI (www.eigi.eu) and the GARUDA grid infrastructure which is now possible using a metascheduler based on Gridway (Huedo 2005).

The TEIN3 link from Europe to Mumbai was commissioned in March 2010. However a number of issues related to the connectivity between the TEIN2 PoP and the WLCG Tier2 at TIFR needed to be solved. Again with the coordinated effort of NIC and EU-IndiaGrid2 partners it was possible since fall 2010 to exploit the TEIN3 links for LHC data transfers. Considering that the Academia Sinica Computing Centre acts as reference Tier1 for CMS Tier2 at TIFR both TEIN3 links (to Europe for CERN and to Singapore for Academia Sinica Tier1) are crucial for WLCG operation in India. In addition the commissioning of the 1 Gbps NKN connectivity from Kolkata to Mumbai makes the international connectivity available also for the ALICE experiment.

Finally the collaboration between Bhabha Atomic Research Centre (BARC) in Mumbai and Commissariat pour l' Energie Atomique (CEA) in Grenoble represents an excellent showcase for the usage of NKN-TEIN3-géant connectivity for remote control and data collection at the Grenoble beam facility. The BARC and CEA research groups collaborate in experiments dedicated to the study of crystallography of biological macromolecules. using protein crystallography beamlines. Two facilities have been set-up in India allowing to operate remotely the beamline FIP on ESRF, Grenoble. Good X-ray diffraction data has been collected on crystals of drug resistant HIV-1 protease enzyme. Both BARC and CEA are EU-IndiaGrid2 partners and this activity is fully supported by the EU-IndiaGrid2 project.
