35
© Copyright IBM Corporation 2015. Technical University/Symposia materials may not be reproduced in whole or in part without the prior written permission of IBM. pPL3090 Architecting Virtualized Environments: A Business Perspective Charlie Cler Enterprise Architect Forsythe Technology [email protected]

Architecting with power vm

Embed Size (px)

Citation preview

© Copyright IBM Corporation 2015. Technical University/Symposia materials may not be reproduced in whole or in part without the prior written permission of IBM.

pPL3090

Architecting Virtualized Environments: A Business Perspective

Charlie ClerEnterprise ArchitectForsythe [email protected]

2

Agenda

§ Introduction

§ Self assessment

§ PowerVM virtualization toolkit

§ Virtualized architectures and examples

3

Hardware

Operating System

Middleware

Applications

PowerVM – Hardware Virtualization

AIX V6.1 Workload Partitions

WebSphere Application Server

Application specific instances

Server Examples

Virtualization can occur in many places

4

Things to consider when implementing virtualization

Simplification Need to figure out how to do more with reduced staff … AUTOMATION!!!

Reduce Cost Improved utilization, fewer serversFacilitates consolidationReduce energy costsImproved data center managementReduce software licensingReduce hardware & software maintenance

Improve Service Agility, FasterReduce/eliminate outages: Scheduled Maintenance / Resource BalancingImprove productivity of staff, properly deployed with structure, non-structured can lead to

worse service Use of standards simplify the training/skills requiredImprove service levelsFocus is on providing service to end-user, not just installing their serverEasier to EOL of assets, re-host via Live Partition Mobility

Manage Risk Install once, reduce physical activity in the data centerReduce weekend/late night workReduce/eliminate scheduled outagesPotentially faster DR recovery when using VIO Server System Planning Tool to rebuild systemIn-house DR between sitesReduce Complexity - Avoid one of everything purchasing

Business Challenges Operational integration: SAN, network, server, dc managementPurchasing model/budgeting: Dedicated has simpler chargeback modelLots of LPARS per server ... More difficult to plan a scheduled outageSoftware licensing mechanisms can be complexNeed to invoke SLAs - New concept, must deal with very good to average performance

Chart source: Unknown author (IBM)

5

How much does it really cost to run a server?

• Software and support costs dominate TCO

• Improving utilization can help to reduce all of these costs, not just hardware

5 year TCO distribution

Chart source: Brad McCredie (IBM)

Cost Relative Cost

Hardware 1x

Power & Facilities 0.25x

Network & Storage 0.8x

Software 2-8x

Support 5-10x

6

Availability• High availability• Hours of operation

Backup / Restore / Site Recovery• Backup / Restore• Disaster Scenario• Restore• Effort for Complete Site Recovery• SAN effort

Infrastructure Cost• Space• Power• Network Infrastructure• Storage Infrastructure• Initial Hardware Costs• Software Costs• Maintenance Costs

Additional development and implementation

• Investment for one platform –reproduction for others

Controlling and Accounting• Analyzing the systems• Cost

Operations Effort• Monitoring, Operating• Problem Determination• Server Management Tools• Integrated Server Management –

Enterprise Wide

A Range of IT Cost Factors - Frequently Not Considered

Integration– Integrated Functionality vs. Functionality to

be implemented (possibly with 3rd party tools)

– Balanced System– Integration of / into Standards

Further Availability Aspects– Planned outages– Unplanned outages– Automated Take Over– Uninterrupted Take Over (especially for DB) – Workload Management across physical

borders– Business continuity– Availability effects for other applications /

projects– End User Service– End User Productivity– Virtualization

Skills and Resources– Personnel Education– Availability of Resources

Security– Authentication / Authorization– User Administration– Data Security– Server and OS Security– RACF vs. other solutions

Deployment and Support – System Programming

• Keeping consistent OS and SW Level• Database Effort

– Middleware

• SW Maintenance• SW Distribution (across firewall)

– Application

• Technology Upgrade• System Release change without

interrupts

Operating Concept– Development of an operating procedure

– Feasibility of the developed procedure

– Automation

Resource Utilization and Performance– Mixed Workload / Batch

– Resource Sharing• shared nothing vs. shared everything

– Parallel Sysplex vs. Other Concepts

– Response Time

– Performance Management

– Peak handling / scalability

Chart source: IBM RACE Conference

7

Virtualization Assessment

From an enterprise infrastructure perspective, many customers grapple with the same questions, growing pains, business

process challenges, etc related to embracing virtualization.

The pace at which a customer adopts virtualization varies by

customer. Some elements that affect this pace are customer internal organizations, politics, business drivers, business

conditions and other contributors.

The pace of a customer’s virtualization adoption at a specific “snapshot” in time can be viewed as the customer’s unique

“Virtualization Adoption Maturity” at that time.

Adoption = 10% technology + 90% customer processes

Text source: Unknown author (IBM)

8

Rack ‘n Stack

Dedicated LPARs

Virtual Processors

Virtual I/OServer

Live Partition Mobility

PowerVC

Cloud

Centralized

Pooled

Individual

Acquisition and Virtualization SpectrumS

erve

r P

urch

ase

Mod

el

Level of Virtualization Deployed

Individual Servers are purchased for specific projects. Typically results in low utilization.

Pooled Server purchases for multiple projects are grouped together. Still allows businesses to directly associate server purchase with project dollars.

Centralized Cloud-like purchasing model. Potential for highest utilization and flexibility.

Most are here

9

Capacity on Demand (CoD) & Related Options

*Trials are also available for Active Memory Expansion (AME) and Live Partition Mobility (LPM)

Model Core

Deconfiguration

Capacity on Demand (CoD)

PermanentActivation Trial*

Elastic(On/Off) Utility

Power Enterprise

Pools

Power 710 n

Power 720 n

Power 730 n

Power 740 n

Power 750 n

Power 760 n

Power 770 n n n n n

Power 780 n n n n n

Power 795 n n n n n

10

Shared Production and Development Server

Concept: Intermix development, test, staging and production LPARs on the same servers

Why?• Improve overall utilization of resources

• Ability to shift development resources to production to satisfy peak workloads

• Improve software license utilization

Common Objections• Development systems (LPARS) might bring production down

• Re-cabling of development systems may affect production

11

Separate Production and Development Servers

DB Appl Server

Virtual Processors

4 + 2 = 6 2 + 2 + 2 = 6

License Count 6 6

LPAR 1

DB Prod

LPAR 3

Appl Server Prod

LPAR 4

Appl Server Prod

LPAR 5

Appl Server

Dev

Processor Pool

V V V V V V

LPAR 2

DB Dev

V V

Processor Pool

Hypervisor Hypervisor

Core CoreCoreCoreCoreCore

V VV V

CoreCore Core Core

Production Dev

12

Shared Production and Development Server

DB Appl Server

Virtual Processors

4 + 2 = 6 2 + 2 + 2 = 6

License Count

4* 4*

Run Dev and Prod on the same server to improve processor utilization• Development servers typically have lower levels of utilization• Shared server => Fewer processor cores required• Improve processor utilization => Reduced hardware costs• Reduced software licensing cost

LPAR 1

DB Prod

LPAR 2

DB Dev

LPAR 3

Appl Server Prod

LPAR 4

Appl Server Prod

LPAR 5

Appl Server

Dev

Hypervisor

Core Core Core CoreCoreCoreCoreCore

Sub Pool Max=4

V V V V V V V V

DB Appl Server

Virtual Processors

4 + 2 = 6 2 + 2 + 2 = 6

License Count 6 6

LPAR 1

DB Prod

LPAR 3

Appl Server Prod

LPAR 4

Appl Server Prod

LPAR 5

Appl Server

Dev

Processor Pool

V V V V V V

LPAR 2

DB Dev

V V

Processor Pool

Hypervisor Hypervisor

Core CoreCoreCoreCoreCore

V VV V

CoreCore Core Core

V VV V

Sub Pool Max=4

*Using license count based on the smaller of total virtual processors and pool size

Production Dev Production and Dev LPARs intermixed

13

Staging and Production Role Swap

ProdDB

ProdAppl

Servern

ProdAppl

Servern

ProdAppl

Servern

StageAppl

Servern+1

StageAppl

Servern+1

StageAppl

Servern+1

Design based on dedicated CPU LPARs4 4 4 4 4 4 16

Implementation: New applications released to production based on swapping roles of staging and production servers.

Tim

e =

n

Environment running code release level “n”

14

Staging and Production Role Swap

ProdDB

ProdAppl

Servern

ProdAppl

Servern

ProdAppl

Servern

StageAppl

Servern+1

StageAppl

Servern+1

StageAppl

Servern+1

ProdDB

ProdAppl

Servern+1

ProdAppl

Servern+1

ProdAppl

Servern+1

StageAppl

Servern+2

StageAppl

Servern+2

StageAppl

Servern+2

Design based on dedicated CPU LPARs4 4 4 4 4 4 16

4 4 4 4 4 4 16

24 Licensed Appl Server Cores

Implementation: New applications released to production based on swapping roles of staging and production servers.

Staging servers and productions roles are reversed with each new application release. This environment requires that all staging and application servers be the same size for both CPUs and memory.

Tim

e =

nTi

me

= n

+ 1

Environment running code release level “n+1”

Environment running code release level “n”

15

Staging and Production Swap

Shared Processor Pool

ProdDB

ProdAppl

Servern

ProdAppl

Servern

ProdAppl

Servern

Stage A

pplS

tage Appl

Stage A

ppl

.2 .2 .2 4 4 4 12

ProdDB

ProdAppl

Servern

ProdAppl

Servern

ProdAppl

Servern

StageAppl

Servern+1

StageAppl

Servern+1

StageAppl

Servern+1

ProdDB

ProdAppl

Servern+1

ProdAppl

Servern+1

ProdAppl

Servern+1

StageAppl

Servern+2

StageAppl

Servern+2

StageAppl

Servern+2

Design based on dedicated CPU LPARs Using Shared Processor Pool4 4 4 4 4 4 16

4 4 4 4 4 4 16

24 Licensed Appl Server Cores

Implementation: New applications released to production based on swapping roles of staging and production servers.

Use of the shared processor pool allows the Staging and Prod appl servers to have their core allocations match their requirements.

Tim

e =

nTi

me

= n

+ 1

Tim

e =

n

Staging servers and productions roles are reversed with each new application release. This environment requires that all staging and application servers be the same size for both CPUs and memory.

16

Staging and Production Swap

Shared Processor Pool

ProdDB

ProdAppl

Servern

ProdAppl

Servern

ProdAppl

Servern

Stage A

pplS

tage Appl

Stage A

ppl

.2 .2 .2 4 4 4 12

Shared Processor Pool

ProdDB

ProdAppl

Servern+1

ProdAppl

Servern+1

ProdAppl

Servern+1

Stage A

pplS

tage Appl

Stage A

ppl

4 4 4 .2 .2 .2 12

ProdDB

ProdAppl

Servern

ProdAppl

Servern

ProdAppl

Servern

StageAppl

Servern+1

StageAppl

Servern+1

StageAppl

Servern+1

ProdDB

ProdAppl

Servern+1

ProdAppl

Servern+1

ProdAppl

Servern+1

StageAppl

Servern+2

StageAppl

Servern+2

StageAppl

Servern+2

Design based on dedicated CPU LPARs Using Shared Processor Pool4 4 4 4 4 4 16

4 4 4 4 4 4 16

24 Licensed Appl Server Cores

Implementation: New applications released to production based on swapping roles of staging and production servers.

Use of the shared processor pool allows the Staging and Prod appl servers to change size when a new release is deployed. Use of DLPAR or AMS can reduce memory requirements, too.

13 Licensed Appl Server Cores (46% reduction)

Tim

e =

nTi

me

= n

+ 1

Tim

e =

nTi

me

= n

+ 1

Staging servers and productions roles are reversed with each new application release. This environment requires that all staging and application servers be the same size for both CPUs and memory.

17

Live Partition Mobility – Potential Uses

1. Scheduled maintenanceEvacuate a server so that you can perform scheduled maintenance activities like firmware upgrades, non hot-plug service activities, etc.

2. Workload rebalancingMove an LPAR to another server that has more CPU or memory resources required to satisfy application demands.

3. End of service life migrationMigrate all running LPARs from an older system to a new system so that the older system can be retired.

18

Live Partition Mobility: Architecture Examples

1 2 3 4Perform Service1

2

3 4

2 2

N + 1

Partial Capacity

Mixed Environments

Transition to space made available by

suspending development work

Transition to multiple servers

Transition to spare server

Prod

Prod

Dev Dev

ProdProd

Perform Service

Spare1 2 3 1 23Perform Service

Note: Implementation of VMControl Enterprise Edition (VMControl Systems Pools) can simplify all of this.

19

Live Partition Mobility: N+1

Entire server’s LPARs are moved as a group• Provides level of comfort that CPU/memory resources will be adequate for the mix of

running LPARs

How big can “N” be?• Depends on frequency of scheduled maintenance x number of servers.

Spare server• Can be used for sandbox work• Must be configured with CPUs and memory to match the largest active server. If all

servers are identical, software licensing issues are minimized.• Spare capacity can’t be used to satisfy peak workloads

Typical servers• 720, 740, S822, S824

Spare1 2 3 1 23Perform Service

20

Live Partition Mobility: Partial Capacity

Server’s LPARs are distributed among the other servers in the group• Resources (white space) are reserved on each server to support the evacuation of

the largest server in the group.

• As more servers are added, no addition reserve will be required (see server #5).

• Requires planning to figure out where to migrate LPARs to. Future automation might simplify this task

Benefits• Under normal conditions, excess CPU/memory capacity on servers can be used to

handle peak workloads

• White space can be CoD that is activated via On/Off during an active migration

Typical servers• 750, 770, 780, S824, E850, E870, E880

1 2 3 4Perform Service1

2

3 4

2 2

5 5

21

Live Partition Mobility: Mixed Environment

Mixing development and production• No spare/reserved resources are required

• Drive higher levels of server and software utilization

Server Sizing• Distribute production between the two servers. Make sure that each server

is sized so that it can handle all of production by itself.

• Development will be down during migration/service period. Requires migration event to be coordinated with development teams.

Typical servers• 770, 780, 795, E870, E880

Prod

Prod

Dev Dev

ProdProd

Perform Service

22

Live Partition Mobility: Power Enterprise Pools

Normal Operation• Active workload on both servers

• Adequate COD resources on both servers to permit all workload to run on either server.

Schedule LPM Operation• Schedule planned event with IBM

• Use Power Systems Pools to temporarily transfer processor and memory activations. Requires interaction with IBM for every LPM event

• Use of Power Enterprise Pools and new mobile activations allows you to reassign cores and memory within servers in a pool. No interaction with IBM required for reassignment – user controlled.

Servers Supported• Power Systems Pools: 780, 795

• Power Enterprise Pools: 770, 780, 795

COD CODProd

Prod ProdProd

Perform Service

23

Live Partition Mobility Trial is Available

How do I activate Trial Live Partition Mobility feature?

1) New order - You can order feature code "ELPM" in your new Power server configuration.

2) MES upgrade existing configuration by adding ELPM as a No Charge MES upgrade.

The trial period is 60 days. If an MES upgrade to PowerVM Enterprise Edition is not done at the end of the trial period, your system automatically returns to PowerVM Standard Edition.

All pre-requisites for Live Partition Mobility still applies for the Trial to work.

24

Active Memory Sharing – PowerHA 5-to-1 Cluster

Design considerations• PowerHA cluster with many-to-1 cluster backup

• Shared Memory Pool reduces memory requirement on backup server

PowerHA

Hot Standby

PowerHA

Hot Standby

PowerHA

Hot Standby

PowerHA

Hot Standby

PowerHA

Hot Standby

Hypervisor

Shared Memory Pool (80 GB)

4 GB4 GB4 GB 4 GB 4 GB

VIO #2VIO #1

2 GB 2 GB

PowerHA

Production Server

64 GB

PowerHA

Production Server

64 GB

PowerHA

Production Server

64 GB

PowerHA

Production Server

64 GB

PowerHA

Production Server

64 GB

824 Production Servers 824 Stand-by Server

Dedicated Memory

25

Active Memory Sharing – PowerHA 5-to-1 Cluster

Design considerations• Production server fails

• Workload is shifted to hot-standby server

• Required memory automatically acquired from shared memory pool

• Similar story applies for shared processor pool

PowerHA

Hot Standby

PowerHA

Hot Standby

PowerHA

Hot Standby

PowerHA

Hot Standby

PowerHA

Hot Standby

Hypervisor

Shared Memory Pool (80 GB)

4 GB4 GB4 GB 64 GB 4 GB

VIO #2VIO #1

2 GB 2 GB

PowerHA

Production Server

64 GB

PowerHA

Production Server

64 GB

PowerHA

Production Server

64 GB

PowerHA

Production Server

64 GB

PowerHA

Production Server

64 GB

824 Production Servers 824 Stand-by Server

Dedicated Memory

26

Traditional DR Options

PrimaryData Center

3rd Party Disaster Recovery Service

Hot Standby Data Center

Dual Purpose Data Center

• High potential for quick recovery

• Expensive resources mostly idle

• May not cover all required applications due to cost

• Generalized recovery servers provided by 3rd party may not match your specific server models

• Testing typically limited to a one or two per year

• Recovery hardware used for Dev/Test under normal conditions• Improved utilization of resources• Becoming more common as company mergers make

additional data centers available

1

2

3

27

• Recovery at a remote location

• Recovery hardware likely to be different from your production environment

• 1 or 2 tests per year, typically negotiated into contract, provides limited ability to verify full recovery of production systems

• Improve recovery procedure using the System Planning tool

§ Works best for fully virtualized environment (VIO based)§ Periodically create System Plan from the HMC§ During DR event, deploy System Plan at recovery center to

configure all LPARs and virtual adapters. § Final recovery customization for physical disk / network

interfaces on the VIO servers.

3rd Party Disaster Recovery Service

System Planning Tool http://www-947.ibm.com/systems/support/tools/systemplanningtool/

URLs

http://www.ibmsystemsmag.com/aix/enewsletterexclusive/24624p1.aspxUsing the System Planning tool for disaster recovery, by Jaqui Lynch

System Planning Tool on IBM Developerworks http://www.ibm.com/developerworks/aix/library/au-spt/index.html

28

Hot-Standby Data Center – DR Recovery LPARS in Standby Mode

Dal

las

–LP

AR

1D

alla

s –

LPA

R 2

Dal

las

–LP

AR

3D

alla

s –

LPA

R …

Dal

las

–LP

AR

nD

enve

r –LP

AR

1D

enve

r –LP

AR

2D

enve

r –LP

AR

3D

enve

r –LP

AR

STL

–LP

AR

nP

hoen

ix –

LPA

R 1

Pho

enix

–LP

AR

2P

hoen

ix –

LPA

R 3

Pho

enix

–LP

AR

…P

hoen

ix –

LPA

R n

Den

ver –

LPA

R n

Spare Capacity

Shared CPU, Memory, SAN, and Network Resources

DR server runs with thin recovery LPARs, always ready for a DR event from one of the production locations.

Dallas

Denver

Phoenix

ProductionData Centers

(Prod & Dev/Test)

Hot-Standby Data Center(DR Recovery LPARs)

SANNetwork

29

Hot-Standby Data Center – Active DR

Dallas

Denver

Phoenix

Dal

las

–LP

AR

1D

alla

s –

LPA

R 2

Dal

las

–LP

AR

3D

alla

s –

LPA

R …

STL

–LP

AR

nP

hoen

ix –

LPA

R 1

Pho

enix

–LP

AR

2P

hoen

ix –

LPA

R 3

Pho

enix

–LP

AR

…P

hoen

ix –

LPA

R n

Dal

las

–LP

AR

n

Den

ver –

LPA

R 1

Den

ver –

LPA

R 2

Den

ver –

LPA

R 3

Den

ver –

LPA

R …

Den

ver –

LPA

R n

Spa

re C

apac

ity

Shared CPU, Memory, SAN, and Network Resources

In the event of a disaster, LPARs automatically have access to CPU, Memory, SAN, and network resources required to support the recovered workload. No manual re-configuration is required.

SANNetwork

Production Data Centers

(Prod & Dev/Test)

Hot-Standby Data Center(DR Recovery LPARs)

Active DR for Denver

30

Den

ver –

1D

enve

r –2

Den

ver –

…D

enve

r –n

Den

ver –

Dev

Den

ver –

Dev

Den

ver –

Dev

Den

ver –

Dev

Den

ver –

Test

Den

ver –

3

Den

ver –

1D

enve

r –2

Den

ver –

…D

enve

r –n

Den

ver –

Dev

Den

ver –

Dev

Den

ver –

Dev

Den

ver –

Dev

Den

ver –

Test

Den

ver –

3

Dual Purpose Data Center – Normal Operation

Production data center servers can be smaller since they are no longer running Dev / Test.

At Dual Purpose data center, under normal operation, spare capacity is used to run Dev and Test LPARs. Thin, Live recovery LPARs continue to run.

SANNetwork

Shared CPU, Memory, SAN, and Network Resources

Pho

enix

–LP

AR

1P

hoen

ix –

LPA

R 2

Pho

enix

–LP

AR

…P

hoen

ix –

LPA

R n

Pho

enix

–D

ev

Pho

enix

–D

ev

Pho

enix

–D

ev

Pho

enix

–D

ev

Pho

enix

–Te

st

Pho

enix

–LP

AR

3

Server primarily used for Dev/test workloads

DR Recovery LPARs in standby mode

Dallas

Denver

Phoenix

ProductionData Center

(Prod & Dev)

Dallas

Denver

Phoenix

31

Den

ver –

1D

enve

r –2

Den

ver –

…D

enve

r –n

Den

ver –

Dev

Den

ver –

Dev

Den

ver –

Dev

Den

ver –

Dev

Den

ver –

Test

Den

ver –

3

Den

ver –

1D

enve

r –2

Den

ver –

…D

enve

r –n

Den

ver –

Dev

Den

ver –

Dev

Den

ver –

Dev

Den

ver –

Dev

Den

ver –

Test

Den

ver –

3

Dual Purpose Data Center – Active DR

Dallas

Denver

Phoenix

SANNetwork

ProductionData Center

(Prod & Dev)

Shared CPU, Memory, SAN, and Network Resources

Dallas

Denver

Phoenix

Dev/Test resources are reduced

Pho

enix

–D

ev

Pho

enix

–D

ev

Pho

enix

–D

ev

Pho

enix

–D

ev

Pho

enix

–te

st

Phoenix production data center has experienced a disaster.

Phoenix workload is transferred to the Dual Purpose data center where the Phoenix recovery LPARs automatically have access to required CPU, Memory, SAN, and network resources.

Phoenix development LPAR resources are reduced.

Pho

enix

–LP

AR

1

Pho

enix

–LP

AR

2

Pho

enix

–LP

AR

3

Pho

enix

–LP

AR

Pho

enix

–LP

AR

n

LPARs automatically expand to handle DR workload

32

Next steps…

33

Where is your business headed?

Is your company bringing together the application development, enterprise architecture, procurement, and facilities departments?

Which layer(s) of virtualization is your company currently using or planning to investigate?

34

Additional Information

www.ibm.com/redbooks

• PowerVM Virtualization on IBM System p Introduction and Configuration (SG24-7940)

• PowerVM Virtualization on IBM System p Managing and Monitoring (SG24-7590)

• IBM PowerVM Live Partition Mobility (SG24-7460)

• PowerVM Virtualization Active Memory Sharing (REDP-4470)

35

Trademarks

The following are trademarks of the International Business Machines Corporation in the United States, other countries, or both.

The following are trademarks or registered trademarks of other companies.

* All other products may be trademarks or registered trademarks of their respective companies.

Notes: Performance is in Internal Throughput Rate (ITR) ratio based on measurements and projections using standard IBM benchmarks in a controlled environment. The actual throughput that any user will experience will vary depending upon considerations such as the amount of multiprogramming in the user's job stream, the I/O configuration, the storage configuration, and the workload processed. Therefore, no assurance can be given that an individual user will achieve throughput improvements equivalent to the performance ratios stated here. IBM hardware products are manufactured from new parts, or new and serviceable used parts. Regardless, our warranty terms apply.All customer examples cited or described in this presentation are presented as illustrations of the manner in which some customers have used IBM products and the results they may have achieved. Actual environmental costs and performance characteristics will vary depending on individual customer configurations and conditions.This publication was produced in the United States. IBM may not offer the products, services or features discussed in this document in other countries, and the information may be subject to change without notice. Consult your local IBM business contact for information on the product or services available in your area.All statements regarding IBM's future direction and intent are subject to change or withdrawal without notice, and represent goals and objectives only.Information about non-IBM products is obtained from the manufacturers of those products or their published announcements. IBM has not tested those products and cannot confirm the performance, compatibility, or any other claims related to non-IBM products. Questions on the capabilities of non-IBM products should be addressed to the suppliers of those products.Prices subject to change without notice. Contact your IBM representative or Business Partner for the most current pricing in your geography.

Adobe, the Adobe logo, PostScript, and the PostScript logo are either registered trademarks or trademarks of Adobe Systems Incorporated in the United States, and/or other countries.Cell Broadband Engine is a trademark of Sony Computer Entertainment, Inc. in the United States, other countries, or both and is used under license therefrom. Java and all Java-based trademarks are trademarks of Sun Microsystems, Inc. in the United States, other countries, or both. Microsoft, Windows, Windows NT, and the Windows logo are trademarks of Microsoft Corporation in the United States, other countries, or both.Intel, Intel logo, Intel Inside, Intel Inside logo, Intel Centrino, Intel Centrino logo, Celeron, Intel Xeon, Intel SpeedStep, Itanium, and Pentium are trademarks or registered trademarks of Intel Corporation or its subsidiaries in the United States and other countries.UNIX is a registered trademark of The Open Group in the United States and other countries. Linux is a registered trademark of Linus Torvalds in the United States, other countries, or both. ITIL is a registered trademark, and a registered community trademark of the Office of Government Commerce, and is registered in the U.S. Patent and Trademark Office.IT Infrastructure Library is a registered trademark of the Central Computer and Telecommunications Agency, which is now part of the Office of Government Commerce.

For a complete list of IBM Trademarks, see www.ibm.com/legal/copytrade.shtml:

*, AS/400®, e business(logo)® , DBE, ESCO, eServer, FICON, IBM®, IBM (logo)®, iSeries®, MVS, OS/390®, pSeries®, RS/6000®, S/30, VM/ESA®, VSE/ESA, WebSphere®, xSeries®, z/OS®, zSeries®, z/VM®, System i, System i5, System p, System p5, System x, System z, System z9®, BladeCenter®

Not all common law marks used by IBM are listed on this page. Failure of a mark to appear does not mean that IBM does not use the mark nor does it mean that the product is not actively marketed or is not significant within its relevant market.Those trademarks followed by ® are registered trademarks of IBM in the United States; all others are trademarks or common law marks of IBM in the United States.