Muokkaa

Jaa


Deploy IBM Maximo Application Suite on Azure

Azure Files
Azure Load Balancer
Azure Red Hat OpenShift
Azure Virtual Machines
Azure Virtual Network

IBM Maximo Application Suite (MAS) 8.x and up runs on OpenShift, and it's beneficial to familiarize yourself with OpenShift and the suggested patterns for installation on Azure. For more information, see Preparing to install on Azure. This architecture illustrates an OpenShift cluster. It doesn't go into detail on how to install MAS. To learn more about the installation process, see Installing Maximo Application Suite.

Architecture

Architecture diagram that shows the components and services that support deployment of IBM Maximo Application Suite on Azure.

Download a Visio file of this architecture.

The workload can be deployed internally or externally facing, depending on your requirements.

Workflow

From the perspective of infrastructure, this architecture provides the following:

  • A container hosting platform to deploy highly available workloads across availability zones
  • A privatized deployment of worker and control nodes that are integrated with storage
  • Azure Files premium and standard files for storage (OpenShift Data Foundation not required)
  • SQL Server on Azure VMs or container-based IBM Db2 Warehouse
  • Azure DNS for DNS management of OpenShift and its containers
  • Microsoft Entra ID for single sign-on (SSO) into MAS

Components

  • Azure Virtual Machines to host the OpenShift platform and run the Maximo containers. Virtual Machines is an infrastructure-as-a-service (IaaS) offering. You can use Virtual Machines to deploy on-demand, scalable computing resources.

  • Red Hat Enterprise Linux CoreOS to provide a custom VM image for OpenShift.

  • Azure Load Balancers to provide connectivity into the cluster. Azure Load Balancer is a high-performance, ultra low-latency Layer 4 load-balancing service (inbound and outbound) for all UDP and TCP protocols. It's built to handle millions of requests per second while ensuring your solution is highly available. Azure Load Balancer is zone-redundant, ensuring high availability across Availability Zones.

  • Virtual Network for communication between nodes, Azure services, and hybrid connectivity needs. Virtual Network is the fundamental building block for private networks in Azure.

  • Azure Files to host the stateful data for the databases and systems inside the cluster. Azure Files provides fully managed file shares in the cloud that are accessible via the SMB and Network File System (NFS) protocols.

  • Azure DNS to manage DNS resolution for the containers inside and outside of the solution. Azure DNS supports all common DNS records and provides high availability.

  • Azure Bastion (optional) and a subnet for enhanced-security access to any of the worker nodes or optional JumpBox machines. Azure Bastion is a fully managed service that provides seamless enhanced-security RDP and SSH access to VMs without any exposure through public IP addresses.

  • SQL Server on Azure Virtual Machines (optional) to provide data services to MAS. The database can also be another, like Oracle Exadata or IBM Db2 Warehouse. Azure SQL Database and Azure SQL Managed Instance aren't supported right now.

  • Twilio Send Grid (optional) to send emails from MAS to your consumers.

  • Linux virtual machines in Azure (optional) to provide a jump box for installation of OpenShift. You can also use this VM to connect and manage the OpenShift cluster because it contains the Kubernetes configuration file after installation. If you have network connectivity into your Azure environment, you can perform the installation from an existing machine.

Alternatives

The following services typically aren't necessary, but they're effective alternatives:

Scenario details

IBM's Maximo Application Suite (MAS), also known as Maximo, is an enterprise asset management platform with AI-based asset maintenance. MAS focuses on operational resiliency and reliability. The suite consists of a core application platform, MAS, and applications and industry-specific solutions on top of the platform. Each application provides a specific benefit:

  • Manage. Reduce down time and costs by using asset management to improve operational performance.
  • Monitor. Use IoT for advanced AI-powered monitoring of remote assets at scale.
  • Health. Manage asset health by using IoT data from sensors, asset data, and maintenance history.
  • Visual inspection. Train machine learning models to use visual inspection for visual analysis of emerging issues.
  • Predict. Predict future failures by using machine learning and data analytics.
  • Assist. Assist technicians by providing AI-powered guidance to a knowledge base of equipment maintenance data and by giving them remote access to experts.
  • Safety. Collect and analyze data from sensors, provide contextual data, and derive meaningful analytics.
  • Civil. Integrate inspection, defect tracking, and maintenance activities to help improve asset life, keep critical systems operating, and lower total costs of ownership of civil infrastructure.

These applications and MAS 8.x and up are tested for use on Azure. Microsoft and the IBM Maximo team partnered to ensure this solution is configured to run optimally on Azure. This article provides a design for running MAS 8.x and up on Azure for customers who have support from IBM and a partner for installation. Contact your IBM team for product-specific questions. The Azure Marketplace offers an alternative installation for MAS that supports bringing your own license. For more information, see IBM Maximo Application Suite (bring your own license (BYOL)). This guide details how to install Maximo manually.

Potential use cases

Many industries and sectors use the solutions in MAS, such as:

  • Energy and utilities
  • Oil and gas
  • Manufacturing
  • Travel, automotive, and transportation
  • Public sector

Find more information about use cases for MAS on IBM's website at IBM Maximo Application Suite.

Recommendations

We recommend installing the latest stable version of MAS because it provides the best integration options with Azure. Pay close attention to the versions of OpenShift that are supported, because the supported versions vary with the specific version of MAS.

Use of earlier or later major versions of OpenShift can result in falling out of official support for MAS. Before building out your own deployment, we recommend to thoroughly read through the installing on Azure and planning for Azure documentation so that you understand how the deployment and configuration works. Knowing the installation details speeds up the creation of the design requirements for your implementation.

Microsoft works with IBM and other partners to ensure that the documentation, architecture, and guidance gives you the best experience on Azure. They follow the best practices as outlined in the Microsoft Azure Well-Architected Framework. Contact your IBM account team for support beyond this documentation.

Before you proceed with your deployment, you need to answer the following questions about design:

  • What MAS applications do you need?
  • What dependencies do your applications have?
  • What version of OpenShift is required?
  • Which method of installation of OpenShift should you use?
  • What databases are needed?
  • What number and sizes of VMs do you need?
  • Will users connect from external networks?

Maximo Application Suite

Microsoft has tested MAS versions 8.7 and later on Azure. Our recommendation is to use the latest version of MAS, which currently is version 9.0. If you are on earlier versions of Maximo Application Suite, it is recommended to upgrade to benefit from better integration with Azure.

Review the MAS applications that you need for your complete business scenario, and then review the requirements for each of the applications. For more information, see IBM Maximo Application Suite system requirements. Each of the applications might need separate databases. We have tested and support the following databases on Azure:

You might also choose to run Oracle Exadata on a VM or on Oracle Cloud Infrastructure by using interconnection, but this isn't a tested configuration. For more information about interconnection, see Interconnecting Oracle Cloud with Microsoft Azure. Currently, Azure SQL Database, Azure SQL Managed Instance, and Azure Cosmos DB aren't supported.

Note

In some cases, you can't reuse a database for multiple MAS applications because of conflicting database settings. For example, you can't use the same IBM Db2 Warehouse for Health and Manage in combination with Monitor. However, you can mix different database products, such as using SQL Server for one application and IBM Db2 Warehouse for another.

For more information about database requirements for the Health application, see Configuring the database for Maximo Health.

MAS and some of its applications have dependencies on MongoDB and Kafka. Decide how to deploy these solutions based on considerations of performance and operations. The defaults are to deploy MongoDB Community Edition and Strimzi Kafka inside the clusters. Some of the prerequisites of MAS, for example BAS, use databases that can't be externalized but that require persistent storage to be provided to the OpenShift cluster.

For state-based services that run inside of the OpenShift cluster, frequently backing up data and moving the backups into another region is necessary. Design and plan out a recovery strategy in case of disaster and decide accordingly, especially when running Kafka or MongoDB inside of OpenShift.

For services that retain state, use external Azure platform as a service (PaaS) offerings when possible. Doing so improves supportability during an outage.

Some of the services might require other IBM tools and services, such as IBM Watson Machine Learning and IBM App Connect. You can deploy all the tools and services on the same OpenShift cluster.

OpenShift

Note

IBM Maximo Application Suite supports Azure Red Hat OpenShift, provided that the underlying versions of OpenShift and Cloud Pak for Data (CP4D) align.

Before you install OpenShift, you need to determine which method you'll be using:

  • Installer Provisioned Infrastructure (IPI). This method uses an installer to deploy and configure the OpenShift environment on Azure. IPI is the most common method for deploying on Azure, and you should use IPI unless your security requirements are too strict to do so.

  • User Provisioned Infrastructure (UPI). This method allows fine-grained control over your deployment. UPI requires more steps and considerations to build your environment. Use UPI if IPI doesn't meet your needs. A common use case for UPI is for private, air-gapped installation. Choose UPI when you have no outbound internet access when building the environment.

We recommend using IPI whenever possible, because it significantly reduces the amount of work that's required to complete installation of OpenShift.

Note

After you install OpenShift, the owner of the control plane is responsible for maintaining and scaling the worker nodes on Azure. You increase the cluster size by using machine sets in the admin console, not through the Azure portal. For more information, see Creating a machine set on Azure.

When installing OpenShift, you must resolve the following considerations:

  • Region selection. We recommend using a region with availability zones. During deployment, OpenShift automatically attempts to create nodes across zones based on the configuration in the configuration file, install-config.yaml. By default, OpenShift balances workloads across all available nodes and across the availability zones. If there's an outage in a zone, your solution can continue functioning by having nodes in other zones that can take over the work.

  • Backup & recovery. You can use the instructions for Azure Red Hat OpenShift for backup and recovery. For more information, see Create an Azure Red Hat OpenShift 4 cluster Application Backup. If you use this method for back-up and recovery, you must provide another method of disaster recovery for the database.

  • Failover. Consider deploying OpenShift in two regions and using Red Hat Advanced Cluster Management. If your solution has public endpoints, you can place Azure Traffic Manager between them and the internet to redirect traffic to the appropriate cluster when there's an outage of a region. In such a situation, you must also migrate your applications' states and persistent volumes.

Air-gapped installations

In some cases, such as for regulatory compliance, you might require an air-gapped installation of MAS on Azure. Air gapped means that there's no inbound or outbound internet access. Without an internet connection, your installation can't retrieve the installation dependencies at run time for the installation of MAS or OpenShift.

Note

Air-gapped deployments require UPI for installation. However, they have not been fully tested.

We don't recommend that you do an air-gapped installation unless that's a security requirement. An air gap adds significant complexity to the operations of your solution. Activities such as installing software, mirroring containers, updating a mirror to protect against security vulnerabilities, and managing a firewall can become very time consuming.

For more information about air-gapped installations, see the following OpenShift documentation:

After you've installed OpenShift, see the MAS documentation for similar guidance.

Sizing your environment

For all workloads (except visual inspection), we recommend using the latest Ds series VMs as your worker nodes. Examples are the Dsv3, Dasv4, Dsv4, Dasv5, or Dsv5. We recommend using the latest versions, when possible, because they provide better performance. Only use VMs that have premium storage.

Maximo Visual Inspection requires GPU nodes to perform its machine learning. The solution uses CUDA and only supports NVIDIA GPUs. The recommended types of VMs are NCv3 and NCasT4_v3. If you need to train by using YOLOv3, you'll need Ampere-based GPUs. Use the NVadsA10 v5 or NC A100 v4 for larger training tasks.

For the GPU machines, we recommend starting with the smallest node and scaling up as your requirements increase.

Warning

If you need GPU machines, you need OpenShift 4.8.22 as a minimum version to enable the GPUs through the NVIDIA GPU Operator.

For all other machines, we recommend configuring VMs across availability zones to support high availability. Configure the nodes as follows:

  • Control nodes. A minimum of one VM per availability zone within the selected region. We recommended a vCPU count of at least 4. Our reference uses 3x Standard_D8s_v4 nodes.

  • Worker nodes. A minimum of two machines per availability zone within the selected region. We recommend a vCPU count of at least 8. Our reference uses 6x Standard_D8s_v4 nodes.

MAS core requires 13 vCPUs for a standard-sized base installation. Sizing for the worker nodes varies based on which MAS applications your configuration deploys and the load on your environment. For example, Manage for 10 users requires another 2 vCPUs. We recommend that you review the IBM Maximo Application Suite system requirements to get a good sizing estimate.

Try to keep the types of VMs similar to each other to provide proximity with each of the availability zones between worker and control nodes. That is, if you use a v4 VM for your control nodes, also use a v4 VM for your worker nodes.

If you need a jump box to use the OpenShift command-line interface (oc) or to install MAS, deploy a VM that's running Red Hat Enterprise Linux version 8.4.

Network

With OpenShift, we use the default container network interface (CNI) provider of OpenShift's software-defined networking (SDN). For more information about the default OpenShift CNI, see Understanding Networking in OpenShift Container Platform. You must size your network for the number of OpenShift control and worker nodes that you need, and also for any other requirements, such as databases and storage accounts.

For a standard MAS production installation, we recommend a virtual network with the address space that a classless inter-domain routing (CIDR) prefix of /24 provides. The virtual network has three or four subnets (for Azure Bastion). For OpenShift, the subnet for the worker nodes has a CIDR prefix of /25, and the control nodes have a prefix of /27. A subnet for endpoints and an optional external database server should have a prefix of /27. If you're deploying Azure Bastion, which is optional, you need a subnet named AzureBastionSubnet with a prefix of /26. For more information about the requirements for Azure Bastion, see Architecture.

If you're short on IP addresses, you can implement a highly available configuration with a minimum prefix of /27 for the subnet of control nodes and /27 for the subnet of worker nodes.

If you want to use a different CNI, size your networks accordingly. MAS with some standard applications deploys over 800 pods, which likely require a CIDR prefix of /21 or larger.

Database specifics

Various components of MAS use MongoDB as a metadata store. The default guidance is to deploy MongoDB Community Edition inside of the cluster. If you deploy it by using that method, ensure that you have in place a proper procedure for backing up and restoring the database. Consider using MongoDB Atlas on Azure, because it provides an externalized store, backups, scaling, and more. Azure doesn't currently support using MongoDB APIs with Azure Cosmos DB.

If you deploy IoT services, you're required to also provide a Kafka endpoint. The default guidance is to use Strimzi to deploy Kafka inside the OpenShift cluster. During a disaster recovery, data inside Strimzi will most likely be lost. If data loss within Kafka is unacceptable, you should consider using Confluent Kafka on Azure. Currently, Azure Event Hubs with Kafka endpoints aren't supported.

MAS comes packed with many databases inside its pods, and those databases retain their states on the file system that's provided for MAS. We recommend using a zone-redundant storage (ZRS) mechanism to retain the states outside of your clusters to be able to absorb zone failures. Our recommended pattern is to use Azure File Storage with the following configurations:

  • Standard. Provides Server Message Block (SMB) shares for lower throughput and ReadWriteOnce (RWO) workloads. Standard is a great fit for parts of the application that don't write to storage often and require a single persistent volume (for example, IBM single-level storage).

  • Premium. Provides Network File System (NFS) shares for higher throughput and ReadWriteMany (RWX) workloads. Volumes like these are used throughout the cluster for RWX workloads, such as the Db2 Warehouse in Cloud Pak for Data or Postgres in Manage.

Be sure to disable policies for enforcing secure transfer on the Azure Blob Storage or exempt the accounts from such policies. Azure Premium Files with NFS requires that secure transfer be disabled. Be sure to use a private endpoint to guarantee private connectivity to your shares.

By default, Db2 Warehouse deploys on top of OpenShift Data Foundation (previously known as OpenShift Container Storage). For reasons of cost, performance, scaling, and reliability, we recommended using Azure Premium Files with NFS instead of OpenShift Data Foundation.

Don't use Azure Blob Storage with CSI drivers, because it doesn't support hard links, which are required. Some pods can't run without hard links.

Considerations

These considerations implement the pillars of the Azure Well-Architected Framework, which is a set of guiding tenets that you can use to improve the quality of a workload. For more information, see Microsoft Azure Well-Architected Framework.

Security

Security provides assurances against deliberate attacks and the abuse of your valuable data and systems. For more information, see Overview of the security pillar.

Maintaining access and visibility into the maintenance lifecycle of your assets can be one of your organization's greatest opportunities to operate efficiently and maintain uptime. To improve the security posture of your environment, it's important to use secure authentication and to keep your solutions up to date. Use encryption to help protect all data that moves in and out of your architecture.

Azure delivers MAS by using the models of infrastructure as a service (IaaS) and PaaS. Microsoft builds security protections into the service at the following levels:

  • Physical datacenter
  • Physical network
  • Physical host
  • Hypervisor

Carefully evaluate the services and technologies that you select for the areas above the hypervisor, such as the latest patched version of OpenShift for a major release. Be sure to provide the proper security controls for your architecture. You're responsible for patching and maintaining the security of the IaaS systems. Microsoft takes that role for the PaaS services.

Use network security groups to filter network traffic to and from resources in your virtual network. With these groups, you can define rules that grant or deny access to your MAS services. Examples include:

  • Allow SSH access into the OpenShift nodes for troubleshooting
  • Block access to all other parts of the cluster
  • Control which locations can have access to MAS and the OpenShift cluster

If you need access to your VMs for some reason, you can connect through your hybrid connectivity or through the OpenShift admin console. If you have an online deployment or don't want to rely on connectivity, you can also access your VMs through Azure Bastion (which is optional). For security reasons, you shouldn't expose VMs to a network or the internet without configuring network security groups to control access to them.

Server-side encryption (SSE) of Azure Disk Storage protects your data. It also helps you meet organizational security and compliance commitments. With Azure managed disks, SSE encrypts the data at rest when persisting it to the cloud. This behavior applies by default to both OS and data disks. OpenShift uses SSE by default.

Authentication

MAS currently supports single sign-on (SSO) with Security Assertion Markup Language (SAML) in Microsoft Entra ID. This authentication method requires an enterprise application within Microsoft Entra ID and permissions to modify the application. For more information, see Microsoft Entra SSO integration with Maximo Application Suite.

Before you set up SAML-based authentication, we recommend that you go through the IBM configuration and the Azure configuration. For information about SAML with MAS, see SAML in the documentation for MAS. For information about SAML with Azure, see Quickstart: Enable single sign-on for an enterprise application.

You should also configure Open Authorization (OAuth) for OpenShift. For more information, see Overview of authentication and authorization in the OpenShift documentation.

Protect your infrastructure

Control access to the Azure resources that you deploy. Every Azure subscription has a trust relationship with a Microsoft Entra tenant. Use Azure role-based access control (RBAC) to grant users within your organization the correct permissions to Azure resources. Grant access by assigning Azure roles to users or groups at a certain scope. The scope can be a subscription, a resource group, or a single resource. Be sure to audit all changes to infrastructure. For more information about auditing, see Azure Monitor activity log.

Cost optimization

Cost optimization is about looking at ways to reduce unnecessary expenses and improve operational efficiencies. For more information, see Overview of the cost optimization pillar.

A standard deployment of MAS consists of the following components:

  • 3 control VMs
  • 6 worker VMs
  • 3 worker VMs for Db2 Warehouse
    • You can substitute SQL Server on Azure VMs in some configurations, rather than use Db2 Warehouse.
  • 2 Azure Storage accounts
  • 2 DNS zones
  • 2 Load balancers
  • Azure Bastion
  • 1 Visual Inspection VM
    • This isn't required unless you're planning to run Visual Inspection inside of MAS.

You can review an example estimate by using our cost calculator. Configurations vary, and you should verify your configuration with your IBM sizing team before finalizing your deployment.

Reliability

OpenShift has built-in capabilities for self-healing, scaling, and resilience to make sure OpenShift and MAS work successfully. OpenShift and MAS have been designed for parts that fail and recover. A key requirement for self-healing to work is that there are enough worker nodes. To recover from a zone failure within an Azure region, your control and worker nodes must be balanced across availability zones.

MAS and OpenShift use storage to persist state outside of the Kubernetes cluster. To ensure that the storage dependencies continue to work during a failure, you should use zone-redundant storage whenever possible. This type of storage remains available when a zone fails.

Because human error is common, you should deploy MAS by using as much automation as possible. In our quickstart guide, we provide some sample scripts for setting up full, end-to-end automation.

Deploy this scenario

Before you start, we recommend that you review the IBM Maximo Application Suite system requirements. Be sure that you have the following resources available before starting the deployment:

  • Access to an Azure Subscription with Reader permission
  • Application Registration or service principal name that has Contributor and User Access Administrator permissions to the subscription
  • Domain or delegated subdomain to an Azure DNS zone
  • Pull secret from Red Hat to deploy OpenShift
  • MAS entitlement key
  • MAS license file (created after MAS installation)
  • IBM-recommended cluster sizing
  • Existing virtual network or a new virtual network created by IPI, depending on your requirements
  • High availability and disaster recovery requirements for your specific deployment
  • Configuration file, install-config.yaml, for the installer

For a step-by-step guide for installing OpenShift and MAS on Azure, including how to address the prerequisites, see our quickStart guide on GitHub.

Note

QuickStart Guide: Maximo Application Suite on Azure includes an example of an install-config.yaml file in /src/ocp/.

Deployment considerations

It's best to deploy workloads by using infrastructure as code (IaC) rather than manually deploying workloads, because manual deployment can result in misconfiguration. Container-based workloads can be sensitive to misconfiguration, which can reduce productivity.

Before building your environment, review the planning for Azure documentation provided by IBM to develop an understanding of the design parameters. The quickstart guide isn't intended for a production-ready deployment, but you can use the guide's assets to get to a production-grade mechanism for deployment.

IBM offers specialist services to help you with installation. Contact your IBM team for support.

Contributors

This article is maintained by Microsoft. It was originally written by the following contributors.

Principal authors:

Other contributor:

To see non-public LinkedIn profiles, sign in to LinkedIn.

Next steps

For help with getting started, see the following resources:

To learn more about the featured technologies, see the following resources: