Geospatial analysis for the telecommunications industry

Azure Data Factory
Azure Data Lake
Azure Databricks
Azure Machine Learning
Azure Maps

The focus of this article is to showcase a practical architecture that uses Azure Cloud Services to process large volumes of geospatial data. It provides a path forward when on-premises solutions don't scale. It also allows for continued use of the current geospatial analysis tools.

Apache®, Apache Spark®, GeoSpark®, and Sedona® are either registered trademarks or trademarks of the Apache Software Foundation in the United States and/or other countries. No endorsement by The Apache Software Foundation is implied by the use of these marks.

GeoPandas®, QGIS®, and ArcGIS® are trademarks of their respective companies. No endorsement is implied by the use of these marks.

Architecture

Diagram for an architecture that uses Azure Cloud Services to process large volumes of geospatial data.

Download a Visio file of this architecture.

Workflow

  1. Azure Data Factory ingests geospatial data into Azure Data Lake Storage. The source of this data is geospatial databases such as Teradata, Oracle Spatial, and PostgreSQL.
  2. Azure Key Vault secures passwords, credentials, connection strings, and other secrets.
  3. Data is placed in various folders and file systems in Data Lake Storage according to how it has been processed. The diagram shows a multi-hop architecture. The bronze container holds raw data, the silver container holds semi-curated data, and the gold container holds fully curated data.
  4. Data is stored in formats such as GeoJson, WKT and Vector tiles. Azure Databricks and the GeoSpark / Sedona package can convert formats and efficiently load, process, and analyze large-scale spatial data across machines.
  5. Azure Databricks and Apache Sedona do various kinds of processing at scale:
    1. Joins, intersections, and tessellations
    2. Spatial sampling and statistics
    3. Spatial indexing and partitioning
  6. GeoPandas exports data in various formats for use by third-party GIS applications such as QGIS and ArcGIS.
  7. Azure Machine Learning extracts insights from geospatial data, determining, for example, where and when to deploy new wireless access points.
  8. Power BI and Azure Maps Power BI visual (Preview) render a map canvas to visualize geospatial data. Power BI uses an Azure Databricks native connector to connect to an Azure Databricks cluster.
  9. Log Analytics, a tool in the Azure portal, runs queries against data in Azure Monitor Logs to implement a robust and fine-grained logging system to analyze events and performance.

Components

  • Azure Data Lake Storage is a scalable and secure data lake for high-performance analytics workloads. You can use Data Lake Storage to manage petabytes of data with high throughput. It can accommodate multiple, heterogeneous sources, and data that's in structured, semi-structured, or unstructured formats.
  • Azure Databricks is a data analytics platform that uses Spark clusters. The clusters are optimized for the Azure cloud platform.
  • Azure Data Factory is a fully managed, scalable, and serverless data integration service. It provides a data integration and transformation layer that works with various data stores.
  • Microsoft Power BI is a collection of software services, apps, and connectors that work together to turn multiple sources of data into coherent, visually immersive, and interactive insights.
  • Azure Maps is a collection of geospatial services and SDKs that use fresh mapping data to provide geographic context to web and mobile applications.
  • Azure Machine Learning is a fully managed cloud service that's used to train, deploy, and manage machine learning models at scale.
  • Azure Key Vault is a service that can be used to securely store, manage, and tightly control access to tokens, credentials, certificates, API Keys, and other secrets.
  • Azure Monitor is a comprehensive solution for collecting, analyzing, and acting on telemetry from your cloud and on-premises environments. You can use it to maximize the availability and performance of your applications and services.

Alternatives

  • You can use Synapse Spark Pools for geospatial analytics instead of Azure Databricks, using the same open-source frameworks.
  • Instead of using Data Factory to ingest data, you can use Azure Event Hubs. It can receive massive amounts of data directly or from other event streaming services such as Kafka. Then you can use Azure Databricks to process the data. For more information, see Stream Processing with Azure Databricks.
  • Instead of Azure Databricks, you can use Azure SQL Database or Azure SQL Managed Instance to query and process geospatial data. These databases provide the familiar T-SQL language, which you can use for geospatial analysis. For more information, see Spatial Data (SQL Server).
  • Like Event Hubs, Azure IoT Hub can ingest large amounts of data from sensor and telecom IoT devices. You can use the IoT Hub bidirectional capability to communicate securely with devices and potentially manage and control them from a centralized platform in the cloud.
  • You can use Azure Maps to provide geographic context to your web and mobile applications. In addition to location intelligence, Azure Maps can search services to locate addresses, places, and points of interest to get real-time traffic information. Azure Maps Power BI Visual provides the same capabilities in both Power BI Desktop and the Power BI service.

Scenario details

Location intelligence and geospatial analytics can uncover important regional trends and behaviors that affect telecommunications companies. The companies can use such knowledge to enhance their radio signal and wireless coverage, and thus gain competitive advantage.

Telecommunications companies have large volumes of geographically dispersed asset data, most of which is user telemetry. The data comes from radio networks, IoT sensing devices, and remote sensing devices that capture geospatial data. It's in various structured and semi-structured formats such as imagery, GPS, satellite, and textural. Making use of it requires aggregating it and joining it with other sources such as regional maps and traffic data.

After the data is aggregated and joined, the challenge is to extract insights from it. Historically, telecommunications companies relied on legacy systems such as on-premises databases with geospatial capabilities. Eventually such systems hit scalability limits due to the ever-increasing amount of data. Also, they require third-party software to perform tasks that the geospatial database systems can't.

Potential use cases

This solution is ideal for the telecommunications industry, and it applies to the following scenarios:

  • Analyzing signal information across locations to assess network quality
  • Analyzing real-time network infrastructure data to guide maintenance and repair
  • Analyzing market segmentation and market demand
  • Identifying relationships between customer locations and company marketing campaigns
  • Creating capacity and coverage plans to ensure connectivity and quality of service

Considerations

These considerations implement the pillars of the Azure Well-Architected Framework, which is a set of guiding tenets that can be used to improve the quality of a workload. For more information, see Microsoft Azure Well-Architected Framework.

Consider following the Microsoft Azure Well-Architected Framework when you implement this solution. The framework provides technical guidance across five pillars: cost optimization, security, reliability, performance efficiency, and operational excellence.

Performance

  • Follow Apache Sedona programming guides on design patterns and performance tuning best practices.
  • Geospatial indexing is crucial for processing large-scale geospatial data. Apache Sedona and other open-source indexing frameworks such as H3 provide this capability.
  • The GeoPandas framework doesn't have the distributed features of GeoSpark / Apache Sedona. Therefore, as much as possible, use Sedona framework for geospatial processing.
  • Consider using Sedona's built-in functions to validate geometry formatting before processing.

Security

Security provides assurances against deliberate attacks and the abuse of your valuable data and systems. For more information, see Overview of the security pillar.

For better security, consider following this guidance:

Cost optimization

Cost optimization is about looking at ways to reduce unnecessary expenses and improve operational efficiencies. For more information, see Overview of the cost optimization pillar.

  • To estimate the cost of implementing this solution, use the Azure Pricing Calculator for the services mentioned above.
  • Power BI comes with various licensing offerings. For more information, see Power BI pricing.
  • Your costs increase if you have to scale your Azure Databricks cluster configurations. This depends on the amount of data and the complexity of the analysis. For best practices on cluster configuration, see Azure Databricks Best practices: Cluster configuration.
  • See Overview of the cost optimization pillar for ways to minimize costs.
  • For the third-party components such as QGIS and ArcGIS, see the vendor websites for pricing information.
  • The frameworks mentioned in this solution, such as Apache Sedona and GeoPandas, are free open-source frameworks.

Contributors

This article is maintained by Microsoft. It was originally written by the following contributors.

Principal author:

Next steps