Transform Logs into Actionable Insights with Mezmo Pipelines & Dashboards
Dive into the key steps for reshaping or transforming raw, messy, data to yield more effectiveness with analytics tools, observability tools, and AI models.
What is Data Transformation?
Data transformation is the process of converting raw data from one format, structure, or value state into another so that it becomes more useful, consistent, and compatible with downstream systems. It’s a key step in data pipelines, analytics, observability, and AI workflows.
Data transformation is about reshaping data for purpose. Raw data is often messy, inconsistent, or not structured in a way that analytics tools, observability pipelines, or AI models can use effectively. Transformation makes it standardized, enriched, and optimized.
There are four main types of data transformation:
- Structural Transformation
- Changing the data schema (like rows to columns or JSON to CSV).
- Normalizing or denormalizing data tables.
- Aggregating or pivoting data.
- Content Transformation
- Cleaning (removing duplicates, fixing errors).
- Standardizing formats .
- Masking or obfuscating sensitive information.
- Converting codes or labels into human-readable form.
- Enrichment
- Adding contextual information (geo lookups, user metadata, log enrichment).
- Joining data from multiple sources.
- Optimization
- Filtering irrelevant data.
- Sampling or summarizing to reduce volume.
- Compression or encoding for storage/performance.
To sum up, data transformation turns raw data into useful data. It’s the bridge between messy input and actionable insight.
What is the purpose of log transformation in data analysis?
Log transformation is the process of applying a logarithm function (commonly log base 10, natural log, or log base 2) to each data point in a dataset. Its purpose is not about logs from observability (system logs) but rather about mathematical transformation of numerical data for analysis. The purpose of log transformation is to normalize distributions, stabilize variance, make multiplicative patterns linear, reduce outlier influence, and improve interpretability in data analysis.
Log transformation can reduce skewness and normalized distributions, stabilize variance, handle multiplicative relationships, improve interpretability and make outliers less influential. H2: The basics of telemetry data transformation
Basic filtering
Telemetry data is often high-volume, noisy, and redundant. Before it reaches your backend (like a log store, monitoring platform, or AI system), you can transform it to control cost, performance, and usefulness.
One of the simplest but most impactful transformation techniques is filtering. Filtering means deciding which telemetry data to keep, drop, or route differently before ingestion. Instead of sending everything downstream, you only send the signals that matter. Basic filtering provides cost and noise reduction, boosts performance and improves compliance and security. Basic filtering in telemetry transformation is about cutting out the noise - dropping irrelevant, redundant, or sensitive data so that only high-value telemetry flows into your observability systems.
Adding or Deleting Attributes
Telemetry data is most useful when it carries the right contextual attributes (a.k.a. fields, tags, or labels). Transformation allows you to enrich or streamline that context before ingestion. Attributes contribute context, searchability, cost efficiency and compliance. Adding attributes enriches telemetry with useful context (making it more valuable). Deleting attributes reduces noise, cost, and risk (making it more efficient and compliant).
Renaming Metrics or Metric Labels
Telemetry data comes from many sources, each with its own naming conventions. Left unchecked, this creates inconsistency, duplication, and confusion. Renaming is a transformation step that ensures consistency, clarity, and usability. Renaming metrics and labels is about standardization. It makes telemetry consistent, queryable, and compatible across systems - preventing duplication, improving usability, and ensuring observability data is reliable.
Enriching Telemetry with Resource Attributes
One of the most powerful (and widely used) forms of telemetry data transformation is enrichment with resource attributes. This step turns raw logs, metrics, and traces into context-rich signals that make debugging, monitoring, and cost optimization far easier.
In observability, resource attributes describe the entity that produced the telemetry.
They are key–value pairs that add context about the source system, environment, or infrastructure. They are defined in standards like OpenTelemetry Semantic Conventions, making them portable and interoperable. Enriching telemetry with resource attributes adds vital context about the system, environment, and infrastructure that produced the data. It transforms raw telemetry into actionable signals that can be searched, correlated, and cost-attributed across the observability pipeline.
Setting a span status
A span’s status summarizes the outcome of the span’s operation. Status is not the same as log level, exceptions, or HTTP codes - it’s a normalized success/failure signal for traces. A span status should be set at instrumentation time and/or in the pipeline. Span status matters for triage and SLOs, determining root cause speed, and for improving cost and clarity.
Mezmo Telemetry Pipeline: What is it?
The Mezmo Telemetry Pipeline is a real-time, cloud-native pipeline for ingesting, processing, transforming, and routing telemetry data - logs, metrics, traces, and events - before it reaches your observability, security, or analytics tools. It acts as a control plane for telemetry: instead of sending raw, unfiltered, and expensive data directly to multiple backends, you use Mezmo’s pipeline to shape, enrich, reduce, and distribute telemetry where it delivers the most value.
The Mezmo Telemetry Pipeline has a number of core capabilities including;
1. Ingestion
- Supports a wide range of sources (applications, services, cloud infra, Kubernetes, OpenTelemetry, Fluentd/Fluent Bit, syslog, etc.).
- Handles high-volume, real-time data at scale.
2. Transformation
- Filtering: Drop irrelevant data (e.g., debug logs, health checks).
- Attribute management: Add, delete, or rename fields/labels.
- Enrichment: Append metadata such as service name, region, environment, trace IDs.
- Normalization: Standardize metric/log naming, formats, and schemas.
- Redaction: Remove sensitive/PII data before it leaves your environment.
3. Optimization
- Sampling & aggregation: Reduce volume by keeping only representative data.
- Compression & batching: Improve throughput and lower cost.
- Cost control knobs: Shape telemetry before ingestion into expensive systems (e.g., Splunk, Datadog, Elastic).
4. Routing
- Dynamically send different data streams to different destinations.
- Example:
- Send enriched, high-value traces to APM (Datadog, New Relic).
- Send security logs to SIEM (Splunk, Sumo Logic).
- Send raw data to low-cost object storage (S3, GCS).
- Support for multi-sink delivery ensures one copy of data serves many teams.
The Mezmo Telemetry Pipeline offers a number of benefits including cost efficiency, data quality, flexibility, governance, and future-proofing.
The Mezmo Telemetry Pipeline is the smart traffic controller for observability data. It gives teams control, efficiency, and flexibility over how telemetry is collected, transformed, and routed, ensuring you get maximum insight while keeping costs in check.
Mezmo Log Analysis: What is it?
Mezmo Log Analysis is the capability within the Mezmo observability platform that lets engineers and operators search, query, visualize, and derive insights from log data.
It takes the raw or transformed logs that flow through the Mezmo Telemetry Pipeline and makes them actionable for debugging, monitoring, and business intelligence.
Think of it as the “front-end lens” on top of your telemetry data.
Mezmo Log Analysis has a number of core capabilities such as centralized log collection, search and query, parsing and structure, visualization and dashboards, alerting and monitoring, and collaboration and troubleshooting. The benefits of Mezmo Log Analysis are wide ranging and include faster troubleshooting, data-driven monitoring, cost efficiency and cross-team value. Mezmo Log Analysis is the search, query, and visualization layer of Mezmo’s observability platform. It turns logs into actionable insights by letting you collect, parse, search, visualize, and alert on them, all in real time and at scale.
How can Mezmo Telemetry Pipeline and Log Analysis help transform your telemetry data?
Telemetry data can be transformed in a variety of ways using the Mezmo Telemetry Pipeline.
- Data ingestion and normalization: You no longer wrangle inconsistent data — it arrives standardized and ready to analyze.
- Attribute enrichment and context: Telemetry gains business and operational context, making root-cause analysis faster.
- Transformation and optimization: less telemetry volume = lower cost, faster queries and clearer insights.
- Routing and multi-use distribution: the same telemetry powers observability, security, compliance, and business intelligence — without duplication.
- Actionable insights: Transformation doesn’t stop at shaping the data — it enables faster decisions and remediation.
Remove data noise
- Filtering and noise reduction: Cuts out noise, reduces storage cost, and speeds up troubleshooting.
Real-Time Insights
- Data shaping at ingestion: Real-time visibility into relevant events with full context (service, environment, location).
The Mezmo Telemetry Pipeline and Log Analysis transform raw telemetry into real-time insights that cut noise, add context, and speed up incident response.
Data visualization
Clean and structured data for visuals, noise-free visual signals and context-enriched dashboards transform raw telemetry into visual insights your teams can act on.
Related Articles
Share Article
Ready to Transform Your Observability?
- ✔ Start free trial in minutes
- ✔ No credit card required
- ✔ Quick setup and integration
- ✔ Expert onboarding support