Skip to main content

VictoriaMetrics Releases VictoriaLogs

VictoriaMetrics announced the GA release of its logging solution - VictoriaLogs.

The easy-to-use, open source log management solution combines a powerful query language for easy log searching with minimal resource requirements. It’s perfect for managing and analyzing large volumes of log data, especially in containerized environments such as Kubernetes.

Also announced is the upcoming preview-release of VictoriaLogs Cluster, with the full release scheduled for 2025.

VictoriaLogs - Key Highlights:

- Improves query performance for haystack searches by up to 1000 times

- Uses up to 30 times less RAM and 15 times less disk space than comparable solutions

- Accepts logs from popular log collectors

- Supports ingestion directly via Syslog protocol, removing the need for proxies and converters

“VictoriaLogs addresses the major challenges of traditional log management tools. It significantly reduces memory usage and infrastructure costs, making it a game-changer for those frustrated by slow searches in large log volumes. By using bloom filters, VictoriaLogs accelerates haystack search times and quickly pinpoints relevant data, making queries up to 1,000 times faster than competitors. Even the single-node version of VictoriaLogs is capable of replacing a comparable solutions’ cluster of up to 30 nodes. It’s a high-performance tool that simplifies log management,” said Aliaksandr Valialkin, Co-founder and CTO at VictoriaMetrics.

VictoriaLogs integrates seamlessly with the broader observability ecosystem. By allowing users to correlate logs with metrics, it provides a holistic view of system performance and behavior. Support for popular log shippers, enables easy adoption without disrupting existing logging workflows. This integrated approach not only simplifies the observability stack, but also enhances troubleshooting capabilities by allowing users to swiftly navigate between related logs and metrics.

VictoriaLogs uses automatically adjusted bloom filters instead of inverted indexes, allowing for accurate word searches providing definitive yes or no answers. By minimizing CPU time and disk read IO spent on unpacking, parsing, and reading logs, VictoriaLogs significantly saves computing resources at large scale.

Offering robust support for Syslog ingestion, VictoriaLogs simplifies this transition by allowing direct ingestion over Syslog without intermediaries. While Syslog remains a widely used standard, many existing solutions require additional proxies or converters, complicating the ingestion process. Users can easily migrate to VictoriaLogs by updating just the address or URL for their existing log shippers, ensuring a smooth transition.

“VictoriaLogs sets a new standard in log management performance, addressing the demands of today's data-intensive environments. For example, where a comparable solution creates new log streams each time a detail (such as an IP address or user ID) changes, VictoriaLogs treats these details as regular information within each log entry - resulting in fewer log streams and improving system performance and resource utilization. VictoriaLogs also achieves this while using up to 30 times less RAM and 15 times less disk space compared to popular alternatives, making it an ideal choice for organizations dealing with massive log volumes,” said Aliaksandr Valialkin, Co-founder and CTO at VictoriaMetrics.

VictoriaLogs is optimized for efficient storing and querying of wide events containing hundreds of fields, accepting wide events with different sets of fields without the need to configure. The solution’s LogsQL query language simplifies querying wide events' stats at high speed.

VictoriaLogs' efficient data management also extends to its disk I/O performance, addressing a critical bottleneck in log analytics. By requiring substantially less disk space, the solution dramatically reduces the volume of data read during resource-intensive queries. Industry tests have shown that this approach can accelerate query performance by up to two orders of magnitude, with heavy queries executing up to 100 times faster than on comparable solutions.

The Latest

As businesses increasingly rely on high-performance applications to deliver seamless user experiences, the demand for fast, reliable, and scalable data storage systems has never been greater. Redis — an open-source, in-memory data structure store — has emerged as a popular choice for use cases ranging from caching to real-time analytics. But with great performance comes the need for vigilant monitoring ...

Kubernetes was not initially designed with AI's vast resource variability in mind, and the rapid rise of AI has exposed Kubernetes limitations, particularly when it comes to cost and resource efficiency. Indeed, AI workloads differ from traditional applications in that they require a staggering amount and variety of compute resources, and their consumption is far less consistent than traditional workloads ... Considering the speed of AI innovation, teams cannot afford to be bogged down by these constant infrastructure concerns. A solution is needed ...

AI is the catalyst for significant investment in data teams as enterprises require higher-quality data to power their AI applications, according to the State of Analytics Engineering Report from dbt Labs ...

Misaligned architecture can lead to business consequences, with 93% of respondents reporting negative outcomes such as service disruptions, high operational costs and security challenges ...

A Gartner analyst recently suggested that GenAI tools could create 25% time savings for network operational teams. Where might these time savings come from? How are GenAI tools helping NetOps teams today, and what other tasks might they take on in the future as models continue improving? In general, these savings come from automating or streamlining manual NetOps tasks ...

IT and line-of-business teams are increasingly aligned in their efforts to close the data gap and drive greater collaboration to alleviate IT bottlenecks and offload growing demands on IT teams, according to The 2025 Automation Benchmark Report: Insights from IT Leaders on Enterprise Automation & the Future of AI-Driven Businesses from Jitterbit ...

A large majority (86%) of data management and AI decision makers cite protecting data privacy as a top concern, with 76% of respondents citing ROI on data privacy and AI initiatives across their organization, according to a new Harris Poll from Collibra ...

According to Gartner, Inc. the following six trends will shape the future of cloud over the next four years, ultimately resulting in new ways of working that are digital in nature and transformative in impact ...

2020 was the equivalent of a wedding with a top-shelf open bar. As businesses scrambled to adjust to remote work, digital transformation accelerated at breakneck speed. New software categories emerged overnight. Tech stacks ballooned with all sorts of SaaS apps solving ALL the problems — often with little oversight or long-term integration planning, and yes frequently a lot of duplicated functionality ... But now the music's faded. The lights are on. Everyone from the CIO to the CFO is checking the bill. Welcome to the Great SaaS Hangover ...

Regardless of OpenShift being a scalable and flexible software, it can be a pain to monitor since complete visibility into the underlying operations is not guaranteed ... To effectively monitor an OpenShift environment, IT administrators should focus on these five key elements and their associated metrics ...

VictoriaMetrics Releases VictoriaLogs

VictoriaMetrics announced the GA release of its logging solution - VictoriaLogs.

The easy-to-use, open source log management solution combines a powerful query language for easy log searching with minimal resource requirements. It’s perfect for managing and analyzing large volumes of log data, especially in containerized environments such as Kubernetes.

Also announced is the upcoming preview-release of VictoriaLogs Cluster, with the full release scheduled for 2025.

VictoriaLogs - Key Highlights:

- Improves query performance for haystack searches by up to 1000 times

- Uses up to 30 times less RAM and 15 times less disk space than comparable solutions

- Accepts logs from popular log collectors

- Supports ingestion directly via Syslog protocol, removing the need for proxies and converters

“VictoriaLogs addresses the major challenges of traditional log management tools. It significantly reduces memory usage and infrastructure costs, making it a game-changer for those frustrated by slow searches in large log volumes. By using bloom filters, VictoriaLogs accelerates haystack search times and quickly pinpoints relevant data, making queries up to 1,000 times faster than competitors. Even the single-node version of VictoriaLogs is capable of replacing a comparable solutions’ cluster of up to 30 nodes. It’s a high-performance tool that simplifies log management,” said Aliaksandr Valialkin, Co-founder and CTO at VictoriaMetrics.

VictoriaLogs integrates seamlessly with the broader observability ecosystem. By allowing users to correlate logs with metrics, it provides a holistic view of system performance and behavior. Support for popular log shippers, enables easy adoption without disrupting existing logging workflows. This integrated approach not only simplifies the observability stack, but also enhances troubleshooting capabilities by allowing users to swiftly navigate between related logs and metrics.

VictoriaLogs uses automatically adjusted bloom filters instead of inverted indexes, allowing for accurate word searches providing definitive yes or no answers. By minimizing CPU time and disk read IO spent on unpacking, parsing, and reading logs, VictoriaLogs significantly saves computing resources at large scale.

Offering robust support for Syslog ingestion, VictoriaLogs simplifies this transition by allowing direct ingestion over Syslog without intermediaries. While Syslog remains a widely used standard, many existing solutions require additional proxies or converters, complicating the ingestion process. Users can easily migrate to VictoriaLogs by updating just the address or URL for their existing log shippers, ensuring a smooth transition.

“VictoriaLogs sets a new standard in log management performance, addressing the demands of today's data-intensive environments. For example, where a comparable solution creates new log streams each time a detail (such as an IP address or user ID) changes, VictoriaLogs treats these details as regular information within each log entry - resulting in fewer log streams and improving system performance and resource utilization. VictoriaLogs also achieves this while using up to 30 times less RAM and 15 times less disk space compared to popular alternatives, making it an ideal choice for organizations dealing with massive log volumes,” said Aliaksandr Valialkin, Co-founder and CTO at VictoriaMetrics.

VictoriaLogs is optimized for efficient storing and querying of wide events containing hundreds of fields, accepting wide events with different sets of fields without the need to configure. The solution’s LogsQL query language simplifies querying wide events' stats at high speed.

VictoriaLogs' efficient data management also extends to its disk I/O performance, addressing a critical bottleneck in log analytics. By requiring substantially less disk space, the solution dramatically reduces the volume of data read during resource-intensive queries. Industry tests have shown that this approach can accelerate query performance by up to two orders of magnitude, with heavy queries executing up to 100 times faster than on comparable solutions.

The Latest

As businesses increasingly rely on high-performance applications to deliver seamless user experiences, the demand for fast, reliable, and scalable data storage systems has never been greater. Redis — an open-source, in-memory data structure store — has emerged as a popular choice for use cases ranging from caching to real-time analytics. But with great performance comes the need for vigilant monitoring ...

Kubernetes was not initially designed with AI's vast resource variability in mind, and the rapid rise of AI has exposed Kubernetes limitations, particularly when it comes to cost and resource efficiency. Indeed, AI workloads differ from traditional applications in that they require a staggering amount and variety of compute resources, and their consumption is far less consistent than traditional workloads ... Considering the speed of AI innovation, teams cannot afford to be bogged down by these constant infrastructure concerns. A solution is needed ...

AI is the catalyst for significant investment in data teams as enterprises require higher-quality data to power their AI applications, according to the State of Analytics Engineering Report from dbt Labs ...

Misaligned architecture can lead to business consequences, with 93% of respondents reporting negative outcomes such as service disruptions, high operational costs and security challenges ...

A Gartner analyst recently suggested that GenAI tools could create 25% time savings for network operational teams. Where might these time savings come from? How are GenAI tools helping NetOps teams today, and what other tasks might they take on in the future as models continue improving? In general, these savings come from automating or streamlining manual NetOps tasks ...

IT and line-of-business teams are increasingly aligned in their efforts to close the data gap and drive greater collaboration to alleviate IT bottlenecks and offload growing demands on IT teams, according to The 2025 Automation Benchmark Report: Insights from IT Leaders on Enterprise Automation & the Future of AI-Driven Businesses from Jitterbit ...

A large majority (86%) of data management and AI decision makers cite protecting data privacy as a top concern, with 76% of respondents citing ROI on data privacy and AI initiatives across their organization, according to a new Harris Poll from Collibra ...

According to Gartner, Inc. the following six trends will shape the future of cloud over the next four years, ultimately resulting in new ways of working that are digital in nature and transformative in impact ...

2020 was the equivalent of a wedding with a top-shelf open bar. As businesses scrambled to adjust to remote work, digital transformation accelerated at breakneck speed. New software categories emerged overnight. Tech stacks ballooned with all sorts of SaaS apps solving ALL the problems — often with little oversight or long-term integration planning, and yes frequently a lot of duplicated functionality ... But now the music's faded. The lights are on. Everyone from the CIO to the CFO is checking the bill. Welcome to the Great SaaS Hangover ...

Regardless of OpenShift being a scalable and flexible software, it can be a pain to monitor since complete visibility into the underlying operations is not guaranteed ... To effectively monitor an OpenShift environment, IT administrators should focus on these five key elements and their associated metrics ...