Skip to main content

The Rise of AI Will Actually Add to the Data Scientist's Plate

Sijie Guo
CEO
StreamNative

The data scientist was coined "the sexiest job of the 21st century" not that long ago. Harvard Business Review reported in 2012 that, "capitalizing on big data depends on hiring scarce data scientist." Fast forward to 2024, and we're in the era of generative Artificial Intelligence (AI) and large language models (LLMs) where one might assume that the role of data scientists would simplify or even diminish. Yet, the reality is quite the opposite. As AI becomes more prevalent across all industries, it's expanding the scope and responsibilities of data scientists, particularly in terms of building and managing real-time AI infrastructure.

Traditionally, data scientists focused primarily on analyzing existing datasets, deriving insights, and building predictive models. This included a unique skill set of communicating those findings to leaders within the organization and identifying strategic business recommendations based on their findings. Their toolbox typically included programming languages like Python and R, along with various statistical and machine learning (ML) techniques. The rise of AI is dramatically reshaping this landscape.

Today's data scientists are increasingly required to step beyond their traditional analytical roles. They're now tasked with designing and implementing the very infrastructure that powers AI systems. This shift is driven by the need for real-time data processing and analysis, which is critical for many AI applications.

Real-Time AI Infrastructure: A New Challenge

The demand for real-time AI capabilities is pushing data scientists to develop and manage infrastructure that can handle massive volumes of data in motion. This includes streaming data pipelines, edge computing, scalable cloud architecture, and data quality and governance. These new responsibilities require data scientists to expand their skill sets significantly; They now need to be well-versed in cloud technologies, distributed systems, and data engineering principles.

Organizations are increasingly recognizing the competitive advantage that real-time AI can provide. This is resulting in pressure on data science teams to deliver insights and predictions at unprecedented speeds. The ability to make split-second decisions based on current data is becoming crucial in many industries, from finance and healthcare to retail and manufacturing.

This shift towards real-time AI is not just about speed; it's about relevance and accuracy. By processing data as it's generated, organizations can respond to changes in their environment more quickly and make more informed decisions.

As data scientists take on these new challenges, they're no longer siloed in analytics departments, but instead are becoming integral to various aspects of business operations. This expansion of responsibilities includes:

1. Collaboration with IT and DevOps: Working closely with infrastructure teams to ensure AI systems are robust, scalable, and integrated with existing IT ecosystems.

2. Product Development: Embedding AI capabilities directly into products and services, requiring data scientists to work alongside product teams.

3. Ethical Considerations: Addressing the ethical implications of AI systems, including bias detection and mitigation in real-time environments.

The Emergence of DataOps Engineers

As the complexity of data ecosystems grows, a role has emerged to support data scientists: the DataOps Engineer. This role parallels the DevOps evolution in software development, focusing on creating and maintaining the infrastructure necessary for efficient data operations. DataOps Engineers bridge the gap between data engineering and data science, ensuring that data pipelines are robust, scalable, and capable of supporting advanced AI and analytics initiatives. Their emergence is a direct response to the increasing demands placed on data infrastructure by AI applications.

The rise of DataOps has significant implications for data scientists. In large enterprises, organizations with the resources to employ dedicated DataOps teams can significantly streamline their data pipelines. This allows data scientists to focus more on developing advanced models and extracting actionable insights, rather than getting bogged down in infrastructure management. Smaller companies, which may not have the budget for dedicated DataOps teams, often require data scientists to take on dual roles. This can naturally lead to bottlenecks, with data scientists dividing their time between infrastructure management and actual data analysis.

As a result of these changes, data scientists are now expected to have a broader skill set that includes proficiency in cloud infrastructure (AWS, Azure, GCP), an understanding of modern analytics tools, familiarity with data pipeline tools like Apache Spark and Hadoop, and knowledge of containerization and orchestration platforms like Kubernetes. While not all data scientists need to be experts in these areas, a basic understanding is becoming increasingly important for effective collaboration with DataOps teams and for navigating complex data ecosystems.

The Opportunity Ahead for Data Scientists

While AI is undoubtedly making certain aspects of data analysis more efficient, it's simultaneously expanding the role of data scientists in profound ways. The rise of AI is adding complexity to the data scientist's plate, requiring them to become architects of real-time AI infrastructure in addition to their traditional analytical roles.

This evolution presents both challenges and opportunities. Data scientists who can successfully navigate this changing landscape will be invaluable to their organizations, driving innovation and competitive advantage in the AI-driven future. The rise of AI isn't simplifying the role of data scientists — it's elevating it to new heights of importance and complexity, while also fostering the growth of supporting roles and teams.

Sijie Guo is CEO at StreamNative

Hot Topics

The Latest

Like most digital transformation shifts, organizations often prioritize productivity and leave security and observability to keep pace. This usually translates to both the mass implementation of new technology and fragmented monitoring and observability (M&O) tooling. In the era of AI and varied cloud architecture, a disparate observability function can be dangerous. IT teams will lack a complete picture of their IT environment, making it harder to diagnose issues while slowing down mean time to resolve (MTTR). In fact, according to recent data from the SolarWinds State of Monitoring & Observability Report, 77% of IT personnel said the lack of visibility across their on-prem and cloud architecture was an issue ...

In MEAN TIME TO INSIGHT Episode 23, Shamus McGillicuddy, VP of Research, Network Infrastructure and Operations, at EMA discusses the NetOps labor shortage ... 

Technology management is evolving, and in turn, so is the scope of FinOps. The FinOps Foundation recently updated their mission statement from "advancing the people who manage the value of cloud" to "advancing the people who manage the value of technology." This seemingly small change solidifies a larger evolution: FinOps practitioners have organically expanded to be focused on more than just cloud cost optimization. Today, FinOps teams are largely — and quickly — expanding their job descriptions, evolving into a critical function for managing the full value of technology ...

Enterprises are under pressure to scale AI quickly. Yet despite considerable investment, adoption continues to stall. One of the most overlooked reasons is vendor sprawl ... In reality, no organization deliberately sets out to create sprawling vendor ecosystems. More often, complexity accumulates over time through well-intentioned initiatives, such as enterprise-wide digital transformation efforts, point solutions, or decentralized sourcing strategies ...

Nearly every conversation about AI eventually circles back to compute. GPUs dominate the headlines while cloud platforms compete for workloads and model benchmarks drive investment decisions. But underneath that noise, a quieter infrastructure challenge is taking shape. The real bottleneck in enterprise AI is not processing power, it is the ability to store, manage and retrieve the relentless volumes of data that AI systems generate, consume and multiply ...

The 2026 Observability Survey from Grafana Labs paints a vivid picture of an industry maturing fast, where AI is welcomed with careful conditions, SaaS economics are reshaping spending decisions, complexity remains a defining challenge, and open standards continue to underpin it all ...

The observability industry has an evolving relationship with AI. We're not skeptics, but it's clear that trust in AI must be earned ... In Grafana Labs' annual Observability Survey, 92% said they see real value in AI surfacing anomalies before they cause downtime. Another 91% endorsed AI for forecasting and root cause analysis. So while the demand is there, customers need it to be trustworthy, as the survey also found that the practitioners most enthusiastic about AI are also the most insistent on explainability ...

In the modern enterprise, the conversation around AI has moved past skepticism toward a stage of active adoption. According to our 2026 State of IT Trends Report: The Human Side of Autonomous AI, nearly 90% of IT professionals view AI as a net positive, and this optimism is well-founded. We are seeing agentic AI move beyond simple automation to actively streamlining complex data insights and eliminating the manual toil that has long hindered innovation. However, as we integrate these autonomous agents into our ecosystems, the fundamental DNA of the IT role is evolving ...

AI workloads require an enormous amount of computing power ... What's also becoming abundantly clear is just how quickly AI's computing needs are leading to enterprise systems failure. According to Cockroach Labs' State of AI Infrastructure 2026 report, enterprise systems are much closer to failure than their organizations realize. The report ... suggests AI scale could cause widespread failures in as little as one year — making it a clear risk for business performance and reliability.

The quietest week your engineering team has ever had might also be its best. No alarms going off. No escalations. No frantic Teams or Slack threads at 2 a.m. Everything humming along exactly as it should. And somewhere in a leadership meeting, someone looks at the metrics dashboard, sees a flat line of incidents and says: "Seems like things are pretty calm over there. Do we really need all those people?" ... I've spent many years in engineering, and this pattern keeps repeating ...

The Rise of AI Will Actually Add to the Data Scientist's Plate

Sijie Guo
CEO
StreamNative

The data scientist was coined "the sexiest job of the 21st century" not that long ago. Harvard Business Review reported in 2012 that, "capitalizing on big data depends on hiring scarce data scientist." Fast forward to 2024, and we're in the era of generative Artificial Intelligence (AI) and large language models (LLMs) where one might assume that the role of data scientists would simplify or even diminish. Yet, the reality is quite the opposite. As AI becomes more prevalent across all industries, it's expanding the scope and responsibilities of data scientists, particularly in terms of building and managing real-time AI infrastructure.

Traditionally, data scientists focused primarily on analyzing existing datasets, deriving insights, and building predictive models. This included a unique skill set of communicating those findings to leaders within the organization and identifying strategic business recommendations based on their findings. Their toolbox typically included programming languages like Python and R, along with various statistical and machine learning (ML) techniques. The rise of AI is dramatically reshaping this landscape.

Today's data scientists are increasingly required to step beyond their traditional analytical roles. They're now tasked with designing and implementing the very infrastructure that powers AI systems. This shift is driven by the need for real-time data processing and analysis, which is critical for many AI applications.

Real-Time AI Infrastructure: A New Challenge

The demand for real-time AI capabilities is pushing data scientists to develop and manage infrastructure that can handle massive volumes of data in motion. This includes streaming data pipelines, edge computing, scalable cloud architecture, and data quality and governance. These new responsibilities require data scientists to expand their skill sets significantly; They now need to be well-versed in cloud technologies, distributed systems, and data engineering principles.

Organizations are increasingly recognizing the competitive advantage that real-time AI can provide. This is resulting in pressure on data science teams to deliver insights and predictions at unprecedented speeds. The ability to make split-second decisions based on current data is becoming crucial in many industries, from finance and healthcare to retail and manufacturing.

This shift towards real-time AI is not just about speed; it's about relevance and accuracy. By processing data as it's generated, organizations can respond to changes in their environment more quickly and make more informed decisions.

As data scientists take on these new challenges, they're no longer siloed in analytics departments, but instead are becoming integral to various aspects of business operations. This expansion of responsibilities includes:

1. Collaboration with IT and DevOps: Working closely with infrastructure teams to ensure AI systems are robust, scalable, and integrated with existing IT ecosystems.

2. Product Development: Embedding AI capabilities directly into products and services, requiring data scientists to work alongside product teams.

3. Ethical Considerations: Addressing the ethical implications of AI systems, including bias detection and mitigation in real-time environments.

The Emergence of DataOps Engineers

As the complexity of data ecosystems grows, a role has emerged to support data scientists: the DataOps Engineer. This role parallels the DevOps evolution in software development, focusing on creating and maintaining the infrastructure necessary for efficient data operations. DataOps Engineers bridge the gap between data engineering and data science, ensuring that data pipelines are robust, scalable, and capable of supporting advanced AI and analytics initiatives. Their emergence is a direct response to the increasing demands placed on data infrastructure by AI applications.

The rise of DataOps has significant implications for data scientists. In large enterprises, organizations with the resources to employ dedicated DataOps teams can significantly streamline their data pipelines. This allows data scientists to focus more on developing advanced models and extracting actionable insights, rather than getting bogged down in infrastructure management. Smaller companies, which may not have the budget for dedicated DataOps teams, often require data scientists to take on dual roles. This can naturally lead to bottlenecks, with data scientists dividing their time between infrastructure management and actual data analysis.

As a result of these changes, data scientists are now expected to have a broader skill set that includes proficiency in cloud infrastructure (AWS, Azure, GCP), an understanding of modern analytics tools, familiarity with data pipeline tools like Apache Spark and Hadoop, and knowledge of containerization and orchestration platforms like Kubernetes. While not all data scientists need to be experts in these areas, a basic understanding is becoming increasingly important for effective collaboration with DataOps teams and for navigating complex data ecosystems.

The Opportunity Ahead for Data Scientists

While AI is undoubtedly making certain aspects of data analysis more efficient, it's simultaneously expanding the role of data scientists in profound ways. The rise of AI is adding complexity to the data scientist's plate, requiring them to become architects of real-time AI infrastructure in addition to their traditional analytical roles.

This evolution presents both challenges and opportunities. Data scientists who can successfully navigate this changing landscape will be invaluable to their organizations, driving innovation and competitive advantage in the AI-driven future. The rise of AI isn't simplifying the role of data scientists — it's elevating it to new heights of importance and complexity, while also fostering the growth of supporting roles and teams.

Sijie Guo is CEO at StreamNative

Hot Topics

The Latest

Like most digital transformation shifts, organizations often prioritize productivity and leave security and observability to keep pace. This usually translates to both the mass implementation of new technology and fragmented monitoring and observability (M&O) tooling. In the era of AI and varied cloud architecture, a disparate observability function can be dangerous. IT teams will lack a complete picture of their IT environment, making it harder to diagnose issues while slowing down mean time to resolve (MTTR). In fact, according to recent data from the SolarWinds State of Monitoring & Observability Report, 77% of IT personnel said the lack of visibility across their on-prem and cloud architecture was an issue ...

In MEAN TIME TO INSIGHT Episode 23, Shamus McGillicuddy, VP of Research, Network Infrastructure and Operations, at EMA discusses the NetOps labor shortage ... 

Technology management is evolving, and in turn, so is the scope of FinOps. The FinOps Foundation recently updated their mission statement from "advancing the people who manage the value of cloud" to "advancing the people who manage the value of technology." This seemingly small change solidifies a larger evolution: FinOps practitioners have organically expanded to be focused on more than just cloud cost optimization. Today, FinOps teams are largely — and quickly — expanding their job descriptions, evolving into a critical function for managing the full value of technology ...

Enterprises are under pressure to scale AI quickly. Yet despite considerable investment, adoption continues to stall. One of the most overlooked reasons is vendor sprawl ... In reality, no organization deliberately sets out to create sprawling vendor ecosystems. More often, complexity accumulates over time through well-intentioned initiatives, such as enterprise-wide digital transformation efforts, point solutions, or decentralized sourcing strategies ...

Nearly every conversation about AI eventually circles back to compute. GPUs dominate the headlines while cloud platforms compete for workloads and model benchmarks drive investment decisions. But underneath that noise, a quieter infrastructure challenge is taking shape. The real bottleneck in enterprise AI is not processing power, it is the ability to store, manage and retrieve the relentless volumes of data that AI systems generate, consume and multiply ...

The 2026 Observability Survey from Grafana Labs paints a vivid picture of an industry maturing fast, where AI is welcomed with careful conditions, SaaS economics are reshaping spending decisions, complexity remains a defining challenge, and open standards continue to underpin it all ...

The observability industry has an evolving relationship with AI. We're not skeptics, but it's clear that trust in AI must be earned ... In Grafana Labs' annual Observability Survey, 92% said they see real value in AI surfacing anomalies before they cause downtime. Another 91% endorsed AI for forecasting and root cause analysis. So while the demand is there, customers need it to be trustworthy, as the survey also found that the practitioners most enthusiastic about AI are also the most insistent on explainability ...

In the modern enterprise, the conversation around AI has moved past skepticism toward a stage of active adoption. According to our 2026 State of IT Trends Report: The Human Side of Autonomous AI, nearly 90% of IT professionals view AI as a net positive, and this optimism is well-founded. We are seeing agentic AI move beyond simple automation to actively streamlining complex data insights and eliminating the manual toil that has long hindered innovation. However, as we integrate these autonomous agents into our ecosystems, the fundamental DNA of the IT role is evolving ...

AI workloads require an enormous amount of computing power ... What's also becoming abundantly clear is just how quickly AI's computing needs are leading to enterprise systems failure. According to Cockroach Labs' State of AI Infrastructure 2026 report, enterprise systems are much closer to failure than their organizations realize. The report ... suggests AI scale could cause widespread failures in as little as one year — making it a clear risk for business performance and reliability.

The quietest week your engineering team has ever had might also be its best. No alarms going off. No escalations. No frantic Teams or Slack threads at 2 a.m. Everything humming along exactly as it should. And somewhere in a leadership meeting, someone looks at the metrics dashboard, sees a flat line of incidents and says: "Seems like things are pretty calm over there. Do we really need all those people?" ... I've spent many years in engineering, and this pattern keeps repeating ...