With a focus on GenAI, industry experts offer predictions on how AI will evolve and impact IT and business in 2024. Part 3 covers the technologies that will drive AI.
Start with: 2024 AI Predictions - Part 1
Start with: 2024 AI Predictions - Part 2
Go to: predictions about AIOps
The promise of greater computing power is on businesses doorsteps and, in 2024, has the potential to take digital investments to the next level. Case in point: the question on every leader's mind in 2023 was: how soon will I see the return on my AI investment? The answer may lie in quantum computing, which enables computer-based applications to produce reliable results even with smaller data sets. In 2024, we will see greater progress made to allow these applications to perform far more complex operations and help solve previously "unsolvable" problems faster than conventional solutions.
Global AI and US Innovation Technology Leader, PwC
The only way to scale AI will be to distribute it, with the help of edge computing. I predict that the convergence of edge and cloud AI is the way to deliver AI at scale with the cloud and edge offloading computational tasks to the other side as needed. For instance, the edge can handle model inferences while the cloud may handle model training or the edge may offload queries to the cloud depending on the length of a prompt and so on. When it comes to a successful AI strategy, it's not practical to have a cloud-only approach. Companies need to consider an edge computing strategy — in tandem with the cloud — to enable low-latency, real-time AI predictions in a cost effective way without compromising on data privacy and sovereignty.
Director of Product Management, Couchbase
As generative AI models are trained and use cases expand, in 2024 we will enter the next generation of edge and scaled computing through the demands of inference (putting the generative AI models to work locally). AI-driven supercomputing environments and databases will not stop or get smaller as they require 24/7 runtime, but they may seek to be closer to the end user in contrast to the large model training locations that can be asynchronously located. While this evolution is certainly something to watch, the next generation computing at the edge is still underway and likely another one to two years from materializing and understanding what it will actually look like, but we believe modern, state of the art power dense data centers will be required to support.
SVP of Strategy, EdgeCore Digital Infrastructure
Neuromorphic computing has the potential to launch AI's capabilities outside the realm of traditional computing's binary principles. But before the technology can be actualized for business use, professionals and researchers must band together to design hardware and software standards for the technology. In 2024, we will see businesses upskilling key employees and connecting with researchers to align on enterprise applications. This foundation will be critical for businesses to uplevel AI applications in coming years.
Global AI and US Innovation Technology Leader, PwC
Retrieval Augmented Generation
RAG (Retrieval Augmented Generation) will be the rage in 2024. Companies will take foundational LLMs (Large Language Models), train and tune them with their own data, and churn out chatbots, copilots and similar utilities for external and internal productivity gains.
Mark Van de Wiel
Field CTO, Fivetran
The hallucination problem will be largely solved, removing a major impediment to the adoption of generative AI. In 2023, "hallucinations" by large language models were cited as a major barrier to adoption. If generative AI models can simply invent facts, how can they be trusted in enterprise settings? I predict however, that several technical advances will all but eliminate hallucinations as an issue. One such innovation is Retrieval Augmented Generation (RAG), which primes the large language model with true, contextually relevant information right before prompting it with a user query. This technique, still in its infancy, has been shown to dramatically decrease hallucinations and is already making waves. According to a recent State of LLM Apps 2023 report, 20% of the apps use vector retrieval — a key ingredient of RAG. In the coming year, I predict that RAG, along with better trained models, will begin to rapidly solve the hallucination problem — ultimately paving the way for widespread adoption of generative AI within enterprise workflows.
Director of Product Management and Head of Streamlit, Snowflake
There is going to be a rapid shift from infrastructure-based GenAI to local GenAI because right now, that's not really possible. The average startup doesn't have thousands of dollars to throw at a cloud provider and it will prove almost impossible to run by yourself but that is changing quickly with the innovation around local generative AI. With it going local, you will have a complete RAG stack under your control with your access controls. That way, you won't have to expose your proprietary data in any way. When we go from centralized, API-based LLMs to local LLMs, it will happen quickly. The ones that will work will be adopted like wildfire. Just be mindful of the downside as de-centralized LLMs introduce the concept of bad actors in the loop.
VP of Developer Relations, DataStax
LARGE GRAPHICAL MODELS
LGMs become the next household gen AI tech in the enterprise — Today, nearly every organization is experimenting with LLMs in some way. Next year, another major AI technology will emerge alongside LLMs: Large Graphical Models (LGMs). An LGM is a probabilistic model that uses a graph to represent the conditional dependence structure between a set of random variables. LGMs are probabilistic in nature, aiming to capture the entire joint distribution between all variables of interest. They are particularly suitable for modeling tabular data, such as data found in spreadsheets or tables. LGMs are useful for analyzing time series data. By analyzing time series data through the novel lens of tabular data, LGMs are able to forecast critical business trends, such as sales, inventory levels and supply chain performance. These insights help guide enterprises to make better decisions. This is game changing because existing AI models have not adequately addressed the challenge of analyzing tabular, time-series data (which accounts for the majority of enterprise data). Instead, LLMs and other models were created to analyze text documents. That's limited the enterprise use cases they're really capable of supporting: LLMs are great for building chatbots, but they're not designed to support detailed predictions and forecasting. Those sort of use cases offer organizations the most business value today — and LGMs are the only technology that enables them. Enterprises already have tons of time-series data, so it'll be easy for them to begin getting value from LGMs. As a result, in 2024, LGM adoption will take off, particularly in retail and healthcare.
Co-CEO and Co-Founder, Ikigai Labs and AI Professor at MIT
Explainable AI will play a key role in the broader acceptance and trust of AI systems as adoption continues to increase. The next frontier in AI for physical operations lies in the synergy between AI, IoT, and real-time insights across a diversity of data. In 2024, we'll see substantial advancements in predictive maintenance, real-time monitoring, and workflow automation. We may also begin to see multimodal foundation models that combine not just text and images, but equipment diagnostics, sensor data, and other sources from the field. As leaders seek new ways to gain deeper insights into model predictions and modernize their tech stack, I expect organizations to become more interested in explainable AI (XAI). XAI is essential for earning trust among AI users — it sheds light on the black-box nature of AI systems by providing deeper insights into model predictions and it will afford users a better understanding of how their AI systems are interacting with their data. Ultimately, this will foster a greater sense of reliability and predictability. In the context of AI Assistants, XAI will reveal more of the decision-making process and empower users to better steer the agent toward desired behaviors. In the new year, I anticipate XAI will advance both the functionality of AI Agents and the trust of AI systems.
Head of AI and Data, Samsara
OPEN SOURCE TOOLS
LLM providers will increase data transparency and embrace open-source tools. The term "open source" has evolved over time into something differentiated from its initial meaning. Traditionally, open-source tools gave insight into code used to build something. Now, with the rapid emergence of LLMs, we're able to accelerate past the code and get right to the end result without needing full observability of how a model was produced. This is where we see fine-tuning of models using domain-specific data for models being critical, and I foresee companies putting a concerted effort around transparency around how open-source models are trained heading into 2024. In my opinion, LLMs are easier to build than one may think given the availability of open source tools. That said, the important element that cannot be overlooked for building generative AI models is understanding how you uplevel the model to be specific to each use case, and giving users visibility into where training data was derived and how it'll be used in the future. One way to ensure data transparency is to encourage your team to embrace data lineage best practices, which I think we'll see become more of a norm heading into 2024. Incorporating data lineage into your data pipelines allows for heightened tracking and observability, catalyzing the potential for increased transparency.
Go to: 2024 AI Predictions - Part 4, covering AI challenges.
Generative AI will usher in advantages within various industries. However, the technology is still nascent, and according to the recent Dynatrace survey there are many challenges and risks that organizations need to overcome to use this technology effectively ...
In today's digital era, monitoring and observability are indispensable in software and application development. Their efficacy lies in empowering developers to swiftly identify and address issues, enhance performance, and deliver flawless user experiences. Achieving these objectives requires meticulous planning, strategic implementation, and consistent ongoing maintenance. In this blog, we're sharing our five best practices to fortify your approach to application performance monitoring (APM) and observability ...
In MEAN TIME TO INSIGHT Episode 3, Shamus McGillicuddy, VP of Research, Network Infrastructure and Operations, at Enterprise Management Associates (EMA) discusses network security with Chris Steffen, VP of Research Covering Information Security, Risk, and Compliance Management at EMA ...
In a time where we're constantly bombarded with new buzzwords and technological advancements, it can be challenging for businesses to determine what is real, what is useful, and what they truly need. Over the years, we've witnessed the rise and fall of various tech trends, such as the promises (and fears) of AI becoming sentient and replacing humans to the declaration that data is the new oil. At the end of the day, one fundamental question remains: How can companies navigate through the tech buzz and make informed decisions for their future? ...
We increasingly see companies using their observability data to support security use cases. It's not entirely surprising given the challenges that organizations have with legacy SIEMs. We wanted to dig into this evolving intersection of security and observability, so we surveyed 500 security professionals — 40% of whom were either CISOs or CSOs — for our inaugural State of Security Observability report ...