The infancy of software models and tools around generative AI and the huge requirements of energy and computation are moving this innovation away from on-premises platforms into the cloud says Priyanka Roy at ManageEngine.
IT has long been in a fascinating position within the organisation. Tasked with, as they say, keeping the lights on while also enabling innovation and modernisation, the IT department has always had its work cut out for it. The last few years, especially, could not have been easy, as we witnessed one wave of disruption after the other.
Be it blockchain or pandemic, grappling with disruptors has become an everyday job for IT folks. Now, there is a new disruptor on the block, breaking all the rules and changing the game as we know it: Generative AI.
In this ever-evolving realm of technology, amidst the relentless pursuit of operational excellence, cloud adoption has long been considered as a guaranteed path to gaining that much coveted competitive edge. Matters, however, took a turn for the worse during the past year as organisations slashed cloud budgets amidst worldwide economic slowdown, and cloud deals took almost 50% longer than normal to close.
But we may be able to anticipate a more favourable turn of events thanks to generative AI. As companies, big and small, fervently rush to capitalise on its benefits, this new-gen disruptor might just be able to revive the cloud market.
GenAI is not on-premises
If it is not clear already, what organisations are slowly coming to realise is that those who rely solely on on-premises technology might be left out of the generative AI revolution. To leverage the many benefits of generative technology effectively, the cloud is the way forward.
An undeniable significant change which levels the playing field for all kinds of businesses, the cloud’s scalability, flexibility, and cost-efficiency makes it the best platform for generative AI.
While this breakthrough technology is making headlines across the world, the fact remains that it is still in very nascent stages, which means that it is still evolving and going through multiple iterations every day. Take ChatGPT, for example.
When the conversational engine was launched, it was based on the generative pre-trained transformer, GPT 3. But we had GPT 2 before that, launched in early 2019, which was not even released to the public due to its potential to be misused. This reinforces the fact that new technologies of this nature are bound to be in a state of flux for the future.
Complexity, fast moving innovation
To further add to this mix, the science behind large language models is complicated itself. Languages like Python and Rust and frameworks like Tensorflow and Torch are not standard in most corporate environments, and when you put all of this together, you will see that what we have is new and complex software with many moving parts, which makes it an unlikely on-premises candidate.
Investing in something of this nature, which comes with no guarantees of long-term value, on-premises is not ideally the soundest decision for an organisation to make.
Before diving headfirst into the world of transformer technology, it is also vital to consider the financial implications. Building and running generative AI models requires specialised hardware investments, notably GPUs. Famous for their significant computational power, GPUs are expensive processors that are imperative if organisations want to run generative technology on-premises.
The downside to building such a niche infrastructure on-premises is that organisations will not always be using this computational power. Only when a generative-AI-specific service request comes in will this infrastructure be put into use. It also bears mentioning that organisations will have to incur the added costs of significantly higher energy consumption and cooling requirements.
Thus, the cost versus benefit analysis of such an investment looks quite bleak. From this perspective, the cloud model of paying only when you use the service appears to be a much more feasible option.
Large scale datasets, security
When you take the technical and financial challenges away from this mix, the cloud gives you the opportunity to enable pure innovation. Organisations do not have to worry about the infrastructure requirements, maintenance costs, or how they can keep it up and running; all they need to do is figure out what value they want to extract from generative AI and start running it in the cloud, thereby prioritising innovation.
However, one factor that organisations should be careful about is ensuring security.
Generative AI requires immense amounts of data to be rapidly processed in an environment which might not have clear boundaries or established controls. The nature and quality of data also needs to be monitored along with having clear policies around data jurisdiction.
This can cause unprecedented cybersecurity and data privacy gaps which organisations need to consider before embarking on their generative technology journey.
For example, if an organisation builds a model specific to a certain customer using their environment’s data, they need to ensure that the training data used for this customer’s requirement is not used for any other customer as that would lead to a privacy breach.
The good news is that most major cloud providers have well-established protocols in place regarding cybersecurity, data residency, jurisdiction, and more. With the advent of generative AI, well-known providers have already or are in the process of launching dedicated services for generative AI.
Generative AI seems well set to accelerate the next wave of cloud adoption. In many ways, this seems like a natural progression after the IT modernisation and innovation we have been witnessing in the last decade. Organisations which realised the long-term value of the cloud early on and laid the foundation will reap the benefits now, while those who were late adopters will have to step up their game.
While this new world holds limitless potential, we are yet to see where we will land. But one thing is for sure: The future is intelligent, and it is in the cloud.