Blog

Home / Blog

The looming battle over where generative ai systems will run

Jason Li
Sr. Software Development Engineer
Skilled Angular and .NET developer, team leader for a healthcare insurance company.
September 03, 2023


AI on cloud versus on-premises systems

Although deciding whether to use on-premises systems or the cloud for AI may appear straightforward, it's actually far more difficult. It is increasingly clear that decisions are still being made about where the majority of generative AI systems will be housed (public cloud platforms versus on-premises and edge-based platforms). Numerous articles make the point that on-premises and public cloud platforms for AI systems are competing head-to-head. This is predicated on the idea that using the public cloud carries some risk, such as IP leakage or when the rival draws stronger conclusions from your data.

In addition, businesses still store a lot of data on edge computing or in conventional data centres rather than on the cloud. Data silos are widespread in most modern businesses, which can create issues if the data cannot be moved quickly to the cloud. Hosting the AI systems closest to the data may make sense because data is necessary for AI systems to be useful.

Many scholars contend that silos of data should not exist and that by doing so, you are facilitating an already existing issue. However, given the expense of correcting such problems, many businesses might not have other, more practical options. Most businesses prioritise generative AI, even if it means operating on suboptimal infrastructure that they are hesitant to modify or cannot afford. This implies that generative AI may cause many organisations to incur yet another layer of technical debt.

In the early days of cloud computing, businesses made many of the same errors. However, instead of storing items on the cloud this time, businesses are storing them inside data centres. Another issue is when programmes and data are moved to the cloud without sufficient planning and preparation. You end up with underoptimized solutions at both extremes.

Many benefits of the cloud cannot be matched by conventional legacy platforms. The accessibility of tools and technology on public clouds, as well as the speed at which these solutions may be deployed, cannot be matched. They already excel at generative AI, and they have the infrastructure to scale and change with new technology.

In the early days of cloud computing, businesses made many of the same errors. However, instead of storing items on the cloud this time, businesses are storing them inside data centres. Another issue is when programmes and data are moved to the cloud without sufficient planning and preparation. You end up with underoptimized solutions at both extremes.

Many benefits of the cloud cannot be matched by conventional legacy platforms. The accessibility of tools and technology on public clouds, as well as the speed at which these solutions may be deployed, cannot be matched. They already excel at generative AI, and they have the infrastructure to scale and change with new technology.

On public clouds, maintaining these platforms is someone else's responsibility. This is another rack of servers that does all the horrible things that physical servers you own and control do, despite the fact that the majority of businesses already have support. On-premises systems, however, might cost half as much as public cloud platforms, much like the increase in repatriations, if the generative AI systems are actually collocated with the training data and the use of that data is going to be pretty simple to forecast.

It varies

Generative AI systems are primarily designed with specific tasks in mind, such as intelligently processing supply chains, automating repetitive manual labour to decrease staff, providing marketing insight, etc. The characteristics of that generative AI system and the type of problem you're trying to address largely determine where the systems operate. That response is unpopular, but it is accurate. It is similar to every other system you develop and deploy in many ways.

We do become worried about on-premises or cloud assumptions that are only occasionally accurate. We don't know if "IP leakage" is actually more dangerous in the cloud because many essential systems, like security, operations, and scalability, perform better there. Public clouds can be more expensive than on-premises systems, but they may be a good fit depending on the use case. Pre-solving issues ("cloud-only" or "cloud-never") without fully comprehending the scenario has before caused us troubles. With brand-new generative AI systems, we keep committing the same blunders.

Where should you host your AI applications: on-premise or in the cloud?

Similar to selecting between buying and renting a home, deploying AI applications requires a decision between on-premise and cloud infrastructure. To make the best decision on where to host AI applications, it is crucial to understand every aspect of cloud vs. on-premise implementation.

Since most of its earlier problems have been resolved, a whole new generation of sophisticated AI applications has emerged. Because AI applications are being used by businesses of all sizes, the rapid rise of AI technology is becoming apparent. The question of whether to install applications in the cloud or on-premises will become increasingly important in the thoughts of business leaders as AI develops further and more customers start to request and use its applications. Pay as you go is a viable alternative with cloud hosting. On-premise hosting, however, offers more freedom because there is only one fee and you own the gear. Cloud computing does appear to be the greatest option for hosting AI applications due to the pay as you go option's availability. Both cloud and on-premise hosting have advantages and disadvantages, but a company's needs are the deciding factor in where to host its AI applications.

AI Applictions: On-Premise vs. Cloud Hosting

Like renting a house, cloud hosting. AI applications may be used for however long the contract specifies. Additionally, the hosting company is in charge of maintaining the hardware. In contrast, on-premise hosting is comparable to purchasing a house. If the business needs it, the application may remain on the hardware. The following are a few of the many considerations when choosing between the two:

Cost

Massive computer power is needed to build machine learning algorithms and train neural networks. Additionally, in order for AI networks to learn and enhance the effectiveness of their services, as well as to keep operation-critical data current, updates are occasionally required. The cost to develop and maintain an AI application may increase when new neural networks and algorithms are added.

On-premise hosting can ultimately lower costs in the long run, while being expensive at the time of adoption. AI application deployment on-premises will do away with the requirement for contract term renewals. With an on-premise cloud built around hardware storage, businesses can deploy as many AI applications as they require. A slight error in the hardware or software implementation can result in a significant cost since on-premise hosting demands a full one-time payment at the beginning.

Both small and large businesses can benefit from the reduced implementation costs that cloud services provide when experimenting with the use of AI applications. With cloud hosting, there is no need for up-front hardware or software installation or maintenance costs because the hosting provider takes care of everything. The problem with adopting cloud services is that the cost can go up over time because contracts must be renewed and software licencing can offset hardware savings.

Scalability

Enterprises require scalable technology and software that can be readily upgraded and efficiently managed in order to keep up with the changing needs of customers and business requirements.

With full control over the hardware provided by on-premise hosting, the company's administrators could exercise close control over its updates. Because the hardware is scalable, businesses can add additional AI applications to run on the same hardware. However, a deployment strategy that has been properly thought out in advance would be needed to add another software or enhance an existing software. It takes a lot of time to gather the necessary data, develop or upgrade the software, and deploy the software.

Companies adopting on-premise deployment may fall behind if no advance planning is made because they won't be able to react fast to unexpected changes in demand. Cloud resources, in contrast to on-premise hosting, can be quickly modified to satisfy particular needs. However, when employing public cloud services, the hardware stacks have too much software clutter, which lowers the scalability. Although the software offered by cloud providers is often customisable, some organisations may want further customization.

Security

An enterprise's data collection efforts may yield important information about their clients or rival companies. A business's reputation could be damaged, clients might start to doubt its dependability, and there could be other serious negative effects from losing important data. Therefore, regardless of where they are placed, every organisation wants their data to be secure.

Enterprises have complete control over their locally stored data thanks to on-premise hosting. No third party can access the data because it is kept on the enterprise's property, barring hacking. The data would be more secure because only employees of that company would understand how the gear works.

On-premise AI application data security demands a team of committed and knowledgeable workers. The firm runs serious risks with the data if it lacks the necessary resources. Data security is the responsibility of the hosting companies; the organisation is not liable for it. To prevent breaches, hosting companies always keep their systems up to date and have the data secured. The hosting companies in the case of centralised computing or anyone in the case of decentralised computing can access the data stored in clouds. The business might not be aware of where and how frequently its data is backed up. Hackers are also actively after cloud data.

Conclusion

The decision about whether to host generative AI systems on-premises or in the cloud comes down to a complicated interplay of elements including cost, scalability, and security. Organisations must base this choice on their unique demands for AI applications rather than using a one-size-fits-all strategy. Making hasty decisions might result in less than ideal results, highlighting the value of making well-informed choices in an environment where AI is constantly changing.