How are large professional services organisation deploying GPT4 (or similar LLMs) within their own organisation for internal use, particularly when using sensitive/confidential/PII data? For example we're a UK based accountancy firm so have created our own internal app based on Microsoft's Open AI GPT4 model deployed to our private Azure hosting. We have built a custom app designed to "preprogramme" personas, i.e. pre-created prompts, for different areas of the profession, so internal users can have GPT4 act in different ways without prior knowledge of prompt engineering. I would be interested to see how others are building or buying solutions, particularly those based in EU/UK who can't use OpenAI/ChatGPT for with customer/confidential information.

4.3k viewscircle icon3 Upvotescircle icon3 Comments
Sort by:
Global Head of AI, Data & Analytics in Software2 years ago

Assume this question is a bit older, but you did outline a standard approach, host / access a model as a service internally through Azure OpenAI, Azure Machine Learning / VM, or Amazon Sagemaker, Amazon Bedrock
Though now Enterprise ChatGPT will make things easier if you don't care about federating models

Director of Other in Software2 years ago

We are building solutions and enable our customers to bring their own LLMs or integrate with any open and closed-source LLMs. In case like yours, we encourage our customers to integrate with Azure OpenAI to take advantage of Azure's enterprise security. 

Lightbulb on1
CEO in Services (non-Government)2 years ago

One of the answers is to use Microsoft Azure GPT-4 if we want GPT-4 kind of performance. Another option is to host open source models such as LLama2 in the cloud/in-premise environment and serve it for inference or consumption purposes. We see such requirements coming in from multiple clients. For one of the India Govt client for developer experience, we have hosted a code LLM and built visual studio plugin for developer to consume it. For a Italian financial client we are about to make this happen. This involve 1 time hardware cost if the client is looking for on-premise solution, for cloud based there will be recurrent GPU usage cost based on the number of requests that needs to be served.

Content you might like

Cost concerns21%

Fear of disruption32%

Lack of awareness 42%

Compatibility concerns33%

Lack of strategic alignment / vision28%

Productivity and downtime fears23%

Trusting on current implementation18%

New technologies aren't the basement of the innovation 2%

Other (use the comment)1%

View Results

Equal access to opportunity11%

Remote work; with enhanced infrastructure and technology to support48%

Better flexibility, mental health, and work/life balance27%

Humans + machines working together or intersecting further12%

Other (comment)

View Results