AWS AI for Startups – The Early Innings

We recently hosted a call with Karthik Bharathy, Director of AI/ML services at AWS, who spoke on recent strategic announcements and AWS capabilities – specifically on how AWS’ AI/ML offerings look, how you can leverage them, and how to best work with the AWS team as a startup.

Key Takeaways

  • There are four key considerations for startups when spinning up ML capabilities:
    • What is the easiest way for your company to build?
    • How can you differentiate with your data?
    • How can you get the highest performance infrastructure at the lowest cost?
    • How can applications and services vastly improve your productivity?
  • Amazon Bedrock is a fully managed service that offers high performing foundation models from companies like AI21 Labs, Cohere, Anthropic, Stability AI, etc. including foundation models coming from Amazon (Titan models).
  • Foundation models are just one piece of the puzzle. From a process perspective, there’s a lot more to the orchestration piece. Users typically want a task to be accomplished (vs. just interacting with data). There’s a process in and of itself which involves more than just interaction with the model – you must also interact with the data, a bunch of APIs on the back-end, and so on
  • Differentiating with your data is key – It’s pretty evident, but while foundation models can do a lot of things out of the box, their impact is vastly amplified once they are fine-tuned with your data sources.
  • Getting Educated – Amazon has heavily leveraged Coursera to provide a comprehensive suite of offerings on the topic. You can choose the service that you want to learn about + your role in your organization, and it provides several resources to learn more. And should you have specific asks on a Bedrock Workshop, or you want to learn about Sagemaker Jumpstart, AWS has hands-on workshops where their specialists will engage with you and help set things up.

The Tipping Point for Generative AI

At AWS, the primary goal today is figuring out how the business can take advantage of generative AI, beyond just simple text and chat use cases. What led to this moment was the massive proliferation of data and compute: it became available at low cost and a very large scale. So, machine learning has experienced a ton of innovation over the 2-3 years, which has accelerated the prior efforts tremendously.

Generative AI is a fundamental paradigm shift from the AI of the past – you’re trying to generate new content, powered by foundation models, which are pre-trained ML models leveraging vast amounts of data. What’s important is that these foundation models can be customized for specific uses cases, giving them more power and relevance

From an AWS standpoint, ML innovation is in their DNA – going all the way back to e-commerce recommendations, or picking the routes where packages can be stored, Alexa, Prime Air, or even Amazon Go where you have a physical retail experience. These products already incorporate machine learning and are backed by foundation models

Today, there are over 100K customers across different geographies and verticals, all using AWS for machine learning, and all of these customers are already in production.

There are four key considerations for startups when spinning up ML capabilities:

  • How should you be building? What is the easiest way for your company to build – this probably comes down to where are you on your ML journey as a startup
  • How can you differentiate with your data? What are the specific capabilities or functionalities you’re looking at? You want to customize what’s already out there so that you can get the best of what’s already being provided
  • How can you get the highest performance infrastructure at the lowest cost? This applies to startups building their own foundational models, as well as those looking at fine-tuning existing models and leveraging the infrastructure for their application(s)
  • How can applications and services vastly improve your productivity?

Easiest Way to Build / Getting Started with Amazon Bedrock

Amazon Bedrock is a fully managed service that offers high performing foundation models from companies like AI21 Labs, Cohere, Anthropic, Stability AI, etc. including foundation models coming from Amazon (Titan models).

This is a serverless offering, meaning you don’t have to manage any infrastructure when you access these models – all you have to do is use APIs to interact with them. You can also customize these models with your own data (e.g. fine-tuning or RAG)

You have the choice of taking advantage of an on-demand mode where you look at the input and output tokens, and then essentially index on what your pricing will be, so that you can project based on your current application needs (vs. future application needs). Whether this is coming from NVIDIA, AWS Inferentia, etc. – it’s all under the hood – you’re not exposed to instances. There are even capabilities like instance recommender that can suggest what’s the best instance for a given model – it really just comes down to the use case

There are many different models today, and this list will continue to expand. You can try them all in a sandbox environment via the AWS console.

The Bedrock service is HIPPA compliant so you can use the models in GA along with Bedrock for your production use cases

Getting started with Bedrock:

  • Head to the AWS console and choose which foundation model you want to start with (there are many predefined templates that let you get started on the set of prompts that you can issue)
  • Once you narrow down your choice on a given foundation model from a provider, you can use prompt engineering to get the best output POS, or, you can fine-tune the model and create something that’s very specific to your use-case
  • Once you have the right response, you can have that model deployed in your environment, and there are options on how to deploy it
  • Finally, you can integrate it with the rest of your gen AI application

In terms of pricing there are three different dimensions:

  • On-Demand – Pay as you go, no usage commitments. You’re charged based on how many input tokens are processed and how many output tokens are generated. A token is a sequence of characters, and you pay based on the number of tokens that have been processed
  • Provision Throughput – In some cases, where you want to refer to an earlier question and you need a certain level of performance (consistent inference workloads), you will get pro throughput guarantees on when a model is available for inference, and then the pricing dimension is based on the number of model units
  • Fine-Tuning – You will be charged for fine-tuning the model, which is based off of a training Epoch, this is nothing but a run that’s trying to fine-tune the model based on a given dataset. And there are of course storage charges for the new model that’s being generated with fine-tuning. Additionally, fine-tuning fine-tuned models requires a provision throughput which is more of a dollars per hour cost

If you’re an ML practitioner who wants to try out an open-source model, and have access to the instances, you can use Sagemaker JumpStart. This is a model hub where you can access HuggingFace models (or other models) directly within Sagemarker and deploy them to an inference endpoint. Fine-tuning is different from the Bedrock experience – you work via a notebook where you make changes to the model in a more hands-on way, so if you’re an ML practitioner who is very familiar with the different techniques of fine-tuning, it gives you a lot of knobs and flexibility on how you can build, train, and deploy models on Sagemarker

Foundation models, however, are just one piece of the puzzle. From a process perspective, there’s a lot more to the orchestration piece. Users typically want a task to be accomplished (vs. just interacting with data), so if you have an application, but you just want to book your vacation, that’s going to involve a series of steps (e.g. understanding the different prices, selecting the different options, etc.). So, it’s a process in and of itself and that involves more than just interaction with the model – you must also interact with the data, and a bunch of APIs on the back-end, and so on. And at the same time, you want to ensure that security is tight, because while there’s orchestration, you also want to meet your enterprise cloud policies. This can take a number of weeks if you do this on your own, and Amazon Bedrock has just announced “Agents” to make this a lot simpler

How do Amazon Bedrock Agents work to enable generative AI applications to complete tasks in just a few clicks?

  • Breaks down and orchestrates tasks – You can break down a set of activities into different tasks, and you can then orchestrate them within the Bedrock environment
  • Securely accesses and retrieves company data – You can connect your company’s data sources, convert them into a format that can be understood by the model, and have relevant outputs coming off the application
  • Takes action by executing API calls on your behalf – You have the capability to execute APIs on behalf of your application, and orchestrate the entire end to end flow
  • Provides fully managed infrastructure support – You can complete entire tasks in a matter of clicks rather than composing an entire end-to-end application

Differentiate with your Data

Differentiating with your data is key – It’s pretty evident, but while foundation models can do a lot of things out of the box, their impact is vastly amplified once they are fine-tuned with your data sources. So, net data is your differentiator. You’ll get higher accuracy for the specific tasks that you’re after. All you need to do is point to the repository of your custom data, then point that to the foundation models. The foundation models will do the training run and produce a fine-tuned version of the model that you can use in your application

The customer data you provide to fine-tune the model is only used for your own, newly made, fine-tuned model. It won’t ever be used to improve the underlying model that Amazon is providing. AWS can’t access your data and don’t intend to use it for improving their own service. Everything is being generated in your VPC, so there are enough guardrails in place on who can access the data or the model. In fact, whenever a model is being used, e.g. a proprietary model, the model weights are protected so that consumers of the model don’t get access to the model. At the same time, the data that’s being used to fine-tune the model is not available to AWS or the model provider

It’s important to have a comprehensive data strategy that augments your gen AI applications. You have a variety of different data sources and in the case of structured or vector data, in some cases you may want to label the data. There are services in AWS which can be used for labeling the data which will give you more accurate results when you fine-tune your model. Then of course, you also may need to integrate multiple datasets. There are capabilities for ETL, so you can connect all your different data sources. Data and ML governance are also available as you build out your application

Most Performant, Low Cost

When you’re trying to run foundation models, it’s important that you have the more performant infrastructure at the lowest cost. AWS silicon supports many of the foundation models – you have the choice of using the GPUs when it comes to hosting and training foundation models like the A100s and the H100s, there is also custom silicon, AWS Inferentia 2, and Tranium available

  • Performance – Optimized infrastructure for training large models using model parallelism and data parallelism. SageMaker Distributed Training is up to 40
  • Resilience – AWS also offers resiliency against hardware failure, which becomes a big concern when you want to minimize downtime and focus purely on model-building. There is monitoring of failed nodes – how do you replace them? This can save up to 21% in model training time
  • Cost Efficiency – Dynamically run multiple models on GPU-based instances using multi-model endpoints to reduce cost. Cost efficiency is so important because GPUs are at a premium – and you want to get maximum utilization of the instances that you’re using
  • Responsible AI – When it comes to detecting buyers, explaining predictions, providing alerts when there is model drift, etc. – there are capabilities available today that you can take advantage of. AWS also offers ML governance, which lets you enforce a set of standard policies that apply to all of your data scientists and devs in the organization (who can build a model, who can deploy a model, etc.). There’s a model dashboard that provides you with all these key metrics

Increase Productivity

Another fantastic way to leverage Amazon, is via targeted applications that enhance productivity. These are a few popular options:

  • Amazon Quicksight – New FM-powered capabilities for business users to extract insights, collaborate, and visualize data. Author, fine-tune, and add visuals to dashboards (e.g. you write in natural language to generate a sales forecast and add that goes directly to your dashboard)
  • AWS HealthScribe – This is specific to the healthcare vertical, it’s a HIPAA-eligible service that’s targeted towards healthcare vendors for building clinical applications. The service can automatically generate clinical notes by analyzing patient and clinician convos. You can validate the notes that you pointed to and have it generate a summary for you. It also supports responsible AI by including references to the original patient transcripts for anything that’s generated by AI
  • Amazon CodeWhisperer – This is targeted towards app developers who are writing undifferentiated code. It can automatically generate the code, allowing developers to focus on the creative aspects of coding. Whisperer integrates with all the popular IDUs including Sagemaker Studio

Learn the Fundamentals of Generative AI for Real-World Applications

Amazon has heavily leveraged Coursera to provide a comprehensive suite of offerings on the topic.

You can choose the service that you want to learn about + your role in your organization, and it gives you a bunch of resources to learn more. And should you have specific asks on a Bedrock Workshop or you want to learn about Sagemaker Jumpstart, AWS has hands-on workshops where their specialists will engage with you and help set things up. AWS is also investing a large amount of money into their generative AI innovation center, which will connect you with ML and AI experts, so that if you have an idea, they can help you transform that into a generative AI solution

Karthik is an product and engineering leader with experience in product management, product strategy, execution and launch. He manages a team of product managers, TPMs, UX designers, Business Development Managers, Solution Specialists and engineers on Amazon SageMaker. He has incubated and grown several businesses including Amazon Neptune at AWS, PowerApps, Windows Azure BizTalk Services, SQL Server at Microsoft and shipped across all layers of the technology stack – graphs, relational databases, middleware, and low-code/no-code app development.

Karthik holds a masters degree in business and an undergraduate in computer engineering.

Srini Koushik, President of Technology and Sustainability, Rackspace

Today we welcome Srini Koushik, Chief Technology Officer for Rackspace Technology to our CXO of the Future Podcast. He is currently leading a significant new initiative at Rackspace. However, he began his career in healthcare and IBM, and also conducted research at NTT.

Question 1: First Job: What was your first job, and how did it help you build your career?

Well, I think it’s actually a case of following your passion. I have always been someone who likes to solve problems, and what comes with problem-solving is curiosity. I was always one of those guys growing up who wanted to know how things worked when I saw something happening. I wasn’t just caught up with the outcome: I would wonder, ‘How did they do that?’ This curiosity has been a common theme throughout my career.

Technology was the right place for me. It was the mid-eighties, and I was in India. India hadn’t gone through a lot of the technological revolution that you’ve seen over the last 20 or 30 years. We were learning on punch cards and COBOL, knockoffs of IBM mainframes that we got from the Russians. It was fundamentally about how you learn, but it taught you quite a lot. I had to learn how to bootstrap a TDC 316, which was a box 11 machine tape reader. It made me wonder, ‘Why do they call that a bootstrap?’ So, you start looking at those types of things. Technology was almost a natural fit for what I like to do. If I were to trace one thread from the time I started to today, it would be that I’m a technologist at my core. I love technology and I love solving problems using technology. If someone asked me to define my core competency with a gun to my head, that’s it. Everything in my career after that has been adjacencies that I’ve picked up to be able to function in multiple roles.

I began my career by delving into the realm of writing compilers and developing operating systems—essentially, software. This was where my journey began. One of my earliest mentors was Nick Donofrio, who held the position of CTO at IBM at that time. Nick was an extraordinary individual, always taking technologies under his wing and nurturing them. He imparted to me a valuable perspective—that technology is like a trade; you need to consider its practical application for people. If it doesn’t serve a purpose, what’s the point? Nick initiated programs that transitioned people from research into client engagements. He made significant contributions during the Gerstner era at IBM, and I was fortunate to benefit from his guidance. Consequently, I transitioned from being a hardcore technologist to tackling some of the most challenging problems. These challenges ranged from ensuring the scalability of IBM’s pioneering U.S. Open website during the late nineties to handling a myriad of transactions. Alternatively, it involved assisting customers like X, Y, and Z, who had signed major Y2K deals and were in dire straits.

Looking back, I realize that despite initially questioning ‘why me?’ during those moments, I’m grateful for the opportunities that were presented to me. They played a crucial role in shaping
who I am today. My message to my children—and anyone else—is that in the midst of challenges, we may question our circumstances, wondering why we were chosen. However, with time, we
come to appreciate the experiences that have defined us. IBM was a remarkable chapter in my life, providing an exceptional learning experience in a premier technology company. Even today, IBM continues to excel in the realm of technology. My time spent at IBM during the Gerstner era, under the leadership of Nick Donofrio and others, was focused on propelling IBM into relevance during the era of e-business and beyond. It was an incredible period, and IBM remains a truly outstanding organization.

Question 2: Leadership: What is the most important leadership skill that you have learned over your career, that has a positive impact, and can you explain with an example?

If I had to narrow it down to a couple of key lessons, the first one would be never stop learning, right? This is a field that has constantly evolved. I’ve been a part of it since around 1986, giving me
38 years of experience in this industry. Over the years, I’ve witnessed technology change numerous times.

At its core, it’s always been about solving problems for people and persistently using that approach. I’ve observed many individuals who initially started in this field move away from the ethos of continuous learning. For me, this continuous learning process has been incredibly valuable. As I approach 60 years old, venturing into a startup focused on generative AI, conditions may not be the conventional choice. However, I believe that age is just a number. It’s a cliche, but I genuinely feel that way. Almost everyone involved in generative AI likely has around two years of experience. Well, I also have those initial two years, but I’ve additionally gained the benefit of 35 years of learning how to do things, and perhaps more importantly, how not to do things. So, the aspect of continuous learning is an incredibly important part of this journey. I would advise anyone at any stage in their career to never give that up because once you stop evolving, you risk becoming obsolete. That’s the essential principle to follow.

I’m curious. Srini, how do you afford time to do continuous learning?

You have a really heavy plate of responsibilities. Doing research, reading and staying up on content is an afforded time. Yes, indeed. I engage in a variety of activities. During my walks, I immerse myself in podcasts across diverse domains. I particularly enjoy listening to Freakonomics Radio and similar shows. These activities often relate to my line of work, prompting intriguing thoughts and considerations. It sparks questions like, ‘How might this apply in my professional sphere?’ This auditory approach suits my learning style, as I identify as an auditory learner rather than a visual one. Consequently, I tend to absorb knowledge through listening.

Additionally, I consciously allocate time for personal growth. Every weekend, I make it a priority to set aside 3 to 4 hours each weekend to do some training. Over the past five to ten years, especially with the advent of Massive Open Online Courses (MOOCs), an abundance of educational content is readily available. The challenge lies in being discerning and selective amidst this wealth of information. In the last seven months, I’ve actively consumed nearly every piece of content relevant to AI. Whenever something new is introduced, like a release from Google or DeepMind, I make it a structured effort to delve into it. This structured, incremental learning approach has proven to be highly effective. I frequently receive inquiries on LinkedIn about how I manage to find time for certifications. The key is to treat learning like a compound interest; the more you invest, the greater the returns. It’s akin to saving for retirement, where the accrued knowledge compounds over time.

This approach is fundamental to my learning philosophy. Another pivotal lesson for me revolves around my mentors. I previously mentioned Nick Donofrio, but another significant mentor in my life has been Mike Keller. He formerly served as the enterprise CIO at Nationwide and played a pivotal role in my career transition from IBM to Nationwide. During my early months at Nationwide, gave me advice that, in hindsight, proved immensely valuable. He said “Srini, you’re used to running the Green Berets and the Delta Force at IBM because everywhere you look around, you got somebody who was as good as you, if not better than you. It’s different when you’re commanding the army, right? You’re going to have to learn additional skills. So just being that technologist alone is not enough. You’re going to have to figure out how to scale, how to get work done with teams, and how to bring people along.” And again, as I said, that didn’t resonate with me at that time, but it proved to be a phenomenal lesson throughout my career. I realized I’m one of those guys who actually goes and can drive a team to go take the hill. And I am not the person who’d sit there, and smell the roses.

I realized I’ve got to surround myself with people who compliment me, not deflect what I do. So I’ve had some amazing people that have worked for me and they’ve worked for me to be with
me. Their guidance and collaboration have shaped my leadership style, reinforcing that success is a journey of collective accomplishments and shared growth.

Question 3: Prediction: Do you have a prediction around the core technology and core changes that are happening in the industry that we should all spend more time learning about?

Yeah, reflecting on it, we’re currently in the early stages of a significant era—the age of AI. AI is 75-plus years old. The term was coined in 1956. Back then, the tools and techniques were complex, requiring highly intelligent individuals to bring them together, making scaling challenging.

Generative AI and, more specifically, large language models, have been instrumental in democratizing these tools. We’re reaching a point where I don’t necessarily need to understand the intricate math behind parameter-efficient functions or feature tuning. I can benefit from the advancements in my field without diving deep into the details. I often describe this transformation as a kind of Maslow’s hierarchy of technology. We’ve moved past worrying about basic needs like food, shelter, and clothing, allowing us to focus on higher pursuits, inching closer to self-actualization. This, to me, marks the beginning of an age of enlightenment—an incredibly exciting prospect.

However, our excitement needs to be tempered with responsibility. At Fair and Rackspace, we emphasize guiding the responsible use of AI. I believe AI should be symbiotic, coexisting and augmenting human capabilities rather than replacing them. Security is paramount, ensuring safety for individuals and protecting data in accordance with privacy and copyright laws. And lastly, sustainability is key—ensuring a diverse and reliable supply chain of data to prevent bias and uphold accuracy.

Srini serves as President of Technology and Sustainability at Rackspace Technology® and is responsible for technical strategy, product strategy, thought leadership and content marketing. Prior to joining Rackspace Technology, Srini was Vice President, GM, and Global Leader for Hybrid Cloud Advisory Services at IBM where he worked with CIOs on their hybrid cloud strategy and innovation. Before that, he was the Chief Information Officer for Magellan Health where helped double the company’s revenue in just four years. Prior to Magellan, he was the President and CEO of NTT Innovation Institute Inc., a Silicon Valley-based startup focused on building multi-sided platforms for digital businesses. Srini also serves on the advisory boards for Sierra Ventures, Mayfield Ventures and Clarigent Health.

Srini is an innovative and dynamic executive with a track record of leading organizations to deliver meaningful business results through digital technologies, design thinking, agile methods, lean processes, and unique data-driven insights for the last two decades

Srini serves as President of Technology and Sustainability at Rackspace Technology® and is responsible for technical strategy, product strategy, thought leadership and content marketing.

Prior to joining Rackspace Technology, Srini was Vice President, GM, and Global Leader for Hybrid Cloud Advisory Services at IBM where he worked with CIOs on their hybrid cloud strategy and innovation. Before that, he was the Chief Information Officer for Magellan Health where helped double the company’s revenue in just four years. Prior to Magellan, he was the President and CEO of NTT Innovation Institute Inc., a Silicon Valley-based startup focused on building multi-sided platforms for digital businesses. Srini also serves on the advisory boards for Sierra Ventures, Mayfield Ventures and Clarigent Health.

Srini is an innovative and dynamic executive with a track record of leading organizations to deliver meaningful business results through digital technologies, design thinking, agile methods, lean processes, and unique data-driven insights for the last two decades.

Generative AI – From Big Vision to Practical Execution

For our latest CXO Insight Call: “Generative AI – From Big Vision to Practical Execution,” we were joined by Sreekar Krishna, US National Leader for Artificial Intelligence at KPMG and Erik Bovee, Head of Business Development at MindsDB, who spoke to what a pragmatic generative AI path to success will look like in today’s large organizations.

Even before ChatGPT, one-third of CIOs said their organization had already deployed artificial intelligence (AI) technologies, and 15% more believed they would deploy AI within the next year, according to the 2023 Gartner CIO and Technology Executive Survey. But deciding how best to proceed means factoring AI into business value, risk, talent, and investment priorities. As a next step, CIOs and other organizational leaders must create an AI strategy roadmap that synthesizes the enterprise’s vision for the future – outlining potential benefits, while mitigating risk, capturing KPIs, and implementing best practices for value creation.

Sreekar spoke in depth on how businesses should be thinking responsibly about AI best practices and making them foundational to AI strategy, as opposed to just an afterthought.

Key Takeaways

  • Generative AI is being widely adopted. 64% of Fortune 500 executives polled by KPMG plan to implement their first generative AI solution within the next two years. This is regardless of where these companies are on their AI journey today.
  • Generative AI requires a different treatment when contrasted with the AI practices of the past. Teams no longer need a data scientist in the room just to talk about AI.
  • AI has become an ecosystem play. The data players who provide data for companies today, either third party or internal data, can now all use LLMs to offer not just data, but insights, becoming part of a broader ecosystem
  • When getting started, it’s important to prioritize business use cases. This is a very important piece of the puzzle. When considering use cases, consider effort vs. use case area. The low hanging fruit will be quick wins in the back and middle office.
  • A Tiger team is needed in order to address foundational needs like legal, risk, IP, etc. Consider setting up a cross-functional team with a single, empowered rep from each function (e.g. IT, Data, Risk, Legal, Cyber…)
  • Literacy and digital enablement of GenAI capabilities needs to be front and center, particularly executive literacy, in order to encourage proper investment and resource allocation
  • Enabling employee access across the entire institution will allow the exploration of back, middle, and front office capabilities equally. Enabling the function for only a small handful of people within the firm doesn’t provide the mass needed to harvest future use cases
  • Infrastructure matters once more. Ensure that the infrastructure is ready around genAI models, data gravity and cross-infra/cloud leverage. Investment is required to develop an enterprise architecture that can accommodate LLM buy, build, and integration. Focusing on only the buy or build options is a mistake.
  • There are four major LLM use cases in the mix today:
    • Embedded LLMs – These are the large language models embedded within business platforms. Think: Salesforce EinsteinGPT, Github CoPilot, etc. These will be used by line of business professionals, who are already using these tools today. This is more of a buy program – the LLM is being bought through the provider
    • Self-Service LLMs – These are things like ChatGPT, OpenAI, Claude2 that are self-serve. They will be mostly low-code and no-code platforms and business analysts will be the key user here
    • Specialized LLMs – These are function-specific e.g. for cybersecurity. Cyber LLMs are emerging already and require deep domain knowledge for them to be fine-tuned. Domain data scientists will be the key players here
    • Proprietary LLMs – These are built from scratch, and subsequently fine tuned. BloombergGPT is an example here. Power data scientists will be the users
      AI will be a huge gamechanger. Humans and machines working together outperform either humans or machines working on their own.

 

AI: Where is the Momentum?

Over the last couple of decades, businesses have been implementing a variety of different strategies around AI and data science – but many of these efforts are now dated, or even in direct conflict with the new wave of generative AI solutions coming to market. So the question now arises: How can companies transition their prior AI efforts over to a generative AI approach, with value creation as the top priority?

KPMG sends out a global survey every 3 months to C-level executives in the Fortune 500, and their most recent survey focused on generative AI specifically.

Over three quarters of the executives who participated expect that generative AI will have a large impact on their organization over the next 3-5 years, as measured by increasing workflow productivity. And 60% plan to implement their first generative AI solution within the next two years. This is regardless of where these companies are on their AI journey today. Advancing initiatives in this space is seen as increasing the moat around the business.

IT and tech are the definitive early adopters as this new wave is coming in the form of SaaS products that have fantastic API integration. However, finance, HR, procurement, sales, and many other areas that have traditionally relied on unstructured data are starting to make a lot of impact.

AI: Where are the Challenges?

That’s not to say there won’t be challenges: Over 90% of respondents believe that generative AI introduces moderate to high-risk concerns – especially in areas such as IP, legal and security. Furthermore, over two-thirds of companies have not set up a centralized person or team to organize their response. This is still the case even in very large organizations that haven’t been traditional power users of AI, particularly in the back or middle offices. This is difficult for the business which will ultimately run into major challenges like: “How do we make this work?” “Who do we go to?” “Do we need permission for X?” etc.” Engineers and data scientists are already dedicated to front office functions and building the product, so there needs to be some sort of ownership around the back and middle office.

Furthermore, talent is somewhat of a challenge both in terms of availability and how teams are organized. Data engineers and data scientists are hard to recruit. But on the upside, IT has been completely embracing generative AI. It’s much more exciting than the previous generation of AI, which was positioned as just one more thing to help the business with. Every business unit would come to them with a new AI model that worked a different way.

Now, what we’re seeing is data engineers, software engineers and integration engineers actually bringing generative AI use cases to light. This is happening within the IT function itself, so there’s a lot of adoption bubbling up from the grassroots of tech players within large companies. Data scientists are now becoming SMEs to IT specialists, with IT using data scientists as subject matter experts, in order to better understand how xyz model is working. This is in contrast to the prior generation, where IT and data scientists were peers in taking a model and making it production-ready. Data scientists were expected to answer the data engineers’ questions and IT integration questions, and they struggled quite a lot with that. Now data scientists can be good at what they’re good at: interpreting models and outcomes.

However, there are emergent funding problems here: heads of data science are starting to question if they are being proportionately funded relative to the IT function’s spend around generative AI. Do they have enough funding and support to really address all of these new use cases? It could be that data science’s funding model will be integrated within the business function where they are supporting all the growth that’s happening there. For companies that are sticking to the center of excellence model, they are proportionately beginning to fund data scientists more as they are proportionately more involved in these new generative AI initiatives. Data scientists aren’t necessarily needed to start an AI engagement, but they are needed to help support ongoing AI engagements.

Finally, there are definitely risks in place around these new technologies like cybersecurity, data privacy, etc. And when thinking about the regulatory landscape, it’s important that these new concerns are addressed in a methodical manner.

KPMG’s High-Level Takeaways

There are four key takeaways that the KPMG team has run up against working on various client initiatives. First, generative AI requires a different treatment when contrasted with the AI practices of the past. You no longer need a data scientist in the room just to talk about AI.

This leads to point two, which is that the path to adoption for different business units within each organization is different. For some teams, just enabling and white labeling GPT is enough for them to get started and get value out of generative AI. For others, an AI app may be required to produce a software that can integrate with the system and have data flowing through, which is a different adoption curve.

The third piece is that generative AI has become an ecosystem play. Previously, companies had a whole conversation around buy vs build, e.g. “Should I go with a certain environment depending on my use case?” Now, we’re in a place where certain kinds of AI must be bought, and certain kinds must be built. It has become a buy and build game and that’s where your ecosystem is going to matter a lot. Your data players who are providing your data today, either third party or internal data, can now all use LLMs to offer you not just data, but insights. And at that point, they become part of your ecosystem, the same way you could be part of someone else’s ecosystem, giving them insights yourself. Raw data is still important, which is why data engineering has been so huge the last few years, but now we’re starting to see insights moving between business groups and integration points.

Finally, a responsibility-first approach is going to be so important. We all know the risks with generative AI. Unlike with the previous waves where it was easy to do a small POC, and teams could think about the responsibility side later – that aspect is now becoming more front and center. It’s important to address the risk, the IP issues, the licensing issues. This could be even broader like an employee entering PII information accidentally, or bad actors poisoning the prompts, or a consultant using hallucination output, etc. This must all be addressed in stage one of the MVP before it’s taken further. Some thoughts here include focusing heavily on employee literacy and education (e.g. mandatory LinkedIn Learning or other programs), avoiding the use of client data in the early innings, or having a controls layer within the API that tracks things like social security numbers, names of people, phone numbers, etc. Prompts can be blocked if they are getting into PII. Furthermore, other products can be brought into the pipeline to help block prompts that create problematic exposure.

The 4R Model

So, in response to all these quandaries, KPMG has come up with what they’re calling the 4R model to harness the power of generative AI.

  • Represent Responsibility – It’s crucial to establish and maintain trust and transparency while operationalizing privacy and data ownership rights. In the past, the business told everyone where the data is, and then technology went and worked with the data. But now there’s an accountability and responsibility element in play. Take privacy as an example: If a customer comes and says “delete my data,” do all the AI models that trained on their data need to be deleted? And then retrained? How will this actually work in practice?
  • Reshaping Your Organization – Enterprise AI can reshape not just the front office, but also the middle and back office, driving huge value to today’s organizations. How will this reshape the organization and subsequently the workforce?
  • Rebuilding Your Ecosystem – Who will be the buy and build players within the ecosystem? The data players, the technology players, the AI model providers?
  • Re-Imagining Your Clients and Customer Engagement – This is a big portion of what we’re seeing in the early innings today. Over the next 1-2 years, large organizations will be thinking about where generative AI will directly interact with their clients and customers. What new models of interaction will clients across a variety of industries require? How can that be incorporated into the product? Or into an organization’s way of thinking?

The Evolution of Generative AI

Using the horizons model, it’s clear that we’re already seeing companies beginning to traverse the curve. Horizon 1 is well underway within the back office and middle office functions of many businesses today. Generative AI is going to production, particularly in financial service institutions like insurance. Traditionally, they have been slow to adopt these kinds of technologies, but now, they are moving quite fast. Many F500s and otherwise are putting generative AI models into production as we speak.

Horizon 2 will start to happen in late 2024, or early 2025, and we’ll begin seeing generative AI models interacting with clients and customers. On the B2B side, some of this is already starting to take shape within sales and client interaction models. Then, in 2025 and beyond, new and disruptive business models will start to emerge. Private equity is already trying to figure out how to spin off portions of their portfolio into automated generative AI business models that will go to market in the 2025 timeframe.

So, with all this in mind, what does the roadmap really look like?

There is an emergent “Crawl, Walk, Run” strategy for getting started with generative AI without wasting too much effort on strategy.

When getting started, it’s important to prioritize business use cases. This is a very important piece of the puzzle. When considering use cases, consider effort vs. use case area. The diagram below gives a good sense of where businesses should start: clustered around the blue line. Crawl for back office and middle office, while progressing towards a front office run function as a strategic play.

Some things are quick wins, some things are experiments, and some things should be held off on given how rapidly this environment is changing, and how many new players are entering the market. But at the same time, staying quiet and assuming that something is going to emerge in the future is not a good strategy either.

When it comes to “Crawl / Back Office” – just enabling white label GPT for the back-office (call centers, internal audit, etc.) may already add a lot of value to the mundane tasks they’re completing. But coming to the middle of the function: “Walk / Back Office” – it could be smarter to just buy a foundational model and fine-tune it in order to solve a particular set of tasks. A good example here would be a cognitive search engine custom-built for the call center, an assistive agent they can ask quick questions to about product, customers, call histories, etc.

Finally, when thinking about “Run / Front Office,” the strategic play, the idea is to envision how to completely disrupt operations. For example, futurizing call centers where interactions are entirely with generative AI models, rather than with humans, as a first step in the process.

The Crawl

“The Crawl” requires two important elements to be addressed: foundational enablement within the organization around how things are going to happen, and the change management that goes with it.

  • Is there a Tiger team to address foundational needs like legal, risk, IP? Consider setting up a cross-functional team with a single, empowered rep from each function (e.g. IT, Data, Risk, Legal, Cyber…)
  • Getting started with the technical enablement of a white-labeled LLM can be a great start. Focus on high-level usage patterns (as opposed to individual use cases) to drive quick experimentation. Use cases that fit within the usage patterns are approved automatically
  • Literacy and digital enablement of GenAI capabilities needs to be front and center, particularly executive literacy, in order to encourage proper investment and resource allocation
  • Enabling access for the entire institution will allow the exploration of back, middle, and front office capabilities equally. Enabling the function for only a small handful of people within the firm doesn’t provide the mass needed to harvest future use cases
  • Develop use cases both centrally and through crowdsourcing: Ideally small ideas with an outsized impact (take a task that 500 employees do 100x a month and make it >20% faster – operations, call centers, data analysis, summarization, are all low-hanging fruits). Roll out MVP use cases quickly and expand those that gain traction
  • Ensure that infrastructure is ready – genAI models, data gravity and cross-infra/cloud leverage

KPMG As a Case Study

KPMG has created an API control layer within which they are launching a lot of foundational AI models: LaMDA, Claude, ChatGPT, etc. These models have all been opened up for any employee to access. ~80,000 US employees now have access to AdvisoryGPT within the advisory function and the adoption has been phenomenal. Within two weeks, they had an active user group of 10-11K employees on a weekly basis, using this new functionality for many different questions and prompts – all of which are then being harvested by the data team. Now, KPMG can look at what kinds of questions people are asking. What kinds of use cases are they looking for? Where do they sit in the organization? How can they be better enabled?

These are all log-level details coming through PowerBI, with a dashboard that functional leaders can view. These kinds of details help develop a roadmap around what further capabilities should be built out. Ergo, the crawl program can help to create a map around what the walk and run programs ought to look like over the next 2-3 years. And this can be cheap, <$10,000 a month.

The architecture above was built with ChaptGPT as the first initial cut, using Azure OpenAI services. KPMG created a private endpoint with Microsoft, and that private endpoint is controlled by IT layering and preventing any kind of external access to it. So, it’s not connected to the internet (aside from just their team and connections into Microsoft).

In essence, this is a private, securitized, endpoint within Azure where this OpenAI/ChatGPT endpoint is running and does not connect to the internet. It’s a static model that is left within the four walls of Microsoft’s network. Anything that is sent as a prompt lives inside of KPMG’s Azure instance, and the rules are set in such a way that Microsoft cannot access the data. These are the same security controls that they are using for their Gov Cloud. Any data produced in the OpenAI environment is subsequently pulled into their Azure environment, so that the data doesn’t live within OpenAI, but instead within their Azure controlled environment. That’s where all the PowerBI and log details are coming into the network.

Getting started can really be as simple as using a consumer-grade ChatGPT virtualized into an enterprise environment. Fine-tuning is not necessary to start crawling. Without enabling this secure access for employees, those same employees are going to explore on their own.

The Walk

“The Walk” requires two important elements to be addressed: data mobilization and workforce acceleration. This is where data starts to become the key to success. Data must be enabled to flow to the right places at the right times. This is going to be a huge piece of the puzzle. Different data components will need to be integrated into a seamless channel in order to feed whatever different models are being built. The LLM portion needs to be plug and play on top of a fixed and static data piece.

But at a high level, organizations are going to be looking at:

  • How to best develop a roadmap for buying, tuning, and building LLMs into the functional processes of the organization
  • The integration of bought and built LLMs for quicker wins and easier operationalization
  • The creation and curation of prioritized use cases across various business units
  • Identifying third party data sources that are critical for business development, and buy or build capabilities around these data sources
  • The prioritization and mobilization of institutional data for knowledge discovery
  • The operationalization of LLMs into the middle and back office workflow

The buy and build part of the ecosystem comes into play once organizations want to take things to the next level.

There are, generally speaking, four major types of use cases in the mix today.

  • Embedded LLMs – These are the large language models embedded within business platforms. Think: Salesforce EinsteinGPT, Github CoPilot, etc. These will be used by line of business professionals, who are already using these tools today. This is more of a buy program – the LLM is being bought through the provider
  • Self-Service LLMs – These are things like ChatGPT, OpenAI, Claude2 that are self-serve. They will be mostly low-code and no-code platforms and business analysts will be the key user here
  • Specialized LLMs – These are function-specific e.g. for cybersecurity. Cyber LLMs are emerging already and require deep domain knowledge for them to be fine-tuned. Domain data scientists will be the key players here
  • Proprietary LLMs – These are built from scratch, and subsequently fine tuned. BloombergGPT is an example here. Power data scientists will be the users

When considering this landscape: the “Buy” end of the spectrum is on the left, and the “Build” end of the spectrum is on the right. In between are use cases that may require a little bit of fine tuning. Taking only the buy or only the build options is a mistake. Companies must be engaged in the information and decision integration process – how can all of these new tools be integrated in a secure manner?

Furthermore, many companies are emerging to help simplify all the commotion. One of our portfolio companies, MindsDB, has created sophisticated middleware to help connect any data source to any AI model, including a business’ own models that are in production. All this while providing an abstraction layer that allows developers to still use their preferred developer framework to build sophisticated AI applications, move them into production, and maintain/update them more easily. The goal here is to allow any developer to implement quick TTV use cases, while still being able to support the most sophisticated use cases. The best way to take advantage of the opportunity AI presents, is to place that opportunity into the hands of every developer.

The Run

The focus is going to be all about ensuring that architecture can enable enterprise-wide, responsible use of LLMs. This will involve:

  • Investing in the development of an enterprise architecture than can accommodate LLM buy, build, and integration
  • Focusing on the responsible use of AI apps that both business and tech buy into
  • Data governance for authentic vs. synthetic data
  • Privacy engineering to ensure end to end data and model compliance management
  • Never losing sight of quick wins

Certain elements are going to be absolutely necessary to reach the “Run” phase in terms of enterprise architecture. Privacy Ops (RBAC, ABAC), security & IAM programs, as well as good devSecOps, cloud engineering and integration will be key factors.

Structured and unstructured data will also be huge. Unstructured data will be growing really fast and the governance of synthetic vs. authentic data will be key. Registering what type of LLMs are being brought into the company (are they coming from a startup, cloud provider, or are they internally built?) will need to be tracked, as will AI apps. Harvesting prompts and prompt analytics will be an important piece that also must be enabled within the architecture. LLM orchestration will be huge, vector and graph databases will be huge. How to enable all these things will be a huge piece of the program that will need to be built out.

In Closing

It’s an exciting time for the modern organization, the activities that promote scale in AI create a virtuous cycle. Interdisciplinary teams will bring together diverse skills and perspectives, and employees will start to think bigger—moving from discrete solutions to completely reimagining business and operating models. Innovation will accelerate as teams across large organizations get a taste for the test-and-learn approaches that successfully propelled these early AI pilots.

Companies that are able to implement AI throughout their organizations will retain a vast advantage in a world where humans and machines working together outperform either humans or machines working apart.

Q&A

1. Would you mind talking more about PE firms thinking through GAI spin-offs, are they trying to launch companies with a scaled down GAI only operations?

PE firms are looking into using Gen AI to optimize operations in their new acquisitions. They are also looking inside their own portfolio companies to see if there are candidate companies with proprietary industry-specific knowledge which can be packaged into LLMs, and be spun off as companies of their own.

2. On C’s description of RAG (retrieval augmented generation), what level of success do people see with this and do you expect to continue to build it internally, or buy a solution for this?

RAG is showing a lot of promise, but it’s yet another advancement in the Gen AI algorithmic growth. We’ll definitely see more advancements come through in other techniques. Hence the suggestion that when it comes to the “Run” model, it’s important to build an enterprise architecture that allows efficient orchestration of LLMs and algorithmic techniques around LLMs.

3. Aren’t some of the quickest wins and low hanging fruits for generative AI in the “Front Office” (e.g. Customer Interactions)? So, could one “Run”” without the “Crawl” and “Walk”? Or, am I reading the matrix wrong?

Given some of the concerns with Hallucinations and the need to monitor the outputs from Gen AI, our recommendation is to have some thin veil of controls before exposing everything to front office. Granted, if all you’re doing is opening an experimental experience for your customers, or clients, you can provide them with enough fine print, and just open things up. But if the goal is to aid in real decision-making, especially critical decisions like financial, legal, health etc., it’s important to consider front office applications as being more risky than back and middle office use cases.

4. Where would you put using AI to write code in, re: Crawl/Walk/Run?

It depends on how integrated the code generator is within the enterprise. If it’s just allowing coders to use ChatGPT to write better code, it’s definitely a “Crawl.” However, integrated technologies within the dev environment would be more like a “Walk.” Finally, fine tuning foundational models with your enterprise code base and creating very specific code that was custom delivered for your applications is more like a “Run.”

5. What about the talent challenge, is it reasonable to assume organizations can hire this talent or should we be looking to leverage experts from consulting and/or solution providers?

This goes back to building your ecosystem. If you are able to build an ecosystem that has vendors, consultants, data providers, product companies, etc. then the conversation is not just around talent, it’s also around process and technology. You may be able to solve some aspects of the talent challenge with process or tech (startups). Also, we’re noticing that as Gen AI technologies mature, companies are wanting to invest more in their Data Analysts, because those are the SME who understand the business better and are the core source for “growing the business.”

6. How much effort was needed to create the underlying data or cleanse/normalize the data to make these LLMs effective?

This depends on where one starts from. Some of the proprietary LLMs are very good out of the box. In fact, there is evidence now that one could deteriorate the outputs from some of these very large LLMs if you try to tune them too much into a narrow domain. On the other hand, smaller LLMs can do much better in specific tasks when trained with even very small volumes of data.

7. Can you cover what type of risks you’re seeing in KPMG (internally and with clients)? How are you addressing these?

IP is one of the biggest ones. There are many questions around what is considered unique knowledge to the firm, especially if Gen AI was used to develop a writeup, or code, or image, etc. Hallucinations is another big one. We want to make sure models don’t lie about facts. Cybersecurity around models to ensure no one is injecting bad data into the training process (Insider Threat) or choking the LLMs (e.g. DDoS). Legal risks still are very high, especially controlling the legality of use. The best mitigation has been a prudent and measured approach to the use of Gen AI in specific use cases. Creating the Tiger Team that is able to spearhead conversations ahead of the use cases will be important.

8. Are you getting a lot of requests / needs from clients around VectorDB? If so, how do you deal with embedding vector in the raw data?

Yes, but VectorDBs have become a natural extension to the use of LLMs. Now the focus is shifting to the use of other novel approaches, including ‘context aware chunking,’ ‘graph embedded contextual vectors,’ etc.

9. Are you seeing any best of breed products for Fine tuning, RAG, ETL, etc. outside of the cloud vendors and OpenAI?

10. What is the average time from first discussions and ideas to PoC/MVP in your experience?

Two weeks is the fastest turn around we’ve done with our clients, but the client’s tech stack and data must be ready for the use case in order to move at that speed.

11. Since everyone is doing crawl/walk/run, how can you build competitive differentiation?

In our experience with the Fortune 100 to 500 companies, very few have executed even the “crawl” program well. Further, this differs by industry also. The tech industry is definitely ahead of the curve, but even within the top tech companies, the use of AI is restricted to their product side. When it comes to their middle and back office, even the most well known AI companies struggle with enterprise AI.

12. What happens to AI when all content owners (like NYT) mandate that their content be removed from all LLMs, AI engines, etc.?

Very good question. Maybe there are startups who are looking to license all data to train their models, and such an LLM would truly be able to provide content that is completely licensed.

13. How are you (or how do you plan to) measure quality and effectiveness of implementations for these use cases?

This depends truly on the use case at hand. For example, finance use cases must have numerical quantitative measures of success, while generated documents, or reports, need to be measured with some of the more NLG metrics for measurement. Human feedback is an expensive, but time tested way to measure the outputs also.

14. Why did you pick OpenAI / Microsoft vs Google or Open Source (e.g. Llama2)?

We were already an Azure stack, the decision was easy. Further, looking into the security model, privacy model, etc. Azure OpenAI offered the best first step. Having said that, we have already enabled Llama 2, Claude 2, Dolly, Falcon etc.

15. How do you see the landscape evolving between Proprietary LLMs vs Open-source and is Multi-LLM going to be important?

Open-source and proprietary-source conversation will continue to dominate the same way it has with other technologies we have experienced in the past, e.g. Operating system, database technologies, cloud technologies, etc. LLMs are facing a similar trajectory. It will be important to see how enterprises are able to stand up capabilities to orchestrate LLMs within their four walls. If they are able to manage the operational overhead, open-source will have strong adoption.

16. Are there any other foundational tools/applications (e.g. Langchain, Github Copilot etc.) that you see as critical to broad GenAI adoption at KPMG or amongst your clients?

Yes, we consider these technologies to be important pillars within the ecosystem. The ecosystem varies depending on the industry and the maturity of the client, but we see some of these technologies as being foundational.

17. How did KPMG internally build a business case for setting up the Advisory LLM ?

We are committed to the Generative AI future all the way up to the board level. In fact, we just established an internal executive role that reports to the CEO/COO that focuses on evolving KPMG’s consulting services backed by Gen AI. Further, the leader chosen to run the new function previously ran the entire advisory consulting organization.

18. Which LLM models are better for which purpose in your experience?

See question #9

19. How are you measuring the impact/ROI of these solutions? Are teams tracking time savings, cost savings, etc., to get a sense of the value created?

We definitely have to create measures of success depending on the use case at hand, and the business case at hand.

The Voice of the CIO – Insights on Generative AI Adoption in the Enterprise

We recently hosted a number of local CIOs here at our offices to chat a bit about where we’re at with generative AI, and when it will start being operationalized. The F500 has already kicked off a plethora of enterprise POCs in the space and are hard at work extracting the more important enterprise use case needs. While generative AI may not be 100% enterprise ready yet, in the next 6-12 month we’re going to start seeing more serious (potentially customer-facing) deployments. Budgets are being defined for larger-scale 2024 deployments, with the rest of 2023 being the experimentation phase. Our group chimed in to provide some color on where we’re heading.

Where is Enterprise Generative AI at Today?

  • Back to Basics People, Process, Technology hasn’t really changed
  • Architecture Matters Once More There is so much at stake and the pieces need to come together properly
  •  A Long-Term Vision is Necessary How software is built in the next 20 years will completely change – it won’t just be efficiency, it will eventually become true innovation. The big transformations of the last 30 years (for the enterprise) will be the internet -> mobile -> LLMs – so a cohesive game plan for the 2020s will be required
  • AI Will Initially Look a Lot Like RPA++ 
  • AI for Internal Use & AI in the Product – CEOs of every company (not just tech companies) will want to know how this is embedded in the product (“we have AI technology”). Over the next few years, companies will need to be embedding generative AI features into all of their products to compete – that planning starts now
  • There Will Be More Losers Than Winners – so maybe hold off on any major investments – no contracts of >1 year. People are asking for cost premiums today on things that will be commodities tomorrow – now is the time for small bets and exploration
  • Work Hard to Keep Things Contained Until This is More Enterprise-Ready – Avoid IP and sensitive data leakage
  • Begin Early Exploration with Large Partners Like Salesforce and Microsoft – What brings benefit and what doesn’t? It’s easy to piggyback on their IP
  • There Will Be New Legal and Regulatory Hurdles to Tackle – How do you explain how these new tools came to their conclusions? How do you explain how a 30B parameter model came to its conclusion?
  • It’s Time to Review Vendor Renewals With a Closer Lens – Are you opting out of sharing your data? Are you opting out of your data being used for training? Especially if that data is proprietary. Many existing software providers are attempting to use your data to train their models. Make vendors prove that they’re retiring your data
  • Data Will be the Key for Impact and Utility, Not Technology (Which Will be a Me-Too Product) – (1) The provenance of your data is critical in regulated industries (e.g. in healthcare, if one data source is wrong, people could die), (2) Data quality will be the key to advancing AI efficiencies – you need something that will validate the data that you’re getting, (3) The ecosystem around enterprise data warehouses is going to continue to explode , (4) The Chief Data Officer will be replaced by the Chief AI Officer, (5) Early use cases should be lightweight and not break anything major
  • It’s Crucial to Keep and Eye Towards Real Deployment – Don’t leave things in the sandbox forever while your competitors pull ahead
  • Generative AI Will Ultimately Be Rolled Into Standard Product Offerings – It won’t be something that you pay for separately as it becomes increasingly commoditized – competition will drive the shift
  • AI Startups Will Need to do a Better Job Complying to Enterprise Standards – Play by the rules, be transparent, and build trust
  • There Will Be Many Competitive Opportunities for Vertical Plays – Open source is going to be great for industry-specific or niche build use cases (vector in your own data etc.)
  • Companies Will Need to Gain Comfort Around Which LLMs Are the Right LLMs for Them What should they be looking for? How should they be evaluating? VCs and other thought leaders can help evangelize here
  • Real Production Use Cases Will Begin to Emerge in 2024

How Will Generative AI Change the Security Landscape?

  • Teams are Excited and No One is Thinking About Security
  • Data Security is a Rising Issue – As soon as the data leaves to a third party it’s a huge risk (company recommendation for this: https://www.private-ai.com/). The security team responsible for third party risk is going to have to change their process a little bit (show me you are deleting my data, insist on positive consent in contracts)
  • Threat Vectors Will Be Smarter Than They Used to Be – Security solutions will have to be smarter too
  • More Available Data -> Increase in Cyber Crime

How Should Leadership Be Handling the People Element?

  • Education, Training, and Communication is Key – You have to focus on educating your teams, educating the board, educating your customers and educating regulators – the language people use today is inconsistent or inaccurate
  • Try to Enable Internal Teams – Otherwise they are going to be doing these things on their own. Employees are not cautious. Collect internal use cases by division or persona
  • Establish Guard Rails and Weekly AI Office Hours – Figure out how to partner with the business, not just block them
  • Create a Sandbox for Employee Use Cases

How Effective Will Early Generative AI Use Cases Be?

  • The Tech Looks Interesting, but the Deployment Isn’t There Yet – There are issues with data, security, leakage, etc. So the opportunity is present, but still nascent
  • The Type of Use Case Will Matter A Lot – Regulated industries and areas like security are going to require a lot more caution
  • Code Generation is Still a Challenge – It impacts IP directly (is AI generated code your code?)
  • KPIs are Needed for Each Use Case – Figure out how it’s improving things either via TTV, eliminating/freezing staff, etc.
  • There is More Scrutiny from Regulators Now – This will slow things down

What Early Use Cases Are Looking Strong?

  • The best use cases today are the ones where we already know what good looks like (e.g. RPA)
  • RFP processes – E.g. Transforming/automating the RFP, RFI, DDQ and security questionnaire processes
  • Customer service/customer facing efficiency and knowledge sharing and enterprise search (e.g. Glean++)
  • Many companies are seeing modest, but measurable, productivity gains with copilot (15-25%)
  • Document (job descriptions, emails, etc.), communications, and content generation
  • QA
  • Forecasting
  • New product development (e.g. a genAI tutor for math)
  • Code conversion (vs. generation) – e.g. Gosu into Python
  • Contract management and summarization
  • Translation
  • Internal security

AI Council Best Practices

  • Who Should Be On It? Head of Data/AI, Head of Product, R&D, Legal, CIO (representing GTM, finance, HR, etc.), Engineering, Program Manager
  • Have people submit needs, group them, and give one team a chance to drive things forward for the whole company (vs. having 100 siloed conversations)
  • This could also be a group that drives what evolution in the product will be (maybe have a “working group” for product/eng)

Buy or Build?

  • Hybrid model, or buy and train
  • Take generic models and fine tune them – but the value has to be there relative to the cost (costs will come down over time)

Will CIOs Get More Budget?

  • No
  • Good news: GPT-3 cost $15M to build, Mosaic built the equivalent with $500K

Generative AI – The Enterprise Primer

For our latest CXO Insight Call: “Generative AI with Google – The Enterprise Primer,” we were joined by Josh Gwyther, Global Generative AI Lead for Google, who shared an expert’s POV on where LLMs are at today (and where they’ll be tomorrow), what’s possible in the enterprise today with generative AI on Google Cloud, and what some of the best practices and pitfalls are for the enterprise. New opportunities emerge every day in operational efficiencies, cost savings, and value creation. This includes how companies can build, tune, and deploy foundation models with Vertex AI, personalize marketing content with Gen AI Studio, and improve conversational experiences with Gen App Builder.

Wide ranging concerns persist around the security and maturity of today’s AI offerings, and enterprises, unlike consumers, have many important needs around controlling their data, avoiding fraud and security breaches, controlling costs, integrating existing data and applications, and ensuring outcomes are accurate and explainable. Read on to learn about how the Google Cloud AI Portfolio will be tackling enterprise readiness, AI data governance, and open & scalable AI infrastructure.

Key Takeaways

What’s Happening in the Market? (Enterprise)

  • Only a small percentage of enterprises have moved in any significant way – Companies are still coming up to speed, mainly being pushed by boards. There is a much greater concern around legality, IP, who owns what, etc.
  • Data security remains a huge concern – Sending IP to shared models isn’t something companies want to do. Where is the data going? How is it being used? Why did your model do what it did?
  • Lack of tooling (security, alignment, inference logic, etc.) is a struggle
  • Most enterprises are leaning towards buy vs. build when it comes to models due to lack of internal expertise
  • The focus is on data corpuses and what value they have versus model specifics at the moment. Every company will have the opportunity to bring some niche functionality to market (but this won’t be a trivial task)

What’s Happening in the Market? (Startups)

  • Building on functions: hard to defend as everyone is gaining access
  • Foundational models are great for MVPs, but expensive at scale. You must be able to pass significant value along to customers, to the extent that they are willing to pay a premium for it
  • Model capacity is increasingly becoming hard to come by – computation is powering this generative AI renaissance, but there is definitely a scarcity of capacity right now. This is another forcing function that will push the industry towards smaller models
  • In just 8 months we’ve come full circle of build-buy-build – Seed funding for giant foundational models was very popular, followed by everyone saying they’re going to consume API endpoints around foundational models, then back to build with the open-source movement meeting enterprise demands via a much smaller footprint (cheaper inference, better scale)
  • Open-source has become a viable option
  • We’ve moved from single API calls to “bundles” of API calls due to the chaining of LLMs in application logic. With the advent of LangChain and a bunch of other cool new libraries, LLM calls are being strung together in a higher-level function, and we’re starting to see some performance degradation. If you have an application stack using an API call multiple times throughout every function, you’re going to pay the price if your application doesn’t sit next to the foundational model. This is a classic problem in the cloud space – customers had data sitting on one cloud and applications on another, and when you make all those calls across the internet you pay for egress with performance
  • Data corpuses have become the moat
  • Sending everything to OpenAI has become a hard sell for their enterprise customers

How We Got Here

Generative AI is not new to Google, in fact, it dates back over 20 years prior. AI was always intended to be the mechanism by which Google would scale search, and the components were all there, back as early as 2000. The mindset was that eventually the team would utilize generative AI to facilitate search at scale.

Very early on, Google put the pieces of the puzzle together and realized that the core components of AI for search are:

  • A massive amount of computation
  • The knowledge base
  • Comprehension of that knowledge base

Compute and massive data centers are required to train the models. Then, as much data as possible is fed into them (in Google’s case, all the data of the internet). And finally, understanding is required in order for the models to recognize patterns in that deep set of data.

If you move forward along the timeline above, these are the pivotal events in the formation of where we are today with generative AI. A big leap forward took place in 2006 when nVidia released CUDA. This is a pivotal moment in generative AI history because it really democratized the supercomputer. This allowed everybody – universities, companies like Google, etc. to create supercomputers at scale, cost-effectively.

The other pivotal moment relative to Google’s march down this path was in 2009 when ImageNet was created for vision learning. Back in 2009 and 2010 there was a lot of work attempting to figure out vision classification with machine learning. This seems trivial now, but the epiphany moment that Fei-Fei Li had at Stanford was that maybe we just need a larger training set, and that with a well-labeled, massive dataset of images, vision models would be able to classify images at the same level as a human. This was a hard task at the time.

So, Google figured out a more scalable way for training sets to be labeled. Every image on the web has an alt tag, which was supposed to be there for accessibility. So if you had a blind user surfing the web, the tools they would use would actually communicate to them via that alt tag what they were looking at. In essence, the work was already done before it began. That was a huge step forward in proving out that a large amount of labeled data enables vision classification.

The next challenge that Google had was language. How do you translate sentiment across multiple languages? You need a machine learning model that can translate between any language on the globe, and derive context correctly. Labeling images is straightforward, but context around written language is very tough. Phrasing and word placement is tricky. In 2017, the Brain team came up with a paper called “Attention Is All You Need” which was really the invention of the Transformer Model. So when you hear the T in GPT, that’s what it’s based on. Instead of trying to label text to understand translation and natural language processing, the team just took all the text that was on the internet, and transformed it into numeric variables, and then looked for patterns. And the attention part of that is not just looking for patterns for the next word, but also paying attention to the words up front that it’s being compared to.

Think of old school Mad Libs: you had to fill in the verbs, or the nouns, and make funny sentences. What the team did was take all the data’s conversational text information, transform that into numeric values, then run it through the model over and over again, taking portions of the phrases out, until the model was positioning the right phrase in those sentences.

This was essentially the birth of the generative AI movement, because they found out that even without context, giving enough examples of how humans communicate symbolically, we can understand human context, fulfilling the third requirement necessary in order to enable generative AI.

One of the reasons Google has been slower to market despite this breakthrough back in 2017, is because they knew about the problems with hallucination and alignment right off the bat, and as a large company, they were very cautious, as they did not want to erode any trust in Google search.

What Does Google’s Platform Look Like Today?

One of the major areas that differentiates Google in this market is that they have to invest in generative AI. It’s the core of how they scale their products. They aren’t building models like Bard and PaLM to create a new revenue stream, these are core to their capabilities for their direct to consumer products. So investment in Google from an AI perspective is not new, and nor will it slow down, what is new is that now Google is allowing companies to gain access to these models via Google Cloud. This is accomplished via their platform, Vertex AI.

This has been around for a while, it’s the machine learning portion of the Google Cloud platform. But, new components have been added around generative AI, e.g. Generative AI App Builder, Generative AI Studio, Model Garden, etc. This is intended to be an easy-to-consume marketplace for deploying and leveraging models on Google Cloud.

Right now, Google has foundational models available in text, code completion and generation, image generation, dialogue or chat, audio and music (soon), and video (soon). These are being invested in from both a B2B and B2C standpoint.

Customizing Your Models: Fine-Tuning vs. Embedding

When considering how to best customize your enterprise models, it’s important to first understand the difference between fine-tuning vs. embedding.

Fine-tuning is the persona of your model – if you want your model to respond in a certain way, almost like giving it an accent, that’s where fine-tuning comes in. This is how you have a model respond the way you want. In the case of PaLM, for example, there is a version for the medical profession, Med-PaLM, where you really want the model to have the vernacular of a medical professional. That’s not trivial, however, it took weeks and millions of dollars to do. Cybersecurity is another area where fine-tuning is often needed. But typically, unless there’s a very specific need to have a model be very precise in terms of how it responds and functions, usually fine-tuning doesn’t have an awesome ROI.

Embedding is augmenting a model’s dataset with an additional form of data (often data that is frequently changing, which doesn’t necessarily need to be a part of the foundational model itself).

For example: If you wanted PaLM to glean insights on internal documents within your organization, you wouldn’t fine-tune the model on those documents, that would just make it “talk” like the document, while still using the data from your general corpus. With embedding, you can take that document, run it through the embeddings API, and ask questions based on that specific data. First, the document is queried, then that’s circumvented to the foundational model to say “hey, summarize this!” The foundational model wouldn’t be able to get that data on its own because it doesn’t have access, and it would just wind up hallucinating a nonsense answer.

Architected for Data Security

There have been a number of early concerns around data security with regards to foundational models. For example, is data being sent to foundational models being collected and used to better train them? For Google, this is segmented completely in Google Cloud. Customer data is not being used to train foundational models. Those weights are frozen, and if you want to fine-tune that model, they create an adaptive layer of neurons (a caching layer) that sit within your projects, like a DMZ between you and the foundational model. So, you can augment the skill set of that foundational model with your own data and your own training that will only live within your project. There obviously is a shared foundational model behind it all, and your question does get sent to that foundational model for interpretation, but it’s not stored and it’s not used for training.

Even then, that can sometimes be a sticking point for companies (people get into the weeds on how the memory works, and how long that memory state lasts etc.). One of the big questions that comes up is “Can I just run an entire version of PaLM inside my own project (because I don’t want anything being shared)?” That is likely going to be the direction in which things evolve. The reality is that if you wanted to run PaLM within your own project today, that’s possible, it’s just a very compute-intensive and cost-prohibitive thing to do right now. The new hotness is going to wind up being how small you can get these models while still performing the base level functions you want. Google is super focused on that, all the way down to being able to run localized models on Android. So eventually, depending on the level of functionality of the model, you’ll have the option to run the whole thing within your project.

Any company that is using customer data to train their own foundational models is a red flag, because fundamentally, what goes in can come out. And it’s usually just in bits and pieces, but if you’re actually training a shared model, there’s a probability that whatever it’s being learned on, a fraction of that can resurface as an answer. So there are very real concerns around IP in that circumstance. Google is being very explicit that that’s not the direction they want to move in.

What is Google’s POV on Open Source?

Google has always been a huge investor in AI, regardless of the current market, because of the business need in their direct to consumer products. These new foundational models are just internal services being exposed to enterprises in a consumable format on the cloud.

So fundamentally, Google is embracing third party and open source models. For them, it’s not about competing, but about bringing them to the platform and allowing you to access them on Google Cloud. Google doesn’t care if you’re consuming their model or someone else’s. There will be some big announcements on this very shortly.

How Should Companies Think About Partnerships?

There will be a lot of experimentation over the next year or two with generative AI. Foundational models are going to be widely available to anyone who wants to use them, so the focus right now should be on private data and the security aspects of that. If you’re experimenting with different cloud providers, you want to make sure that the cloud provider will support 1st party, 3rd party, and open source models, because that’s just going to be the future.

You’ll have an application stack that’s making calls to multiple models – larger foundational models for high-level tasks, smaller more efficient models for some base language processing, etc. So what you should be looking for in the early days is ensuring that you have a choice of models, and that you can land your application stack next to those models.

There will be a lot of pressure for open-source solutions, because depending on who your end customers are, they will ask you to be on xyz as a constraint of them using you. Open-source can be a great answer to that. If you want to be multi-cloud, you’re going to have to be multi-stack. If you have a customer that’s on GCP and a customer that’s on Azure, and you want to use generative AI, then your stack is going to look different, because the way these models respond to prompting is very different. You’ll need some portion of the stack that’s interpreting and responding with PaLM, and then on the Azure side, a completely different portion of the stack will be working with OpenAI. Customers today are already trying to figure this out – load balancing between foundational models is a hard problem to solve. Open-source allows you consistency across cloud providers.

What Will The Pricing Look Like?

Models charge for words in and words out (processing). That’s hard to provide a rough estimate for, because it all depends on what you’re asking the model for and how you’re asking it. When you hear about prompt engineering, teams are basically surrounding a question with context in the prompt, and some of those prompts can get very elaborate – up to thousands of characters long. So depending on how much you need prompting for, or how much you use prompting vs. embedding or fine-tuning, it can sway your costs on a generative AI application pretty heavily.

When most of us as laymen are directly inputting questions to a model, our prompt looks like the question that we asked. But in application development, when you really want to get these models to do complex functions, you’re surrounding it with pages of prompting information to get it to act. That’s not editable by your customer – that’s in the application stack itself. Every one of those characters has a cost associated with it, and depending on how often your application is asking for data from the foundational model, that can exponentially increase the cost. That’s the starting point: look at how you’re going to interact with that model, how often, and how big is the prompt (+ how big is the response back to the model)? Design is very important.

Deepak Seth, Product Director – Trading Architecture Platform Strategy & Support, Charles Schwab

Today we welcome Deepak Seth, a Technology, Innovation and Digital Transformation expert, a deep thinker and a futurist who has handled diverse and progressively increasing responsibilities at leading consulting organizations and with several Fortune 500 companies.

With a background in engineering and management he went back to Post Graduate “school” recently to further hone his knowledge with the latest developments in Data Science, AI and ML. He is a prolific writer and speaker and has authored numerous blog posts, papers, articles, webcast, podcasts and is also actively engaged with the Start-Up ecosystem (that’s how our paths crossed for the first time).

He currently works in the area of TAPSS (Trading Architecture Platform Strategy and Support) with a leading financial services company – Charles Schwab.

Question 1: First Job: What was your first job, and how did it help you build your career?

My basic undergraduate degree was in industrial engineering. Many people used to say industrial engineering is all common sense, and we used to say, ‘Hey! Common sense is not so common.’ So afterwards, I additionally earned postgraduate qualifications in management. These were from the Indian Institute of Technology and Indian Institute of Management.

But, unlike most people who went straight into cushy corporate jobs, I journeyed to the hinterlands of India to join the Family Products Division of what was then Glaxo India Limited. My task was improving the productivity of their operations, but I wore many hats. Part of my role even included riding the milk truck to collect milk (this was a dairy products-based unit). While there, I kicked off an employee suggestions program in order to try and find ways to improve productivity. If any value was derived from the suggestions, those employees were given a share of the resulting profits. This wound up becoming highly successful, and I got an award from the chairman of the company, as, at the time, this was the most popular employee engagement and suggestion program in Glaxo’s history.

That first role was truly foundational when looking across the rest of my career, as I got to experience a huge learning opportunity in driving and engaging people to deliver common objectives. A lot of what I’ve done after, delivering projects and innovations for companies like Xerox, Charles Schwab, or Accenture, has not been about myself, but about harnessing the power of many different people to complete our objectives. It’s all about the team and making sure you’re creating the right environment so that each member can deliver their best results.

Question 2: Leadership: What is the most important leadership skill that you have learned over your career, that has a positive impact, and can you explain with an example?

One thing I’ve learned, which has served me in good stead, is the ability to create the vision, while aligning the stakeholders to deliver key objectives against that vision. Oftentimes, technical roles can force a lot of focus on narrow tasks: e.g. we are upgrading to the next version, or we are installing this software, etc. But people are more engaged when they know they’re not just laying the bricks, but are instead building a cathedral.

This is a very important part of what any leader does, and sometimes, creating that vision means drawing from your own personal experiences. An interesting example that I’ve often used is my experience of tandem skydiving. It really relates closely to the way a project gets executed as a company. Everyone starts out with a lot of fear and uncertainty, but then every subsequent stage of the experience can be linked to the experience of executing a project. This creates some really powerful imagery and messaging.
So that’s the grand vision of building the cathedral, but then you have to go down to the next level of granularity. One thing which I have observed in projects over the years is that failure happens at the seams. So, failure happens at the point where two systems, two individuals, or two processes are interfacing with one another.

The US has one of the greatest track teams in the world, whether it’s the 100M or 400M. But we always seem to miss winning the gold medal in the relays. Why? The teams will frequently drop the baton. So in terms of execution, I always try to reinforce that when moving things from one team to another, or one process to another, the baton has to fully change hands. Your work is not done when you have run your piece, but when the other guy picks things up and runs with them.

This all sounds very basic, but what I have found is that often teams don’t pick up the baton (and don’t say anything), or they have the baton, and it’s dropping, (and they don’t say anything), etc. Small things add up and creating a whole mindset around this has been a very powerful execution tool.

You start with the planning, and then you take it all the way through execution. And you build muscle memory so that it becomes almost second nature that you must communicate. So, it’s a planning thing as well as a communication thing.

Question 3: Prediction: Do you have a prediction around the core technology and core changes that are happening in the industry that we should all spend more time learning about?

I saw the AI bus coming a few years ago. In fact, I went back to school and did some postgraduate qualifications in AI because I didn’t want to be caught off guard when this thing came down the pipe. So, there are two parts to the question:

  1. AI is the biggest thing right now. In the short term we will see intermediaries help companies figure out how to implement generative AI, what to do with it, etc. This could be the big consulting companies, or it could be smaller players who do some prompt engineering stuff. In the medium term we will see some government regulations. Traffic lights, controls, equity, and other aspects will come into play, and companies, institutions, and governments will get a better handle on it. But, I’m most excited about the long term which will be the human and AI interface. We’re seeing things like neural implants and that make impaired humans work effectively (via interaction with AI), but there is room for so much more. We have a lot of redundant capacity in our brains which is unutilized or underutilized (we use less than 5% of the supercomputer in our skulls), and there could be a future where we will leverage human intelligence in conjunction with artificial intelligence. Think of the immense potential that could be unleashed if we were able to harness our brain and link it with AI – definitely far reaching stuff.
  2. AI is the flavor of today and generative AI is the special flavor of today. But bigger than that, I think, is the general discussion around disruption. How can companies be prepared for disruption? Most of the disruptions are not going to be super slow. Companies that figure things out now and get prepared will be able to ride the bus, while others will be left by the wayside. The underlying technology beyond all of these AI developments has been around for years, but the hardware has finally caught up. In a big corporation, it takes some time for everyone to wrap their heads around things. So leaders will need to build internal momentum, identify and prove early use cases, run POCs and establish case studies even if someone else is already doing it. Every company has their own unique set of requirements. Privacy and controls springs to mind – there is a lot of variability across sectors.

Bonus: Is there a certain principle or rule that you live by that you’d like to share as a takeaway for our audience today?

I’ll give you two. The first one is 100% DO IT! This came from one of my bosses at Bausch & Lomb. Focus on 100% Delivery, 100% Ownership, 100% Integrity, and 100% Teamwork. We all used to carry a card that said: ‘100% DO IT! ’

The other one is something from Charles Schwab, and it comes from Mr. Schwab himself: Seeing through the client’s eyes. Whatever we do, we like to frame it in terms of the impact it will have on the client. Always put the client first, No matter what.

This all ties back to what I was saying earlier about creating the vision and the cathedral. We must frame whatever we do in terms of the value we’re delivering for our clients. So when we start looking at everything that way, it becomes a very powerful way of operating.

Deepak Seth is a Technology, Innovation and Digital Transformation expert, a deep thinker and a futurist who has handled diverse and progressively increasing responsibilities at leading consulting organizations and with several Fortune 500 companies.

With a background in engineering and management he went back to Post Graduate “school” recently to further hone his knowledge with the latest developments in Data Science, AI and ML. He is a prolific writer and speaker and has authored numerous blog posts, papers, articles, webcast, podcasts and is also actively engaged with the Start-Up ecosystem (that’s how our paths crossed for the first time).

He currently works in the area of TAPSS (Trading Architecture Platform Strategy and Support) with a leading financial services company – Charles Schwab.

Deepak has been a passionate advocate at Fortune 500 and leading consulting companies for innovation and adoption of emerging/ nascent/disruptive technologies into the enterprise to develop new opportunities and applications. In line with this he is currently actively engaged with Generative AI initiatives and publishes a very popular newsletter DEEPakAI: AI Demystified on LinkedIn.

Geeta Pyne, Chief Enterprise Architect of the Intuit Platform at Intuit

Today we welcome Geeta Pyne, Chief Enterprise Architect of the Intuit Platform at Intuit to our CXO of the Future Podcast. She joined Intuit in January 2021. She started her career as a research scientist developing algorithms for Satellite Image Processing and takes pride in holding IP in India’s first parallel computer PARAM and Image Processing system ISROVISION. She is also a highly experienced Chief Enterprise Architect and Engineering Leader with over 25 years of industry experience and has a proven track record of transforming companies with next-generation architecture strategies, visions, execution roadmaps, and metrics to create and accelerate SaaS and Data platforms.

Question 1: First Job: What was your first job, and how did it help you build your career?

First of all, thank you so much; I think it’s very humbling to come to your podcast. I got to my role through a lot of grit and hard work. 25 plus years in the industry started as a software engine, and I’ll talk a little bit about that, but kind of navigated from engineer to becoming an architect. So really understanding not just that technology – I’m a tech geek – but why we do it. Why does it matter to connect technology with business? Which is why what I do is the enterprise architecture which is truly the glue between strategy and execution. I started, believe it or not, very humble. I started as an engineer, scientist engineering, I was in the India Space research organization for a little bit. You know, I grew up in India. And I studied actually with an India government scholarship, so I was very lucky to have a scholarship from the Government of India all my middle school, high school, and so on.

I did my engineering commuter science, and in the final year everybody goes to a campus recruitment. So you’re getting all these offers. That was good, I was the top one in the class, so I was able to choose what I wanted to do. And I chose this company which is under the department of space.

And I think it was my, I would say, main reason for coming back to the country. I will never be able to pay back. I wouldn’t be talking to you if I did not get that scholarship to go to that school. It was a residential school, but I think about so many things I learned, and I truly believe it really shaped me. Foundation is everything. And I think I learned a lot in this first job. You know, not only being able to work on very hard problems. But also dealing with satellite data before you even knew what Cloud was, before you had all this generative AI, and all that.

I usually like 3, the number 3. I think number one is first principle thinking. If anything seems too complex, you know, challenge every assumption convention. Really go back to this, go back to the drawing board, figure it out, and think about different ways of looking at the problem. So if I draw the analogy. How did I look? You think about the satellite imaging the earth, ultimately think about the lens and the focal slit point, and the point being image, or actually line, a simple straight line. How beautiful is that? It’s like a pinhole. Geometry. I mean, it’s such an easy way to think about it, or visualize it as well. Right? So you start breaking it down, and I think that was one of the foundation things I learned, and I kind of apply every day now.

Secondly, when you’re building software. What great looks like? How do you make sure that you are building things that are going to perform, that are going to scale, that are going to be resilient? Because when things are up in the ear, it doesn’t matter. Very little thing. You can change, right? So really, there is no easy path to getting to the final outcome. You really have to be thought. We have to understand deep take while also solving the right problem and do in a way that is going to give you that beautiful ultimately thing that you are trying to achieve. My first project was radiometric and geometry connection of Noah satellite. I mean, it was a mouthful for me then. I didn’t know anything other than somebody has evolved that ongoing, and I could break it down to a good software.

The third thing is really: knowledge makes you humble. When you are surrounded by all these great people and also multi discipline, you need to have all different aspects of it, whether you are computer science, whether you’re a physicist, whether you are, you know, a chemistry, or, you know, a mechanical engineer. All things have to come together in order to truly make something beautiful. So I think those 2 takeaways have really shaped me. I still think about that first principle thinking. Make sure that the knowledge is making you humble. And then to make sure you are thinking about solving. You know, something that is really going to work. And it’s a durable solution.

Question 2: Leadership: What is the most important leadership skill that you have learned over your career, that has a positive impact, and can you explain with an example?

There are a lot of leadership principles that you follow, whether you look at into it or out of electronics. I think number one is really leading with clarity, having a very clear vision and meeting with that clarity of thoughts and vision is so important. No, it is not about the technology. It is not about getting people aligned with that. And every day I’m trying to get better. I can never say that I have mastered it, because there are more variables that come up.

Also, if you all have a vision. How do you clearly articulate it for all sorts of audiences? And this is also in an architect role, as we say, you have to be able to travel all the flows of the elevator, whether it is the C-suite to all the way up to the engine room. So, being the very clear thoughts and articulate that to all the types of audiences, that to me is the number one.

Secondly, I would say, how do you rally up a team? How do you build a high performing team? You can have all this vision, you got everybody excited, and you actually create a team behind that.

And thirdly, it’s all about results. Right? How do you drive winding results?

I think these 3 are always common, right? And you know, years back, like in 2,000, I used to work for this company called Arrow Electronics, and I was just getting of one project which I was doing for a global transformation. And I said, ‘Okay, I’m kind of done and burnt out. Can you get me off my CIO of in Melvin and put me in something else?’ I didn’t know. I was probably getting from fire to a frying pan or from a frying pan to a fire, whichever way you think about. I was brought into a project to lead the global e-commerce for the components business. When I came in somebody had already made an assumption about how strategy and architecture should be for commerce. And I had to actually really come in as like, ‘Okay, it doesn’t really make sense’. But how do I articulate that very quickly? And I think I’ve learned that, you know, 15 years back, you get it by writing it down clearly, articulating it to my CIO, to all the business stakeholders, and all the way up. It’s not just about my opinion. ‘Do you have evidence to support your ideas? So I think it was a powerful thing. So what happened was I could convince the global CIO of Arrow that you cannot build this, because here are the 3 things that are not going to work. When I articulated the vision, I understood the power of clarity. When I was preaching that globally, the global business leader, Alistair, I remember, called out: “This is, for the record, the first time we have clarity around what we are talking about, and not just the vision, also articulating how will you get that? Not that ‘How might we?’ But what needs to be true in order to achieve that vision.” And I was like, ‘Oh, my God! These people from Australia never says these good things about anyone!’ Being able to articulate that I think is the best. And then, of course, there will be challenges coming later. But that was my execution, right? And then I had to build that team.
Now that you got it, make it happen, right? So I have to scramble, be scrappy, and build a small team, a long-term durable team.

Question 3: Prediction: Do you have a prediction around the core technology and core changes that are happening in the industry that we should all spend more time learning about?

I think, of course, data. Data, data. Did I say data? But, you know, this is just my list. I think spatial data. It had so much importance, and I don’t think we have been leveraging that. Whether it is for security, whether it is for, you know, building something more sustainable and good. I think we got to all try to uncover and learn. What does that outer space have. And how should we leverage the spatial data and then, of course, blend with, amalgamate with other sources and forms of data.

And in order to do that there are basic things like, ‘how are you going to store this? How are you going to put data and business logic together?’ We talk about vector, databases now, but there will be providing the newer versions. And there are ways of storing the logic, everything together with the data and the corresponding language. I think, again, those things will really help us. Whether there are more devices coming, you have to handle that with security and compliance, and ethics. I think we underestimate to know what is the right problem to solve. Are we framing the problem correctly? We are talking about prompt engineering. Before you know, there will be prompt engineers, jobs and Rubik’s everywhere.

But I think is important how you frame a problem. And that will differentiate the people that are solving versus those that are solving the right problem. That is going to be a differentiator, so that blending of critical thinking, understanding the domain, and also multi disciplining, I think the convergence…We have to look at multiple angles and viewpoints. How are you going to leverage data? How are you going to leverage industrial IoT? How are you going to use it for planting the right seed?

Think about how we project on the higher level, start to think about multi-dimensional problems and try to solve them holistically.

Bonus: Is there a certain principle that you live by that you’d like to share as a takeaway for our audience today?

The number first thing I would say is ‘be the change you want to see in the world’. Don’t complain. Just be the change, make it happen. Even if you don’t know whether it is right or wrong. At least make some displacement. I think that’s what I live by every day.

Live, love, and make the world a better place. I think we all have a fully shared responsibility to make the world a better place.

Geeta Pyne, Chief Enterprise Architect of the Intuit Platform, joined Intuit in January 2021.

Geeta is a highly experienced Chief Enterprise Architect and Engineering Leader with over 25 years of industry experience. She has proven track record of transforming companies with next-generation architecture strategies, visions, execution roadmaps, and metrics to create and accelerate SaaS and Data platforms. A Data Strategy & Enterprise Architecture expert who has built global high-performance teams, leading to significant improvements in efficiency, cost savings, and new revenue generation. She started her career as a research scientist developing algorithms for Satellite Image Processing and takes pride in holding IP in India’s first parallel computer PARAM and Image Processing system ISROVISION.

Geeta is also a Board Member/Advisor for Chief Architect Forum & Women in Architecture, SIM, Bay Area, Evanta San Francisco, Gartner company, as well as the Evanta Global CIO, Gartner Peer Insights Ambassador, and GTM Capital Advisory Board.

GenAI in the Developer Landscape

This image was created with Midjourney!

We recently hosted our latest roundtable to hear from our local CTO and VPE peers on how the future of generative AI is going to impact the developer landscape.

In the next few years, will the impact of generative AI on development represent productivity gains or full replacement? Where is it offering the most value to developers today? What are the subsections of generative AI that still have significant gaps? And further: how are engineering leaders currently driving education and experimentation inside of their teams? How are employees being provided with the guidance required to accelerate the desired experimentation while avoiding new risks? What are good leadership practices around this today?

The experts chimed in with some great feedback, and early use cases around what they’re seeing in the market.

Key Takeaways

For Now, It’s Definitely Only Productivity Gains, Developer Jobs are Safe – Github CoPilot and other similar solutions still generate a lot of errors and bugs – they also need more analytics than they have today. Still, productivity gains are already coming down the pipeline. These tools are starting to impact:

  • Dev Iteration
  • Test Automation
  • Code Quality
  • Code Security
  • Documentation

What Kind of Code You’re Writing Matters – GenAI logic is very general. Business logic tends to be different. Segregating code base architecture and separating those two logic structures could make AI solutions much more impactful

Everyone Becomes a Manager – It won’t replace people, it will augment them. Even junior people have to “manage” the AI. This may eventually lead to a restructuring of teams and empowerment of junior employees

This Will Help Ops and Business Teams Leap Into Technology – E.g. product will start contributing to engineering, other teams will lean into these advances. Documentation and text-based work is the most real right now – marketing is a bigger fan than the devs

Expectations from CEOs are High – The reality is a lot more mixed, but we all have to get on the train

The Focus Needs to be on Inexact Use Cases – Great for avoiding the blank canvas problem – gets you 30% of the way there, etc. Things that need binary answers like aspects of dev and security are not good use cases. The tech is still too fuzzy and optimization is tricky

Source Code is Not Your Secret Sauce – It’s not necessarily a big deal to share this with genAI tooling – proprietary or customer data is another matter. Much of this is already in the cloud – is this tooling going to be any different?

On-Prem is a Niche That Will Need to Get Filled – Major competitors like OpenAI and Anthropic are avoiding it

If Models are Commoditized Then Data Becomes the Differentiator – Protect your data

Today’s Guardrails Are Really Soft – Most companies do not centralize access to software through IT – ultimately, developers are going to be downloading and accessing what they want. It’s a speed issue. Larger companies are going to have larger guardrails. LangChain may become the defacto standard to prevent data from leaving the building

Engineering Productivity Will Be Rewritten – DORA 2.0 is going to emerge over the next few years – the companies at the vanguard of generative AI will be reshaping these metrics (e.g. Accuracy, Freshness, etc.)

Code Generation May Not Be the Killer App – Business applications today look a lot more promising. There will also be better communication between business and dev – translation of requirements from what a business person is looking to produce in terms of code, and vice-versa

Early Days Use Cases

Dev and Math: Copilot, Code Interpretation, Source Code Documentation

Generative Art: Midjourney, Adobe Firefly, DAHL-E

Operations: Predictive Maintenance, Access to Ops Data Sooner, Test Case Generation, Logs, Tech Documentation/Internal Knowledge Bases, Document Summarization

Security: Security Operations, Interpretation of Video Footage, Security Design Review, Threat Modeling

Human Resources: Employee/Engineer Onboarding, Resume Upscaling and Generation

Data: Analysis on Disparate Data, Reduction of Manual Work for Data Stewards, Stringing Data Sets Together

Alisa Choong, SVP & CIO of Information & Digital Services and Operations at Shell

Today we welcome Alisa Choong, SVP and CIO of Information & Digital Services and Operations at Shell to our CXO of the Future Podcast. She joined Shell as Executive Vice President of Technical and Competitive IT in August 2015 and was CIO of Projects and Technology since 2018 until she moved to her current position in January 2019, where she’s responsible for end-user strategy for Shell’s employees and contractors globally. Critical to this role is ensuring the delivery of robust, secure, and reliable IT operations that increase productivity and lay the foundations for Shell’s digital transformation.

Question 1: First Job: What was your first job, and how did it help you build your career?

As you know, my name is Alisa Choong. Even though I’m based in the Netherlands, I’m a Malaysian. And I’ve always been a proud Malaysian. So coming to the Netherlands to work in a global environment brings that there’s a lot of diversity that I have to bring in. There’s a lot of inclusiveness for us to work. I graduated with a Bachelor of Economics majoring in Accounting when I left Monash University in Australia. It is quite amazing at that point in time that most of us in Malaysia, for some reason, got overseas education. I was very lucky. My parents believed in educating a woman. And I got sent for my university degree after I finished high school at Monash University. During that time, even in Melbourne, the environment was different.

But my first job was as a tax consultant, can you believe that? And I was a tax consultant in one of the biggest major forces that is still alive today. And what does it actually tell about how it built me to be who I am later in my career? Our test consultants need to be very detail orientated. I got to really understand the businesses of my clients, their activities, and where they operate, because every country has different tax regimes that allow different benefits and deductions that will ensure that we maximize what is available.

More importantly, being a tax consultant also means that I work with a group of specialists, because I may only know the tax corporate part, but I may not know the sales part. And later, they will introduce a lot of goods and services tracks. So always constantly learning. And the greatest continuous learning that finally made me less of a tax consultant is that in every country, every year, there’s a budget. And that means that you must relearn. It is a continuous learning that was torturous at that time, but it’s something that I will always take with me because it forced me to be curious, and it forced me to learn on an annual basis from that.

So, what did it mean to me later in my career? It set the foundation for excellence. Because everything that you do needs to have excellence, redeem the boundaries that are given to you. You must live with those boundaries. There’s no right or wrong and you do not fight it unnecessarily. It also gave me my love for business; understanding the business strategy, how the company makes money, and how companies sustain themselves to be sustainable. It also allowed me to work in a very highly collaborative global team, listening to different advice, because I’m not the specialist for every country on that. And more importantly, it gave me my love for global working and global challenges in complex organizations. I think that’s how it forged my career.

Extra question: What’s your advice to young women as they marched through their careers and try to establish themselves as leaders?

I think it would have been great to have someone tell me this before: ‘Belief in yourself’. You know, my parents believed in me, my father believed in me. But the worst thing is, I did not believe in myself. Every time a new opportunity came, the first idea was, “No, I’m not ready. I can’t do it”. So that not self-belief is something that holds me back for quite a while. To think that you need to be 100% right and sure before you even open your mouth. That is a very ancient trait that I had to adapt later in my life, because I can never be 100%. And now, even if I got 50% right, I’m really happy. So that is an advice that I would give to any young women who is forging a career. Do your best, celebrate your achievement, but more importantly, believe in yourself. Because when people offer you the opportunity, they actually believe in you. So why are you holding yourself back? At what point did you believe in yourself?

I think a lot of it came when I transferred back to Malaysia, after my life in Australia. I was in a car with the managing partner of PricewaterhouseCoopers, we were going to a very important meeting. And he didn’t give me any time to brief him. He gave me in the same car 15 minutes. He was meeting the government officers, meeting CEOs…And he just said, “Give me the crux of what you have done and what I should say”. And I thought, wow, I must be good.

He really believed in me and trusted me to get him ready in 15 minutes in a car. Giving me the voice to give him an alternative opinion in the meeting really struck me when I was in my early 30s. So, it made a lot of difference thereafter.

Question 2: Leadership: What is the most important leadership skill that you have learned over your career, that has a positive impact, and can you explain with an example?

One of the things that always has a bias is that women are not strategic, that women are executors. Somehow, that is still the perception. Especially when you’re doing work that you don’t carry the title for, like strategy or planning…; you know, you can carry other titles, and people think you are the doer. But to me, as a leader, you need to be visionary, because you’re leading a big team. I lead a global team of over 3000. That’s quite awesome, but you need to have a vision. You need to have a roadmap, and more importantly, you need to be able to communicate. For me, that was the major learning. You can’t just tell a vision once. You can’t tell a vision for 30,000 feet and expect that it will go down to every single level. One important thing is making sure that you have different communication channels, and that you communicate differently to different stakeholders so that it touches the heart. For me, that communication is left to right. It basically addresses the needs of people who are doing today’s technical operations, tell them what they’re doing, and how are they setting up a foundation that will lead to the future and the transformation you want to see. Because you cannot ask them to take the jump off the cliff. Now, they don’t want to do that. They want to make sure that they’re safe. I can understand that because I was there once. You must communicate it, so everybody has this great vision. Vision doesn’t drop off the sky. Making the message very specific to different level groups is a major thing that a leader needs to keep in mind.

And to me, the last critical success factor is all about being diverse and inclusive and being equitable. The sharing of different thoughts and opinions. You can only do that when the rubber hits the road. When you go out, you want to see how your vision is being executed. If you have an open environment people will give you feedback, tell you what is doable. What are the challenges, what are the barriers, so that you can make the corrections? But this is only possible if you build an environment where people feel safe to talk and they are recognized for having the courage to tell you the truth. Being diverse, being inclusive is good. And again, sometimes it’s very hard to listen that your brilliant idea is just not working. But you know it in advance so that you don’t know in 24 months, and you can cause correct. Have a vision, communicate. And be patient when you want to drive change on that and create a culture of diversity.

Question 3: Prediction: Do you have a prediction around the core technology and core changes that are happening in the industry that we should all spend more time learning about?

I think everybody knows the biggest topic is Artificial Intelligence. We’re not even at the really innovative stage yet. But already, with this artificial intelligence, we’re at the point of really seeing a lot of challenges. There’s a need for more computing power, there’s a need for speed. So, understanding the trends, and what content computing is going to bring is going to be very important. Most importantly, because decisions are made in a split minute, with the use of AI, bots, and everything. Having a reliable network, having reliable connectivity, that can’t fail. Can you imagine a world without the internet for half a day? I can’t now. I can’t even imagine it. I kind of over imagine walking out from my door, and not having my mobile phone. You know, it’s just so scary on that basis. So, with artificial intelligence and its growth, it also comes with some risks. The whole concept of AI, the faster data processing with great connectivity, comes with some ethical questions. How do you address and really understand that part of it? Because you wouldn’t know in 10 years’ time if the decision that you made with the help of AI might come back and bite you.
And, of course, with the need for more data computing, as I work in an energy company, I’m very worried about climate change. Because the need for more data center that grows at 5x each year means there’s more new global power that needs to be there. And people don’t see that.

Bonus: Is there a certain principle that you live by that you’d like to share as a takeaway for our audience today?

I think as a leader, it took me a while to figure out what it was to be an Asian female technology leader. There are a lot of female energy leaders in the world, but I think the major difference is that I have really adopted an identity because I’m a Malaysian and proud to be a Malaysian. So, it’s hard to actually be adaptable in a global working world and be true to myself and my identity. It may sound like I talk strangely; my sentence construction is different. But it doesn’t mean that I’m not there with the rest of the leaders in the global world. So that is a key learning.

And the second one is, a lot of people ask, what’s your key success factor? I think to me is how do I merge the courage of the West with the tremendous, respectful, and humbleness of the East? You want to be bold, but sometimes it can be conflicting. So, merging the best of the West, which means courage, and also transforming and being very respectful in all your dealings has been one of my critical success factors.

Alisa Choong, SVP and CIO of Information & Digital Services and Operations at Shell to our CXO of the Future Podcast (“The Three Questions” edition). She joined Shell as Executive Vice President of Technical and Competitive IT in August 2015 and was CIO of Projects and Technology since 2018 until she moved to her current position in January 2019, where she’s responsible for end-user strategy for Shell’s employees and contractors globally.

Prior to Shell, Alisa worked for a number of Fortune corporations – PETRONAS, IBM, PricewaterhouseCoopers, National Bank of Australia, and KPMG.
In her current role as CIO of Projects and Technology, in addition to running a reliable and secure IT operations for the business, her focus is partnering with the senior stakeholders to embed digital into the business strategy, driving pervasive replication to unlock the business value thru a strong IT digital backbone across the global breath where the business has a presence. She is a Malaysian citizen currently on assignment in The Hague, Netherlands.

Mayfield Portfolio Contractor Survey 2023

Welcome to our inaugural Mayfield Portfolio Contract Survey. Our goal has been to gather a list of the best agencies or contractors that have supported our extended portfolio, as we know that great referenceable contractors for website, presentation development, CRM, SEO, etc. are critical to early-stage company growth. We asked you and all your peers for your favorite experts.  These are all companies that were specifically endorsed because of very positive experiences. These recommendations also include the name or names of those who recommended them. If you’d like direct introductions, we will put you in touch with the person who nominated them.

CEO & Executive Coaching

Engineering

DevOps

Software Development Contracting Services

Finance

Accounting

 

Fractional CFO

Structure and Best Practices

Hardware

Hardware Design and Prototype Manufacturing

PCB Assembly

Industrial Design

IP Design Services

Micromolding and Tooling

Material Sciences for Specialized Biosensors

Package Design

Legal

409A

Ai-Bias

FedRAMP

Third-Party Whistleblower Hotline

Marketing

Account-Based Marketing

Content

Copywriting

DTC Growth Marketing

 

Brand, Design, Web

Event Booth Production and Storage

Marketing/Product Training and Development

PLG

PR

SEO and Paid Media

Video Production

Recruiting

Sales

APAC Business Development

B2B Leadgen and Outsourced BDR/SDR

Channels Development / Enablement

Content Repository / Sales Deal Rooms

CRM Management

Fractional Sales Leadership

GTM Consulting

RevOps as a Service

Sales Training

Sales Workspace