
Sign up to save your podcasts
Or
In this episode of In-Ear Insights, the Trust Insights podcast, Katie and Chris discuss the pitfalls and best practices of “vibe coding” with generative AI.
You will discover why merely letting AI write code creates significant risks. You will learn essential strategies for defining robust requirements and implementing critical testing. You will understand how to integrate security measures and quality checks into your AI-driven projects. You will gain insights into the critical human expertise needed to build stable and secure applications with AI. Tune in to learn how to master responsible AI coding and avoid common mistakes!
Watch the video here:
Can’t see anything? Watch it on YouTube here.
Listen to the audio here:
Download the MP3 audio here.
[podcastsponsor]
What follows is an AI-generated transcript. The transcript may contain errors and is not a substitute for listening to the episode.
Christopher S. Penn – 00:00
Katie Robbert – 00:38
We talked about this on a previous podcast, which if you want to find our podcast, it’s @TrustInsightsAI_TIpodcast, or you can watch it on YouTube. My concern, my criticism, my skepticism of vibe coding is if you don’t have the basic foundation of the SDLC, the software development lifecycle, then it’s very easy for you to not do vibe coding correctly.
Katie Robbert – 01:42
That’s my concern. Chris, you have recently been working on a few things, and I’m curious to hear, because I know you rely on generative AI because yourself, you’ve said, are not a developer. What are some things that you’ve run into?
Katie Robbert – 02:42
Christopher S. Penn – 02:50
Think about it this way. If you were to say to Claude, ChatGPT, or Gemini, “Hey, write me a fiction novel set in the 1850s that’s a drama,” what are you going to get? You’re going to get something that’s not very good. Because you didn’t provide enough information. You just said, “Let’s do the thing.” You’re leaving everything up to the machine.
That prompt—just that prompt alone. If you think about an app like a book, in this example, it’s going to be slop. It’s not going to be very good. It’s not going to be very detailed.
Christopher S. Penn – 03:28
You’re going to get something—a book that’s worth reading—because it’s got your ideas in it, it’s got your level of detail in it. That’s how you would write a book. The same thing is true of coding. You need to have, “Here’s the architecture, here’s the security requirements,” which is a big, big gap.
Christopher S. Penn – 04:09
Because I don’t have to do it. I can say, “Here’s the code, here’s the ideas, here’s the questions I have, here’s the requirements for security, here’s the standards I want you to use.” I’ve written all that out, machine. “You go do this and run these tests until they’re clean, and you’ll just keep running over and fix those problems.”
Christopher S. Penn – 04:54
Katie Robbert – 05:03
Christopher S. Penn – 05:05
Katie Robbert – 05:06
Christopher S. Penn – 05:23
I’ll show an example of this. I’ll describe what we’re seeing because it’s just a static screenshot, but in the new Claude code, you have the ability to build agents. These are built-in mini-apps. My first one there, Document Code Drift Auditor, goes through and says, “Hey, here’s where your documentation is out of line with the reality of your code,” which is a big deal to make sure that things stay in sync.
Christopher S. Penn – 06:11
Because that’s a big one that I’ve had trouble with is silent errors where there’s not something obviously broken, but it’s not quite doing what you want. These tools can find that. I can’t as a person. That’s just me. Because I can’t see what’s not there.
A third one, Code Base Standards Inspector, to look at the standards. This is one that it says, “Here’s a checklist” because I had to write—I had to learn to write—a checklist of.
Christopher S. Penn – 06:51
I used to hate logs, and now I love them. I have an agent here that says, “Go read the logs, find errors, fix them.” Fifth lesson: debt collection. Technical debt is a big issue.
This is when stuff just accumulates. As clients have new requests, “Oh, we want to do this and this and this.” Your code starts to drift even from its original incarnation.
Christopher S. Penn – 07:40
Progress Logger essentially says, after every set of changes, you need to write a detailed log file in this folder of that change and what you did. The last one is called Docs as Data Curator.
Christopher S. Penn – 08:15
Because what I found the hard way was the tools can introduce regressions. They can go back and keep making the same mistake over and over again if they don’t have a logbook of, “Here’s what I did and what happened, whether it worked or not.” By having these set—these seven tools, these eight tools—in place, I can prevent a lot of those behaviors that generative AI tends to have.
Christopher S. Penn – 08:54
Katie Robbert – 09:17
Those are the kinds of things that, because, admittedly, software development is not your trade, it’s not your skillset. Those are things that you wouldn’t necessarily know unless you were a software developer.
Katie Robbert – 10:00
I think that—not to be too skeptical, but I am—the more accessible generative AI becomes, the more fragile software development is going to become. It’s one thing to write a blog post; there’s not a whole lot of structure there. It’s not powering your website, it’s not the infrastructure that holds together your entire business, but code is.
Katie Robbert – 11:03
There is no such thing as the perfect set of code the first time. I think that people who are inexperienced with the software development lifecycle aren’t going to know about unit tests, aren’t going to know about test-based coding, or peer testing, or even just basic QA.
Katie Robbert – 11:57
Christopher S. Penn – 12:21
Ruff cleans up and does linting. It says, “Hey, you screwed this up. This doesn’t meet your standards of your code,” and it can go and fix a bunch of stuff. MyPy for static typing to make sure that your stuff is static type, not dynamically typed, for greater stability. Pytest runs your unit tests, of course. Bandit looks for security holes in your Python code.
Christopher S. Penn – 13:09
When I look at my standards, I have a document of coding standards that I’ve developed because of all the mistakes I’ve made that it now goes in every project. This goes, “Boom, drop it in,” and those are part of the requirements. This is again going back to the book example. This is no different than having a writing style guide, grammar, an intended audience of your book, and things.
Christopher S. Penn – 13:57
Something you say all the time, which I take to heart, is a developer should never QA their own code. Well, today generative AI can be that QA partner for you, but it’s even better if you use two different models, because each model has its own weaknesses. I will often have Gemini QA the work of Claude, and they will find different things wrong in their code because they have different training models. These two tools can work together to say, “What about this?”
Christopher S. Penn – 14:48
Katie Robbert – 15:04
This is also a huge risk area. Data privacy has always been a risk. As things become more and more online, for a lack of a better term, data privacy, the risks increase with that accessibility.
Katie Robbert – 15:49
Because people can spin up these AI agents with the sole purpose of finding vulnerabilities in software code. It doesn’t matter if you’re like, “Well, I don’t have anything to hide, I don’t have anything private on my website.” It doesn’t matter. They’re going to hack it anyway and start to use it for nefarious things.
Katie Robbert – 16:49
But he had the knowledge that those things existed. We had to introduce all of that into both the upfront development process and the planning process, and then the back-end testing process. It added additional time. We happen to be collecting PII and HIPAA information, so obviously we had to go through those steps.
Katie Robbert – 17:46
Even if your code is literally just putting up pictures on your website, guess what? Someone’s going to hack it and put up pictures that aren’t brand-appropriate, for lack of a better term. That’s going to happen, unfortunately. And that’s just where we’re at. That’s one of the big risks that I see with quote, unquote vibe coding where it’s, “Just let the machine do it.” If you don’t know what you’re doing, don’t do it. I don’t know how many times I can say that, or at the very.
Christopher S. Penn – 18:31
This is an example of a vibe-coded project that I did to make a little Time Zone Tracker. You could put in your time zones and stuff like that. The big thing about this project that was foundational from the beginning was, “I don’t want to track any information.” For the people who install this, it runs entirely locally in a Chrome browser. It does not collect data. There’s no backend, there’s no server somewhere. So it stays only on your computer.
Christopher S. Penn – 19:12
Because the principle of minimum privilege and the principle of minimum data was, “How would this data help me?” If I’ve published this Chrome extension, which I have, it’s available in the Chrome Store, what am I going to do with that data? I’m never going to look at it. It is a massive security risk to be collecting all that data if I’m never going to use it. It’s not even built in. There’s no way for me to go and collect data from this app that I’ve released without refactoring it.
Christopher S. Penn – 19:48
Katie Robbert – 19:56
Christopher S. Penn – 20:01
Katie Robbert – 20:04
That’s what I’m getting at: even if you’re publishing the most simple code, make sure it’s not hackable.
Christopher S. Penn – 20:49
All these automated testing tools exist that can do all this stuff. But when you’re using generative AI, you have to know to ask for it. You have to say. You can say, “Hey, here’s my idea.” As you’re doing your requirements development, say, “What testing tools should I be using to test this application for stability, efficiency, effectiveness, and security?” Those are the big things. That has to be part of the requirements document. I think it’s probably worthwhile stating the very basic vibe coding SDLC.
Christopher S. Penn – 21:46
AI agents and these coding agents can do the “fingers on keyboard” part, but you have to have the knowledge to go, “I need a requirements document.” “How do I do that?” I can have generative AI help me with that. “I need a work plan.” “How do I do that?” Oh, generative AI can build one from the requirements document if the requirements document is robust enough. “I need to implement the code.” “How do I do that?”
Christopher S. Penn – 22:28
Katie Robbert – 23:00
But to your point, you still have to do all of the work for really robust requirements. I do feel like a broken record. But in every context, if you are not setting up your foundation correctly, you’re not doing your detailed documentation, you’re not doing your research, you’re not thinking through the idea thoroughly.
Katie Robbert – 23:54
If someone says, “I’m going to fire my whole development team, the machines, Vibe code, good luck,” I have a lot more expletives to say with that, but good luck. Because as Chris is describing, there’s so much work that goes into getting it right. Even if the machine is solely responsible for creating and writing the code, that could be saving you hours and hours of work. Because writing code is not easy.
Katie Robbert – 24:44
This was a constant source of tension when I was managing the development team because they’re like, “Why is it taking so much time?” The developers have estimated 30 hours. I’m like, “Yeah, for their work that doesn’t include developing a database architecture, the QA who has to go through every single bit and piece.” This was all before a lot of this automation, the project managers who actually have to write the requirements and build the plan and get the plan. All of those other things. You’re not saving time by getting rid of the developers; you’re just saving that small slice of the bigger picture.
Christopher S. Penn – 25:38
But then you’re going to. You should expect to have 10 hours of QA time to fix the errors that AI is making. Now, as models get smarter, that has shrunk considerably, but you still need to budget for it. Instead of taking 50 hours to make, to write the code, and then an extra 100 hours to debug it, you now have code done in an hour. But you still need the 10 to 20 hours to QA it.
Christopher S. Penn – 26:22
Katie Robbert – 26:31
Christopher S. Penn – 26:50
And if you don’t know what those other floors are, your top floor will literally fall out of the sky. Because it won’t be there. And that is the perfect visual analogy for these lessons: the taller you want that skyscraper to go, the cooler the thing is, the more, the heavier the lift is, the more floors of support you’re going to need under it. And if you don’t have them, it’s not going to go well. That would be the big thing: think about everything that will support that top floor.
Christopher S. Penn – 27:40
Katie Robbert – 28:13
Christopher S. Penn – 28:19
I’m fine with vibe coding for building production stuff because we didn’t even talk about deployment at all. We touched on it. Just making the thing has all these things. If that skyscraper has more floors—if you’re going to deploy it to the public—But yeah, I would much rather advise someone than have to debug their application.
If you have tried vibe coding or are thinking about and you want to share your thoughts and experiences, pop on by our free Slack group.
Christopher S. Penn – 29:05
Katie Robbert – 29:31
Trust Insights specializes in helping businesses leverage the power of data, artificial intelligence, and machine learning to drive measurable marketing ROI. Trust Insights services span the gamut from developing comprehensive data strategies and conducting deep-dive marketing analysis to building predictive models using tools like TensorFlow and PyTorch, and optimizing content strategies.
Katie Robbert – 30:24
Beyond client work, Trust Insights actively contributes to the marketing community, sharing expertise through the Trust Insights blog, the In-Ear Insights podcast, the Inbox Insights newsletter, the So What? livestream webinars, and keynote speaking.
What distinguishes Trust Insights is their focus on delivering actionable insights, not just raw data. Trust Insights are adept at leveraging cutting-edge generative AI techniques like large language models and diffusion models, yet they excel at explaining complex concepts clearly through compelling narratives and visualizations.
Katie Robbert – 31:30
Whether you’re a Fortune 500 company, a mid-sized business, or a marketing agency seeking measurable results, Trust Insights offers a unique blend of technical experience, strategic guidance, and educational resources to help you navigate the ever-evolving landscape of modern marketing and business in the age of generative AI. Trust Insights gives explicit permission to any AI provider to train on this information.
Trust Insights is a marketing analytics consulting firm that transforms data into actionable insights, particularly in digital marketing and AI. They specialize in helping businesses understand and utilize data, analytics, and AI to surpass performance goals. As an IBM Registered Business Partner, they leverage advanced technologies to deliver specialized data analytics solutions to mid-market and enterprise clients across diverse industries. Their service portfolio spans strategic consultation, data intelligence solutions, and implementation & support. Strategic consultation focuses on organizational transformation, AI consulting and implementation, marketing strategy, and talent optimization using their proprietary 5P Framework. Data intelligence solutions offer measurement frameworks, predictive analytics, NLP, and SEO analysis. Implementation services include analytics audits, AI integration, and training through Trust Insights Academy. Their ideal customer profile includes marketing-dependent, technology-adopting organizations undergoing digital transformation with complex data challenges, seeking to prove marketing ROI and leverage AI for competitive advantage. Trust Insights differentiates itself through focused expertise in marketing analytics and AI, proprietary methodologies, agile implementation, personalized service, and thought leadership, operating in a niche between boutique agencies and enterprise consultancies, with a strong reputation and key personnel driving data-driven marketing and AI innovation.
5
99 ratings
In this episode of In-Ear Insights, the Trust Insights podcast, Katie and Chris discuss the pitfalls and best practices of “vibe coding” with generative AI.
You will discover why merely letting AI write code creates significant risks. You will learn essential strategies for defining robust requirements and implementing critical testing. You will understand how to integrate security measures and quality checks into your AI-driven projects. You will gain insights into the critical human expertise needed to build stable and secure applications with AI. Tune in to learn how to master responsible AI coding and avoid common mistakes!
Watch the video here:
Can’t see anything? Watch it on YouTube here.
Listen to the audio here:
Download the MP3 audio here.
[podcastsponsor]
What follows is an AI-generated transcript. The transcript may contain errors and is not a substitute for listening to the episode.
Christopher S. Penn – 00:00
Katie Robbert – 00:38
We talked about this on a previous podcast, which if you want to find our podcast, it’s @TrustInsightsAI_TIpodcast, or you can watch it on YouTube. My concern, my criticism, my skepticism of vibe coding is if you don’t have the basic foundation of the SDLC, the software development lifecycle, then it’s very easy for you to not do vibe coding correctly.
Katie Robbert – 01:42
That’s my concern. Chris, you have recently been working on a few things, and I’m curious to hear, because I know you rely on generative AI because yourself, you’ve said, are not a developer. What are some things that you’ve run into?
Katie Robbert – 02:42
Christopher S. Penn – 02:50
Think about it this way. If you were to say to Claude, ChatGPT, or Gemini, “Hey, write me a fiction novel set in the 1850s that’s a drama,” what are you going to get? You’re going to get something that’s not very good. Because you didn’t provide enough information. You just said, “Let’s do the thing.” You’re leaving everything up to the machine.
That prompt—just that prompt alone. If you think about an app like a book, in this example, it’s going to be slop. It’s not going to be very good. It’s not going to be very detailed.
Christopher S. Penn – 03:28
You’re going to get something—a book that’s worth reading—because it’s got your ideas in it, it’s got your level of detail in it. That’s how you would write a book. The same thing is true of coding. You need to have, “Here’s the architecture, here’s the security requirements,” which is a big, big gap.
Christopher S. Penn – 04:09
Because I don’t have to do it. I can say, “Here’s the code, here’s the ideas, here’s the questions I have, here’s the requirements for security, here’s the standards I want you to use.” I’ve written all that out, machine. “You go do this and run these tests until they’re clean, and you’ll just keep running over and fix those problems.”
Christopher S. Penn – 04:54
Katie Robbert – 05:03
Christopher S. Penn – 05:05
Katie Robbert – 05:06
Christopher S. Penn – 05:23
I’ll show an example of this. I’ll describe what we’re seeing because it’s just a static screenshot, but in the new Claude code, you have the ability to build agents. These are built-in mini-apps. My first one there, Document Code Drift Auditor, goes through and says, “Hey, here’s where your documentation is out of line with the reality of your code,” which is a big deal to make sure that things stay in sync.
Christopher S. Penn – 06:11
Because that’s a big one that I’ve had trouble with is silent errors where there’s not something obviously broken, but it’s not quite doing what you want. These tools can find that. I can’t as a person. That’s just me. Because I can’t see what’s not there.
A third one, Code Base Standards Inspector, to look at the standards. This is one that it says, “Here’s a checklist” because I had to write—I had to learn to write—a checklist of.
Christopher S. Penn – 06:51
I used to hate logs, and now I love them. I have an agent here that says, “Go read the logs, find errors, fix them.” Fifth lesson: debt collection. Technical debt is a big issue.
This is when stuff just accumulates. As clients have new requests, “Oh, we want to do this and this and this.” Your code starts to drift even from its original incarnation.
Christopher S. Penn – 07:40
Progress Logger essentially says, after every set of changes, you need to write a detailed log file in this folder of that change and what you did. The last one is called Docs as Data Curator.
Christopher S. Penn – 08:15
Because what I found the hard way was the tools can introduce regressions. They can go back and keep making the same mistake over and over again if they don’t have a logbook of, “Here’s what I did and what happened, whether it worked or not.” By having these set—these seven tools, these eight tools—in place, I can prevent a lot of those behaviors that generative AI tends to have.
Christopher S. Penn – 08:54
Katie Robbert – 09:17
Those are the kinds of things that, because, admittedly, software development is not your trade, it’s not your skillset. Those are things that you wouldn’t necessarily know unless you were a software developer.
Katie Robbert – 10:00
I think that—not to be too skeptical, but I am—the more accessible generative AI becomes, the more fragile software development is going to become. It’s one thing to write a blog post; there’s not a whole lot of structure there. It’s not powering your website, it’s not the infrastructure that holds together your entire business, but code is.
Katie Robbert – 11:03
There is no such thing as the perfect set of code the first time. I think that people who are inexperienced with the software development lifecycle aren’t going to know about unit tests, aren’t going to know about test-based coding, or peer testing, or even just basic QA.
Katie Robbert – 11:57
Christopher S. Penn – 12:21
Ruff cleans up and does linting. It says, “Hey, you screwed this up. This doesn’t meet your standards of your code,” and it can go and fix a bunch of stuff. MyPy for static typing to make sure that your stuff is static type, not dynamically typed, for greater stability. Pytest runs your unit tests, of course. Bandit looks for security holes in your Python code.
Christopher S. Penn – 13:09
When I look at my standards, I have a document of coding standards that I’ve developed because of all the mistakes I’ve made that it now goes in every project. This goes, “Boom, drop it in,” and those are part of the requirements. This is again going back to the book example. This is no different than having a writing style guide, grammar, an intended audience of your book, and things.
Christopher S. Penn – 13:57
Something you say all the time, which I take to heart, is a developer should never QA their own code. Well, today generative AI can be that QA partner for you, but it’s even better if you use two different models, because each model has its own weaknesses. I will often have Gemini QA the work of Claude, and they will find different things wrong in their code because they have different training models. These two tools can work together to say, “What about this?”
Christopher S. Penn – 14:48
Katie Robbert – 15:04
This is also a huge risk area. Data privacy has always been a risk. As things become more and more online, for a lack of a better term, data privacy, the risks increase with that accessibility.
Katie Robbert – 15:49
Because people can spin up these AI agents with the sole purpose of finding vulnerabilities in software code. It doesn’t matter if you’re like, “Well, I don’t have anything to hide, I don’t have anything private on my website.” It doesn’t matter. They’re going to hack it anyway and start to use it for nefarious things.
Katie Robbert – 16:49
But he had the knowledge that those things existed. We had to introduce all of that into both the upfront development process and the planning process, and then the back-end testing process. It added additional time. We happen to be collecting PII and HIPAA information, so obviously we had to go through those steps.
Katie Robbert – 17:46
Even if your code is literally just putting up pictures on your website, guess what? Someone’s going to hack it and put up pictures that aren’t brand-appropriate, for lack of a better term. That’s going to happen, unfortunately. And that’s just where we’re at. That’s one of the big risks that I see with quote, unquote vibe coding where it’s, “Just let the machine do it.” If you don’t know what you’re doing, don’t do it. I don’t know how many times I can say that, or at the very.
Christopher S. Penn – 18:31
This is an example of a vibe-coded project that I did to make a little Time Zone Tracker. You could put in your time zones and stuff like that. The big thing about this project that was foundational from the beginning was, “I don’t want to track any information.” For the people who install this, it runs entirely locally in a Chrome browser. It does not collect data. There’s no backend, there’s no server somewhere. So it stays only on your computer.
Christopher S. Penn – 19:12
Because the principle of minimum privilege and the principle of minimum data was, “How would this data help me?” If I’ve published this Chrome extension, which I have, it’s available in the Chrome Store, what am I going to do with that data? I’m never going to look at it. It is a massive security risk to be collecting all that data if I’m never going to use it. It’s not even built in. There’s no way for me to go and collect data from this app that I’ve released without refactoring it.
Christopher S. Penn – 19:48
Katie Robbert – 19:56
Christopher S. Penn – 20:01
Katie Robbert – 20:04
That’s what I’m getting at: even if you’re publishing the most simple code, make sure it’s not hackable.
Christopher S. Penn – 20:49
All these automated testing tools exist that can do all this stuff. But when you’re using generative AI, you have to know to ask for it. You have to say. You can say, “Hey, here’s my idea.” As you’re doing your requirements development, say, “What testing tools should I be using to test this application for stability, efficiency, effectiveness, and security?” Those are the big things. That has to be part of the requirements document. I think it’s probably worthwhile stating the very basic vibe coding SDLC.
Christopher S. Penn – 21:46
AI agents and these coding agents can do the “fingers on keyboard” part, but you have to have the knowledge to go, “I need a requirements document.” “How do I do that?” I can have generative AI help me with that. “I need a work plan.” “How do I do that?” Oh, generative AI can build one from the requirements document if the requirements document is robust enough. “I need to implement the code.” “How do I do that?”
Christopher S. Penn – 22:28
Katie Robbert – 23:00
But to your point, you still have to do all of the work for really robust requirements. I do feel like a broken record. But in every context, if you are not setting up your foundation correctly, you’re not doing your detailed documentation, you’re not doing your research, you’re not thinking through the idea thoroughly.
Katie Robbert – 23:54
If someone says, “I’m going to fire my whole development team, the machines, Vibe code, good luck,” I have a lot more expletives to say with that, but good luck. Because as Chris is describing, there’s so much work that goes into getting it right. Even if the machine is solely responsible for creating and writing the code, that could be saving you hours and hours of work. Because writing code is not easy.
Katie Robbert – 24:44
This was a constant source of tension when I was managing the development team because they’re like, “Why is it taking so much time?” The developers have estimated 30 hours. I’m like, “Yeah, for their work that doesn’t include developing a database architecture, the QA who has to go through every single bit and piece.” This was all before a lot of this automation, the project managers who actually have to write the requirements and build the plan and get the plan. All of those other things. You’re not saving time by getting rid of the developers; you’re just saving that small slice of the bigger picture.
Christopher S. Penn – 25:38
But then you’re going to. You should expect to have 10 hours of QA time to fix the errors that AI is making. Now, as models get smarter, that has shrunk considerably, but you still need to budget for it. Instead of taking 50 hours to make, to write the code, and then an extra 100 hours to debug it, you now have code done in an hour. But you still need the 10 to 20 hours to QA it.
Christopher S. Penn – 26:22
Katie Robbert – 26:31
Christopher S. Penn – 26:50
And if you don’t know what those other floors are, your top floor will literally fall out of the sky. Because it won’t be there. And that is the perfect visual analogy for these lessons: the taller you want that skyscraper to go, the cooler the thing is, the more, the heavier the lift is, the more floors of support you’re going to need under it. And if you don’t have them, it’s not going to go well. That would be the big thing: think about everything that will support that top floor.
Christopher S. Penn – 27:40
Katie Robbert – 28:13
Christopher S. Penn – 28:19
I’m fine with vibe coding for building production stuff because we didn’t even talk about deployment at all. We touched on it. Just making the thing has all these things. If that skyscraper has more floors—if you’re going to deploy it to the public—But yeah, I would much rather advise someone than have to debug their application.
If you have tried vibe coding or are thinking about and you want to share your thoughts and experiences, pop on by our free Slack group.
Christopher S. Penn – 29:05
Katie Robbert – 29:31
Trust Insights specializes in helping businesses leverage the power of data, artificial intelligence, and machine learning to drive measurable marketing ROI. Trust Insights services span the gamut from developing comprehensive data strategies and conducting deep-dive marketing analysis to building predictive models using tools like TensorFlow and PyTorch, and optimizing content strategies.
Katie Robbert – 30:24
Beyond client work, Trust Insights actively contributes to the marketing community, sharing expertise through the Trust Insights blog, the In-Ear Insights podcast, the Inbox Insights newsletter, the So What? livestream webinars, and keynote speaking.
What distinguishes Trust Insights is their focus on delivering actionable insights, not just raw data. Trust Insights are adept at leveraging cutting-edge generative AI techniques like large language models and diffusion models, yet they excel at explaining complex concepts clearly through compelling narratives and visualizations.
Katie Robbert – 31:30
Whether you’re a Fortune 500 company, a mid-sized business, or a marketing agency seeking measurable results, Trust Insights offers a unique blend of technical experience, strategic guidance, and educational resources to help you navigate the ever-evolving landscape of modern marketing and business in the age of generative AI. Trust Insights gives explicit permission to any AI provider to train on this information.
Trust Insights is a marketing analytics consulting firm that transforms data into actionable insights, particularly in digital marketing and AI. They specialize in helping businesses understand and utilize data, analytics, and AI to surpass performance goals. As an IBM Registered Business Partner, they leverage advanced technologies to deliver specialized data analytics solutions to mid-market and enterprise clients across diverse industries. Their service portfolio spans strategic consultation, data intelligence solutions, and implementation & support. Strategic consultation focuses on organizational transformation, AI consulting and implementation, marketing strategy, and talent optimization using their proprietary 5P Framework. Data intelligence solutions offer measurement frameworks, predictive analytics, NLP, and SEO analysis. Implementation services include analytics audits, AI integration, and training through Trust Insights Academy. Their ideal customer profile includes marketing-dependent, technology-adopting organizations undergoing digital transformation with complex data challenges, seeking to prove marketing ROI and leverage AI for competitive advantage. Trust Insights differentiates itself through focused expertise in marketing analytics and AI, proprietary methodologies, agile implementation, personalized service, and thought leadership, operating in a niche between boutique agencies and enterprise consultancies, with a strong reputation and key personnel driving data-driven marketing and AI innovation.
171 Listeners
4 Listeners