Ed Charbeneau, in a recent discussion, highlighted practical strategies for deploying AI models in production. Known for his work as Principal Developer Advocate at Progress Software, Charbeneau emphasized managing non-deterministic AI behavior, integrating agentic AI into developer workflows, and optimizing tools and context management to enhance productivity and reliability in both mobile and web applications.
Ed Charbeneau: When integrating GenAI models into production, it's important to consider the technology's non-deterministic behavior. While some consistency is possible, there's always some variance from one output to the next. Additionally, upgrading or changing models can introduce unexpected results as new models process prompts differently. When changing models, some prompts may need to be simplified or rewritten.
Ed Charbeneau: AI in disconnected environments such as mobile and web (Edge AI) often use local models. Local models are chosen for their speed, battery efficiency, and ability to run without a connection to the internet. However, the models need to sacrifice accuracy to meet the limitations imposed by the device's memory, battery, and other factors.
Ed Charbeneau: Monitoring AI post-launch effectively is done through multiple channels that include automated testing, telemetry, and human-in-the-loop. Automated testing is ideal for identifying issues quickly, while collecting telemetry data with tools like Telerik Fiddler Everywhere Reporter or OpenTelemetry is useful for diagnosing long-term issues or detecting downtime. With technology like GenAI (where variance is possible), having a human in the loop as a checkpoint is a valuable tool for ensuring generative results are acceptable.
Ed Charbeneau: There is currently a shift happening in traditional coding and programming. The shift is pushing developers from writing every line of code to having more of a supervisory role. With MCP tool calling, agents now focus on cohesive planning and execution rather than isolated tasks. This evolution changes how developers write prompts; for example, agentic tools respond well to punch lists with multiple objectives.
Ed Charbeneau: Agentic AI will likely replace the way CI/CD pipelines are constructed and executed. Agents can use tools to automate the creation of deployment pipelines. In addition, agents are part of the CI/CD pipeline itself, able to initiate processes, respond to errors, and parse logs.
Ed Charbeneau: The junior to senior developer relationship is changing with AI technology. With AI, junior developers can self-start and learn much more rapidly. Senior developers can delegate activities to both agents and junior developers while making high-level decisions on a much broader scale. As roles evolve, experience will be determined by problem-solving abilities rather than knowledge of specific APIs, frameworks, or coding practices.
Ed Charbeneau: With agentic AI and tool calling, prompt engineering is evolving into context engineering. Because agentic AI is better at planning, prompts have become generalized while increasing the value of the AI's context window. Through Context Augmented Generation (CAG) and Retrieval Augmented Generation (RAG), additional context and grounding are supplied to the model. Implementing CAG and RAG can be done through MCP tools/agents. For example, a domain-specific scenario may benefit from an MCP tool grounded in that domain. Tools like the Telerik and Kendo assistants for UI development do this today.
Ed Charbeneau: One thing is absolutely clear: as a Developer Advocate, I do not advise developers to attempt building a RAG system from scratch. While RAG is a powerful tool, it is easy to get the implementation details wrong. This is akin to what we see in the Security sector where DIY solutions are ill-advised. Instead, I recommend Progress Agentic RAG as this solution has a robust ingestion API, out-of-the-box knowledge graph capabilities, and security & compliance measures.
Ed Charbeneau: Managing memory schemas, saving user preferences, and retrieving context using semantic similarity should be done through readily available SDKs. Abstracted layers for storing and retrieving structured data through Redis, Pinecone, and Postgres are available, which integrate directly through libraries like Semantic Kernel.
Ed Charbeneau: The AI coding assistant landscape is highly competitive. However, a lot of the heavy lifting is coming from the LLM ecosystem. What makes them different is the activities the developers are performing. While GitHub Copilot is great for general development and supports a wider range of languages, AWS CodeWhisperer excels at AWS-related tasks including security scanning. When evaluating AI coding assistants, developers should consider platform-specific capabilities offered by each vendor.
Ed Charbeneau: While IDE compatibility, model capability, and privacy guarantees are all important considerations, teams should prioritize selecting AI tools that maximize productivity for their specific tech stack and project objectives. These factors should guide the decision, rather than being the sole criteria.
Ed Charbeneau: Yes, both directly and indirectly. AI assistants are starting to become orchestrators that can delegate tasks to other assistants. One such example is the Telerik Agentic UI Generator, which orchestrates UI, UX, and Design assistants to produce completed application user interfaces while following design guidelines and accessibility best practices.
Ed Charbeneau: AI tools are customizable through various means. For example, GitHub Copilot in Visual Studio is configurable with an MCP tool marketplace. Tools can be chosen individually, grouped, or customized as modes. Modes are a collection of tools and prompt fragments that are enabled through the VS Code panel as they’re needed.
Ed Charbeneau: Prompt libraries are effective when combined with custom instruction files. Custom instructions exist for nearly all AI coding assistants: GitHub Copilot, Cursor, Bolt, and more. Custom instructions are a set of prompts that create additional context on how assistants understand your project and how to build, test, and validate its changes. Each assistant has its own convention such as GitHub Copilot’s copilot-instructions.md file.
Ed Charbeneau: Team culture can absolutely make or break AI adoption. Organizations that discourage the use of AI will ultimately lag behind and remain at a competitive disadvantage. Teams using AI will not only see a boost in productivity but gain experience either directly or indirectly with prompt engineering, context management, MCP tools, and related technologies. Early adopters understand AI workings better than those who adopt later and see it as a black box.
Ed Charbeneau: There is an efficiency gain when using AI tools with developer workflows. Coding assistants are good at completing redundant code and allow developers to focus on business-specific problems. AI assistants provide recommendations on performance, best practices, and early bug detection. They also reduce context switching by referencing documentation directly in the IDE.
Ed Charbeneau: Without sufficient context, AI can sometimes take over a code base and add unnecessary responses. For example, instructing an assistant to create a web API may result in tests, examples, and UI code instead of focusing on the task. Newer models generate and display plans before execution, giving developers an opportunity to adjust prompts.
Ed Charbeneau: Developer Experience (DevEx) metrics work well for evaluating AI ROI. These metrics reveal how smoothly developers can build, ship, and maintain software. Feedback loops, cognitive state, and flow state are dimensions to consider. Quantitative metrics like PR size, merge time, and deployment frequency identify shipping speed, while qualitative surveys and interviews gauge interruptions and developer satisfaction.
Ed Charbeneau: Some models are better at specific tasks, code languages, or platforms than others. GPT-5 is strong at planning tasks and coding in multiple languages and frameworks. Claude Sonnet 4 has an extended context window, which is helpful for refactoring large codebases. Developers should try a “dry run” with multiple models to validate their capabilities before production deployment.
Ed Charbeneau: Local models are interesting but less powerful than hosted models. Hosted models have larger parameter counts, which improve abstraction, reasoning, and pattern recognition. Local setups are usually limited to 7B–13B parameters, while hosted models can exceed 70B. Hybrid approaches may allow scaling between local and hosted models if constraints are met.
Ed Charbeneau: Open models like LLaMA 3 are suitable for Edge AI, privacy-sensitive tasks, and highly customized solutions. LLaMA competes with proprietary models like GPT and Sonnet but may require additional configuration for multimodal capabilities and tool calling.
Photo credit: Ed Charbeneau
Ed Charbeneau is Principal Developer Advocate at Progress Software, specializing in AI integration, developer workflows, and agentic AI tools. He works closely with developers to provide guidance on deploying AI models responsibly, managing context, optimizing tools, and improving productivity in both mobile and web applications. With extensive experience in generative AI and coding assistants, Ed helps teams adopt AI effectively while maintaining reliability, security, and efficiency in production environments.
Address:
1855 S Ingram Mill Rd
STE# 201
Springfield, Mo 65804
Phone: 1-844-277-3386
Fax:417-429-2935
E-Mail: contact@appdevelopermagazine.com