The difference between a proof-of-concept and a production application lies in the details.
Generic chat responses might work for demos, but professional applications need appropriate outputs that align with specific requirements.
In a professional environment code is (ideally) tested, metrics are collected, analytics are displayed somewhere.
AI development can follow these established patterns.
You will hit roadblocks when trying to:
- Implement essential backend infrastructure (databases, caching, auth) specifically for AI-driven applications.
- Debug and understand the âblack boxâ of AI agent decisions, especially when multiple tools are involved.
- Ensure chat persistence, reliable routing, and real-time UI updates for a seamless user experience.
- Objectively measure AI performance moving beyond subjective âvibe checksâ for improvements.
- Manage complex agent logic without creating brittle, monolithic prompts that are hard to maintain and optimize.
In this course you will build out a âDeepSearchâ AI application from the ground up to help you understand and implement these patterns and ensure a production-ready product.
Days 00-02: Getting Started
Youâll start with a project that is already built out using Next.js TypeScript (of course), PostgreSQL through the Drizzle ORM, and Redis for caching.
The first couple of days you will implement fundamental AI app features building out a Naive agent. Youâll start by hooking up an LLM of your choice to the Next.js app using the AI SDK and implement a search tool it can use to supplement its knowledge when conversing with users.
Chats with an LLM donât save themselves so you will also save conversations to the database as well.
Days 03-05: Improve Your Agent through Observability and Evals
The first real differentiator between a vibe-coded side project and a production-ready product you can feel confident putting in front of customers is observability and evals.
You need to know what is going on with your LLM calls as well as need an objective means to judge the output that LLMs are producing.
This is exactly what the next few days are about. Youâll hook up your app to LangFuse and get familiar with looking through traces produced by the application.
Once you can see what your LLM is doing, nowâs the time to test inputs and outputs of your agent using evals. Evals are the unit test of the AI application world and weâll start by wiring up Evalite which is an open-source, vitest-based runner. Youâll learn about what makes good success criteria and build out your evals including implementing LLM-as-a-Judge and custom Datasets specific to your product. Weâll also discuss how you can capture negative feedback intro traces that you can feed into your app to make it better.
Days 06-07: Agent Architecture
Up until this point, your app is driven by an increasingly large prompt that will become unwieldy and impossible to test and iterate on when your app complexity grows.
Weâll take these next two days to revisit the over-all architecture of our application and refactor it to better handle complex multi-step AI processes. The primary idea behind this refactor is called Task Decomposition where you allow a smart LLM model determine the next steps to take based on the current conversation but allow room to hand off actual action to focused or cheaper models.
Days 08-09: Advanced Patterns
The last two days we will evaluate what our deepsearch agent is and how we can further optimize output. Youâll learn the differences between âAgentâ and âWorkflowâ and see how in this use-case weâll lean harder into workflow patterns to build a more reliable product.
In ai-land, this pattern is called the evaluator-optimizer loop which effectively means that if the agent has enough information, it will answer the question presented but if it doesnât it will search for more information. With this pattern defined weâll embrace and optimize around its design.
By the end of this cohort you will be confident in building out AI applications that are reliable and will improve through iteration and user feedback. LLM models and the whole AI field is changing rapidly so understanding these fundamentals will give you a foundation to navigate building applications for years to come.
You must be <a href="https://wislibrary.net/my-account/">logged in</a> to post a review.