Exploring Four Practical AI Engineering Projects: From Brochure Generation to Code Conversion

Have you ever wondered what “AI engineering” really looks like in practice? Not the theoretical concepts or flashy demos, but actual implementations that solve real problems? Today, I want to walk you through four concrete AI projects that demonstrate how large language models can be integrated into practical applications with real-world value.

As someone who’s worked extensively with AI systems, I’ve seen countless examples of technology that looks impressive in a demo but fails to deliver practical value. These projects stand out because they’re not just theoretical exercises—they represent thoughtful implementations that address specific challenges using Large Language Models (LLMs) in combination with other technologies.

Let’s dive into each project and explore what makes them valuable, how they work, and why they matter for anyone interested in practical AI implementation.

Project 1: AI Brochure Generator – Transforming Website Content into Professional Marketing Materials

The Real Problem It Solves

Imagine you’re working in marketing and need to create brochures for different international markets. Traditionally, this would involve manually gathering information from a company website, organizing it, translating it, and formatting it into a brochure. It’s time-consuming, inconsistent, and often requires multiple team members with different skill sets.

This is exactly the problem the AI Brochure Generator solves. Instead of manual processes, this tool automatically transforms a company’s website content into professional, multilingual brochures with minimal human intervention.

How It Actually Works

The AI Brochure Generator follows a thoughtful process that combines several technologies:

  1. Website Analysis: The tool begins by scanning a company’s website using BeautifulSoup, a reliable web scraping library. This isn’t just grabbing everything—it’s intelligent analysis.

  2. Content Selection: Here’s where it gets interesting. Instead of taking all content indiscriminately, the system uses LLM reasoning to identify which pages are actually relevant for a brochure (like “About Us,” “Careers,” or “Services” pages). This mimics how a human marketer would select content.

  3. Information Structuring: The relevant information is organized into a logical structure that makes sense for a brochure format.

  4. Multilingual Generation: Using either OpenAI GPT models or Ollama (for local model execution), the tool generates a polished brochure in Markdown format, ready for any language you specify.

  5. User-Friendly Interface: All of this complex processing happens behind a simple Gradio interface that anyone can use, regardless of technical background.

AI Brochure Generator interface showing website URL input and language selection options

Why This Approach Matters

What makes this project stand out from simple content scrapers? Several key considerations:

  • Intelligent Filtering: Most scrapers just grab everything, creating information overload. This tool uses LLM reasoning to select only what’s truly relevant—just as a skilled human would.

  • Language Flexibility: Need brochures in Spanish, French, or Mandarin? The system handles translation seamlessly without requiring separate translation tools.

  • Consistent Formatting: Marketing teams often struggle with maintaining consistent branding across materials. The Markdown output provides a standardized format that can be easily adapted to specific brand guidelines.

  • Time Savings: What might take a marketing team several days can be accomplished in minutes, freeing up valuable time for strategic work.

For small businesses without dedicated marketing teams, this tool is particularly valuable. It democratizes professional marketing material creation, allowing organizations of any size to produce high-quality brochures without specialized skills or expensive software.

Project 2: FlightAI Assistant – A Comprehensive Airline Customer Service Solution

The Customer Service Challenge

Anyone who’s traveled knows the frustration of navigating airline websites for basic information. Is my flight on time? What are the baggage restrictions? Can I get a seat upgrade? These simple questions often require clicking through multiple pages or waiting on hold with customer service.

FlightAI Assistant addresses this pain point by providing a single, intelligent interface for all airline-related inquiries.

Beyond Basic Chatbots

What sets FlightAI apart from typical airline chatbots? Let’s break down its capabilities:

✈️ Travel Information Expertise: It doesn’t just answer generic questions—it understands airline-specific terminology and processes. Ask about “premium economy” or “connection times,” and it provides accurate, context-aware responses.

🎨 Visual Destination Assistance: One particularly innovative feature is its integration with DALL·E 3 to generate city illustrations. Instead of just describing a destination, it can show you what it looks like—helping travelers visualize their trip.

💰 Real-Time Pricing Information: Unlike static FAQ systems, FlightAI can check current ticket prices, giving travelers up-to-date information for decision making.

🎟️ Seat Availability Insights: Need to know if there are seats available on a specific flight? The system can check availability without requiring users to navigate complex booking interfaces.

FlightAI Assistant interface showing conversation about flight details with generated city image

Why This Implementation Works

The magic happens through thoughtful technical integration:

  • Custom Tool Integration: Rather than trying to be everything at once, FlightAI connects to specialized tools for pricing, availability, and image generation. This modular approach creates a more reliable system.

  • Context-Aware Responses: The system remembers previous questions in the conversation, allowing for natural follow-up questions like “What about tomorrow?” without needing to repeat all details.

  • Concise Communication: Understanding that travelers often need quick answers, the system provides short, direct responses rather than lengthy explanations.

For airlines, this represents a significant improvement over traditional customer service systems. It reduces call center volume for routine inquiries while providing a better customer experience. For travelers, it means getting accurate information quickly without navigating complex websites or waiting on hold.

Project 3: Python/Cobol to C++ High-Performance Code Converter – Bridging the Legacy Gap

The Performance Challenge

Many organizations face a critical dilemma: they have valuable business logic written in older languages like COBOL (which powers much of our banking infrastructure) or Python (great for development but sometimes slow in production), but need the performance benefits of compiled languages like C++.

Rewriting these systems from scratch is prohibitively expensive and risky. This is where the Python/Cobol to C++ High-Performance Code Converter comes in.

How the Conversion Process Works

This tool follows a sophisticated approach to code conversion:

  1. Input Analysis: The system accepts either Python or COBOL source code as input.

  2. Intelligent Translation: Using GPT models, it translates the code into equivalent C++ syntax while preserving the original logic and functionality.

  3. Performance Optimization: Crucially, it doesn’t just create a direct translation—it applies C++ best practices and optimization techniques, including the -Ofast compiler flag for maximum speed.

  4. Validation Capability: The tool includes the ability to execute both the original and converted code, allowing developers to verify that the functionality remains consistent.

  5. User-Friendly Interface: All this complex processing happens behind a Gradio interface that makes it accessible even to developers who aren’t C++ experts.

Code conversion interface showing Python input and C++ output with execution options

Why This Matters for Real-World Systems

Let’s consider why this approach is valuable:

  • Legacy System Modernization: Many critical business systems run on COBOL, but finding developers with these skills is increasingly difficult. This tool helps bridge that gap.

  • Performance Critical Applications: For applications where speed matters (financial trading systems, scientific computing, real-time processing), converting from Python to optimized C++ can yield significant performance improvements.

  • Risk Reduction: Rather than a complete rewrite (which introduces new bugs), this approach preserves the original business logic while gaining performance benefits.

  • Verification Process: The ability to run both versions side by side provides confidence that the conversion hasn’t altered the intended functionality.

This isn’t about replacing human developers—it’s about augmenting their capabilities. The tool handles the tedious translation work, while developers can focus on reviewing and refining the output for their specific use cases.

Project 4: RAG AI Assistant with Chroma & LangChain – Knowledge-Powered Conversational Intelligence

The Information Overload Problem

Most organizations have a wealth of knowledge trapped in documents, PDFs, and internal systems. Finding specific information often feels like searching for a needle in a haystack. Traditional search systems rely on keyword matching, which fails to understand the meaning behind queries.

The RAG (Retrieval-Augmented Generation) AI Assistant solves this by creating a smart interface to your organization’s knowledge base.

How RAG Technology Works

This implementation follows a sophisticated process:

  1. Knowledge Base Construction: Documents are loaded into Chroma, a vector database that stores information in a format that captures meaning rather than just keywords.

  2. Semantic Query Processing: When you ask a question, the system doesn’t just look for matching words—it understands the meaning behind your query.

  3. Contextual Response Generation: Using LangChain, the system retrieves the most relevant information and crafts a response that directly addresses your question.

  4. Visual Knowledge Mapping: One particularly useful feature is the ability to visualize the relationships between documents in 2D and 3D space using TSNE algorithms, helping users understand how information connects.

  5. Conversational Memory: The system remembers previous exchanges in the conversation, allowing for natural follow-up questions without repeating context.

RAG AI Assistant interface showing question input, response, and vector space visualization

Why This Approach Outperforms Traditional Search

The RAG system offers several advantages over conventional search:

  • Meaning-Based Retrieval: It understands that “How do I request time off?” and “What’s the vacation policy?” are asking the same thing, even though the wording differs.

  • Context Preservation: In multi-turn conversations, it maintains context, so you can ask follow-up questions naturally.

  • Source Transparency: Unlike black-box AI responses, this system can show which documents informed its answer, building trust and allowing verification.

  • Knowledge Visualization: The vector space visualizations provide insight into how documents relate to each other, revealing connections you might not have noticed.

For organizations drowning in information but starving for knowledge, this system transforms how employees access and use internal information. It’s particularly valuable for customer support teams, HR departments, and technical support staff who need quick access to accurate information.

Technical Foundations: Why These Specific Tools?

You might wonder why these particular technologies were chosen for each project. Let’s examine the thoughtful engineering decisions behind the tech stack.

Language Selection Strategy

Language Primary Use Case Key Advantages
Python Main application development Rich ecosystem of AI/ML libraries, rapid development cycle
C++ High-performance code generation Execution speed, memory control, industry standard for performance-critical applications
COBOL Legacy system integration Compatibility with existing enterprise systems (especially financial)

The choice of languages wasn’t arbitrary—it reflects careful consideration of where each language excels.

Framework and Library Selection

Each project uses a carefully curated set of frameworks that serve specific purposes:

Gradio: This appears consistently across all projects for good reason. It provides the simplest path to creating interactive web interfaces for AI applications without requiring front-end development expertise. For engineers focused on the AI functionality, Gradio handles the UI layer efficiently.

LangChain: This framework is particularly valuable for the RAG system because it provides modular components for building complex AI workflows. Rather than reinventing the wheel for common patterns like document loading or chain execution, LangChain offers battle-tested implementations.

BeautifulSoup: While newer web scraping tools exist, BeautifulSoup remains an excellent choice for projects like the Brochure Generator because it’s reliable, well-documented, and perfectly suited for the task of parsing HTML content.

ChromaDB: As a vector database, Chroma offers the right balance of simplicity and power for knowledge-based applications. It’s lightweight enough for development and small deployments but scales reasonably well.

OpenAI API & Ollama: This dual approach provides flexibility. Organizations with API budgets can leverage powerful cloud models, while those concerned about privacy or costs can run models locally via Ollama.

The technology choices reflect a practical engineering philosophy: use the right tool for the job, not the newest or shiniest option available.

Getting Started: Practical Implementation Guide

Understanding these projects is one thing—actually using them is another. Let’s walk through the concrete steps to get any of these projects running on your system.

Step 1: Setting Up Your Environment

First, you’ll need to clone the repository:

git clone https://github.com/SionAlin/LLM_Engineering.git

This creates a local copy of all four projects on your machine.

Step 2: Configuring API Access

Most projects require an OpenAI API key for full functionality:

  1. Create a .env file in the project root directory
  2. Add your API key using this format:

    OPENAI_API_KEY=sk-yourkeyhere
    

Note that some projects also support Ollama for local model execution, which doesn’t require an API key.

Step 3: Launching Individual Projects

Each project operates independently, so you can work with just one if that’s all you need:

Option A: Using Jupyter Notebooks (Recommended for Learning)

cd path/to/project_directory
jupyter notebook Main.ipynb

This approach is excellent for understanding how each component works, as notebooks provide a step-by-step execution environment.

Option B: Direct Python Execution (For Deployment)

cd path/to/project_directory
python Main.py

This method is more suitable when you want to run the application as a complete system.

Troubleshooting Common Issues

As with any technical setup, you might encounter some challenges. Here are solutions to common problems:

Problem: “Module not found” errors
Solution: Install required dependencies with pip install -r requirements.txt

Problem: API key not recognized
Solution: Verify that your .env file is in the correct directory and properly formatted

Problem: Gradio interface not launching
Solution: Check if the default port (usually 7860) is already in use; you may need to specify a different port in the code

Problem: Slow responses from OpenAI API
Solution: Consider implementing request caching or switching to a local model via Ollama for development

Remember that each project includes its own README with specific setup instructions—always check there first for project-specific requirements.

Frequently Asked Questions

Let’s address some common questions people have about these projects:

How much programming knowledge do I need to use these projects?

You’ll need basic Python knowledge to set up and customize the projects. However, the Gradio interfaces make the functionality accessible even to non-programmers once the system is running. If you’re comfortable following installation instructions and editing simple configuration files, you should be able to get these projects working.

Do I need to pay for OpenAI API usage?

Yes, using the OpenAI API will incur costs based on your usage. However, all projects support Ollama as an alternative, allowing you to run open-source models locally without API charges. For learning purposes, you can start with a small API budget or use free-tier eligible models.

Can I use these projects in a commercial product?

These projects serve as excellent starting points for commercial applications, but you’ll need to enhance them for production use. Consider adding:

  • User authentication and authorization
  • Comprehensive error handling and logging
  • Performance monitoring
  • Enhanced security measures
  • Professional UI/UX design

The core functionality provides a solid foundation that can be built upon for commercial deployment.

How accurate is the COBOL to C++ conversion?

The conversion works best for well-structured COBOL programs with clear business logic. Complex legacy systems with non-standard extensions may require manual refinement after conversion. Always validate the converted code thoroughly against the original functionality before deploying in production.

How can I expand the RAG system’s knowledge base?

Adding new documents is straightforward:

  1. Place your documents in the documents/ directory
  2. Run the knowledge base rebuilding script
  3. The system will automatically process new content and update the vector database

Supported formats typically include PDF, Word documents, and plain text files.

Are these projects suitable for learning AI engineering concepts?

Absolutely. These projects demonstrate practical implementations of key AI engineering concepts:

  • Tool integration with LLMs
  • User interface design for AI applications
  • Performance optimization techniques
  • Knowledge management systems
  • Real-world application of RAG architecture

Working through these projects will give you hands-on experience with the challenges and solutions in practical AI engineering.

Can I run these projects without an internet connection?

Some functionality requires internet access (particularly when using OpenAI API), but most projects support Ollama for local model execution. The code converter and brochure generator can work offline with local models, while the FlightAI assistant requires internet access for real-time data like flight prices.

Why These Projects Represent Good AI Engineering Practice

In an era of AI hype, it’s important to distinguish between flashy demos and genuinely useful implementations. These projects stand out for several reasons:

Solving Actual Problems

Each project addresses a concrete challenge:

  • Marketing teams needing efficient brochure creation
  • Airlines seeking better customer service tools
  • Organizations modernizing legacy systems
  • Companies struggling with information overload

They’re not “AI for AI’s sake” but thoughtful applications of technology to real problems.

Thoughtful Integration, Not Just API Calls

These projects demonstrate proper integration patterns:

  • Using LLMs for reasoning where appropriate (like selecting relevant web pages)
  • Combining multiple tools to create comprehensive solutions
  • Providing verification mechanisms (like side-by-side code execution)
  • Maintaining transparency about sources and limitations

This represents the difference between simply using AI and engineering with AI.

Practical Engineering Considerations

Good AI engineering requires attention to details beyond the core model:

  • User experience design
  • Error handling and graceful degradation
  • Performance optimization
  • Validation and verification processes
  • Clear documentation

These projects incorporate these considerations, reflecting professional engineering standards.

Learning Value

For those looking to develop AI engineering skills, these projects offer:

  • Complete implementations rather than fragments
  • Realistic complexity levels
  • Multiple integration patterns to study
  • Clear separation of concerns
  • Production-ready architecture patterns

They serve as excellent learning resources that bridge the gap between theoretical knowledge and practical implementation.

Conclusion: The Essence of Practical AI Engineering

These four projects demonstrate what effective AI engineering looks like in practice. They’re not about chasing the latest model releases or creating viral demos—they’re about thoughtfully applying technology to solve real problems.

What makes these implementations valuable isn’t just the technology they use, but how they use it:

  • They focus on specific, well-defined problems rather than trying to be everything
  • They integrate multiple technologies in service of user needs
  • They prioritize usability alongside technical sophistication
  • They provide verifiable results rather than black-box outputs
  • They’re built with practical constraints in mind

In the rapidly evolving AI landscape, it’s easy to get caught up in the hype cycle. But real progress happens in the thoughtful, practical applications—the ones that quietly make work better, solve genuine problems, and create tangible value.

Whether you’re an AI practitioner looking to deepen your engineering skills or a business leader evaluating AI solutions, these projects offer valuable insights into what effective AI implementation looks like. They demonstrate that the most valuable AI applications aren’t necessarily the most complex—they’re the ones that reliably solve real problems in practical ways.

The future of AI belongs not to those who can build the most impressive demos, but to those who can engineer solutions that stand the test of real-world use. These projects point the way forward—a path of thoughtful integration, practical problem-solving, and genuine value creation.

As you explore these implementations, consider how their approaches might apply to your own challenges. What specific problems could benefit from this kind of thoughtful AI integration in your work? The most valuable AI applications often come not from chasing trends, but from solving the specific problems right in front of you.