Generative AI 101 Slides | GenAI Guide

GenAI enable - Creatvity, Automation and Personalization.

We have seen evolution of product development from traditional coding to data-driven machine learning. Today, generative AI is capable of generating both data and code. This shift is set to transform product business models and how software AI driven product will be built. This will change product business model and also product management.

20 Latest Update of Generative AI

Generative AI is rapidly expanding its influence across industries, reshaping how creativity, business, and education intersect. A new project is using AI to restore missing scenes from a 1940s Orson Welles film, showing how the technology can revive lost art. In higher education, Duke University is testing custom AI tools to weigh both the opportunities and risks of AI-assisted learning. Across Europe, companies are adopting generative AI to simplify and update aging mainframe systems while ensuring compliance with strict regulations. Businesses are also beginning to rely on AI not just for efficiency but as a decision-support tool, guiding strategy and risk management. In entertainment, the impact is striking—about one in five new games on Steam now integrate generative AI for visuals, storytelling, or interactive features—highlighting the technology’s growing role in creative production.

Dataknobs is leveraging generative AI to build advanced data products such as stock market signals and equipment health indices, while also creating integrated knowledge solutions that combine knowledge generation, website development, and intelligent chatbot creation.

Gen AI for IT Systems

A CIO building an annual budget for generative AI should start by identifying high-impact, low-risk areas where AI can be applied within IT systems, such as automating service desk requests, accelerating code development, and improving system monitoring. By prioritizing these internal use cases, the CIO can create measurable cost savings and efficiency gains that justify further investment. The budget should allocate funds not only for technology adoption but also for workforce training, governance, and security, ensuring that generative AI is implemented responsibly and delivers sustainable value.

Gen AI offering/update from Vendors

Google, Microsoft, and AWS are all rapidly expanding their generative AI ecosystems to capture enterprise and consumer demand.

  • Google Gemini: Bringing multimodal AI into everyday life with advanced tools like Flash image editing and by replacing Google Assistant across devices, making AI more integrated into both personal and professional environments.
  • Microsoft Azure: Focusing on enterprises by combining OpenAI’s models with Azure AI Foundry, AI Search, and strong governance frameworks, enabling organizations to securely build, customize, and scale generative AI solutions.
  • AWS: Expanding its Bedrock platform with foundation models, the AgentCore framework for AI agents, a third-party AI tools marketplace, and vector-optimized storage—helping developers train, deploy, and scale AI workloads cost-effectively.
Together, these offerings highlight how cloud providers are not only competing but also shaping the future of AI-driven innovation across industries.

Gen AI offering fron Non Hyper Scalers

  • OpenAI
    • Released GPT-5 with stronger reasoning, coding, and multimodal capabilities, adopted at large scale.
    • Introduced open-weight models (gpt-oss-120b, gpt-oss-20b) under Apache 2.0 for flexible, offline-friendly development.
  • Anthropic (Claude)
    • Advanced the Claude family (e.g., Claude 3.7 Sonnet, Claude 4) for deeper reasoning, coding, and tool use.
    • Added Claude Code for enterprises and a Claude agent for Chrome to boost productivity in the browser.
  • Meta Llama (via Hugging Face)
    • Offers Llama 3 and Llama 4 variants (8B–402B), including multimodal and MoE-optimized options.
    • Accessible on Hugging Face for customization, fine-tuning, and open-source innovation.

GenAI vs Predictive AI

Generative AI focuses on creating new content—such as text, images, or code—by learning patterns from data and producing original outputs. Predictive AI, on the other hand, analyzes historical data to forecast future outcomes, trends, or behaviors without generating new content.

Large Language Model

A Large Language Model (LLM) is an advanced AI system trained on vast datasets to understand, process, and generate human-like language. Beyond basic conversation, LLMs are widely applied in code generation, CSS design automation, and a range of NLP tasks such as document analysis, summarization, and entity extraction. They also enable structured data analysis, helping organizations interpret complex datasets, uncover insights, and automate workflows. With these capabilities, LLMs power applications across industries—from intelligent chatbots and virtual assistants to advanced developer tools and enterprise knowledge systems.

How to use LLMs and build applications

Closed-source LLMs, like OpenAI’s GPT-5 or Anthropic’s Claude, are proprietary models where access is limited through APIs or subscriptions, offering high performance but little transparency or customization. Open-source LLMs, such as Meta’s Llama or Mistral, make their model weights publicly available, enabling developers to fine-tune, deploy, and innovate freely, often with strong community support.

  • LLMs building blocks are Data, Architecture, Training, Inference
  • While making decisions about Closed vs Open model you should consider above factors, control needed, IP risks etc
  • Lllama and Hugging Face are 2 famous open source model. ChatGPT, Bard, Palm , Claude are closed models.
  • LLM Comparision Criteria

    Comparison Dimension Description
    Model Size (# of Parameters) Number of parameters (e.g., billions/trillions) affecting complexity, quality, and compute needs.
    Accuracy & Benchmark Performance Scores on MMLU, GSM8K, HumanEval, etc., indicating reasoning, coding, and knowledge strength.
    Open vs. Closed Source Open-weight models allow full customization and self-hosting; closed models offer managed APIs and guardrails.
    Cost & Efficiency Inference and training costs, token pricing, and hardware efficiency (GPU/TPU utilization).
    Latency & Speed Response time and throughput—critical for real-time chat, copilots, and streaming UX.
    Multimodality Support for text, images, audio, and video in input/output versus text-only models.
    Context Window Length Maximum tokens the model can consider at once—important for long documents and multi-turn sessions.
    Fine-Tuning & Customization Options for fine-tuning, adapters, or parameter-efficient methods to fit domain tasks.
    Governance & Safety Content filtering, policy guardrails, auditability, and compliance features.
    Ecosystem & Tool Integration SDKs, plugins, RAG/search tools, and cloud integrations (Azure, AWS, Google Cloud, Hugging Face).

    Guardrails for GenAI LLM and Chatbots

    Guardrails are essentially guidelines and controls that steer the LLM's outputs in the desired direction.

    Here are some ways to keep your LLM on track:



    Input Validation: Set criteria for what kind of information the LLM can process, preventing nonsensical or malicious inputs.
    Output Filtering: Review and potentially edit the LLM's outputs before they are used, catching any biases or factual errors.
    Real-time Monitoring: Continuously track how the LLM is being used and intervene if it generates harmful content.
    Human oversight: Ensure humans are always involved in the LLM interaction
    Guardrail Practices
    Clear use-case & policy scope Define allowed/blocked intents, refusal rules, and escalation paths.
    Privacy & data minimization Strip/obfuscate PII, control retention, encrypt data in transit and at rest.
    Input validation & prompt hardening Templatize prompts, sanitize user/tool inputs, and block jailbreak patterns.
    Prompt-injection defenses Filter untrusted content (RAG docs, web pages) and isolate it from system instructions.
    Grounding to trusted sources Use RAG/knowledge graphs; require citations and confidence signals to curb hallucinations.
    Safety & content filters Toxicity, hate, self-harm, IP/PII leakage, and malware classifiers on inputs and outputs.
    Bias & fairness checks Run bias evaluations, adjust datasets/prompts, and document mitigations.
    Human-in-the-loop gates Require review for high-risk actions (code deploys, financial trades, customer emails).
    Tool/agent safety Least-privilege keys, allow/deny lists, timeouts, cost/iteration budgets, sandboxed execution.
    Access control & auditability RBAC, per-tenant isolation, comprehensive logs, and immutable audit trails.
    Rate limiting & abuse prevention Quotas, anomaly detection, circuit breakers/kill switches.
    Evaluation & red-teaming Task-specific benchmarks, adversarial testing, and pre-prod safety bars.
    Monitoring & drift detection Track quality, safety incidents, latency/cost; alert on regressions and model drift.
    Change management Version prompts/models/tools, run A/B or shadow tests, maintain rollback plans.
    Compliance & provenance Map to SOC2/ISO/GDPR/CCPA; add watermarking or provenance tags where applicable.

    Foundation Model - Pretrain once, Adapt Everywhere

    A foundation model is a big, pre-trained AI you can quickly adapt to many jobs—chat, search, code, document analysis—without starting from scratch. Examples include OpenAI GPT-4o, Anthropic Claude 3.5, Google Gemini 1.5 Pro, Meta Llama 3 (70B), and Mistral Large. From your perspective, you plug one into your app, point it at your knowledge (via RAG or a small fine-tune), set guardrails, and immediately start asking it to draft emails, summarize PDFs, write SQL, or generate UI/CSS—while you monitor cost, latency, and quality with simple knobs.

  • Foundation model work out of box for universal scenarios. As FMs are exposed to internet scale data and various forms and myriad of patterns, FMs learn to apply their knowledge within a wide range of contexts. Reduce the labeling requirements
  • Consideration to Extend Foundation Model

  • Extend and Fine tune for specific task
  • Train custom model for specific domain
  • Build datasets on which you can train model
  • Prompt Engineering

  • Give instruction , Give persona, add delimeter
  • Add examples, provide steps to complete task
  • Specify output format such as HTML, JSON, Table to produce
  • Retrieval-Augmented Generation (RAG)

    Retrieval-Augmented Generation. It's a technique that helps large language models. Before answering your question, the LLM uses RAG to search a vast external knowledge base for relevant information.

    With this extra knowledge, the LLM can provide more accurate and informative answers. RAG combines the vast knowledge of LLMs with your data, enhancing AI's ability to provide contextually rich responses.



    RAG is extremly ueful when you need to answer question on specific domain, latest and up to date information or use internal knowledge. Use RAG is
    Domain-specific knowledge: If your assistant needs to be an expert in a specific domain, RAG can be used to integrate relevant databases or knowledge repositories to enhance its understanding.
    Accuracy is crucial: If your LLM assistant needs to provide highly accurate information, especially on factual topics like science, history, or specific procedures, RAG can ensure responses are grounded in real-world knowledge.
    Combating hallucinations: LLMs can sometimes make up information, called hallucination. RAG combats this by providing verifiable evidence to support the response.
    Building trust: By allowing users to see where the information comes from (think footnotes!), RAG fosters trust and transparency in the assistant's responses.

    AI Assistant Tradeoff Factors

    Trade Off Factors For AI Assistants

    3 Dimensions


    Accuracy : Lead toUser Trust and Effectiveness: Higher accuracy in understanding and responding to user queries builds trust and reliability. Accurate AI assistants can effectively handle complex tasks, providing precise information and solutions.
    Performance : Better Speed and Scale lear to quick response. High-performance AI assistants provide quick responses, improving user experience and efficiency. Efficient performance ensures that the AI can handle numerous requests simultaneously, essential for scaling operations.
    Cost : Lead to adoption and ROI generation. Lower costs can make AI assistants accessible to a broader range of users and businesses. Affordable solutions encourage wider adoption, enabling more industries to benefit from AI.

    Unleash the power of Similarity Search

    Vector DB store date in vector embeddings which are high dimension vectors that represent features or attributes of data. Traditional DB store dat in rows and columns. In traditional db each columen has signle field and each row is a record.

    Traditional DBs are good for extact match vs vector dbs are great for similarity search.


    Vector Dbs have great use in many use cases like:

    Text Search, Question Answer Bot
    Vision search
    Find similar design
    Efficient analysis of audio
    Find visually similar images e.g. blue water

    OpenAI Fine tuning

  • Finetuning and building custom model give you edge
  • With Fine tuning you have your own model, your IP based on data
  • With Fine tune model, you can handle complex scenarios
  • Before Fine tuning - try prompt engineering, few shot learnings.

    Fine tuning Steps

  • One Fine tune - evaluate the results
  • Make a inference calls/li>
  • Do Error Analysis.
  • Tech Stack and modeling architectures

    Generative AI is based on "comprehend existing" data and determine trajectories data can take. It uses it for generation. Generative AI works by comprehending existing data, identifying patterns, and mapping the possible trajectories that data can take. It then leverages these insights to generate new content—such as text, images, or solutions—that aligns with the learned structures while exploring novel variations.

  • Diffusion architecture is suitable for generation
  • Transformer are suitable for language gentration in sequence
  • How to Evaluate Gen AI

    Traditional machine learning model has evaluaition metrics like accuracy. Generative AI creates new data. It evaluation is based on subjecive measures like diversity of data, realism, nobalty, creativity. It is hard to evaluate or benchmark geenrative models.

    Generative AI adoption framework

    Use above dimensions to identify quardant of your use case(s). Low risk and applicability of generic data e.g content writing for sales,travel guide are easy to adopt.

  • Areas where universal data is available but risk of geenrating wrong results are high - it is opportunity for companies that want to train and sell custom models
  • Areas where task specifci fine tuning is needed, is opprtunity for services companies
  • Companies that want to build defendable IP will focus on create new dataset and model training on these dataset for areas where risk is high and universal dataset are not available.
  • Trade off and Conflicts

    Generative AI has potential, but there are many challenges and open questions. One should consider these before using geenrative AI in enterprise(s)

    Uncontrolled data production

    Generative AI is based on "comprehend existing" data and determine trajectories data can take. It uses it for creating new data. However the method of producing new variation of data makes it controllable.

  • Generative Model output is unpredicatble and uncontrollable. Main issue is - how to get confidence if you want to use it in mission critical envioronment.
  • Large language Model (LLM) inherit bias from data they are trained on.
  • There are open questions on who have copy right on generated content. In future there may be new laws that will impact consumers of generative AI
  • LLM or Image model that are trained on universal data and produce new data are compute intensive. There are impact of enviornment one need to consider.
  • Enviornment concern and legal question

    There is significant amount of compute, energy usage. There is significant amount of carbon emission. One need to ensure it is for good cause and not for producing variation(s)

  • Dataknobs has created set of controls to handle this
  • Function Calling

    OpenAI function calling is a powerful feature that enables seamless integration of AI models with structured tasks, APIs, or external systems. By defining specific functions and providing their schemas, developers can guide the AI to trigger those functions only when appropriate, enhancing the model's capability to interact with external data and services. This approach allows for greater control and precision, enabling use cases such as querying databases, fetching live information, executing computations, or even managing workflows. Function calling ensures that responses remain contextually relevant and action-oriented, bridging the gap between conversational AI and real-world applications while maintaining flexibility and scalability in AI-driven solutions.

    GenAI Project Management

    GENAI PROGRAM MANAGEMENT FRAMEWORK

    4 AREAS


    THE GENAI PROJECT MANAGEMENT FRAMEWORK INVOLVES IDENTIFYING OPPORTUNITIES FOR AI INTEGRATION, DEVELOPING A PROOF OF CONCEPT, AND PROGRESSING THROUGH STAGES OF MATURITY INCLUDING PROTOTYPING, PILOTING, SCALING, AND OPTIMIZING. A ROBUST GOVERNANCE STRUCTURE ENSURES SUCCESSFUL IMPLEMENTATION THROUGH RIGOROUS TESTING, COMPLIANCE, CONTINUOUS MONITORING, AND STAKEHOLDER ENGAGEMENT.

    GenAI Budget Planning

    GENAI - How CIOs should allocate Budget

  • Identify Use Cases
  • Understand Tech stack
  • Understanding Costing - Training, Inference, APP
  • Understand how cost change with scale and prod rollout
  • Divide budget - People, Data, Development, API, Off the shelf, Licensing
  • -->

    Governace for GenAI LLM and Chatbots | governance Framework for GenAI

    We recognize the immense potential of LLMs to revolutionize various aspects of our lives. However, we also acknowledge the critical need to ensure their development and deployment are guided by ethical principles and safeguard human values. Above are guiding principles and framework for AI. It is further extended for GenAI. Click to see detail slides for personalziation, automation and creative scenario to specific governance items

    Security Framework - For Gen AI

    Establish AI Governance across enterprise. Have action plan to secure data, infrastructure and Model. See more details by clicking on slides

    GPT 4o vs GPT 4 Turbo vs GPT 3.5

    GPT 4o is latest model released by Open AI. It accept text and image as input and produce text output. It can reason across vision, text and audio

    Use GPT 3.5 Turbo for ordinary task. It is cheapest. Use GPT 4o for complex task that need high quality

    DALLE, TTS, Whisper are model for images, sound

    GPT (Generative Pre-Trained Transformer) is family of large language model. developed by OPEN AI. GPT4 is latest generaral purpose LLM released by OpenAI. ChatGPT4 is chatbot focused LLM.

  • ChatGPT4 has large token length compared to GPT3.5 ChatGPT4 can process 25000 words of context. It is 8 times higher than chatGPT3.5
  • ChatGPT4 can understand and process visual input.
  • ChatGPT 4 has better programming capabilities compared to ChatGPT 3.5
  • ChatGPT 4 has fewer hallunication compared to ChatGPT 3.5
  • Digital Human

    A Digital Human is an AI-powered virtual being that interacts with people naturally, combining advanced speech, vision, and emotional intelligence to simulate human-like conversations. These AI-driven personas can be used in customer service, education, healthcare, and entertainment, providing personalized and engaging interactions. With realistic facial expressions, voice modulation, and contextual awareness, digital humans bridge the gap between technology and human communication, making digital experiences more intuitive and lifelike. As AI and deep learning continue to evolve, digital humans are becoming an integral part of business and everyday life, enhancing engagement and efficiency in human-computer interactions.

    Virtual Agent to Digital Human

  • Virtual Agent/Digital Agent are for one off task.
  • Virtual assistant carry context and are ongoing engagement
  • Digital influencer add experience and emotion into interaction
  • Digtial Human provide experience/emotion for ongoing engagement.
  • AI Assistant vs Digital Human vs Robots

    AI Assistant vs Digital Human vs Robot

    Similarties and differences

    AI Assistant do not have appearance. They are well suited for tasks. These interact thru text and voice. Digital Human has physical appearance but they exist in digital world. They use non verbal cues such as facial expression in addition to text and voice. Robot exist in physical world. They have mobility and can interact with real world enviornment to move things. Digital human can only do digial task such as give information, schedule meeting. Robot on the other hand can do task in physical enviornment.

    GenAI for Technology Domain

    Understand user intent better and improve search with embeddings. With Generative AI - do persona development and create personalized responsed. With GenAI do Proactive threat detection. Generative AI can also simulate cyber attack and help prepare model and advance technqies to handle such attacks.

    Generative AI Applications in Security

  • Generative AI is extremly useful for Cyber Security
  • Simulate phishing attack to check robustmess of security solution
  • Automate the analysis of security logs
  • Generative AI Applications in Payments Industry

  • Payments is highly regulated area
  • Customer Service can be made effective in Payment industry
  • Audit can be made more robust in payment industry
  • Check validation, Fraud detection can be improved
  • Ad

    Gen AI for Mortgage Industry

    GenAI is useful for Mortgage Industry

    It can do document analysis in loan origination process and streamline it. It can help in customer service. Most important it can automate compliance in mortgage industry

    Gen AI for Mobile Development

    GenAI can revolutionze mobile app development.

    New apps willemerge for creativity. Better personalization will be provided in future apps.

    Generative AI Vendors

  • Vendors: OpenAI, Microsoft, GCP, AWS, Anthropic, Dataknobs, Snorkel and more
  • Evaluation criteria : Features, accuracy, flexibility, ability to fine tune model, cost of inference, how reliable reults are
  • li>
  • Open AI and Micorosoft model is most used
  • li>
  • GCP bard provide up to date information. GCP also has TPU
  • AWS has large cloud share
  • Anthropic released caluse with 100K tokens
  • Hugging Face provide many smaller models
  • ChatGPT3.5 ChatGPT4 and Bard

  • Open AI has chatGPT3.5 and chatGPT4.0
  • Microsoft Azure provide OPEN AI services integrated with Azure
  • Google has Bard, Vertex AI with generative AI studio. In addition google has TPUs
  • AWS provide cloud to use existing capabilities

  • AWS bedrock enable using Hugging face, anthropic or other model on AWS cloud
  • Hugging Face is providing various small model like BERT, GPT-3, ROBERTA, XLNET
  • Anthropic has build Claude. Available to use at "poe . com". There are 3 flavors even with 100K tokens.
  • Build Data Products With Agentic AI

    Agent AI is a powerful, autonomous assistant designed to streamline tasks, provide real-time insights, and enhance decision-making across industries. Whether you're managing customer interactions, analyzing data, or automating workflows, Agent AI adapts to your needs using advanced AI models and contextual understanding. With seamless integrations, continuous learning, and personalized recommendations, Agent AI empowers businesses to operate more efficiently while delivering exceptional user experiences.

    build Data Products using GenAI

    Dataknobs capablities - KREATE, KONTROLS and KNOBS.
    KREATE focus on creatibvity and generation
    KONTROLS provide guardrails, lineage, compliance, privacy and security.
    KNOBS enable experimentation and diagnosis