Category: Machine learning
-
Retrieval-Augmented Generation(RAG) – An Introduction
Introduction Retrieval-Augmented Generation (RAG) is an advanced approach that combines the power of large language models (LLMs) with external knowledge sources. Traditional LLMs are trained on vast datasets from the internet, books, and other general sources. They are powerful in generating coherent text, summarizing information, and answering general questions. However, they have significant limitations when…
-
Building a Streamlit SQL Chatbot Using Gemini + LangChain + Oracle DB
1. Introduction This tutorial walks you through a Python-based Streamlit application that uses Google Gemini (via LangChain) to: Following is the code for the SQL Chatbot. 2. Prerequisites Ensure you have the following Python packages installed: Also, make sure: 3. Import Required Libraries Explanation: 4. Define a Helper Function to Clean LLM Output Why it’s…
-
LangChain- An Introduction
1. What is LangChain? LangChain is an open-source framework designed to build powerful applications powered by Large Language Models (LLMs) like OpenAI’s GPT, Google’s Gemini, Claude, Mistral, Cohere, and more. While LLMs can generate impressive results on their own, LangChain lets you chain multiple components together, such as prompts, memory, tools, retrievers, and agents to…
-
Understanding Core Deep Learning Architectures for Sequence Data
In the realm of deep learning, particularly for handling sequential data like text, audio, or time series, several fundamental architectures have emerged as dominant paradigms. This tutorial will demystify three key categories: Autoencoding Models (Encoder-Only), Autoregressive Models (Decoder-Only), and Sequence-to-Sequence Models (Encoder-Decoder). We’ll explore their core concepts, typical applications, and how they differ. 1. The…
-
Understanding Generative Configuration Inference Parameters for LLMs
1. Introduction When interacting with large language models (LLMs) to generate text, you’re not just providing a prompt and getting a fixed output. Modern LLM APIs and frameworks offer a suite of parameters that allow you to control the generation process, influencing the creativity, randomness, coherence, and length of the generated text. These parameters are…
-
Machine Learning Setup – Anaconda, Jupyter Notebook and Virtual Environment
1. Introduction In this session, we’ll walk through everything you need to set up your machine learning environment from scratch. This includes installing essential tools like Anaconda, working with Jupyter Notebook, and setting up a virtual environment to keep your project organized and dependency-free. 2.Tools Needed for Machine Learning Projects Before diving into coding, there…
-
Understanding the Transformer Architecture
1. Introduction The reason it is called the Transformer architecture is because the inputs gothrough a series of “transformationsˮ. 2. What Does a Transformer Do? A Transformer model performs several important tasks that enable it to understand and generate human-like language effectively: 3. Key Features of Transformer Architecture Explained The Transformer model brings together several…
-
Prompt Engineering with OpenAI – An Introduction
1. What is Prompt Engineering? Prompt Engineering is the strategic process of designing and refining input instructions (called prompts) to effectively communicate with AI models such as ChatGPT, Claude, Gemini, or others. It involves crafting clear, specific, and context-aware text inputs that guide the model to generate outputs that are accurate, relevant, and aligned with…
-
Transformers in Machine Learning – An Introduction
Transformers have become the backbone of modern machine learning models, especially in Natural Language Processing (NLP). From Google’s BERT to OpenAI’s GPT, transformers power state-of-the-art models used for text classification, translation, question answering, and even image processing. Introduced in the 2017 paper “Attention Is All You Need” by Vaswani et al., Transformers address limitations of…
-
Introduction to Prompt Engineering
1. What is Prompt Engineering? Prompt engineering is the process of designing precise and structured inputs (known as prompts) that are given to generative AI models, particularly large language models (LLMs) such as OpenAI’s ChatGPT, Google’s Gemini, Anthropic’s Claude, or Meta’s LLaMA. These prompts serve as the instructions or queries that guide the model to…
