Skip to content

This repository contains the backend for a Conversational AI system, built using FastAPI and Gemini LLM, with security layers, query processing, and embeddings for intelligent database interactions.

Notifications You must be signed in to change notification settings

theguywithcode/AgenticAI

Repository files navigation

🚀 Conversational AI - LLM Backend

This repository contains the backend for a Conversational AI system, built using FastAPI and Gemini LLM, with security layers, query processing, and embeddings for intelligent database interactions.

📌 Project Overview

🔥 Key Features

Natural Language Query Processing: Users can submit queries in plain language, which the system converts into structured SQL queries.

Secure Query Handling: A security layer ensures queries are sanitized and prevents malicious attacks.

Real-Time Query Execution: The backend processes queries efficiently, fetching results from the database in real time.

Embedding-Based Search: Advanced embeddings improve the accuracy of database searches.

Scalable & Fast: Built with FastAPI for high performance and scalability.

Flexible Deployment: Can be deployed on cloud servers or on-premise solutions.

🛠️ Tech Stack

Framework: FastAPI (Python)

LLM Model: Gemini LLM

Database: PostgreSQL

Security: Custom security layer for SQL validation

Hosting: Cloud or local server deployment

🏗️ System Architecture

System Arcitecture

Process Flow

sequenceDiagram
    participant Client
    participant FastAPI
    participant SecurityLayer
    participant QueryEngine
    participant GeminiLLM
    participant CustomEmbedding
    participant Database

    Client->>FastAPI: POST /query/ with natural language query
    FastAPI->>SecurityLayer: Apply @secure_query decorator
    SecurityLayer->>SecurityLayer: Check for suspicious patterns
    SecurityLayer->>QueryEngine: Pass sanitized query
    
    QueryEngine->>CustomEmbedding: Generate query embeddings
    CustomEmbedding->>GeminiLLM: Convert text to embeddings
    
    QueryEngine->>GeminiLLM: Generate SQL with context
    GeminiLLM->>QueryEngine: Return SQL query
    
    QueryEngine->>SecurityLayer: Validate SQL (read-only)
    SecurityLayer->>Database: Execute safe SELECT query
    Database->>FastAPI: Return results
    FastAPI->>Client: JSON response with results

Loading

Developer Notes

Updated Requirements.txt

Do Not Change this README and IMPORT Statments from the main app 
They are written in a way to not cause Import issue
user python -m uvicorn main1:app --reload to run the app

Query Body --->
query:{
    "Write your prompt here"
}
(⌐■_■)   < I code, therefore I am! >
     ╭───────────────╮
    / theguywithcode  \
   |    /////\\\\\\    |
   |   ///  o  o \\\   |
    \  ||    ∆    ||  /
     '--\   ---   /--'
         |_______|
        //       \\
       //         \\
     _//           \\_
    (__)           (__)


This is still under-development & not a complete build.

Release 1.0 will be comming soon

About

This repository contains the backend for a Conversational AI system, built using FastAPI and Gemini LLM, with security layers, query processing, and embeddings for intelligent database interactions.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages