| | import streamlit as st |
| | import torch |
| | from transformers import LLMForConditionalGeneration, LLMTokenizer |
| | import sqlite3 |
| |
|
| | |
| | model_name = "microsoft/CodeGPT-small-py" |
| | tokenizer = LLMTokenizer.from_pretrained(model_name) |
| | model = LLMForConditionalGeneration.from_pretrained(model_name) |
| |
|
| | |
| | def generate_sql_query(text): |
| | input_ids = tokenizer.encode(text, return_tensors="pt") |
| | outputs = model.generate(input_ids, max_length=100, do_sample=False) |
| | generated_sql = tokenizer.decode(outputs[0], skip_special_tokens=True) |
| | return generated_sql |
| |
|
| | |
| | def execute_query(sql_query): |
| | conn = sqlite3.connect('C:/Users/Chovatiya.Parth/Desktop/SQL/superstore Creation.sql') |
| | cursor = conn.cursor() |
| | cursor.execute(sql_query) |
| | results = cursor.fetchall() |
| | conn.close() |
| | return results |
| |
|
| | |
| | def main(): |
| | st.title("SQL Chatbot") |
| |
|
| | user_query = st.text_input("Enter your query:") |
| |
|
| | if st.button("Submit"): |
| | sql_query = generate_sql_query(user_query) |
| | st.write("Generated SQL query:", sql_query) |
| |
|
| | try: |
| | results = execute_query(sql_query) |
| | st.write("Results from the database:") |
| | for row in results: |
| | st.write(row) |
| | except Exception as e: |
| | st.error("An error occurred while executing the SQL query.") |
| | st.error(e) |
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|