Skip to content

Commit

Permalink
final code
Browse files Browse the repository at this point in the history
  • Loading branch information
joking-parrot committed Dec 23, 2023
0 parents commit 1dbe0ad
Show file tree
Hide file tree
Showing 40 changed files with 4,031 additions and 0 deletions.
9 changes: 9 additions & 0 deletions .env
Original file line number Diff line number Diff line change
@@ -0,0 +1,9 @@
#export OPENAI_API_KEY=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
#export OPENAI_API_MODEL=gpt-4-0314

export OPENAI_API_KEY=36280dcf174a4410bac4b6784367bc23
export OPENAI_API_MODEL=gpt-4
export OPENAI_API_DEPLOYMENT=gpt-4-deployment
export OPENAI_API_BASE=https://msrenergy-openai.openai.azure.com
export OPENAI_API_VERSION=2023-03-15-preview
export OPENAI_API_TYPE=azure
11 changes: 11 additions & 0 deletions .env.template
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@
#if OPENAI deployment else comment the below
export OPENAI_API_KEY=<--key-->
export OPENAI_API_MODEL=<--modelname-->

#if Azure deployment else comment the below
export OPENAI_API_KEY=<--key-->
export OPENAI_API_BASE=<--base url if required-->
export OPENAI_API_VERSION=<--api version if required-->
export OPENAI_API_TYPE=<--api type if required -->
export OPENAI_API_MODEL=<--modelname-->
export OPENAI_API_DEPLOYMENT=<--deploymentname-->
66 changes: 66 additions & 0 deletions .github/workflows/dev_backend_swissnyf-backend.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,66 @@
# Docs for the Azure Web Apps Deploy action: https://github.com/Azure/webapps-deploy
# More GitHub Actions for Azure: https://github.com/Azure/actions
# More info on Python, GitHub Actions, and Azure App Service: https://aka.ms/python-webapps-actions

name: SwissNYF-backend

on:
push:
branches:
- dev_backend
workflow_dispatch:

jobs:
build:
runs-on: ubuntu-latest

steps:
- uses: actions/checkout@v4

- name: Set up Python version
uses: actions/setup-python@v1
with:
python-version: "3.11"

- name: Create and start virtual environment
run: |
python -m venv venv
source venv/bin/activate
pip install -r requirements.txt
# Optional: Add step to run tests here (PyTest, Django test suites, etc.)
- name: Zip artifact for deployment
run: zip release.zip ./* -r

- name: Upload artifact for deployment jobs
uses: actions/upload-artifact@v3
with:
name: python-app
path: |
release.zip
!venv/
deploy:
runs-on: ubuntu-latest
needs: build
environment:
name: "Production"
url: ${{ steps.deploy-to-webapp.outputs.webapp-url }}

steps:
- name: Download artifact from build job
uses: actions/download-artifact@v3
with:
name: python-app

- name: Unzip artifact for deployment
run: unzip release.zip

- name: "Deploy to Azure Web App"
uses: azure/webapps-deploy@v2
id: deploy-to-webapp
with:
app-name: "SwissNYF-backend"
slot-name: "Production"
publish-profile: ${{ secrets.AZUREAPPSERVICE_PUBLISHPROFILE_9438DD4D8ECA43E1AFE87D081D5C3C28 }}
package: .
1 change: 1 addition & 0 deletions .gitignore
Original file line number Diff line number Diff line change
@@ -0,0 +1 @@
**_pycache__**
189 changes: 189 additions & 0 deletions .ipynb_checkpoints/main-checkpoint.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,189 @@
import random
from fastapi import FastAPI, Response, Form
from fastapi.responses import StreamingResponse
from fastapi.middleware.cors import CORSMiddleware
import time
from pipeline import *
from retriever import *
from utils import *
from configs import *
from utils.tools_util import *
import argparse
import os
import sys
import copy
import pickle
import json
import re
from llama_index.embeddings import OpenAIEmbedding
from llama_index.agent import ReActAgent
import re
from tqdm import tqdm
from llama_index.tools.function_tool import FunctionTool
from typing import List
from llama_index.agent.react.formatter import get_react_tool_descriptions
from llama_index.llms.base import LLM
from functools import wraps
from collections.abc import Iterable
import inspect, itertools
from tqdm import tqdm
from typing import List

from functools import wraps
from collections.abc import Iterable
from abc import abstractclassmethod
from typing import Optional, Dict, List, Tuple
from sentence_transformers import SentenceTransformer, util
from llama_index.llms import AzureOpenAI, OpenAI
from llama_index.embeddings import OpenAIEmbedding
from sentence_transformers import SentenceTransformer, util
from collections import Counter
import os
import inspect
from io import BytesIO
from fastapi import Request

currentdir = os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))

from dotenv import load_dotenv, find_dotenv
dirname = os.path.join(currentdir, '.env')
load_dotenv(dirname)

# model = os.environ["OPENAI_API_MODEL"]
# llm = OpenAI(model=model, temperature=0.01)
# print("LLM is initalised")
model = os.environ["OPENAI_API_MODEL"]
deployment = os.environ["OPENAI_API_DEPLOYMENT"]
llm = AzureOpenAI(deployment_id=deployment, model=model, engine=deployment, temperature=0.01)

app = FastAPI()

app.add_middleware(
CORSMiddleware,
allow_origins=["*"], # Allow all origins
allow_credentials=True,
allow_methods=["*"], # Allow all methods
allow_headers=["*"], # Allow all headers
)
class DummyRetriever:
def __init__(self, all_tool_names):
self.all_tool_names = all_tool_names
def filter(self, query):
return self.all_tool_names

tools = Tools("./data/tools.yaml")
# tools = Tools(load_raw="./data/local_tool_set.json")
all_tools_list = None
all_tools_names = None #python-app
all_tools_desc = None
retriever = DummyRetriever(tools)
pipeline_topgun = TopGun(filter_method=retriever, llm=llm)
pipeline_reverse = ReverseChain(filter_method=retriever, llm=llm)


from pydantic import BaseModel


class StreamPayload(BaseModel):
query: str = ""
planner: str = "topgun"

def set_tools_ret():
global tools, all_tools_list, all_tools_names, all_tools_desc, retriever, pipeline_topgun, pipeline_reverse
if tools.tools_raw is not None:
pipeline_topgun.tool_defs = tools.tools_raw
pipeline_topgun._topgun_corpus = tools.tools_raw
pipeline_reverse.tool_defs = tools.tools_raw
pipeline_reverse._reverse_chain_corpus = tools.tools_raw
pipeline_reverse.set_raw_tools("\n\n".join(tools.tools_raw))
else:
tools.load_tools_yaml()
all_tools_list = tools.get_tools_list()
all_tools_names = tools.get_tool_names()
all_tools_desc = tools.get_tools__desc_str()
retriever = DummyRetriever(tools)
pipeline_topgun.set_tools(all_tools_desc, all_tools_names)
pipeline_reverse.set_tools(all_tools_desc, all_tools_names)

def run_query(query,planner):
if all_tools_desc is None or all_tools_names is None:
set_tools_ret()

if planner== "reversechain":
pipeline = pipeline_reverse
elif planner == "topgun":
pipeline = pipeline_topgun

yield from pipeline.query(query)


@app.get('/')
async def hello_world():
return {'message': 'Hello, World!'}

@app.post('/get_tools')
async def get_tools():
return {'message': tools.get_tools() }

@app.post('/set_tools')
async def set_default_tools():

return {'message': tools.get_tools()}

@app.get("/build_tools")
def get_build_tools(tools):
if tools.tools is None:
tools.load_tools_via_api(tools)
else:
curr_tool_names = tools.get_tool_names()
for tool in tools:
if tool['name'] not in curr_tool_names:
tools.tools[tool['name']] = tool
all_tools_list = tools.get_tools_list()
all_tools_names = tools.get_tool_names()
all_tools_desc = tools.get_tools__desc_str()

# retriever = GearRet(top_k = 9, verbose = False)
# retriever = InstructRet(top_k = 9, verbose = False)
retriever = DummyRetriever(all_tools_list)
# print("Retriever is initalised")
# retriever.set_tool_def(all_tools_list)

return {"message": "All tools are loaded and set up"}


@app.post("/stream_response")
async def stream_functions(body: StreamPayload):
# print("this is the body", body)
query = body.query
planner = body.planner
return StreamingResponse(run_query(query,planner), media_type="text/markdown")
# return StreamingResponse(run_query(query,planner), media_type="text/markdown")

@app.get("/markdown")
def generate_markdown():
markdown_content = """
# FastAPI Streaming Response Example
This is a streaming response example using FastAPI.
## Streaming Markdown
- Bullet point 1
- Bullet point 2
- Bullet point 3
"""

# Convert the Markdown content to bytes
markdown_bytes = markdown_content.encode("utf-8")

# Create a BytesIO object to stream the content
stream = BytesIO(markdown_bytes)

# Return a StreamingResponse with the proper content type
return StreamingResponse(stream, media_type="text/markdown")


if __name__ == "__main__":
import uvicorn
uvicorn.run(python-app, host="0.0.0.0", port=8000)
Loading

0 comments on commit 1dbe0ad

Please sign in to comment.