update
This commit is contained in:
@@ -1,9 +1,7 @@
|
||||
from fastapi import APIRouter, HTTPException, Security, File, UploadFile
|
||||
from app.models import GoogleFile
|
||||
from app.dependencies import dbEngine
|
||||
from fastapi import APIRouter, Security
|
||||
from app.sql_models import SqlSong
|
||||
from app.dependencies import session
|
||||
from app.routers.user import get_current_user
|
||||
import gspread
|
||||
from gspread.urls import DRIVE_FILES_API_V3_URL
|
||||
import pandas as pd
|
||||
import numpy as np
|
||||
|
||||
@@ -13,87 +11,38 @@ router = APIRouter(
|
||||
responses={404: {"description": "Not found"}},
|
||||
)
|
||||
|
||||
gc = gspread.service_account(filename="service_account.json") # type: ignore
|
||||
|
||||
google_files: dict[str, GoogleFile] = {}
|
||||
|
||||
spreadsheets: dict[str, gspread.Spreadsheet] = {} # type: ignore
|
||||
|
||||
selected_worksheets: dict[tuple[str, int], bool] = {}
|
||||
|
||||
|
||||
def fetch_files():
|
||||
if not (google_files):
|
||||
for s in gc.http_client.request("get", DRIVE_FILES_API_V3_URL).json()["files"]: #
|
||||
google_files[s["id"]] = GoogleFile(
|
||||
file_id=s["id"], file_name=s["name"])
|
||||
|
||||
|
||||
def load_spreadsheet(file_id):
|
||||
if file_id not in spreadsheets:
|
||||
spreadsheets[file_id] = gc.open_by_key(file_id)
|
||||
|
||||
|
||||
# Route to get google files (sheets)
|
||||
@router.get("/files")
|
||||
async def get_files() -> dict[str, GoogleFile]:
|
||||
fetch_files()
|
||||
return google_files
|
||||
|
||||
|
||||
@router.get("/files/{file_id}")
|
||||
def get_worksheets(file_id: str) -> dict[int, str]:
|
||||
fetch_files()
|
||||
if file_id not in google_files:
|
||||
raise HTTPException(status_code=404, detail="Spreadsheet not found.")
|
||||
|
||||
load_spreadsheet(file_id)
|
||||
|
||||
# type: ignore
|
||||
return {ws.id: ws.title for ws in spreadsheets[file_id].worksheets()}
|
||||
|
||||
|
||||
@router.get("/selected_worksheets")
|
||||
def list_selected_worksheet() -> dict[tuple[str, int], bool]:
|
||||
return selected_worksheets
|
||||
|
||||
|
||||
@router.post("/selected_worksheets")
|
||||
def select_worksheet(file_id: str, worksheet_id: int, ignore_arrangement: bool = False):
|
||||
selected_worksheets[(file_id, worksheet_id)] = ignore_arrangement
|
||||
|
||||
|
||||
@router.delete("/selected_worksheets")
|
||||
def deselect_worksheet(file_id: str, worksheet_id: int):
|
||||
del selected_worksheets[(file_id, worksheet_id)]
|
||||
|
||||
|
||||
@router.post("/process_worksheets")
|
||||
def process_worksheets():
|
||||
fetch_files()
|
||||
|
||||
song_list = []
|
||||
|
||||
for (file_id, worksheet_id), ignore_arrangement in selected_worksheets.items():
|
||||
load_spreadsheet(file_id)
|
||||
worksheet = spreadsheets[file_id].get_worksheet_by_id(worksheet_id)
|
||||
worksheet_df = pd.DataFrame(worksheet.get_all_records())
|
||||
|
||||
last_column = np.where(worksheet_df.columns == "Kommentare")[0][0]
|
||||
|
||||
worksheet_df = worksheet_df.iloc[:, 0:last_column+1]
|
||||
|
||||
worksheet_df["Ignore Arrangement"] = ignore_arrangement
|
||||
|
||||
song_list.append(worksheet_df)
|
||||
|
||||
song_list = pd.concat(song_list)
|
||||
|
||||
song_list.to_sql(name='songs', con=dbEngine,
|
||||
index=False, if_exists='append')
|
||||
# song_list.to_csv("song-list.csv")
|
||||
def get_main_category(categories) -> int:
|
||||
if np.sum(categories != None) == 1:
|
||||
return np.argmax(categories != None, axis=0)
|
||||
elif "h" in categories:
|
||||
return np.argmax(categories == "h", axis=0)
|
||||
else:
|
||||
return np.argmax(categories != None, axis=0)
|
||||
|
||||
|
||||
@router.post("/process_file")
|
||||
async def create_upload_file(file: UploadFile):
|
||||
return {"filename": file.filename}
|
||||
async def create_upload_file(link_share: str):
|
||||
song_list = pd.read_excel(link_share)
|
||||
song_list = song_list.replace({np.nan: None})
|
||||
song_list = song_list.replace({"n/a": None})
|
||||
|
||||
category_names = list(song_list.iloc[0][6:19])
|
||||
|
||||
for row in song_list[1:].iterrows():
|
||||
row = np.array(row[1])
|
||||
|
||||
s = SqlSong(og_artist=row[0],
|
||||
aca_artist=row[1],
|
||||
title=row[2],
|
||||
yt_url=row[3],
|
||||
is_aca=row[4] == "ja",
|
||||
arng_url=row[5],
|
||||
categories={n: v for n, v in zip(
|
||||
category_names, row[6:19] != None)},
|
||||
main_category=category_names[get_main_category(row[6:19])],
|
||||
singable=row[19] != "nein"
|
||||
)
|
||||
|
||||
session.add(s)
|
||||
session.commit()
|
||||
|
||||
@@ -1,6 +1,7 @@
|
||||
from fastapi import APIRouter, HTTPException, Security
|
||||
from app.models import Song
|
||||
from app.dependencies import dbEngine, Base, dbSession
|
||||
from app.sql_models import SqlSong, SqlVote
|
||||
from app.dependencies import session
|
||||
from app.routers.user import get_current_user, User
|
||||
from typing import Annotated
|
||||
|
||||
@@ -10,7 +11,20 @@ router = APIRouter(
|
||||
responses={404: {"description": "Not found"}},
|
||||
)
|
||||
|
||||
|
||||
@router.get("/")
|
||||
async def get_songs() -> list[dict]:
|
||||
return dbSession.query(Base.songs).all()
|
||||
async def get_songs(user_id : str = "") -> list[Song]:
|
||||
sqlsongs = session.query(SqlSong).filter(SqlSong.singable == True).all()
|
||||
votes = session.query(SqlVote).filter(SqlVote.user_id == user_id).all()
|
||||
votes = {v.song_id : v.vote for v in votes}
|
||||
|
||||
return [Song(**s.__dict__, vote=votes.get(s.id, None)) for s in sqlsongs] # type: ignore
|
||||
|
||||
@router.post("/{song_id}/vote")
|
||||
async def vote(song_id : str, user_id : str, vote : int):
|
||||
vote_entry = session.query(SqlVote).filter((SqlVote.user_id == user_id) & (SqlVote.song_id == song_id)).first()
|
||||
if vote_entry:
|
||||
vote_entry.vote = str(vote) # type: ignore
|
||||
else:
|
||||
vote_entry = SqlVote(song_id=song_id, user_id=user_id, vote=vote)
|
||||
session.add(vote_entry)
|
||||
session.commit()
|
||||
|
||||
Reference in New Issue
Block a user