mirror of
https://github.com/tcsenpai/pensieve.git
synced 2025-06-06 19:25:24 +00:00
655 lines
24 KiB
Python
655 lines
24 KiB
Python
import asyncio
|
|
import os
|
|
import time
|
|
import logging
|
|
from datetime import datetime, timezone
|
|
from pathlib import Path
|
|
from typing import List, Tuple
|
|
|
|
import httpx
|
|
import typer
|
|
from .server import run_server
|
|
from .read_metadata import read_metadata
|
|
from .schemas import MetadataSource
|
|
from tabulate import tabulate
|
|
from tqdm import tqdm
|
|
from enum import Enum
|
|
from magika import Magika
|
|
|
|
IS_THUMBNAIL = "is_thumbnail"
|
|
|
|
app = typer.Typer(context_settings={"help_option_names": ["-h", "--help"]})
|
|
|
|
lib_app = typer.Typer()
|
|
plugin_app = typer.Typer()
|
|
|
|
app.add_typer(plugin_app, name="plugin")
|
|
app.add_typer(lib_app, name="lib")
|
|
|
|
file_detector = Magika()
|
|
|
|
BASE_URL = "http://localhost:8080"
|
|
|
|
ignore_files = [".DS_Store"]
|
|
|
|
# Configure logging
|
|
logging.basicConfig(
|
|
level=logging.WARNING, # Set the logging level to WARNING or higher
|
|
format="%(asctime)s - %(name)s - %(levelname)s - %(message)s",
|
|
)
|
|
|
|
# Optionally, you can set the logging level for specific libraries
|
|
logging.getLogger("httpx").setLevel(logging.ERROR)
|
|
logging.getLogger("typer").setLevel(logging.ERROR)
|
|
|
|
|
|
class FileStatus(Enum):
|
|
UPDATED = "updated"
|
|
ADDED = "added"
|
|
|
|
|
|
def format_timestamp(timestamp):
|
|
if isinstance(timestamp, str):
|
|
return timestamp
|
|
return datetime.fromtimestamp(timestamp).replace(tzinfo=None).isoformat()
|
|
|
|
|
|
def get_file_type(file_path):
|
|
file_result = file_detector.identify_path(file_path)
|
|
return file_result.output.ct_label, file_result.output.group
|
|
|
|
|
|
def display_libraries(libraries):
|
|
table = []
|
|
for library in libraries:
|
|
table.append(
|
|
[
|
|
library["id"],
|
|
library["name"],
|
|
"\n".join(
|
|
f"{folder['id']}: {folder['path']}" for folder in library["folders"]
|
|
),
|
|
"\n".join(
|
|
f"{plugin['id']}: {plugin['name']} {plugin['webhook_url']}"
|
|
for plugin in library["plugins"]
|
|
),
|
|
]
|
|
)
|
|
|
|
print(
|
|
tabulate(table, headers=["ID", "Name", "Folders", "Plugins"], tablefmt="plain")
|
|
)
|
|
|
|
|
|
@app.command()
|
|
def serve():
|
|
run_server()
|
|
|
|
|
|
@lib_app.command("ls")
|
|
def ls():
|
|
response = httpx.get(f"{BASE_URL}/libraries")
|
|
libraries = response.json()
|
|
display_libraries(libraries)
|
|
|
|
|
|
@lib_app.command("create")
|
|
def add(name: str, folders: List[str]):
|
|
|
|
absolute_folders = [str(Path(folder).resolve()) for folder in folders]
|
|
response = httpx.post(
|
|
f"{BASE_URL}/libraries",
|
|
json={"name": name, "folders": absolute_folders},
|
|
)
|
|
if 200 <= response.status_code < 300:
|
|
print("Library created successfully")
|
|
else:
|
|
print(f"Failed to create library: {response.status_code} - {response.text}")
|
|
|
|
|
|
@lib_app.command("add-folder")
|
|
def add_folder(library_id: int, folders: List[str]):
|
|
absolute_folders = [str(Path(folder).resolve()) for folder in folders]
|
|
response = httpx.post(
|
|
f"{BASE_URL}/libraries/{library_id}/folders",
|
|
json={"folders": absolute_folders},
|
|
)
|
|
if 200 <= response.status_code < 300:
|
|
print("Folders added successfully")
|
|
library = response.json()
|
|
display_libraries([library])
|
|
else:
|
|
print(f"Failed to add folders: {response.status_code} - {response.text}")
|
|
|
|
|
|
@lib_app.command("show")
|
|
def show(library_id: int):
|
|
response = httpx.get(f"{BASE_URL}/libraries/{library_id}")
|
|
if response.status_code == 200:
|
|
library = response.json()
|
|
display_libraries([library])
|
|
else:
|
|
print(f"Failed to retrieve library: {response.status_code} - {response.text}")
|
|
|
|
|
|
async def loop_files(library_id, folder, folder_path, force, plugins):
|
|
updated_file_count = 0
|
|
added_file_count = 0
|
|
scanned_files = set()
|
|
semaphore = asyncio.Semaphore(8)
|
|
async with httpx.AsyncClient(timeout=60) as client:
|
|
tasks = []
|
|
for root, _, files in os.walk(folder_path):
|
|
with tqdm(
|
|
total=len(files), desc=f"Scanning {folder_path}", leave=False
|
|
) as pbar:
|
|
candidate_files = []
|
|
for file in files:
|
|
file_path = Path(root) / file
|
|
absolute_file_path = file_path.resolve() # Get absolute path
|
|
if file in ignore_files:
|
|
continue
|
|
|
|
scanned_files.add(str(absolute_file_path))
|
|
candidate_files.append(str(absolute_file_path))
|
|
|
|
batching = 100
|
|
for i in range(0, len(candidate_files), batching):
|
|
batch = candidate_files[i : i + batching]
|
|
|
|
# Get batch of entities
|
|
get_response = await client.post(
|
|
f"{BASE_URL}/libraries/{library_id}/entities/by-filepaths",
|
|
json=batch,
|
|
)
|
|
|
|
if get_response.status_code == 200:
|
|
existing_entities = get_response.json()
|
|
else:
|
|
print(
|
|
f"Failed to get entities: {get_response.status_code} - {get_response.text}"
|
|
)
|
|
continue
|
|
|
|
existing_entities_dict = {
|
|
entity["filepath"]: entity for entity in existing_entities
|
|
}
|
|
|
|
for file_path in batch:
|
|
absolute_file_path = Path(file_path).resolve()
|
|
file_stat = absolute_file_path.stat()
|
|
file_type, file_type_group = get_file_type(absolute_file_path)
|
|
|
|
new_entity = {
|
|
"filename": absolute_file_path.name,
|
|
"filepath": str(absolute_file_path),
|
|
"size": file_stat.st_size,
|
|
"file_created_at": format_timestamp(file_stat.st_ctime),
|
|
"file_last_modified_at": format_timestamp(
|
|
file_stat.st_mtime
|
|
),
|
|
"file_type": file_type,
|
|
"file_type_group": file_type_group,
|
|
"folder_id": folder["id"],
|
|
}
|
|
|
|
is_thumbnail = False
|
|
|
|
if file_type_group == "image":
|
|
metadata = read_metadata(absolute_file_path)
|
|
if metadata:
|
|
if "active_window" in metadata and "active_app" not in metadata:
|
|
metadata["active_app"] = metadata["active_window"].split(" - ")[0]
|
|
new_entity["metadata_entries"] = [
|
|
{
|
|
"key": key,
|
|
"value": str(value),
|
|
"source": MetadataSource.SYSTEM_GENERATED.value,
|
|
"data_type": "number" if isinstance(value, (int, float)) else "text",
|
|
}
|
|
for key, value in metadata.items() if key != IS_THUMBNAIL
|
|
]
|
|
if "active_app" in metadata:
|
|
new_entity.setdefault("tags", []).append(metadata["active_app"])
|
|
is_thumbnail = metadata.get(IS_THUMBNAIL, False)
|
|
|
|
existing_entity = existing_entities_dict.get(
|
|
str(absolute_file_path)
|
|
)
|
|
if existing_entity:
|
|
existing_created_at = format_timestamp(
|
|
existing_entity["file_created_at"]
|
|
)
|
|
new_created_at = format_timestamp(
|
|
new_entity["file_created_at"]
|
|
)
|
|
existing_modified_at = format_timestamp(
|
|
existing_entity["file_last_modified_at"]
|
|
)
|
|
new_modified_at = format_timestamp(
|
|
new_entity["file_last_modified_at"]
|
|
)
|
|
|
|
# Ignore file changes for thumbnails
|
|
if is_thumbnail:
|
|
new_entity["file_created_at"] = existing_entity["file_created_at"]
|
|
new_entity["file_last_modified_at"] = existing_entity["file_last_modified_at"]
|
|
new_entity["file_type"] = existing_entity["file_type"]
|
|
new_entity["file_type_group"] = existing_entity["file_type_group"]
|
|
new_entity["size"] = existing_entity["size"]
|
|
|
|
# Merge existing metadata with new metadata
|
|
if new_entity.get("metadata_entries"):
|
|
new_metadata_keys = {entry["key"] for entry in new_entity["metadata_entries"]}
|
|
for existing_entry in existing_entity["metadata_entries"]:
|
|
if existing_entry["key"] not in new_metadata_keys:
|
|
new_entity["metadata_entries"].append(existing_entry)
|
|
|
|
if (
|
|
force
|
|
or existing_created_at != new_created_at
|
|
or existing_modified_at != new_modified_at
|
|
):
|
|
tasks.append(
|
|
update_entity(
|
|
client,
|
|
semaphore,
|
|
plugins,
|
|
new_entity,
|
|
existing_entity,
|
|
)
|
|
)
|
|
elif not is_thumbnail: # Ignore thumbnails
|
|
tasks.append(
|
|
add_entity(
|
|
client, semaphore, library_id, plugins, new_entity
|
|
)
|
|
)
|
|
pbar.update(len(batch))
|
|
|
|
# Process all tasks after they've been created
|
|
for future in tqdm(
|
|
asyncio.as_completed(tasks),
|
|
desc=f"Processing {folder_path}",
|
|
total=len(tasks),
|
|
leave=False,
|
|
):
|
|
file_path, file_status, succeeded, response = await future
|
|
if file_status == FileStatus.ADDED:
|
|
if succeeded:
|
|
added_file_count += 1
|
|
tqdm.write(f"Added file to library: {file_path}")
|
|
else:
|
|
error_message = "Failed to add file"
|
|
if hasattr(response, "status_code"):
|
|
error_message += f": {response.status_code}"
|
|
if hasattr(response, "text"):
|
|
error_message += f" - {response.text}"
|
|
else:
|
|
error_message += " - Unknown error occurred"
|
|
tqdm.write(error_message)
|
|
elif file_status == FileStatus.UPDATED:
|
|
if succeeded:
|
|
updated_file_count += 1
|
|
tqdm.write(f"Updated file in library: {file_path}")
|
|
else:
|
|
error_message = "Failed to update file"
|
|
if hasattr(response, "status_code"):
|
|
error_message += f": {response.status_code}"
|
|
elif hasattr(response, "text"):
|
|
error_message += f" - {response.text}"
|
|
else:
|
|
error_message += f" - Unknown error occurred"
|
|
tqdm.write(error_message)
|
|
|
|
return added_file_count, updated_file_count, scanned_files
|
|
|
|
|
|
@lib_app.command("scan")
|
|
def scan(
|
|
library_id: int,
|
|
force: bool = False,
|
|
plugins: List[int] = typer.Option(None, "--plugin", "-p"),
|
|
folders: List[int] = typer.Option(None, "--folder", "-f"),
|
|
):
|
|
|
|
response = httpx.get(f"{BASE_URL}/libraries/{library_id}")
|
|
if response.status_code != 200:
|
|
print(f"Failed to retrieve library: {response.status_code} - {response.text}")
|
|
return
|
|
|
|
library = response.json()
|
|
total_files_added = 0
|
|
total_files_updated = 0
|
|
total_files_deleted = 0
|
|
|
|
# Filter folders if the folders parameter is provided
|
|
if folders:
|
|
library_folders = [
|
|
folder for folder in library["folders"] if folder["id"] in folders
|
|
]
|
|
else:
|
|
library_folders = library["folders"]
|
|
|
|
for folder in library_folders:
|
|
folder_path = Path(folder["path"])
|
|
if not folder_path.exists() or not folder_path.is_dir():
|
|
tqdm.write(f"Folder does not exist or is not a directory: {folder_path}")
|
|
continue
|
|
|
|
added_file_count, updated_file_count, scanned_files = asyncio.run(
|
|
loop_files(library_id, folder, folder_path, force, plugins)
|
|
)
|
|
total_files_added += added_file_count
|
|
total_files_updated += updated_file_count
|
|
|
|
# Check for deleted files
|
|
limit = 100
|
|
offset = 0
|
|
total_entities = 0 # We'll update this after the first request
|
|
with tqdm(
|
|
total=total_entities, desc="Checking for deleted files", leave=True
|
|
) as pbar2:
|
|
while True:
|
|
existing_files_response = httpx.get(
|
|
f"{BASE_URL}/libraries/{library_id}/folders/{folder['id']}/entities",
|
|
params={"limit": limit, "offset": offset},
|
|
timeout=60,
|
|
)
|
|
if existing_files_response.status_code != 200:
|
|
pbar2.write(
|
|
f"Failed to retrieve existing files: {existing_files_response.status_code} - {existing_files_response.text}"
|
|
)
|
|
break
|
|
|
|
existing_files = existing_files_response.json()
|
|
if not existing_files:
|
|
break
|
|
|
|
# Update total if this is the first request
|
|
if offset == 0:
|
|
total_entities = int(
|
|
existing_files_response.headers.get(
|
|
"X-Total-Count", total_entities
|
|
)
|
|
)
|
|
pbar2.total = total_entities
|
|
pbar2.refresh()
|
|
|
|
for existing_file in existing_files:
|
|
if existing_file["filepath"] not in scanned_files:
|
|
# File has been deleted
|
|
delete_response = httpx.delete(
|
|
f"{BASE_URL}/libraries/{library_id}/entities/{existing_file['id']}"
|
|
)
|
|
if 200 <= delete_response.status_code < 300:
|
|
pbar2.write(
|
|
f"Deleted file from library: {existing_file['filepath']}"
|
|
)
|
|
total_files_deleted += 1
|
|
else:
|
|
pbar2.write(
|
|
f"Failed to delete file: {delete_response.status_code} - {delete_response.text}"
|
|
)
|
|
pbar2.update(1)
|
|
|
|
offset += limit
|
|
|
|
print(f"Total files added: {total_files_added}")
|
|
print(f"Total files updated: {total_files_updated}")
|
|
print(f"Total files deleted: {total_files_deleted}")
|
|
|
|
|
|
async def add_entity(
|
|
client: httpx.AsyncClient,
|
|
semaphore: asyncio.Semaphore,
|
|
library_id,
|
|
plugins,
|
|
new_entity,
|
|
) -> Tuple[FileStatus, bool, httpx.Response]:
|
|
async with semaphore:
|
|
MAX_RETRIES = 3
|
|
RETRY_DELAY = 2.0
|
|
for attempt in range(MAX_RETRIES):
|
|
try:
|
|
post_response = await client.post(
|
|
f"{BASE_URL}/libraries/{library_id}/entities",
|
|
json=new_entity,
|
|
params={"plugins": plugins} if plugins else {},
|
|
timeout=60,
|
|
)
|
|
if 200 <= post_response.status_code < 300:
|
|
return new_entity["filepath"], FileStatus.ADDED, True, post_response
|
|
else:
|
|
return (
|
|
new_entity["filepath"],
|
|
FileStatus.ADDED,
|
|
False,
|
|
post_response,
|
|
)
|
|
except Exception as e:
|
|
logging.error(
|
|
f"Error while adding entity (attempt {attempt + 1}/{MAX_RETRIES}): {e}"
|
|
)
|
|
if attempt < MAX_RETRIES - 1:
|
|
await asyncio.sleep(RETRY_DELAY)
|
|
else:
|
|
return new_entity["filepath"], FileStatus.ADDED, False, None
|
|
|
|
|
|
async def update_entity(
|
|
client: httpx.AsyncClient,
|
|
semaphore: asyncio.Semaphore,
|
|
plugins,
|
|
new_entity,
|
|
existing_entity,
|
|
) -> Tuple[FileStatus, bool, httpx.Response]:
|
|
MAX_RETRIES = 3
|
|
RETRY_DELAY = 2.0
|
|
async with semaphore:
|
|
for attempt in range(MAX_RETRIES):
|
|
try:
|
|
update_response = await client.put(
|
|
f"{BASE_URL}/entities/{existing_entity['id']}",
|
|
json=new_entity,
|
|
params={
|
|
"trigger_webhooks_flag": "true",
|
|
**({"plugins": plugins} if plugins else {}),
|
|
},
|
|
timeout=60,
|
|
)
|
|
if 200 <= update_response.status_code < 300:
|
|
return (
|
|
new_entity["filepath"],
|
|
FileStatus.UPDATED,
|
|
True,
|
|
update_response,
|
|
)
|
|
else:
|
|
return (
|
|
new_entity["filepath"],
|
|
FileStatus.UPDATED,
|
|
False,
|
|
update_response,
|
|
)
|
|
except Exception as e:
|
|
logging.error(
|
|
f"Error while updating entity {existing_entity['id']} (attempt {attempt + 1}/{MAX_RETRIES}): {e}"
|
|
)
|
|
if attempt < MAX_RETRIES - 1:
|
|
await asyncio.sleep(RETRY_DELAY)
|
|
else:
|
|
return new_entity["filepath"], FileStatus.UPDATED, False, None
|
|
|
|
|
|
@lib_app.command("index")
|
|
def index(
|
|
library_id: int,
|
|
folders: List[int] = typer.Option(None, "--folder", "-f"),
|
|
):
|
|
print(f"Indexing library {library_id}")
|
|
|
|
# Get the library
|
|
response = httpx.get(f"{BASE_URL}/libraries/{library_id}")
|
|
if response.status_code != 200:
|
|
print(f"Failed to get library: {response.status_code} - {response.text}")
|
|
return
|
|
|
|
library = response.json()
|
|
scanned_entities = set()
|
|
|
|
# Filter folders if the folders parameter is provided
|
|
if folders:
|
|
library_folders = [
|
|
folder for folder in library["folders"] if folder["id"] in folders
|
|
]
|
|
else:
|
|
library_folders = library["folders"]
|
|
|
|
# Iterate through folders
|
|
for folder in library_folders:
|
|
tqdm.write(f"Processing folder: {folder['id']}")
|
|
|
|
# List all entities in the folder
|
|
limit = 200
|
|
offset = 0
|
|
total_entities = 0 # We'll update this after the first request
|
|
with tqdm(total=total_entities, desc="Indexing entities", leave=True) as pbar:
|
|
while True:
|
|
entities_response = httpx.get(
|
|
f"{BASE_URL}/libraries/{library_id}/folders/{folder['id']}/entities",
|
|
params={"limit": limit, "offset": offset},
|
|
timeout=60,
|
|
)
|
|
if entities_response.status_code != 200:
|
|
pbar.write(
|
|
f"Failed to get entities: {entities_response.status_code} - {entities_response.text}"
|
|
)
|
|
break
|
|
|
|
entities = entities_response.json()
|
|
if not entities:
|
|
break
|
|
|
|
# Update total if this is the first request
|
|
if offset == 0:
|
|
total_entities = int(
|
|
entities_response.headers.get("X-Total-Count", total_entities)
|
|
)
|
|
pbar.total = total_entities
|
|
pbar.refresh()
|
|
|
|
# Index each entity
|
|
batch_size = 40
|
|
for i in range(0, len(entities), batch_size):
|
|
batch = entities[i : i + batch_size]
|
|
entity_ids = [entity["id"] for entity in batch]
|
|
index_response = httpx.post(
|
|
f"{BASE_URL}/entities/batch-index",
|
|
json=entity_ids,
|
|
timeout=60,
|
|
)
|
|
if index_response.status_code == 204:
|
|
pbar.write(f"Indexed batch of {len(batch)} entities")
|
|
else:
|
|
pbar.write(
|
|
f"Failed to index batch: {index_response.status_code} - {index_response.text}"
|
|
)
|
|
|
|
scanned_entities.update(str(entity["id"]) for entity in batch)
|
|
pbar.update(len(batch))
|
|
|
|
offset += limit
|
|
|
|
# List all indexed entities in the folder
|
|
offset = 0
|
|
while True:
|
|
index_response = httpx.get(
|
|
f"{BASE_URL}/libraries/{library_id}/folders/{folder['id']}/index",
|
|
params={"limit": 200, "offset": offset},
|
|
timeout=60,
|
|
)
|
|
if index_response.status_code != 200:
|
|
tqdm.write(
|
|
f"Failed to get indexed entities: {index_response.status_code} - {index_response.text}"
|
|
)
|
|
break
|
|
|
|
indexed_entities = index_response.json()
|
|
if not indexed_entities:
|
|
break
|
|
|
|
# Delete indexes for entities not in scanned_entities
|
|
for indexed_entity in tqdm(
|
|
indexed_entities, desc="Cleaning up indexes", leave=False
|
|
):
|
|
if indexed_entity["id"] not in scanned_entities:
|
|
delete_response = httpx.delete(
|
|
f"{BASE_URL}/entities/{indexed_entity['id']}/index"
|
|
)
|
|
if delete_response.status_code == 204:
|
|
tqdm.write(f"Deleted index for entity: {indexed_entity['id']}")
|
|
else:
|
|
tqdm.write(
|
|
f"Failed to delete index for entity {indexed_entity['id']}: {delete_response.status_code} - {delete_response.text}"
|
|
)
|
|
|
|
offset += 200
|
|
|
|
print("Indexing completed")
|
|
|
|
|
|
def display_plugins(plugins):
|
|
table = []
|
|
for plugin in plugins:
|
|
table.append(
|
|
[plugin["id"], plugin["name"], plugin["description"], plugin["webhook_url"]]
|
|
)
|
|
print(
|
|
tabulate(
|
|
table,
|
|
headers=["ID", "Name", "Description", "Webhook URL"],
|
|
tablefmt="plain",
|
|
)
|
|
)
|
|
|
|
|
|
@plugin_app.command("ls")
|
|
def ls():
|
|
response = httpx.get(f"{BASE_URL}/plugins")
|
|
plugins = response.json()
|
|
display_plugins(plugins)
|
|
|
|
|
|
@plugin_app.command("create")
|
|
def create(name: str, webhook_url: str, description: str = ""):
|
|
response = httpx.post(
|
|
f"{BASE_URL}/plugins",
|
|
json={"name": name, "description": description, "webhook_url": webhook_url},
|
|
)
|
|
if 200 <= response.status_code < 300:
|
|
print("Plugin created successfully")
|
|
else:
|
|
print(f"Failed to create plugin: {response.status_code} - {response.text}")
|
|
|
|
|
|
@plugin_app.command("bind")
|
|
def bind(
|
|
library_id: int = typer.Option(..., "--lib", help="ID of the library"),
|
|
plugin_id: int = typer.Option(..., "--plugin", help="ID of the plugin"),
|
|
):
|
|
response = httpx.post(
|
|
f"{BASE_URL}/libraries/{library_id}/plugins",
|
|
json={"plugin_id": plugin_id},
|
|
)
|
|
if 200 <= response.status_code < 300:
|
|
print("Plugin bound to library successfully")
|
|
else:
|
|
print(
|
|
f"Failed to bind plugin to library: {response.status_code} - {response.text}"
|
|
)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
app()
|