Auto-update: Mon Jul 29 15:43:01 PDT 2024
This commit is contained in:
parent
1c2812ce43
commit
2354fb1588
1 changed files with 121 additions and 73 deletions
|
@ -387,86 +387,132 @@ class APIConfig(BaseModel):
|
||||||
dest_id = os.environ.get('TS_ID')
|
dest_id = os.environ.get('TS_ID')
|
||||||
dest_ip = self.local_db['ts_ip']
|
dest_ip = self.local_db['ts_ip']
|
||||||
|
|
||||||
info(f"Starting sync from source {source_id} ({source_ip}) to destination {dest_id} ({dest_ip})")
|
info(f"Starting comprehensive sync from source {source_id} ({source_ip}) to destination {dest_id} ({dest_ip})")
|
||||||
|
|
||||||
async with self.get_connection(source_pool_entry) as source_conn:
|
try:
|
||||||
async with self.get_connection() as dest_conn:
|
async with self.get_connection(source_pool_entry) as source_conn:
|
||||||
tables = await source_conn.fetch("""
|
async with self.get_connection() as dest_conn:
|
||||||
SELECT tablename FROM pg_tables
|
# Compare tables
|
||||||
WHERE schemaname = 'public'
|
source_tables = await self.get_tables(source_conn)
|
||||||
""")
|
dest_tables = await self.get_tables(dest_conn)
|
||||||
|
|
||||||
for table in tables:
|
|
||||||
table_name = table['tablename']
|
|
||||||
inserts = 0
|
|
||||||
updates = 0
|
|
||||||
error_count = 0
|
|
||||||
last_synced_version = await self.get_last_synced_version(table_name, source_id)
|
|
||||||
|
|
||||||
# Get primary key columns for the table
|
tables_only_in_source = set(source_tables) - set(dest_tables)
|
||||||
primary_keys = await self.get_primary_keys(dest_conn, table_name)
|
tables_only_in_dest = set(dest_tables) - set(source_tables)
|
||||||
|
common_tables = set(source_tables) & set(dest_tables)
|
||||||
changes = await source_conn.fetch(f"""
|
|
||||||
SELECT * FROM "{table_name}"
|
|
||||||
WHERE version > $1 AND server_id = $2
|
|
||||||
ORDER BY version ASC
|
|
||||||
""", last_synced_version, source_id)
|
|
||||||
|
|
||||||
for change in changes:
|
|
||||||
columns = list(change.keys())
|
|
||||||
values = [change[col] for col in columns]
|
|
||||||
|
|
||||||
debug(f"Syncing data for table: {table_name} from {source_id} to {dest_id}")
|
|
||||||
|
|
||||||
if not primary_keys:
|
|
||||||
# If no primary key, just insert
|
|
||||||
insert_query = f"""
|
|
||||||
INSERT INTO "{table_name}" ({', '.join(columns)})
|
|
||||||
VALUES ({', '.join(f'${i+1}' for i in range(len(columns)))})
|
|
||||||
"""
|
|
||||||
else:
|
|
||||||
# If primary key exists, use it for conflict resolution
|
|
||||||
insert_query = f"""
|
|
||||||
INSERT INTO "{table_name}" ({', '.join(columns)})
|
|
||||||
VALUES ({', '.join(f'${i+1}' for i in range(len(columns)))})
|
|
||||||
ON CONFLICT ({', '.join(primary_keys)}) DO UPDATE SET
|
|
||||||
{', '.join(f"{col} = EXCLUDED.{col}" for col in columns if col not in primary_keys)}
|
|
||||||
"""
|
|
||||||
|
|
||||||
try:
|
|
||||||
result = await dest_conn.execute(insert_query, *values)
|
|
||||||
if 'UPDATE' in result:
|
|
||||||
updates += 1
|
|
||||||
else:
|
|
||||||
inserts += 1
|
|
||||||
except Exception as e:
|
|
||||||
if error_count < 10: # Limit error logging
|
|
||||||
err(f"Error syncing data for table {table_name} from {source_id} to {dest_id}: {str(e)}")
|
|
||||||
error_count += 1
|
|
||||||
elif error_count == 10:
|
|
||||||
err(f"Suppressing further errors for table {table_name}")
|
|
||||||
error_count += 1
|
|
||||||
|
|
||||||
if changes:
|
|
||||||
await self.update_sync_status(table_name, source_id, changes[-1]['version'])
|
|
||||||
|
|
||||||
total_inserts += inserts
|
|
||||||
total_updates += updates
|
|
||||||
table_changes[table_name] = {'inserts': inserts, 'updates': updates}
|
|
||||||
|
|
||||||
info(f"Synced {table_name} from {source_id} to {dest_id}: {inserts} inserts, {updates} updates")
|
|
||||||
if error_count > 10:
|
|
||||||
info(f"Total of {error_count} errors occurred for table {table_name}")
|
|
||||||
|
|
||||||
info(f"Sync complete from {source_id} ({source_ip}) to {dest_id} ({dest_ip})")
|
info(f"Tables only in source: {tables_only_in_source}")
|
||||||
info(f"Total changes: {total_inserts} inserts, {total_updates} updates")
|
info(f"Tables only in destination: {tables_only_in_dest}")
|
||||||
info("Changes by table:")
|
info(f"Common tables: {common_tables}")
|
||||||
for table, changes in table_changes.items():
|
|
||||||
info(f" {table}: {changes['inserts']} inserts, {changes['updates']} updates")
|
for table in common_tables:
|
||||||
|
await self.compare_table_structure(source_conn, dest_conn, table)
|
||||||
|
inserts, updates = await self.compare_and_sync_data(source_conn, dest_conn, table, source_id)
|
||||||
|
|
||||||
|
total_inserts += inserts
|
||||||
|
total_updates += updates
|
||||||
|
table_changes[table] = {'inserts': inserts, 'updates': updates}
|
||||||
|
|
||||||
|
info(f"Comprehensive sync complete from {source_id} ({source_ip}) to {dest_id} ({dest_ip})")
|
||||||
|
info(f"Total changes: {total_inserts} inserts, {total_updates} updates")
|
||||||
|
info("Changes by table:")
|
||||||
|
for table, changes in table_changes.items():
|
||||||
|
info(f" {table}: {changes['inserts']} inserts, {changes['updates']} updates")
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
err(f"Error during sync process: {str(e)}")
|
||||||
|
|
||||||
return total_inserts + total_updates
|
return total_inserts + total_updates
|
||||||
|
|
||||||
|
|
||||||
|
async def get_tables(self, conn):
|
||||||
|
tables = await conn.fetch("""
|
||||||
|
SELECT tablename FROM pg_tables
|
||||||
|
WHERE schemaname = 'public'
|
||||||
|
""")
|
||||||
|
return [table['tablename'] for table in tables]
|
||||||
|
|
||||||
|
async def compare_table_structure(self, source_conn, dest_conn, table_name):
|
||||||
|
source_columns = await self.get_table_structure(source_conn, table_name)
|
||||||
|
dest_columns = await self.get_table_structure(dest_conn, table_name)
|
||||||
|
|
||||||
|
columns_only_in_source = set(source_columns.keys()) - set(dest_columns.keys())
|
||||||
|
columns_only_in_dest = set(dest_columns.keys()) - set(source_columns.keys())
|
||||||
|
common_columns = set(source_columns.keys()) & set(dest_columns.keys())
|
||||||
|
|
||||||
|
info(f"Table {table_name}:")
|
||||||
|
info(f" Columns only in source: {columns_only_in_source}")
|
||||||
|
info(f" Columns only in destination: {columns_only_in_dest}")
|
||||||
|
info(f" Common columns: {common_columns}")
|
||||||
|
|
||||||
|
for col in common_columns:
|
||||||
|
if source_columns[col] != dest_columns[col]:
|
||||||
|
warn(f" Column {col} has different types: source={source_columns[col]}, dest={dest_columns[col]}")
|
||||||
|
|
||||||
|
async def get_table_structure(self, conn, table_name):
|
||||||
|
columns = await conn.fetch("""
|
||||||
|
SELECT column_name, data_type
|
||||||
|
FROM information_schema.columns
|
||||||
|
WHERE table_name = $1
|
||||||
|
""", table_name)
|
||||||
|
return {col['column_name']: col['data_type'] for col in columns}
|
||||||
|
|
||||||
|
async def compare_and_sync_data(self, source_conn, dest_conn, table_name, source_id):
|
||||||
|
inserts = 0
|
||||||
|
updates = 0
|
||||||
|
error_count = 0
|
||||||
|
|
||||||
|
try:
|
||||||
|
primary_keys = await self.get_primary_keys(dest_conn, table_name)
|
||||||
|
if not primary_keys:
|
||||||
|
warn(f"Table {table_name} has no primary keys. Skipping data sync.")
|
||||||
|
return inserts, updates
|
||||||
|
|
||||||
|
last_synced_version = await self.get_last_synced_version(table_name, source_id)
|
||||||
|
|
||||||
|
changes = await source_conn.fetch(f"""
|
||||||
|
SELECT * FROM "{table_name}"
|
||||||
|
WHERE version > $1 AND server_id = $2
|
||||||
|
ORDER BY version ASC
|
||||||
|
""", last_synced_version, source_id)
|
||||||
|
|
||||||
|
for change in changes:
|
||||||
|
columns = list(change.keys())
|
||||||
|
values = [change[col] for col in columns]
|
||||||
|
|
||||||
|
insert_query = f"""
|
||||||
|
INSERT INTO "{table_name}" ({', '.join(columns)})
|
||||||
|
VALUES ({', '.join(f'${i+1}' for i in range(len(columns)))})
|
||||||
|
ON CONFLICT ({', '.join(primary_keys)}) DO UPDATE SET
|
||||||
|
{', '.join(f"{col} = EXCLUDED.{col}" for col in columns if col not in primary_keys)}
|
||||||
|
"""
|
||||||
|
|
||||||
|
try:
|
||||||
|
result = await dest_conn.execute(insert_query, *values)
|
||||||
|
if 'UPDATE' in result:
|
||||||
|
updates += 1
|
||||||
|
else:
|
||||||
|
inserts += 1
|
||||||
|
except Exception as e:
|
||||||
|
if error_count < 10: # Limit error logging
|
||||||
|
err(f"Error syncing data for table {table_name}: {str(e)}")
|
||||||
|
error_count += 1
|
||||||
|
elif error_count == 10:
|
||||||
|
err(f"Suppressing further errors for table {table_name}")
|
||||||
|
error_count += 1
|
||||||
|
|
||||||
|
if changes:
|
||||||
|
await self.update_sync_status(table_name, source_id, changes[-1]['version'])
|
||||||
|
|
||||||
|
info(f"Synced {table_name}: {inserts} inserts, {updates} updates")
|
||||||
|
if error_count > 10:
|
||||||
|
info(f"Total of {error_count} errors occurred for table {table_name}")
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
err(f"Error processing table {table_name}: {str(e)}")
|
||||||
|
|
||||||
|
return inserts, updates
|
||||||
|
|
||||||
|
|
||||||
async def get_primary_keys(self, conn, table_name):
|
async def get_primary_keys(self, conn, table_name):
|
||||||
primary_keys = await conn.fetch("""
|
primary_keys = await conn.fetch("""
|
||||||
SELECT a.attname
|
SELECT a.attname
|
||||||
|
@ -526,6 +572,7 @@ class APIConfig(BaseModel):
|
||||||
WHERE table_name = $1 AND server_id = $2
|
WHERE table_name = $1 AND server_id = $2
|
||||||
""", table_name, server_id) or 0
|
""", table_name, server_id) or 0
|
||||||
|
|
||||||
|
|
||||||
async def update_sync_status(self, table_name, server_id, version):
|
async def update_sync_status(self, table_name, server_id, version):
|
||||||
async with self.get_connection() as conn:
|
async with self.get_connection() as conn:
|
||||||
await conn.execute("""
|
await conn.execute("""
|
||||||
|
@ -535,6 +582,7 @@ class APIConfig(BaseModel):
|
||||||
SET last_synced_version = EXCLUDED.last_synced_version
|
SET last_synced_version = EXCLUDED.last_synced_version
|
||||||
""", table_name, server_id, version)
|
""", table_name, server_id, version)
|
||||||
|
|
||||||
|
|
||||||
async def sync_schema(self):
|
async def sync_schema(self):
|
||||||
source_entry = self.local_db
|
source_entry = self.local_db
|
||||||
source_schema = await self.get_schema(source_entry)
|
source_schema = await self.get_schema(source_entry)
|
||||||
|
|
Loading…
Add table
Reference in a new issue