Auto-update: Tue Jul 30 14:29:20 PDT 2024
This commit is contained in:
parent
7fb82cf08c
commit
297c3454df
1 changed files with 77 additions and 12 deletions
|
@ -352,10 +352,18 @@ class APIConfig(BaseModel):
|
||||||
try:
|
try:
|
||||||
async with self.get_connection(pool_entry) as conn:
|
async with self.get_connection(pool_entry) as conn:
|
||||||
info(f"Starting sync initialization for {pool_entry['ts_ip']}...")
|
info(f"Starting sync initialization for {pool_entry['ts_ip']}...")
|
||||||
|
|
||||||
|
# Check PostGIS installation
|
||||||
|
postgis_installed = await self.check_postgis(conn)
|
||||||
|
if not postgis_installed:
|
||||||
|
warn(f"PostGIS is not installed on {pool_entry['ts_id']} ({pool_entry['ts_ip']}). Some spatial operations may fail.")
|
||||||
|
|
||||||
|
# Continue with sync initialization
|
||||||
tables = await conn.fetch("""
|
tables = await conn.fetch("""
|
||||||
SELECT tablename FROM pg_tables
|
SELECT tablename FROM pg_tables
|
||||||
WHERE schemaname = 'public'
|
WHERE schemaname = 'public'
|
||||||
""")
|
""")
|
||||||
|
|
||||||
all_tables_synced = True
|
all_tables_synced = True
|
||||||
for table in tables:
|
for table in tables:
|
||||||
table_name = table['tablename']
|
table_name = table['tablename']
|
||||||
|
@ -366,11 +374,27 @@ class APIConfig(BaseModel):
|
||||||
info(f"Sync initialization complete for {pool_entry['ts_ip']}. All tables now have version and server_id columns with appropriate triggers.")
|
info(f"Sync initialization complete for {pool_entry['ts_ip']}. All tables now have version and server_id columns with appropriate triggers.")
|
||||||
else:
|
else:
|
||||||
warn(f"Sync initialization partially complete for {pool_entry['ts_ip']}. Some tables may be missing version or server_id columns.")
|
warn(f"Sync initialization partially complete for {pool_entry['ts_ip']}. Some tables may be missing version or server_id columns.")
|
||||||
|
|
||||||
|
# Initialize sync_status table if it doesn't exist
|
||||||
|
await self.initialize_sync_status_table(conn)
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
err(f"Error initializing sync for {pool_entry['ts_ip']}: {str(e)}")
|
err(f"Error initializing sync for {pool_entry['ts_ip']}: {str(e)}")
|
||||||
err(f"Traceback: {traceback.format_exc()}")
|
err(f"Traceback: {traceback.format_exc()}")
|
||||||
|
|
||||||
|
|
||||||
|
async def initialize_sync_status_table(self, conn):
|
||||||
|
await conn.execute("""
|
||||||
|
CREATE TABLE IF NOT EXISTS sync_status (
|
||||||
|
table_name TEXT,
|
||||||
|
server_id TEXT,
|
||||||
|
last_synced_version INTEGER,
|
||||||
|
last_sync_time TIMESTAMP WITH TIME ZONE,
|
||||||
|
PRIMARY KEY (table_name, server_id)
|
||||||
|
)
|
||||||
|
""")
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
async def ensure_sync_structure(self, conn):
|
async def ensure_sync_structure(self, conn):
|
||||||
tables = await conn.fetch("""
|
tables = await conn.fetch("""
|
||||||
|
@ -468,16 +492,24 @@ class APIConfig(BaseModel):
|
||||||
for table in tables:
|
for table in tables:
|
||||||
table_name = table['tablename']
|
table_name = table['tablename']
|
||||||
try:
|
try:
|
||||||
version = await conn.fetchval(f"""
|
result = await conn.fetchrow(f"""
|
||||||
SELECT COALESCE(MAX(version), -1)
|
SELECT MAX(version) as max_version, server_id
|
||||||
FROM "{table_name}"
|
FROM "{table_name}"
|
||||||
|
WHERE version = (SELECT MAX(version) FROM "{table_name}")
|
||||||
|
GROUP BY server_id
|
||||||
|
ORDER BY MAX(version) DESC
|
||||||
|
LIMIT 1
|
||||||
""")
|
""")
|
||||||
info(f"Max version for {pool_entry['ts_id']}, table {table_name}: {version}")
|
if result:
|
||||||
|
version, server_id = result['max_version'], result['server_id']
|
||||||
|
info(f"Max version for {pool_entry['ts_id']}, table {table_name}: {version} (from server {server_id})")
|
||||||
if version > max_version:
|
if version > max_version:
|
||||||
max_version = version
|
max_version = version
|
||||||
most_recent_source = pool_entry
|
most_recent_source = pool_entry
|
||||||
|
else:
|
||||||
|
info(f"No data in table {table_name} for {pool_entry['ts_id']}")
|
||||||
except asyncpg.exceptions.UndefinedColumnError:
|
except asyncpg.exceptions.UndefinedColumnError:
|
||||||
warn(f"Version column does not exist in table {table_name} for {pool_entry['ts_id']}. Attempting to add...")
|
warn(f"Version or server_id column does not exist in table {table_name} for {pool_entry['ts_id']}. Attempting to add...")
|
||||||
await self.ensure_sync_columns(conn, table_name)
|
await self.ensure_sync_columns(conn, table_name)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
err(f"Error checking version for {pool_entry['ts_id']}, table {table_name}: {str(e)}")
|
err(f"Error checking version for {pool_entry['ts_id']}, table {table_name}: {str(e)}")
|
||||||
|
@ -492,6 +524,7 @@ class APIConfig(BaseModel):
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
async def is_server_accessible(self, host, port, timeout=2):
|
async def is_server_accessible(self, host, port, timeout=2):
|
||||||
try:
|
try:
|
||||||
future = asyncio.open_connection(host, port)
|
future = asyncio.open_connection(host, port)
|
||||||
|
@ -530,8 +563,6 @@ class APIConfig(BaseModel):
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
async def pull_changes(self, source_pool_entry, batch_size=10000):
|
async def pull_changes(self, source_pool_entry, batch_size=10000):
|
||||||
if source_pool_entry['ts_id'] == os.environ.get('TS_ID'):
|
if source_pool_entry['ts_id'] == os.environ.get('TS_ID'):
|
||||||
info("Skipping self-sync")
|
info("Skipping self-sync")
|
||||||
|
@ -572,7 +603,7 @@ class APIConfig(BaseModel):
|
||||||
total_changes += changes_count
|
total_changes += changes_count
|
||||||
|
|
||||||
last_synced_version = changes[-1]['version']
|
last_synced_version = changes[-1]['version']
|
||||||
await self.update_last_synced_version(dest_conn, table_name, source_id, last_synced_version)
|
await self.update_sync_status(dest_conn, table_name, source_id, last_synced_version)
|
||||||
|
|
||||||
info(f"Synced batch for {table_name}: {changes_count} changes. Total so far: {total_changes}")
|
info(f"Synced batch for {table_name}: {changes_count} changes. Total so far: {total_changes}")
|
||||||
|
|
||||||
|
@ -584,6 +615,7 @@ class APIConfig(BaseModel):
|
||||||
|
|
||||||
return total_changes
|
return total_changes
|
||||||
|
|
||||||
|
|
||||||
async def apply_batch_changes(self, conn, table_name, changes):
|
async def apply_batch_changes(self, conn, table_name, changes):
|
||||||
if not changes:
|
if not changes:
|
||||||
return 0
|
return 0
|
||||||
|
@ -599,7 +631,10 @@ class APIConfig(BaseModel):
|
||||||
|
|
||||||
# Bulk insert changes into temporary table
|
# Bulk insert changes into temporary table
|
||||||
columns = changes[0].keys()
|
columns = changes[0].keys()
|
||||||
await conn.copy_records_to_table(temp_table_name, records=[tuple(change[col] for col in columns) for change in changes])
|
records = [tuple(change[col] for col in columns) for change in changes]
|
||||||
|
|
||||||
|
# Use copy_records instead of copy_records_to_table
|
||||||
|
await conn.copy_records(temp_table_name, records=records, columns=columns)
|
||||||
|
|
||||||
# Perform upsert
|
# Perform upsert
|
||||||
result = await conn.execute(f"""
|
result = await conn.execute(f"""
|
||||||
|
@ -613,9 +648,17 @@ class APIConfig(BaseModel):
|
||||||
affected_rows = int(result.split()[-1])
|
affected_rows = int(result.split()[-1])
|
||||||
return affected_rows
|
return affected_rows
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
err(f"Error applying batch changes to {table_name}: {str(e)}")
|
||||||
|
err(f"Traceback: {traceback.format_exc()}")
|
||||||
|
return 0
|
||||||
finally:
|
finally:
|
||||||
# Ensure temporary table is dropped
|
# Ensure temporary table is dropped
|
||||||
|
try:
|
||||||
await conn.execute(f"DROP TABLE IF EXISTS {temp_table_name}")
|
await conn.execute(f"DROP TABLE IF EXISTS {temp_table_name}")
|
||||||
|
except Exception as e:
|
||||||
|
warn(f"Failed to drop temporary table {temp_table_name}: {str(e)}")
|
||||||
|
|
||||||
|
|
||||||
async def push_changes_to_all(self):
|
async def push_changes_to_all(self):
|
||||||
for pool_entry in self.POOL:
|
for pool_entry in self.POOL:
|
||||||
|
@ -666,6 +709,15 @@ class APIConfig(BaseModel):
|
||||||
err(f"Error pushing changes to {pool_entry['ts_id']}: {str(e)}")
|
err(f"Error pushing changes to {pool_entry['ts_id']}: {str(e)}")
|
||||||
err(f"Traceback: {traceback.format_exc()}")
|
err(f"Traceback: {traceback.format_exc()}")
|
||||||
|
|
||||||
|
async def update_sync_status(self, conn, table_name, server_id, version):
|
||||||
|
await conn.execute("""
|
||||||
|
INSERT INTO sync_status (table_name, server_id, last_synced_version, last_sync_time)
|
||||||
|
VALUES ($1, $2, $3, NOW())
|
||||||
|
ON CONFLICT (table_name, server_id) DO UPDATE
|
||||||
|
SET last_synced_version = EXCLUDED.last_synced_version,
|
||||||
|
last_sync_time = EXCLUDED.last_sync_time
|
||||||
|
""", table_name, server_id, version)
|
||||||
|
|
||||||
async def get_last_synced_version(self, conn, table_name, server_id):
|
async def get_last_synced_version(self, conn, table_name, server_id):
|
||||||
return await conn.fetchval(f"""
|
return await conn.fetchval(f"""
|
||||||
SELECT COALESCE(MAX(version), 0)
|
SELECT COALESCE(MAX(version), 0)
|
||||||
|
@ -701,6 +753,19 @@ class APIConfig(BaseModel):
|
||||||
END $$;
|
END $$;
|
||||||
""")
|
""")
|
||||||
|
|
||||||
|
async def check_postgis(self, conn):
|
||||||
|
try:
|
||||||
|
result = await conn.fetchval("SELECT PostGIS_version();")
|
||||||
|
if result:
|
||||||
|
info(f"PostGIS version: {result}")
|
||||||
|
return True
|
||||||
|
else:
|
||||||
|
warn("PostGIS is not installed or not working properly")
|
||||||
|
return False
|
||||||
|
except Exception as e:
|
||||||
|
err(f"Error checking PostGIS: {str(e)}")
|
||||||
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue