2022-01-16 18:22:00 +00:00
|
|
|
import click
|
|
|
|
from flask.cli import with_appcontext
|
2022-03-13 02:09:35 +00:00
|
|
|
import random
|
|
|
|
import string
|
|
|
|
import datetime
|
2022-01-16 18:22:00 +00:00
|
|
|
from flask_user import current_app
|
|
|
|
from app import db
|
2022-04-14 03:29:57 +00:00
|
|
|
from app.models import Account, PlayKey, CharacterInfo, Property, PropertyContent, UGC, Mail
|
2022-01-27 05:11:49 +00:00
|
|
|
import pathlib
|
|
|
|
import zlib
|
2022-02-10 18:01:51 +00:00
|
|
|
from wand import image
|
|
|
|
from wand.exceptions import BlobError as BE
|
|
|
|
import app.pylddlib as ldd
|
|
|
|
from multiprocessing import Pool
|
|
|
|
from functools import partial
|
2022-04-14 02:29:08 +00:00
|
|
|
from sqlalchemy import func
|
2022-04-14 03:29:57 +00:00
|
|
|
import time
|
2022-03-13 02:09:35 +00:00
|
|
|
|
2022-05-12 01:37:22 +00:00
|
|
|
|
2022-01-16 18:22:00 +00:00
|
|
|
@click.command("init_db")
|
|
|
|
@click.argument('drop_tables', nargs=1)
|
|
|
|
@with_appcontext
|
|
|
|
def init_db(drop_tables=False):
|
|
|
|
""" Initialize the database."""
|
|
|
|
|
|
|
|
print('Initializing Database.')
|
|
|
|
if drop_tables:
|
|
|
|
print('Dropping all tables.')
|
|
|
|
db.drop_all()
|
|
|
|
print('Creating all tables.')
|
|
|
|
db.create_all()
|
|
|
|
print('Database has been initialized.')
|
|
|
|
return
|
|
|
|
|
|
|
|
|
|
|
|
@click.command("init_accounts")
|
|
|
|
@with_appcontext
|
|
|
|
def init_accounts():
|
|
|
|
""" Initialize the accounts."""
|
|
|
|
|
|
|
|
# Add accounts
|
|
|
|
print('Creating Admin account.')
|
2022-01-27 05:11:49 +00:00
|
|
|
find_or_create_account(
|
2022-01-16 18:22:00 +00:00
|
|
|
'admin',
|
|
|
|
'example@example.com',
|
|
|
|
'Nope',
|
|
|
|
)
|
|
|
|
|
|
|
|
return
|
|
|
|
|
2022-03-13 02:09:35 +00:00
|
|
|
|
2022-04-02 21:33:23 +00:00
|
|
|
@click.command("fix_clone_ids")
|
|
|
|
@with_appcontext
|
|
|
|
def fix_clone_ids():
|
2022-04-14 03:29:57 +00:00
|
|
|
"""
|
|
|
|
Fix incorrect prop_clone_id's
|
|
|
|
Remove duplicate properties
|
|
|
|
Either the one with most models or most recently claimed
|
|
|
|
Retuen Pre-built models via mail
|
|
|
|
(May have errors and need to be run multiple times)
|
|
|
|
"""
|
2022-04-02 21:33:23 +00:00
|
|
|
properties = Property.query.all()
|
|
|
|
count = 0
|
|
|
|
for prop in properties:
|
|
|
|
char = CharacterInfo.query.filter(CharacterInfo.id == prop.owner_id).first()
|
|
|
|
if char.prop_clone_id != prop.clone_id:
|
|
|
|
count += 1
|
|
|
|
prop.clone_id = char.prop_clone_id
|
|
|
|
prop.save()
|
|
|
|
|
2022-04-14 02:29:08 +00:00
|
|
|
print(f"Fixed {count} props where clone id did not match owner's clone id")
|
|
|
|
|
|
|
|
dupes = 0
|
|
|
|
characters = CharacterInfo.query.all()
|
|
|
|
for char in characters:
|
|
|
|
props = Property.query.with_entities(
|
|
|
|
Property.zone_id, func.count(Property.zone_id)
|
|
|
|
).group_by(Property.zone_id).filter(
|
|
|
|
Property.owner_id == char.id
|
|
|
|
).all()
|
|
|
|
for prop in props:
|
|
|
|
if prop[1] != 1:
|
|
|
|
dupes += 1
|
|
|
|
print(f"found dupe on {char.name}'s {prop[0]}")
|
|
|
|
dupe_props = Property.query.filter(
|
|
|
|
Property.owner_id == char.id
|
|
|
|
).filter(
|
|
|
|
Property.zone_id == prop[0]).all()
|
|
|
|
dupe_data = []
|
|
|
|
# id, content_count
|
|
|
|
for dprop in dupe_props:
|
|
|
|
dupe_data.append(
|
|
|
|
[
|
|
|
|
dprop.id,
|
|
|
|
PropertyContent.query.filter(PropertyContent.property_id == dprop.id).count(),
|
|
|
|
dprop.time_claimed
|
|
|
|
]
|
|
|
|
)
|
|
|
|
max_models = max(dupe_data, key=lambda x: x[1])
|
|
|
|
if max_models[1] == 0:
|
|
|
|
newest = max(dupe_data, key=lambda x: x[2])
|
|
|
|
for data in dupe_data:
|
|
|
|
if data[2] != newest[2]:
|
|
|
|
Property.query.filter(Property.id == data[0]).first().delete()
|
|
|
|
else:
|
|
|
|
for data in dupe_data:
|
|
|
|
if data[1] != max_models[1]:
|
|
|
|
contents = PropertyContent.query.filter(PropertyContent.property_id == dprop.id).all()
|
|
|
|
if contents:
|
|
|
|
for content in contents:
|
|
|
|
if content.lot == 14:
|
|
|
|
UGC.query.filter(content.ugc_id).first().delete()
|
2022-04-14 03:29:57 +00:00
|
|
|
content.delete()
|
|
|
|
else:
|
|
|
|
Mail(
|
|
|
|
sender_id=0,
|
|
|
|
sender_name="System",
|
|
|
|
receiver_id=char.id,
|
|
|
|
receiver_name=char.name,
|
|
|
|
time_sent=time.time(),
|
|
|
|
subject="Returned Model",
|
|
|
|
body="This model was returned to you from a property cleanup script",
|
|
|
|
attachment_id=0,
|
|
|
|
attachment_lot=content.lot,
|
|
|
|
attachment_count=1
|
|
|
|
).save()
|
|
|
|
content.delete()
|
|
|
|
time.sleep(1)
|
2022-04-14 02:29:08 +00:00
|
|
|
Property.query.filter(Property.id == data[0]).first().delete()
|
2022-04-02 21:33:23 +00:00
|
|
|
return
|
|
|
|
|
|
|
|
|
2022-01-27 05:11:49 +00:00
|
|
|
@click.command("load_property")
|
|
|
|
@click.argument('zone')
|
|
|
|
@click.argument('player')
|
|
|
|
@with_appcontext
|
|
|
|
def load_property(zone, player):
|
2022-02-25 17:46:13 +00:00
|
|
|
"""shoves property data into db"""
|
2022-01-27 05:11:49 +00:00
|
|
|
char = CharacterInfo.query.filter(CharacterInfo.name == player).first()
|
|
|
|
if not char:
|
|
|
|
print("Character not Found")
|
|
|
|
return 404
|
|
|
|
|
2022-03-13 02:09:35 +00:00
|
|
|
prop = Property.query.filter(Property.owner_id == char.id).filter(Property.zone_id == zone).first()
|
2022-01-27 05:11:49 +00:00
|
|
|
|
|
|
|
if not prop:
|
|
|
|
print(f"Property {zone} not claimed by Character: {char.name}")
|
|
|
|
return 404
|
|
|
|
|
|
|
|
prop_files = pathlib.Path('property_files/')
|
|
|
|
for i in prop_files.glob('**/*'):
|
|
|
|
if i.suffix == '.lxfml':
|
|
|
|
lxfml = ""
|
|
|
|
with open(i, "r") as file:
|
|
|
|
lxfml = file.read()
|
|
|
|
compressed_lxfml = zlib.compress(lxfml.encode())
|
|
|
|
|
|
|
|
new_ugc = UGC(
|
|
|
|
account_id=char.account_id,
|
|
|
|
character_id=char.id,
|
|
|
|
is_optimized=0,
|
|
|
|
lxfml=compressed_lxfml,
|
|
|
|
bake_ao=0,
|
|
|
|
filename=i.name
|
|
|
|
)
|
|
|
|
new_ugc.save()
|
|
|
|
|
|
|
|
new_prop_content = PropertyContent(
|
|
|
|
id=i.stem,
|
|
|
|
property_id=prop.id,
|
|
|
|
ugc_id=new_ugc.id,
|
|
|
|
lot=14,
|
|
|
|
x=0,
|
|
|
|
y=0,
|
|
|
|
z=0,
|
|
|
|
rx=0,
|
|
|
|
ry=0,
|
|
|
|
rz=0,
|
|
|
|
rw=1
|
|
|
|
)
|
|
|
|
new_prop_content.save()
|
|
|
|
|
2022-03-13 02:09:35 +00:00
|
|
|
|
2022-02-10 18:01:51 +00:00
|
|
|
@click.command("gen_image_cache")
|
|
|
|
def gen_image_cache():
|
2022-02-25 17:46:13 +00:00
|
|
|
"""generates image cache"""
|
2022-02-10 18:01:51 +00:00
|
|
|
luclient = pathlib.Path('app/luclient/res')
|
|
|
|
files = [path for path in luclient.rglob("*.dds") if path.is_file()]
|
|
|
|
|
|
|
|
for file in files:
|
|
|
|
cache = get_cache_file(file).with_suffix(".png")
|
|
|
|
if not cache.exists():
|
|
|
|
try:
|
|
|
|
print(f"Convert {file.as_posix()} to {cache}")
|
|
|
|
cache.parent.mkdir(parents=True, exist_ok=True)
|
|
|
|
with image.Image(filename=str(file.as_posix())) as img:
|
|
|
|
img.compression = "no"
|
|
|
|
img.save(filename=str(cache.as_posix()))
|
|
|
|
except BE:
|
|
|
|
return print(f"Error on {file}")
|
|
|
|
|
2022-03-13 02:09:35 +00:00
|
|
|
|
2022-02-10 18:01:51 +00:00
|
|
|
@click.command("gen_model_cache")
|
|
|
|
def gen_model_cache():
|
2022-02-25 17:46:13 +00:00
|
|
|
"""generate model obj cache"""
|
2022-02-10 18:01:51 +00:00
|
|
|
luclient = pathlib.Path('app/luclient/res')
|
|
|
|
files = [path for path in luclient.rglob("*.lxfml") if path.is_file()]
|
|
|
|
pool = Pool(processes=4)
|
|
|
|
pool.map(partial(convert_lxfml_to_obj, lod=0), files)
|
|
|
|
pool.map(partial(convert_lxfml_to_obj, lod=1), files)
|
|
|
|
pool.map(partial(convert_lxfml_to_obj, lod=2), files)
|
|
|
|
|
2022-03-13 02:09:35 +00:00
|
|
|
|
2022-02-10 18:01:51 +00:00
|
|
|
def convert_lxfml_to_obj(file, lod):
|
|
|
|
mtl = get_cache_file(file).with_suffix(f".lod{lod}.mtl")
|
|
|
|
if not mtl.exists():
|
|
|
|
mtl.parent.mkdir(parents=True, exist_ok=True)
|
|
|
|
print(f"Convert LXFML {file.as_posix()} to obj and mtl @ {mtl}")
|
|
|
|
try:
|
2022-03-13 02:09:35 +00:00
|
|
|
ldd.main(str(file.as_posix()), str(mtl.with_suffix("").as_posix()), lod) # convert to OBJ
|
2022-02-10 18:01:51 +00:00
|
|
|
except Exception as e:
|
|
|
|
print(f"ERROR on {file}:\n {e}")
|
|
|
|
else:
|
|
|
|
# print(f"Already Exists: {file} with LOD {lod}")
|
|
|
|
return
|
|
|
|
|
2022-03-13 02:09:35 +00:00
|
|
|
|
2022-02-10 18:01:51 +00:00
|
|
|
def get_cache_file(path):
|
2022-02-25 17:46:13 +00:00
|
|
|
"""helper"""
|
2022-02-10 18:01:51 +00:00
|
|
|
# convert to list so that we can change elements
|
|
|
|
parts = list(path.parts)
|
|
|
|
|
|
|
|
# replace part that matches src with dst
|
|
|
|
parts[parts.index("luclient")] = "cache"
|
|
|
|
del parts[parts.index("res")]
|
|
|
|
|
|
|
|
return pathlib.Path(*parts)
|
2022-01-16 18:22:00 +00:00
|
|
|
|
2022-03-13 02:09:35 +00:00
|
|
|
|
2022-01-16 18:22:00 +00:00
|
|
|
def find_or_create_account(name, email, password, gm_level=9):
|
|
|
|
""" Find existing account or create new account """
|
|
|
|
account = Account.query.filter(Account.email == email).first()
|
|
|
|
if not account:
|
|
|
|
key = ""
|
|
|
|
for j in range(4):
|
|
|
|
key += ''.join(random.choice(string.ascii_uppercase + string.digits) for _ in range(4)) + '-'
|
|
|
|
# Remove last dash
|
|
|
|
key = key[:-1]
|
|
|
|
|
|
|
|
play_key = PlayKey(
|
|
|
|
key_string=key
|
|
|
|
)
|
|
|
|
db.session.add(play_key)
|
|
|
|
db.session.commit()
|
|
|
|
|
|
|
|
play_key = PlayKey.query.filter(PlayKey.key_string == key).first()
|
2022-03-13 02:09:35 +00:00
|
|
|
account = Account(
|
|
|
|
email=email,
|
|
|
|
username=name,
|
|
|
|
password=current_app.user_manager.password_manager.hash_password(password),
|
|
|
|
play_key_id=play_key.id,
|
|
|
|
email_confirmed_at=datetime.datetime.utcnow(),
|
|
|
|
gm_level=gm_level
|
|
|
|
)
|
2022-01-16 18:22:00 +00:00
|
|
|
play_key.key_uses = 0
|
|
|
|
db.session.add(account)
|
|
|
|
db.session.add(play_key)
|
|
|
|
db.session.commit()
|
2022-03-13 02:09:35 +00:00
|
|
|
return # account
|