aboutsummaryrefslogtreecommitdiff
path: root/scripts
diff options
context:
space:
mode:
authorFrederick Muriuki Muriithi2025-06-03 07:00:40 -0500
committerFrederick Muriuki Muriithi2025-06-03 07:29:31 -0500
commit70de586b590e824d7bcc3777830fda323a0e7983 (patch)
tree6a7bf06ff0e902da2e28bc43413d209d8a3aa09e /scripts
parent733d7e89612a7324856611030c0e8faa0b51538c (diff)
downloadgn-uploader-70de586b590e824d7bcc3777830fda323a0e7983.tar.gz
Memory saving: Don't reload saved data
We already have all the data in memory, so we do not need to reload it and keep yet another copy.
Diffstat (limited to 'scripts')
-rw-r--r--scripts/load_phenotypes_to_db.py36
1 files changed, 18 insertions, 18 deletions
diff --git a/scripts/load_phenotypes_to_db.py b/scripts/load_phenotypes_to_db.py
index fdfab4c..d3d9f5a 100644
--- a/scripts/load_phenotypes_to_db.py
+++ b/scripts/load_phenotypes_to_db.py
@@ -242,12 +242,12 @@ def load_data(conn: mysqldb.Connection, job: dict) -> int:
conn, _species["SpeciesId"], _population["Id"])}
# b. Save all the data items (DataIds are vibes), return new IDs
logger.info("Saving new phenotypes data.")
- data = save_pheno_data(conn=conn,
- dataidmap=dataidmap,
- pheno_name2id=pheno_name2id,
- samples=samples,
- control_data=_control_data,
- filesdir=_outdir)
+ _num_data_rows = save_pheno_data(conn=conn,
+ dataidmap=dataidmap,
+ pheno_name2id=pheno_name2id,
+ samples=samples,
+ control_data=_control_data,
+ filesdir=_outdir)
# 4. Cross-reference Phenotype, Publication, and PublishData in PublishXRef
logger.info("Cross-referencing new phenotypes to their data and publications.")
xrefs = cross_reference_phenotypes_publications_and_data(
@@ -255,19 +255,19 @@ def load_data(conn: mysqldb.Connection, job: dict) -> int:
# 5. If standard errors and N exist, save them too
# (use IDs returned in `3. b.` above).
logger.info("Saving new phenotypes standard errors.")
- _data_se = save_phenotypes_se(conn=conn,
- dataidmap=dataidmap,
- pheno_name2id=pheno_name2id,
- samples=samples,
- control_data=_control_data,
- filesdir=_outdir)
+ _num_se_rows = save_phenotypes_se(conn=conn,
+ dataidmap=dataidmap,
+ pheno_name2id=pheno_name2id,
+ samples=samples,
+ control_data=_control_data,
+ filesdir=_outdir)
logger.info("Saving new phenotypes sample counts.")
- _data_n = save_phenotypes_n(conn=conn,
- dataidmap=dataidmap,
- pheno_name2id=pheno_name2id,
- samples=samples,
- control_data=_control_data,
- filesdir=_outdir)
+ _num_n_rows = save_phenotypes_n(conn=conn,
+ dataidmap=dataidmap,
+ pheno_name2id=pheno_name2id,
+ samples=samples,
+ control_data=_control_data,
+ filesdir=_outdir)
# 6. If entirely new data, update authorisations (break this down)
logger.info("Updating authorisation.")
update_auth(_token, _species, _population, _dataset, _phenos)