aboutsummaryrefslogtreecommitdiff
path: root/wqflask
diff options
context:
space:
mode:
Diffstat (limited to 'wqflask')
-rw-r--r--wqflask/base/data_set.py31
1 files changed, 27 insertions, 4 deletions
diff --git a/wqflask/base/data_set.py b/wqflask/base/data_set.py
index 8906ab69..8188d90c 100644
--- a/wqflask/base/data_set.py
+++ b/wqflask/base/data_set.py
@@ -20,7 +20,7 @@
from dataclasses import dataclass
from dataclasses import field
from dataclasses import InitVar
-from typing import Optional, Dict
+from typing import Optional, Dict, List
from db.call import fetchall, fetchone, fetch1
from utility.logger import getLogger
from utility.tools import USE_GN_SERVER, USE_REDIS, flat_files, flat_file_exists, GN2_BASE_URL
@@ -397,7 +397,8 @@ class DatasetGroup:
self.parlist = [maternal, paternal]
def get_study_samplelists(self):
- study_sample_file = locate_ignore_error(self.name + ".json", 'study_sample_lists')
+ study_sample_file = locate_ignore_error(
+ self.name + ".json", 'study_sample_lists')
try:
f = open(study_sample_file)
except:
@@ -447,7 +448,6 @@ class DatasetGroup:
# genotype_1 is Dataset Object without parents and f1
# genotype_2 is Dataset Object with parents and f1 (not for intercross)
-
# reaper barfs on unicode filenames, so here we ensure it's a string
if self.genofile:
if "RData" in self.genofile: # ZS: This is a temporary fix; I need to change the way the JSON files that point to multiple genotype files are structured to point to other file types like RData
@@ -726,7 +726,6 @@ class DataSet:
data_results = self.chunk_dataset(query_results, len(sample_ids))
self.samplelist = sorted_samplelist
self.trait_data = data_results
-
def get_trait_data(self, sample_list=None):
if sample_list:
@@ -1242,3 +1241,27 @@ def geno_mrna_confidentiality(ob):
if confidential:
return True
+
+
+def check_if_dataset_modified(dataset_name, cached_timestamp):
+ """function to check if the dataset has been modified"""
+ last_modified = "query results"
+ return (cached_timestamp == last_modified_timestamp)
+
+
+def cache_dataset_results(dataset_name: str, query_results: List):
+ """function to cache dataset query results to file"""
+ # check if file exists clear if it does
+ # aslo check for the timestamp
+ # hash for unique name ??? are dataset name unique
+ # data computations actions
+ # store the file path on redis
+
+ # hash functiob
+
+ file_path = os.path.join(TMPDIR, f"{dataset_name}.json")
+
+ query_results = [list(results) for result in query_results]
+
+ with open(file_path, "w") as file_handler:
+ json.dump(query_results, file_handler) \ No newline at end of file