-
Notifications
You must be signed in to change notification settings - Fork 1
Expand file tree
/
Copy pathshared.py
More file actions
163 lines (146 loc) · 5.48 KB
/
shared.py
File metadata and controls
163 lines (146 loc) · 5.48 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
import logging
from functools import lru_cache
from pathlib import Path
from typing import Dict, List, Optional
from sqlmodel import select
from werkzeug.utils import secure_filename
import murfey.server.prometheus as prom
from murfey.util import safe_run, sanitise
from murfey.util.config import MachineConfig, from_file, get_machine_config, settings
from murfey.util.db import (
DataCollection,
DataCollectionGroup,
FoilHole,
GridSquare,
ProcessingJob,
RsyncInstance,
Session,
)
logger = logging.getLogger("murfey.server.api.shared")
@lru_cache(maxsize=5)
def get_machine_config_for_instrument(instrument_name: str) -> Optional[MachineConfig]:
if settings.murfey_machine_configuration:
return from_file(Path(settings.murfey_machine_configuration), instrument_name)[
instrument_name
]
return None
def remove_session_by_id(session_id: int, db):
session = db.exec(select(Session).where(Session.id == session_id)).one()
sessions_for_visit = db.exec(
select(Session).where(Session.visit == session.visit)
).all()
# Don't remove prometheus metrics if there are other sessions using them
if len(sessions_for_visit) == 1:
safe_run(
prom.monitoring_switch.remove,
args=(session.visit,),
label="monitoring_switch",
)
rsync_instances = db.exec(
select(RsyncInstance).where(RsyncInstance.session_id == session_id)
).all()
for ri in rsync_instances:
safe_run(
prom.seen_files.remove,
args=(ri.source, session.visit),
label="seen_files",
)
safe_run(
prom.transferred_files.remove,
args=(ri.source, session.visit),
label="transferred_files",
)
safe_run(
prom.transferred_files_bytes.remove,
args=(ri.source, session.visit),
label="transferred_files_bytes",
)
safe_run(
prom.seen_data_files.remove,
args=(ri.source, session.visit),
label="seen_data_files",
)
safe_run(
prom.transferred_data_files.remove,
args=(ri.source, session.visit),
label="transferred_data_files",
)
safe_run(
prom.transferred_data_files_bytes.remove,
args=(ri.source, session.visit),
label="transferred_data_file_bytes",
)
safe_run(
prom.skipped_files.remove,
args=(ri.source, session.visit),
label="skipped_files",
)
collected_ids = db.exec(
select(DataCollectionGroup, DataCollection, ProcessingJob)
.where(DataCollectionGroup.session_id == session_id)
.where(DataCollection.dcg_id == DataCollectionGroup.id)
.where(ProcessingJob.dc_id == DataCollection.id)
).all()
for c in collected_ids:
safe_run(
prom.preprocessed_movies.remove,
args=(c[2].id,),
label="preprocessed_movies",
)
db.delete(session)
db.commit()
logger.debug(f"Successfully removed session {session_id} from database")
return
def get_grid_squares(session_id: int, db):
grid_squares = db.exec(
select(GridSquare).where(GridSquare.session_id == session_id)
).all()
tags = {gs.tag for gs in grid_squares}
res = {}
for t in tags:
res[t] = [gs for gs in grid_squares if gs.tag == t]
return res
def get_grid_squares_from_dcg(session_id: int, dcgid: int, db) -> List[GridSquare]:
grid_squares = db.exec(
select(GridSquare, DataCollectionGroup)
.where(GridSquare.session_id == session_id)
.where(GridSquare.tag == DataCollectionGroup.tag)
.where(DataCollectionGroup.id == dcgid)
).all()
return [gs[0] for gs in grid_squares]
def get_foil_holes_from_grid_square(
session_id: int, dcgid: int, gsid: int, db
) -> List[FoilHole]:
foil_holes = db.exec(
select(FoilHole, GridSquare, DataCollectionGroup)
.where(FoilHole.grid_square_id == GridSquare.id)
.where(GridSquare.name == gsid)
.where(GridSquare.session_id == session_id)
.where(GridSquare.tag == DataCollectionGroup.tag)
.where(DataCollectionGroup.id == dcgid)
).all()
return [fh[0] for fh in foil_holes]
def get_foil_hole(session_id: int, fh_name: int, db) -> Dict[str, int]:
foil_holes = db.exec(
select(FoilHole, GridSquare)
.where(FoilHole.name == fh_name)
.where(FoilHole.session_id == session_id)
.where(GridSquare.id == FoilHole.grid_square_id)
).all()
return {f[1].tag: f[0].id for f in foil_holes}
def get_upstream_tiff_dirs(visit_name: str, instrument_name: str) -> List[Path]:
tiff_dirs = []
machine_config = get_machine_config(instrument_name=instrument_name)[
instrument_name
]
for directory_name in machine_config.upstream_data_tiff_locations:
for p in machine_config.upstream_data_directories:
if (Path(p) / secure_filename(visit_name)).is_dir():
processed_dir = Path(p) / secure_filename(visit_name) / directory_name
tiff_dirs.append(processed_dir)
break
if not tiff_dirs:
logger.warning(
f"No candidate directory found for upstream download from visit {sanitise(visit_name)}"
)
return tiff_dirs