|
15 | 15 | _get_energy_bins, |
16 | 16 | _get_pixel_mask, |
17 | 17 | _get_sub_spectrum_mask, |
18 | | - get_min_uint, |
19 | 18 | rebin_proportional, |
20 | 19 | unscale_triggers, |
21 | 20 | ) |
@@ -203,12 +202,12 @@ def from_levelb(cls, levelb, parent="", keep_parse_tree=True): |
203 | 202 | data["time"] = time |
204 | 203 | data["timedel"] = duration |
205 | 204 | data.add_meta(name="timedel", nix="NIX00405", packets=packets) |
206 | | - data["triggers"] = triggers.astype(get_min_uint(triggers)) |
| 205 | + data["triggers"] = triggers.astype(np.min_scalar_type(triggers)) |
207 | 206 | data.add_meta(name="triggers", nix="NIX00274", packets=packets) |
208 | 207 | data["triggers_comp_err"] = np.float32(np.sqrt(triggers_var)) |
209 | 208 | data["rcr"] = np.hstack(packets.get_value("NIX00276")).flatten().astype(np.ubyte) |
210 | 209 | data.add_meta(name="rcr", nix="NIX00276", packets=packets) |
211 | | - data["counts"] = (counts.T * u.ct).astype(get_min_uint(counts)) |
| 210 | + data["counts"] = (counts.T * u.ct).astype(np.min_scalar_type(counts)) |
212 | 211 | data.add_meta(name="counts", nix="NIX00272", packets=packets) |
213 | 212 | data["counts_comp_err"] = np.float32(np.sqrt(counts_var).T * u.ct) |
214 | 213 |
|
@@ -316,10 +315,10 @@ def from_levelb(cls, levelb, parent="", keep_parse_tree=True): |
316 | 315 | data["time"] = time |
317 | 316 | data["timedel"] = duration |
318 | 317 | data.add_meta(name="timedel", nix="NIX00405", packets=packets) |
319 | | - data["triggers"] = triggers.astype(get_min_uint(triggers)) |
| 318 | + data["triggers"] = triggers.astype(np.min_scalar_type(triggers)) |
320 | 319 | data.add_meta(name="triggers", nix="NIX00274", packets=packets) |
321 | 320 | data["triggers_comp_err"] = np.float32(np.sqrt(triggers_var)) |
322 | | - data["counts"] = (counts.T * u.ct).astype(get_min_uint(counts)) |
| 321 | + data["counts"] = (counts.T * u.ct).astype(np.min_scalar_type(counts)) |
323 | 322 | data.add_meta(name="counts", nix="NIX00278", packets=packets) |
324 | 323 | data["counts_comp_err"] = np.float32(np.sqrt(counts_var).T * u.ct) |
325 | 324 |
|
@@ -438,10 +437,10 @@ def from_levelb(cls, levelb, parent="", keep_parse_tree=True): |
438 | 437 | data.add_meta(name="timedel", nix="NIX00405", packets=packets) |
439 | 438 | data["detector_index"] = detector_index.reshape(-1, 32).astype(np.ubyte) |
440 | 439 | data.add_meta(name="detector_index", nix="NIX00100", packets=packets) |
441 | | - data["spectra"] = (counts.reshape(-1, 32, num_energies) * u.ct).astype(get_min_uint(counts)) |
| 440 | + data["spectra"] = (counts.reshape(-1, 32, num_energies) * u.ct).astype(np.min_scalar_type(counts)) |
442 | 441 | data["spectra"].meta = {"NIXS": "NIX00452", "PCF_CURTX": packets.get("NIX00452")[0].idb_info.PCF_CURTX} |
443 | 442 | data["spectra_comp_err"] = np.float32(np.sqrt(counts_var.reshape(-1, 32, num_energies))) |
444 | | - data["triggers"] = triggers.reshape(-1, num_energies).astype(get_min_uint(triggers)) |
| 443 | + data["triggers"] = triggers.reshape(-1, num_energies).astype(np.min_scalar_type(triggers)) |
445 | 444 | data.add_meta(name="triggers", nix="NIX00484", packets=packets) |
446 | 445 | data["triggers_comp_err"] = np.float32(np.sqrt(triggers_var.reshape(-1, num_energies))) |
447 | 446 | data["num_integrations"] = num_integrations.reshape(-1, num_energies).astype(np.ubyte)[:, 0] |
@@ -556,7 +555,7 @@ def from_levelb(cls, levelb, parent="", keep_parse_tree=True): |
556 | 555 | data["timedel"] = duration |
557 | 556 | data.add_meta(name="timedel", nix="NIX00405", packets=packets) |
558 | 557 | data["control_index"] = control_indices |
559 | | - data["variance"] = variance.astype(get_min_uint(variance)) |
| 558 | + data["variance"] = variance.astype(np.min_scalar_type(variance)) |
560 | 559 | data.add_meta(name="variance", nix="NIX00281", packets=packets) |
561 | 560 | data["variance_comp_err"] = np.float32(np.sqrt(variance_var)) |
562 | 561 |
|
@@ -801,7 +800,7 @@ def from_levelb(cls, levelb, parent="", keep_parse_tree=True): |
801 | 800 | data["timedel"] = duration[unique_time_indices] |
802 | 801 | data.add_meta(name="timedel", nix="NIX00122", packets=packets) |
803 | 802 |
|
804 | | - data["counts"] = (full_counts * u.ct).astype(get_min_uint(full_counts)) |
| 803 | + data["counts"] = (full_counts * u.ct).astype(np.min_scalar_type(full_counts)) |
805 | 804 | data.add_meta(name="counts", nix="NIX00158", packets=packets) |
806 | 805 | data["counts_comp_err"] = (np.sqrt(full_counts_var) * u.ct).astype(np.float32) |
807 | 806 | data["control_index"] = np.arange(len(control)).astype(np.uint16) |
@@ -849,7 +848,7 @@ def from_levelb(cls, levelb, parent="", keep_parse_tree=True): |
849 | 848 | control = Control() |
850 | 849 | control["scet_coarse"] = packets.get("scet_coarse") |
851 | 850 | control["scet_fine"] = packets.get("scet_fine") |
852 | | - control["index"] = np.arange(len(control)).astype(get_min_uint(len(control))) |
| 851 | + control["index"] = np.arange(len(control)).astype(np.min_scalar_type(len(control))) |
853 | 852 |
|
854 | 853 | # When the packets are parsed empty packets are dropped but in LB we don't parse so this |
855 | 854 | # is not known need to compare control and levelb.control and only use matching rows |
|
0 commit comments