""" MSF Zernike Insight - Mid-Spatial Frequency Analysis for Telescope Mirrors Provides detailed analysis of mid-spatial frequency (MSF) content in mirror wavefront error, specifically for gravity-induced support print-through. Key Features: - Band decomposition: LSF (n≤10), MSF (n=11-50), HSF (n>50) - RSS (Root Sum Square) metrics per band - Relative analysis vs reference orientations (20°, 90°) - Higher-order Zernike fitting (100 modes by default) - PSD (Power Spectral Density) visualization - Support print-through identification For GigaBIT and similar telescope mirrors where MSF errors are dominated by gravity-induced support deformation rather than polishing residual. Author: Atomizer Framework """ from pathlib import Path from datetime import datetime from typing import Dict, Any, List, Optional, Tuple import numpy as np from math import factorial from numpy.linalg import LinAlgError from .base import StudyInsight, InsightConfig, InsightResult, register_insight # Lazy imports _plotly_loaded = False _go = None _make_subplots = None _Triangulation = None _OP2 = None _BDF = None def _load_dependencies(): """Lazy load heavy dependencies.""" global _plotly_loaded, _go, _make_subplots, _Triangulation, _OP2, _BDF if not _plotly_loaded: import plotly.graph_objects as go from plotly.subplots import make_subplots from matplotlib.tri import Triangulation from pyNastran.op2.op2 import OP2 from pyNastran.bdf.bdf import BDF _go = go _make_subplots = make_subplots _Triangulation = Triangulation _OP2 = OP2 _BDF = BDF _plotly_loaded = True # ============================================================================ # Band Definitions # ============================================================================ # Default band boundaries (Zernike radial order n) DEFAULT_LSF_MAX = 10 # LSF: n = 1-10 (M2 hexapod correctable) DEFAULT_MSF_MAX = 50 # MSF: n = 11-50 (support print-through) DEFAULT_N_MODES = 100 # Total modes to fit (n ≈ 14) # Band colors for visualization BAND_COLORS = { 'lsf': '#3b82f6', # Blue - low spatial frequency 'msf': '#f59e0b', # Amber - mid spatial frequency 'hsf': '#ef4444', # Red - high spatial frequency } # ============================================================================ # Zernike Mathematics (same as zernike_wfe.py) # ============================================================================ def noll_indices(j: int) -> Tuple[int, int]: """Convert Noll index to (n, m) radial/azimuthal orders.""" if j < 1: raise ValueError("Noll index j must be >= 1") count = 0 n = 0 while True: if n == 0: ms = [0] elif n % 2 == 0: ms = [0] + [m for k in range(1, n//2 + 1) for m in (-2*k, 2*k)] else: ms = [m for k in range(0, (n+1)//2) for m in (-(2*k+1), (2*k+1))] for m in ms: count += 1 if count == j: return n, m n += 1 def noll_to_radial_order(j: int) -> int: """Get radial order n for Noll index j.""" n, _ = noll_indices(j) return n def radial_order_to_first_noll(n: int) -> int: """Get first Noll index for radial order n.""" # Sum of (k+1) for k=0 to n-1, plus 1 return n * (n + 1) // 2 + 1 def zernike_noll(j: int, r: np.ndarray, th: np.ndarray) -> np.ndarray: """Evaluate Zernike polynomial j at (r, theta).""" n, m = noll_indices(j) R = np.zeros_like(r) for s in range((n - abs(m)) // 2 + 1): c = ((-1)**s * factorial(n - s) / (factorial(s) * factorial((n + abs(m)) // 2 - s) * factorial((n - abs(m)) // 2 - s))) R += c * r**(n - 2*s) if m == 0: return R return R * (np.cos(m * th) if m > 0 else np.sin(-m * th)) def zernike_common_name(n: int, m: int) -> str: """Get common name for Zernike mode.""" names = { (0, 0): "Piston", (1, -1): "Tilt X", (1, 1): "Tilt Y", (2, 0): "Defocus", (2, -2): "Astig 45°", (2, 2): "Astig 0°", (3, -1): "Coma X", (3, 1): "Coma Y", (3, -3): "Trefoil X", (3, 3): "Trefoil Y", (4, 0): "Primary Spherical", (4, -2): "Sec Astig X", (4, 2): "Sec Astig Y", (4, -4): "Quadrafoil X", (4, 4): "Quadrafoil Y", (5, -1): "Sec Coma X", (5, 1): "Sec Coma Y", (5, -3): "Sec Trefoil X", (5, 3): "Sec Trefoil Y", (5, -5): "Pentafoil X", (5, 5): "Pentafoil Y", (6, 0): "Sec Spherical", } return names.get((n, m), f"Z(n={n}, m={m})") def zernike_label(j: int) -> str: """Get label for Zernike coefficient J{j}.""" n, m = noll_indices(j) return f"J{j:02d} - {zernike_common_name(n, m)} (n={n}, m={m})" def compute_zernike_coeffs( X: np.ndarray, Y: np.ndarray, vals: np.ndarray, n_modes: int, chunk_size: int = 100000 ) -> Tuple[np.ndarray, float]: """Fit Zernike coefficients to WFE data.""" Xc, Yc = X - np.mean(X), Y - np.mean(Y) R = float(np.max(np.hypot(Xc, Yc))) r = np.hypot(Xc / R, Yc / R).astype(np.float32) th = np.arctan2(Yc, Xc).astype(np.float32) mask = (r <= 1.0) & ~np.isnan(vals) if not np.any(mask): raise RuntimeError("No valid points inside unit disk.") idx = np.nonzero(mask)[0] m = int(n_modes) G = np.zeros((m, m), dtype=np.float64) h = np.zeros((m,), dtype=np.float64) v = vals.astype(np.float64) for start in range(0, len(idx), chunk_size): sl = idx[start:start + chunk_size] r_b, th_b, v_b = r[sl], th[sl], v[sl] Zb = np.column_stack([zernike_noll(j, r_b, th_b).astype(np.float32) for j in range(1, m + 1)]) G += (Zb.T @ Zb).astype(np.float64) h += (Zb.T @ v_b).astype(np.float64) try: coeffs = np.linalg.solve(G, h) except LinAlgError: coeffs = np.linalg.lstsq(G, h, rcond=None)[0] return coeffs, R # ============================================================================ # Band Analysis # ============================================================================ def classify_modes_by_band( n_modes: int, lsf_max: int = DEFAULT_LSF_MAX, msf_max: int = DEFAULT_MSF_MAX ) -> Dict[str, List[int]]: """ Classify Zernike modes into LSF/MSF/HSF bands. Returns: Dict with 'lsf', 'msf', 'hsf' keys containing lists of Noll indices """ bands = {'lsf': [], 'msf': [], 'hsf': []} for j in range(1, n_modes + 1): n = noll_to_radial_order(j) if n <= lsf_max: bands['lsf'].append(j) elif n <= msf_max: bands['msf'].append(j) else: bands['hsf'].append(j) return bands def compute_band_rss( coeffs: np.ndarray, bands: Dict[str, List[int]], filter_modes: int = 4 ) -> Dict[str, float]: """ Compute RSS (Root Sum Square) of coefficients in each band. RSS = sqrt(sum(c_j^2)) for j in band Args: coeffs: Zernike coefficients bands: Dict with band name -> list of Noll indices filter_modes: Number of low-order modes to filter for 'filtered' metric (default 4 = J1-J4) Returns: Dict with 'lsf', 'msf', 'hsf', 'total', and 'filtered' (J5+) RSS values """ rss = {} for band_name, indices in bands.items(): # Convert to 0-indexed band_coeffs = [coeffs[j-1] for j in indices if j-1 < len(coeffs)] rss[band_name] = float(np.sqrt(np.sum(np.array(band_coeffs)**2))) rss['total'] = float(np.sqrt(np.sum(coeffs**2))) # Filtered RSS: exclude first filter_modes (J1-J4 = piston, tip, tilt, defocus) # This is the engineering-relevant metric after M2 alignment filtered_coeffs = coeffs[filter_modes:] if len(coeffs) > filter_modes else np.array([]) rss['filtered'] = float(np.sqrt(np.sum(filtered_coeffs**2))) return rss def compute_band_residual( X: np.ndarray, Y: np.ndarray, W_nm: np.ndarray, coeffs: np.ndarray, R: float, bands: Dict[str, List[int]], keep_bands: List[str] ) -> np.ndarray: """ Compute WFE residual keeping only specified bands. Args: keep_bands: List of band names to keep (e.g., ['msf'] for MSF-only) Returns: WFE array with only specified band content """ Xc = X - np.mean(X) Yc = Y - np.mean(Y) r = np.hypot(Xc / R, Yc / R) th = np.arctan2(Yc, Xc) # Build Zernike basis for modes to keep keep_indices = [] for band in keep_bands: keep_indices.extend(bands.get(band, [])) if not keep_indices: return np.zeros_like(W_nm) # Reconstruct only the kept bands W_band = np.zeros_like(W_nm) for j in keep_indices: if j - 1 < len(coeffs): Z_j = zernike_noll(j, r, th) W_band += coeffs[j-1] * Z_j return W_band def find_dominant_msf_modes( coeffs: np.ndarray, bands: Dict[str, List[int]], top_n: int = 5 ) -> List[Dict[str, Any]]: """ Find the dominant MSF modes by coefficient magnitude. Returns: List of dicts with 'j', 'label', 'coeff_nm', 'n', 'm' """ msf_indices = bands.get('msf', []) mode_data = [] for j in msf_indices: if j - 1 < len(coeffs): n, m = noll_indices(j) mode_data.append({ 'j': j, 'label': zernike_label(j), 'coeff_nm': float(abs(coeffs[j-1])), 'n': n, 'm': m }) # Sort by magnitude mode_data.sort(key=lambda x: x['coeff_nm'], reverse=True) return mode_data[:top_n] def estimate_mesh_resolution(X: np.ndarray, Y: np.ndarray) -> Dict[str, float]: """ Estimate mesh spatial resolution. Returns: Dict with 'node_count', 'diameter_mm', 'avg_spacing_mm', 'max_resolvable_order', 'min_feature_mm' """ n_nodes = len(X) Xc = X - np.mean(X) Yc = Y - np.mean(Y) R = np.max(np.hypot(Xc, Yc)) diameter = 2 * R # Approximate spacing from area area = np.pi * R**2 avg_spacing = np.sqrt(area / n_nodes) # Nyquist: can resolve features > 2x spacing min_feature = 2 * avg_spacing # Max Zernike order max_order = int(diameter / min_feature) return { 'node_count': n_nodes, 'diameter_mm': float(diameter), 'avg_spacing_mm': float(avg_spacing), 'min_feature_mm': float(min_feature), 'max_resolvable_order': max_order } # ============================================================================ # Default Configuration # ============================================================================ DEFAULT_CONFIG = { 'n_modes': 100, 'lsf_max': 10, # n ≤ 10 is LSF 'msf_max': 50, # n = 11-50 is MSF 'amp': 0.5, # Visual deformation scale 'pancake': 3.0, # Z-axis range multiplier 'plot_downsample': 15000, 'colorscale': 'Turbo', 'disp_unit': 'mm', 'show_psd': True, 'show_all_orientations': True, } @register_insight class MSFZernikeInsight(StudyInsight): """ Mid-Spatial Frequency Zernike analysis for telescope mirror optimization. Provides detailed breakdown of spatial frequency content: - LSF (Low): n ≤ 10, correctable by M2 hexapod - MSF (Mid): n = 11-50, support print-through - HSF (High): n > 50, near mesh resolution limit Generates: - Band decomposition table with RSS metrics - MSF-only 3D surface visualization - Coefficient bar chart color-coded by band - Dominant MSF mode identification - Mesh resolution analysis """ insight_type = "msf_zernike" name = "MSF Zernike Analysis" description = "Mid-spatial frequency analysis with band decomposition for support print-through" category = "optical" applicable_to = ["mirror", "optics", "wfe", "msf"] required_files = ["*.op2"] def __init__(self, study_path: Path): super().__init__(study_path) self.op2_path: Optional[Path] = None self.geo_path: Optional[Path] = None self._node_geo: Optional[Dict] = None self._displacements: Optional[Dict] = None def can_generate(self) -> bool: """Check if OP2 and geometry files exist.""" search_paths = [ self.results_path, self.study_path / "2_iterations", self.setup_path / "model", ] for search_path in search_paths: if not search_path.exists(): continue op2_files = list(search_path.glob("**/*solution*.op2")) if not op2_files: op2_files = list(search_path.glob("**/*.op2")) if op2_files: self.op2_path = max(op2_files, key=lambda p: p.stat().st_mtime) break if self.op2_path is None: return False try: self.geo_path = self._find_geometry_file(self.op2_path) return True except FileNotFoundError: return False def _find_geometry_file(self, op2_path: Path) -> Path: """Find BDF/DAT geometry file for OP2.""" folder = op2_path.parent base = op2_path.stem for ext in ['.dat', '.bdf']: cand = folder / (base + ext) if cand.exists(): return cand for f in folder.iterdir(): if f.suffix.lower() in ['.dat', '.bdf']: return f raise FileNotFoundError(f"No geometry file found for {op2_path}") def _load_data(self): """Load geometry and displacement data.""" if self._node_geo is not None: return _load_dependencies() bdf = _BDF() bdf.read_bdf(str(self.geo_path)) self._node_geo = {int(nid): node.get_position() for nid, node in bdf.nodes.items()} op2 = _OP2() op2.read_op2(str(self.op2_path)) if not op2.displacements: raise RuntimeError("No displacement data in OP2") self._displacements = {} for key, darr in op2.displacements.items(): data = darr.data dmat = data[0] if data.ndim == 3 else (data if data.ndim == 2 else None) if dmat is None: continue ngt = darr.node_gridtype.astype(int) node_ids = ngt if ngt.ndim == 1 else ngt[:, 0] isubcase = getattr(darr, 'isubcase', None) label = str(isubcase) if isubcase else str(key) self._displacements[label] = { 'node_ids': node_ids.astype(int), 'disp': dmat.copy() } def _build_wfe_arrays( self, label: str, disp_unit: str = 'mm' ) -> Tuple[np.ndarray, np.ndarray, np.ndarray, np.ndarray]: """Build X, Y, WFE arrays for a subcase.""" nm_per_unit = 1e6 if disp_unit == 'mm' else 1e9 data = self._displacements[label] node_ids = data['node_ids'] dmat = data['disp'] X, Y, WFE = [], [], [] valid_nids = [] for nid, vec in zip(node_ids, dmat): geo = self._node_geo.get(int(nid)) if geo is None: continue X.append(geo[0]) Y.append(geo[1]) wfe = vec[2] * 2.0 * nm_per_unit WFE.append(wfe) valid_nids.append(nid) return (np.array(X), np.array(Y), np.array(WFE), np.array(valid_nids)) def _compute_relative_wfe( self, X1, Y1, WFE1, nids1, X2, Y2, WFE2, nids2 ) -> Tuple[np.ndarray, np.ndarray, np.ndarray]: """Compute WFE1 - WFE2 for common nodes.""" ref_map = {int(nid): (x, y, w) for nid, x, y, w in zip(nids2, X2, Y2, WFE2)} X_rel, Y_rel, WFE_rel = [], [], [] for nid, x, y, w in zip(nids1, X1, Y1, WFE1): nid = int(nid) if nid in ref_map: _, _, w_ref = ref_map[nid] X_rel.append(x) Y_rel.append(y) WFE_rel.append(w - w_ref) return np.array(X_rel), np.array(Y_rel), np.array(WFE_rel) def _analyze_subcase( self, X: np.ndarray, Y: np.ndarray, W_nm: np.ndarray, cfg: Dict ) -> Dict[str, Any]: """ Full MSF analysis for a single subcase. Returns: Dict with coefficients, bands, RSS, dominant modes, mesh info """ n_modes = cfg['n_modes'] lsf_max = cfg['lsf_max'] msf_max = cfg['msf_max'] # Fit Zernike coeffs, R = compute_zernike_coeffs(X, Y, W_nm, n_modes) # Band classification bands = classify_modes_by_band(n_modes, lsf_max, msf_max) # RSS per band rss = compute_band_rss(coeffs, bands) # Band percentages total_rss = rss['total'] pct = { 'lsf': 100 * rss['lsf'] / total_rss if total_rss > 0 else 0, 'msf': 100 * rss['msf'] / total_rss if total_rss > 0 else 0, 'hsf': 100 * rss['hsf'] / total_rss if total_rss > 0 else 0, } # Dominant MSF modes dominant_msf = find_dominant_msf_modes(coeffs, bands, top_n=5) # MSF-only residual W_msf = compute_band_residual(X, Y, W_nm, coeffs, R, bands, ['msf']) # Mesh resolution mesh_info = estimate_mesh_resolution(X, Y) return { 'coefficients': coeffs, 'R': R, 'bands': bands, 'rss': rss, 'rss_pct': pct, 'dominant_msf': dominant_msf, 'W_msf': W_msf, 'mesh_info': mesh_info, 'n_modes': n_modes, 'lsf_max': lsf_max, 'msf_max': msf_max, } def _generate_html( self, analyses: Dict[str, Dict], cfg: Dict, X: np.ndarray, Y: np.ndarray ) -> str: """Generate comprehensive HTML report.""" _load_dependencies() n_modes = cfg['n_modes'] amp = cfg.get('amp', 0.5) pancake = cfg.get('pancake', 3.0) downsample = cfg.get('plot_downsample', 15000) colorscale = cfg.get('colorscale', 'Turbo') # Use first available analysis for visualization primary_key = list(analyses.keys())[0] primary = analyses[primary_key] coeffs = primary['coefficients'] bands = primary['bands'] W_msf = primary['W_msf'] mesh_info = primary['mesh_info'] # Prepare band-colored bar chart data bar_colors = [] for j in range(1, n_modes + 1): if j in bands['lsf']: bar_colors.append(BAND_COLORS['lsf']) elif j in bands['msf']: bar_colors.append(BAND_COLORS['msf']) else: bar_colors.append(BAND_COLORS['hsf']) labels = [zernike_label(j) for j in range(1, n_modes + 1)] coeff_abs = np.abs(coeffs) # Downsample for 3D plot n = len(X) if n > downsample: rng = np.random.default_rng(42) sel = rng.choice(n, size=downsample, replace=False) Xp, Yp, Wp = X[sel], Y[sel], W_msf[sel] else: Xp, Yp, Wp = X, Y, W_msf res_amp = amp * Wp max_amp = float(np.max(np.abs(res_amp))) if res_amp.size else 1.0 # Create figure with subplots fig = _make_subplots( rows=4, cols=1, specs=[[{"type": "scene"}], [{"type": "table"}], [{"type": "table"}], [{"type": "xy"}]], row_heights=[0.40, 0.20, 0.15, 0.25], vertical_spacing=0.03, subplot_titles=[ "MSF Content Only (n=11-50)", "Band Decomposition (RSS)", "Dominant MSF Modes", "Coefficient Magnitude by Band" ] ) # 3D MSF surface try: tri = _Triangulation(Xp, Yp) if tri.triangles is not None and len(tri.triangles) > 0: i, j, k = tri.triangles.T fig.add_trace(_go.Mesh3d( x=Xp, y=Yp, z=res_amp, i=i, j=j, k=k, intensity=res_amp, colorscale=colorscale, opacity=1.0, flatshading=False, lighting=dict(ambient=0.4, diffuse=0.8, specular=0.3, roughness=0.5, fresnel=0.2), lightposition=dict(x=100, y=200, z=300), showscale=True, colorbar=dict(title=dict(text="MSF WFE (nm)", side="right"), thickness=15, len=0.5, tickformat=".1f"), hovertemplate="X: %{x:.1f}
Y: %{y:.1f}
MSF: %{z:.2f} nm" ), row=1, col=1) except Exception: fig.add_trace(_go.Scatter3d( x=Xp, y=Yp, z=res_amp, mode='markers', marker=dict(size=2, color=res_amp, colorscale=colorscale, showscale=True), showlegend=False ), row=1, col=1) # Configure 3D scene fig.update_scenes( camera=dict(eye=dict(x=1.2, y=1.2, z=0.8), up=dict(x=0, y=0, z=1)), xaxis=dict(title="X (mm)", showgrid=True, gridcolor='rgba(128,128,128,0.3)', showbackground=True, backgroundcolor='rgba(240,240,240,0.9)'), yaxis=dict(title="Y (mm)", showgrid=True, gridcolor='rgba(128,128,128,0.3)', showbackground=True, backgroundcolor='rgba(240,240,240,0.9)'), zaxis=dict(title="MSF WFE (nm)", range=[-max_amp * pancake, max_amp * pancake], showgrid=True, gridcolor='rgba(128,128,128,0.3)', showbackground=True, backgroundcolor='rgba(230,230,250,0.9)'), aspectmode='manual', aspectratio=dict(x=1, y=1, z=0.4) ) # Band decomposition table band_headers = ["Band", "Order Range", "Feature Size"] band_values = [ ["LSF (Low)", "MSF (Mid)", "HSF (High)", "Total", "Filtered (J5+)"], [f"n ≤ {cfg['lsf_max']}", f"n = {cfg['lsf_max']+1}-{cfg['msf_max']}", f"n > {cfg['msf_max']}", "All", "J5+ (excl. piston/tip/tilt/defocus)"], [f"> {int(mesh_info['diameter_mm']/cfg['lsf_max'])} mm", f"{int(mesh_info['diameter_mm']/cfg['msf_max'])}-{int(mesh_info['diameter_mm']/(cfg['lsf_max']+1))} mm", f"< {int(mesh_info['diameter_mm']/cfg['msf_max'])} mm", "-", "M2 correctable removed"], ] # Add columns for each analysis for key, analysis in analyses.items(): band_headers.append(f"{key} RSS (nm)") rss = analysis['rss'] pct = analysis['rss_pct'] # Calculate filtered percentage filtered_pct = 100 * rss['filtered'] / rss['total'] if rss['total'] > 0 else 0 band_values.append([ f"{rss['lsf']:.2f} ({pct['lsf']:.1f}%)", f"{rss['msf']:.2f} ({pct['msf']:.1f}%)", f"{rss['hsf']:.2f} ({pct['hsf']:.1f}%)", f"{rss['total']:.2f}", f"{rss['filtered']:.2f} ({filtered_pct:.1f}%)", ]) fig.add_trace(_go.Table( header=dict(values=band_headers, align="left", fill_color='#1f2937', font=dict(color='white')), cells=dict(values=band_values, align="left", fill_color='#374151', font=dict(color='white')) ), row=2, col=1) # Dominant MSF modes table dom_modes = primary['dominant_msf'] if dom_modes: fig.add_trace(_go.Table( header=dict(values=["Rank", "Mode", "|Coeff| (nm)", "Order n"], align="left", fill_color='#1f2937', font=dict(color='white')), cells=dict(values=[ [f"#{i+1}" for i in range(len(dom_modes))], [m['label'] for m in dom_modes], [f"{m['coeff_nm']:.3f}" for m in dom_modes], [str(m['n']) for m in dom_modes], ], align="left", fill_color='#374151', font=dict(color='white')) ), row=3, col=1) # Bar chart with band colors fig.add_trace( _go.Bar( x=coeff_abs.tolist(), y=labels, orientation='h', marker_color=bar_colors, hovertemplate="%{y}
|Coeff| = %{x:.3f} nm", showlegend=False ), row=4, col=1 ) # Add legend for bands for band_name, color in BAND_COLORS.items(): fig.add_trace(_go.Scatter( x=[None], y=[None], mode='markers', marker=dict(size=10, color=color), name=band_name.upper(), showlegend=True )) # Layout fig.update_layout( width=1400, height=1600, margin=dict(t=60, b=20, l=20, r=20), paper_bgcolor='#111827', plot_bgcolor='#1f2937', font=dict(color='white'), title=dict( text=f"Atomizer MSF Analysis | {n_modes} modes | Mesh: {mesh_info['node_count']} nodes, {mesh_info['avg_spacing_mm']:.1f}mm spacing", x=0.5, font=dict(size=16) ), legend=dict( orientation="h", yanchor="bottom", y=1.02, xanchor="right", x=1 ) ) return fig.to_html(include_plotlyjs='cdn', full_html=True) def _generate(self, config: InsightConfig) -> InsightResult: """Generate MSF analysis.""" self._load_data() cfg = {**DEFAULT_CONFIG, **config.extra} disp_unit = cfg['disp_unit'] # Map subcases disps = self._displacements if '1' in disps and '2' in disps: sc_map = {'90°': '1', '20°': '2', '40°': '3', '60°': '4'} elif '90' in disps and '20' in disps: sc_map = {'90°': '90', '20°': '20', '40°': '40', '60°': '60'} else: available = sorted(disps.keys(), key=lambda x: int(x) if x.isdigit() else 0) if len(available) >= 4: sc_map = {'90°': available[0], '20°': available[1], '40°': available[2], '60°': available[3]} else: return InsightResult(success=False, error=f"Need 4 subcases, found: {available}") # Validate subcases for angle, label in sc_map.items(): if label not in disps: return InsightResult(success=False, error=f"Subcase '{label}' (angle {angle}) not found") # Load reference data X_20, Y_20, WFE_20, nids_20 = self._build_wfe_arrays(sc_map['20°'], disp_unit) X_90, Y_90, WFE_90, nids_90 = self._build_wfe_arrays(sc_map['90°'], disp_unit) analyses = {} # Analyze each key orientation for angle in ['40°', '60°']: label = sc_map[angle] X, Y, WFE, nids = self._build_wfe_arrays(label, disp_unit) # Absolute analyses[f"{angle} Abs"] = self._analyze_subcase(X, Y, WFE, cfg) # Relative to 20° X_rel, Y_rel, WFE_rel = self._compute_relative_wfe( X, Y, WFE, nids, X_20, Y_20, WFE_20, nids_20) analyses[f"{angle} vs 20°"] = self._analyze_subcase(X_rel, Y_rel, WFE_rel, cfg) # Relative to 90° (manufacturing) X_rel90, Y_rel90, WFE_rel90 = self._compute_relative_wfe( X, Y, WFE, nids, X_90, Y_90, WFE_90, nids_90) analyses[f"{angle} vs 90°"] = self._analyze_subcase(X_rel90, Y_rel90, WFE_rel90, cfg) # Also analyze 90° absolute (manufacturing baseline) analyses["90° Mfg"] = self._analyze_subcase(X_90, Y_90, WFE_90, cfg) # Generate HTML using 40° vs 20° as primary visualization primary_analysis = analyses["40° vs 20°"] X_40, Y_40, _, _ = self._build_wfe_arrays(sc_map['40°'], disp_unit) X_rel40, Y_rel40, _ = self._compute_relative_wfe( X_40, Y_40, analyses["40° Abs"]['W_msf'], np.arange(len(X_40)), X_20, Y_20, WFE_20, nids_20) html = self._generate_html(analyses, cfg, X_40, Y_40) timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") output_dir = config.output_dir or self.insights_path output_dir.mkdir(parents=True, exist_ok=True) html_path = output_dir / f"msf_zernike_{timestamp}.html" html_path.write_text(html, encoding='utf-8') # Build summary summary = { 'n_modes': cfg['n_modes'], 'lsf_max_order': cfg['lsf_max'], 'msf_max_order': cfg['msf_max'], 'mesh_nodes': primary_analysis['mesh_info']['node_count'], 'mesh_spacing_mm': primary_analysis['mesh_info']['avg_spacing_mm'], 'max_resolvable_order': primary_analysis['mesh_info']['max_resolvable_order'], } # Add key metrics for each analysis for key, analysis in analyses.items(): safe_key = key.replace('°', 'deg').replace(' ', '_') summary[f'{safe_key}_lsf_rss'] = analysis['rss']['lsf'] summary[f'{safe_key}_msf_rss'] = analysis['rss']['msf'] summary[f'{safe_key}_hsf_rss'] = analysis['rss']['hsf'] summary[f'{safe_key}_total_rss'] = analysis['rss']['total'] summary[f'{safe_key}_filtered_rss'] = analysis['rss']['filtered'] summary[f'{safe_key}_msf_pct'] = analysis['rss_pct']['msf'] return InsightResult( success=True, html_path=html_path, summary=summary )