|
|
#!/usr/bin/env python3
|
|
|
# -*- coding: utf-8 -*-
|
|
|
"""
|
|
|
LLM 结果处理脚本 - 合并版本
|
|
|
从 LLM 生成的 Markdown 结果中提取 templates,生成并验证 Alpha 表达式
|
|
|
|
|
|
使用方法:
|
|
|
将 llm_result.md 和 dataset.csv 放在当前目录,然后运行:
|
|
|
python decode_template.py
|
|
|
|
|
|
输出:
|
|
|
在当前目录创建 output/ 文件夹,包含生成的表达式
|
|
|
"""
|
|
|
|
|
|
import json
|
|
|
import itertools
|
|
|
import re
|
|
|
import sys
|
|
|
from pathlib import Path
|
|
|
from typing import List, Dict, Any, Optional, Tuple
|
|
|
|
|
|
|
|
|
# ==================== 表达式处理工具函数 ====================
|
|
|
|
|
|
def detect_dataset_code(dataset_ids: List[str]) -> Optional[str]:
|
|
|
"""从字段ID中检测数据集代码前缀"""
|
|
|
if not dataset_ids:
|
|
|
return None
|
|
|
counts: Dict[str, int] = {}
|
|
|
for fid in dataset_ids:
|
|
|
tok = (str(fid).split("_", 1)[0] or "").strip()
|
|
|
if tok:
|
|
|
counts[tok] = counts.get(tok, 0) + 1
|
|
|
if not counts:
|
|
|
return None
|
|
|
return max(counts.items(), key=lambda kv: kv[1])[0]
|
|
|
|
|
|
|
|
|
def build_allowed_metric_suffixes(field_ids: List[str], max_suffixes: int = 300) -> List[str]:
|
|
|
"""从数据集字段ID中提取实用的占位符候选列表"""
|
|
|
if not field_ids:
|
|
|
return []
|
|
|
|
|
|
# 过滤并转换字段ID
|
|
|
field_ids = [str(fid) for fid in field_ids if fid]
|
|
|
dataset_code = detect_dataset_code(field_ids)
|
|
|
|
|
|
counts: Dict[str, int] = {}
|
|
|
for raw in field_ids:
|
|
|
parts = [p for p in str(raw).split("_") if p]
|
|
|
if len(parts) < 2:
|
|
|
continue
|
|
|
|
|
|
for n in range(1, min(6, len(parts))):
|
|
|
suffix = "_".join(parts[-n:])
|
|
|
if suffix.replace("_", "").isdigit():
|
|
|
continue
|
|
|
if dataset_code and suffix.lower().startswith(dataset_code.lower() + "_"):
|
|
|
continue
|
|
|
if n == 1 and len(suffix) < 8:
|
|
|
continue
|
|
|
if len(suffix) < 6:
|
|
|
continue
|
|
|
counts[suffix] = counts.get(suffix, 0) + 1
|
|
|
|
|
|
ranked = sorted(
|
|
|
counts.items(),
|
|
|
key=lambda kv: (kv[1], kv[0].count("_"), len(kv[0])),
|
|
|
reverse=True,
|
|
|
)
|
|
|
|
|
|
suffixes: List[str] = []
|
|
|
for suffix, _ in ranked:
|
|
|
if suffix not in suffixes:
|
|
|
suffixes.append(suffix)
|
|
|
if len(suffixes) >= max_suffixes:
|
|
|
break
|
|
|
return suffixes
|
|
|
|
|
|
|
|
|
def compress_to_known_suffix(var: str, allowed_suffixes: List[str]) -> Optional[str]:
|
|
|
"""将变量名压缩为已知的后缀形式"""
|
|
|
if not allowed_suffixes:
|
|
|
return None
|
|
|
var_lower = var.lower()
|
|
|
for suffix in allowed_suffixes:
|
|
|
if var_lower.endswith(suffix.lower()):
|
|
|
return suffix
|
|
|
if suffix.lower().endswith(var_lower):
|
|
|
return suffix
|
|
|
return None
|
|
|
|
|
|
|
|
|
def placeholder_is_reasonably_matchable(placeholder: str, dataset_ids: List[str]) -> bool:
|
|
|
"""检查占位符是否可以与数据集字段匹配"""
|
|
|
if not dataset_ids:
|
|
|
return True
|
|
|
v = placeholder.strip()
|
|
|
if not v:
|
|
|
return False
|
|
|
if len(v) <= 3:
|
|
|
pat = re.compile(rf"(^|_){re.escape(v)}(_|$)", flags=re.IGNORECASE)
|
|
|
return any(pat.search(str(fid)) for fid in dataset_ids)
|
|
|
return any(v in str(fid) for fid in dataset_ids)
|
|
|
|
|
|
|
|
|
def normalize_template_placeholders(
|
|
|
template: str,
|
|
|
dataset_ids: List[str],
|
|
|
allowed_suffixes: List[str],
|
|
|
dataset_code: Optional[str],
|
|
|
) -> Tuple[str, bool]:
|
|
|
"""规范化模板中的占位符为后缀形式"""
|
|
|
vars_in_template = re.findall(r"\{([A-Za-z0-9_]+)\}", template)
|
|
|
if not vars_in_template:
|
|
|
return template, False
|
|
|
|
|
|
mapping: Dict[str, str] = {}
|
|
|
for var in set(vars_in_template):
|
|
|
new_var = var
|
|
|
if dataset_code and new_var.lower().startswith(dataset_code.lower() + "_"):
|
|
|
new_var = new_var[len(dataset_code) + 1:]
|
|
|
|
|
|
compressed = compress_to_known_suffix(new_var, allowed_suffixes)
|
|
|
if compressed:
|
|
|
new_var = compressed
|
|
|
|
|
|
mapping[var] = new_var
|
|
|
|
|
|
normalized = template
|
|
|
for src, dst in mapping.items():
|
|
|
normalized = normalized.replace("{" + src + "}", "{" + dst + "}")
|
|
|
|
|
|
vars_after = re.findall(r"\{([A-Za-z0-9_]+)\}", normalized)
|
|
|
ok = all(placeholder_is_reasonably_matchable(v, dataset_ids) for v in vars_after)
|
|
|
return normalized, ok
|
|
|
|
|
|
|
|
|
def extract_template_blocks(markdown_text: str) -> List[Dict[str, str]]:
|
|
|
"""
|
|
|
从 Markdown 文本中解析 **Concept** 块并提取 {template, idea}
|
|
|
|
|
|
返回列表,每个元素包含:
|
|
|
- template: 实现示例模板(包含 {variable} 占位符)
|
|
|
- idea: Concept 的完整描述
|
|
|
"""
|
|
|
concept_re = re.compile(r"^\*\*Concept\*\*\s*:\s*(.*)\s*$")
|
|
|
impl_re = re.compile(r"\*\*Implementation Example\*\*\s*:\s*(.*)$", flags=re.IGNORECASE)
|
|
|
backtick_re = re.compile(r"`([^`]*)`")
|
|
|
boundary_re = re.compile(r"^(?:-{3,}|#{1,6}\s+.*)\s*$")
|
|
|
|
|
|
lines = markdown_text.splitlines()
|
|
|
blocks: List[List[str]] = []
|
|
|
current: List[str] = []
|
|
|
|
|
|
def _flush():
|
|
|
nonlocal current
|
|
|
if current:
|
|
|
while current and not current[0].strip():
|
|
|
current.pop(0)
|
|
|
while current and not current[-1].strip():
|
|
|
current.pop()
|
|
|
if current:
|
|
|
blocks.append(current)
|
|
|
current = []
|
|
|
|
|
|
for line in lines:
|
|
|
if concept_re.match(line.strip()):
|
|
|
_flush()
|
|
|
current = [line]
|
|
|
continue
|
|
|
|
|
|
if current and boundary_re.match(line.strip()):
|
|
|
_flush()
|
|
|
continue
|
|
|
|
|
|
if current:
|
|
|
current.append(line)
|
|
|
|
|
|
_flush()
|
|
|
|
|
|
out: List[Dict[str, str]] = []
|
|
|
for block_lines in blocks:
|
|
|
template: Optional[str] = None
|
|
|
impl_line_idx: Optional[int] = None
|
|
|
|
|
|
for i, raw in enumerate(block_lines):
|
|
|
m = impl_re.search(raw)
|
|
|
if not m:
|
|
|
continue
|
|
|
|
|
|
impl_line_idx = i
|
|
|
tail = (m.group(1) or "").strip()
|
|
|
|
|
|
bt = backtick_re.search(tail)
|
|
|
if bt:
|
|
|
template = bt.group(1).strip()
|
|
|
break
|
|
|
|
|
|
if tail and ("{" in tail and "}" in tail):
|
|
|
template = tail.strip().strip("`")
|
|
|
break
|
|
|
|
|
|
for j in range(i + 1, min(i + 4, len(block_lines))):
|
|
|
nxt = block_lines[j].strip()
|
|
|
if not nxt:
|
|
|
continue
|
|
|
bt2 = backtick_re.search(nxt)
|
|
|
if bt2:
|
|
|
template = bt2.group(1).strip()
|
|
|
break
|
|
|
if "{" in nxt and "}" in nxt:
|
|
|
template = nxt.strip().strip("`")
|
|
|
break
|
|
|
break
|
|
|
|
|
|
# 注释掉跳过逻辑,允许没有 {placeholder} 的模板
|
|
|
# if not template or "{" not in template or "}" not in template:
|
|
|
# continue
|
|
|
|
|
|
# 如果 template 为 None,跳过
|
|
|
if not template:
|
|
|
continue
|
|
|
|
|
|
template = template.strip()
|
|
|
if template.startswith('{') and template.endswith('}'):
|
|
|
template = template[1:-1].strip()
|
|
|
|
|
|
idea_lines: List[str] = []
|
|
|
for i, raw in enumerate(block_lines):
|
|
|
if impl_line_idx is not None and i == impl_line_idx:
|
|
|
continue
|
|
|
idea_lines.append(raw)
|
|
|
|
|
|
idea = "\n".join(idea_lines).strip()
|
|
|
out.append({"template": template.strip(), "idea": idea})
|
|
|
|
|
|
return out
|
|
|
|
|
|
|
|
|
def match_single_horizon_auto(field_ids: List[str], template: str) -> List[str]:
|
|
|
"""
|
|
|
从模板生成表达式,将每个 {variable} 匹配到数据集字段ID
|
|
|
|
|
|
Args:
|
|
|
field_ids: 数据集字段ID列表
|
|
|
template: 包含 {variable} 占位符的模板字符串
|
|
|
|
|
|
Returns:
|
|
|
生成的表达式列表
|
|
|
"""
|
|
|
metrics = re.findall(r'\{([A-Za-z0-9_]+)\}', template)
|
|
|
if not metrics:
|
|
|
return []
|
|
|
|
|
|
metrics = sorted(metrics, key=len, reverse=True)
|
|
|
primary = metrics[0]
|
|
|
|
|
|
if not field_ids:
|
|
|
return []
|
|
|
|
|
|
ids = [str(fid) for fid in field_ids if fid]
|
|
|
|
|
|
def _matches_metric(field_id: str, metric: str) -> bool:
|
|
|
fid = str(field_id)
|
|
|
m = str(metric)
|
|
|
if len(m) <= 3:
|
|
|
return re.search(rf"(^|_){re.escape(m)}(_|$)", fid, flags=re.IGNORECASE) is not None
|
|
|
return m in fid
|
|
|
|
|
|
def _common_prefix_len(a: str, b: str) -> int:
|
|
|
n = min(len(a), len(b))
|
|
|
i = 0
|
|
|
while i < n and a[i] == b[i]:
|
|
|
i += 1
|
|
|
return i
|
|
|
|
|
|
candidates_by_metric = {}
|
|
|
for m in metrics:
|
|
|
cands = [fid for fid in ids if _matches_metric(fid, m)]
|
|
|
seen = set()
|
|
|
uniq = []
|
|
|
for x in cands:
|
|
|
if x not in seen:
|
|
|
seen.add(x)
|
|
|
uniq.append(x)
|
|
|
candidates_by_metric[m] = uniq
|
|
|
|
|
|
if not candidates_by_metric.get(primary):
|
|
|
return []
|
|
|
for m in metrics[1:]:
|
|
|
if not candidates_by_metric.get(m):
|
|
|
return []
|
|
|
|
|
|
MAX_PRIMARY_CANDIDATES = 30
|
|
|
MAX_SECONDARY_CHOICES = 8
|
|
|
MAX_EXPRESSIONS = 5000
|
|
|
|
|
|
results = []
|
|
|
seen_expr = set()
|
|
|
|
|
|
primary_candidates = candidates_by_metric[primary][:MAX_PRIMARY_CANDIDATES]
|
|
|
for primary_id in primary_candidates:
|
|
|
chosen_by_metric = {primary: [primary_id]}
|
|
|
for m in metrics[1:]:
|
|
|
cands = candidates_by_metric[m]
|
|
|
ranked = sorted(cands, key=lambda fid: _common_prefix_len(primary_id, fid), reverse=True)
|
|
|
chosen_by_metric[m] = ranked[:MAX_SECONDARY_CHOICES]
|
|
|
|
|
|
metric_order = metrics
|
|
|
pools = [chosen_by_metric[m] for m in metric_order]
|
|
|
for combo in itertools.product(*pools):
|
|
|
field_map = dict(zip(metric_order, combo))
|
|
|
try:
|
|
|
expr = template.format(**field_map)
|
|
|
except Exception:
|
|
|
continue
|
|
|
if expr in seen_expr:
|
|
|
continue
|
|
|
seen_expr.add(expr)
|
|
|
results.append(expr)
|
|
|
if len(results) >= MAX_EXPRESSIONS:
|
|
|
return results
|
|
|
|
|
|
return results
|
|
|
|
|
|
|
|
|
# ==================== 表达式验证器 ====================
|
|
|
|
|
|
# 尝试导入PLY库,如果不存在则提供安装提示
|
|
|
try:
|
|
|
import ply.lex as lex
|
|
|
import ply.yacc as yacc
|
|
|
except ImportError:
|
|
|
print("错误: 需要安装PLY库。请运行 'pip install ply' 来安装。")
|
|
|
sys.exit(1)
|
|
|
|
|
|
# 1. 定义支持的操作符和函数
|
|
|
supported_functions = {
|
|
|
# Group 类别函数
|
|
|
'group_min': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'category']},
|
|
|
'group_mean': {'min_args': 3, 'max_args': 3, 'arg_types': ['expression', 'expression', 'category']},
|
|
|
'group_median': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'category']},
|
|
|
'group_max': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'category']},
|
|
|
'group_rank': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'category']},
|
|
|
'group_vector_proj': {'min_args': 3, 'max_args': 3, 'arg_types': ['expression', 'expression', 'category']},
|
|
|
'group_normalize': {'min_args': 2, 'max_args': 5, 'arg_types': ['expression', 'category', 'expression', 'expression', 'expression']},
|
|
|
'group_extra': {'min_args': 3, 'max_args': 3, 'arg_types': ['expression', 'expression', 'category']},
|
|
|
'group_backfill': {'min_args': 3, 'max_args': 4, 'arg_types': ['expression', 'expression', 'expression', 'expression'], 'param_names': ['x', 'cat', 'days', 'std']},
|
|
|
'group_scale': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'category']},
|
|
|
'group_count': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'category']},
|
|
|
'group_zscore': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'category']},
|
|
|
'group_std_dev': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'category']},
|
|
|
'group_sum': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'category']},
|
|
|
'group_neutralize': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'category']},
|
|
|
'group_multi_regression': {'min_args': 4, 'max_args': 9, 'arg_types': ['expression'] * 9},
|
|
|
'group_cartesian_product': {'min_args': 2, 'max_args': 2, 'arg_types': ['category', 'category']},
|
|
|
'combo_a': {'min_args': 1, 'max_args': 3, 'arg_types': ['expression', 'expression', 'expression']},
|
|
|
|
|
|
# Transformational 类别函数
|
|
|
'right_tail': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'bucket': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'tail': {'min_args': 1, 'max_args': 4, 'arg_types': ['expression', 'expression', 'expression', 'expression']},
|
|
|
'left_tail': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'trade_when': {'min_args': 3, 'max_args': 3, 'arg_types': ['expression', 'expression', 'expression']},
|
|
|
'generate_stats': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
|
|
|
# Cross Sectional 类别函数
|
|
|
'winsorize': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'expression'], 'param_names': ['x', 'std']},
|
|
|
'rank': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'regression_proj': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'vector_neut': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'regression_neut': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'multi_regression': {'min_args': 2, 'max_args': 100, 'arg_types': ['expression'] * 100},
|
|
|
|
|
|
# Time Series 类别函数
|
|
|
'ts_std_dev': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_mean': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_delay': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_corr': {'min_args': 3, 'max_args': 3, 'arg_types': ['expression', 'expression', 'number']},
|
|
|
'ts_zscore': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_returns': {'min_args': 2, 'max_args': 3, 'arg_types': ['expression', 'number', 'number'], 'param_names': ['x', 'd', 'mode'], 'keyword_only': True},
|
|
|
'ts_product': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_backfill': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number'], 'param_names': ['x', 'd']},
|
|
|
'days_from_last_change': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'last_diff_value': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_scale': {'min_args': 2, 'max_args': 3, 'arg_types': ['expression', 'number', 'number'], 'param_names': ['x', 'd', 'constant'], 'keyword_only': True},
|
|
|
'ts_entropy': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number'], 'param_names': ['x', 'd']},
|
|
|
'ts_step': {'min_args': 1, 'max_args': 1, 'arg_types': ['number']},
|
|
|
'ts_sum': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_co_kurtosis': {'min_args': 3, 'max_args': 3, 'arg_types': ['expression', 'expression', 'number']},
|
|
|
'inst_tvr': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_decay_exp_window': {'min_args': 2, 'max_args': 3, 'arg_types': ['expression', 'number', 'number'], 'param_names': ['x', 'd', 'factor'], 'keyword_only': True},
|
|
|
'ts_av_diff': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_kurtosis': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_min_max_diff': {'min_args': 2, 'max_args': 3, 'arg_types': ['expression', 'number', 'number'], 'param_names': ['x', 'd', 'f'], 'keyword_only': True},
|
|
|
'ts_arg_max': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_max': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_min_max_cps': {'min_args': 2, 'max_args': 3, 'arg_types': ['expression', 'number', 'number'], 'param_names': ['x', 'd', 'f'], 'keyword_only': True},
|
|
|
'ts_rank': {'min_args': 2, 'max_args': 3, 'arg_types': ['expression', 'number', 'number'], 'param_names': ['x', 'd', 'constant'], 'keyword_only': True},
|
|
|
'ts_ir': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_theilsen': {'min_args': 3, 'max_args': 3, 'arg_types': ['expression', 'expression', 'number']},
|
|
|
'hump_decay': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'number'], 'param_names': ['x', 'p'], 'keyword_only': True},
|
|
|
'ts_weighted_decay': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'number'], 'param_names': ['x', 'k'], 'keyword_only': True},
|
|
|
'ts_quantile': {'min_args': 2, 'max_args': 3, 'arg_types': ['expression', 'number', 'string'], 'param_names': ['x', 'd', 'driver'], 'keyword_only': True},
|
|
|
'ts_min': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_count_nans': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_covariance': {'min_args': 3, 'max_args': 3, 'arg_types': ['expression', 'expression', 'number']},
|
|
|
'ts_co_skewness': {'min_args': 3, 'max_args': 3, 'arg_types': ['expression', 'expression', 'number']},
|
|
|
'ts_min_diff': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_decay_linear': {'min_args': 2, 'max_args': 3, 'arg_types': ['expression', 'number', 'boolean'], 'param_names': ['x', 'd', 'dense'], 'keyword_only': True},
|
|
|
'jump_decay': {'min_args': 2, 'max_args': 4, 'arg_types': ['expression', 'number', 'number', 'number'], 'param_names': ['x', 'd', 'sensitivity', 'force'],
|
|
|
'keyword_only': True},
|
|
|
'ts_moment': {'min_args': 2, 'max_args': 3, 'arg_types': ['expression', 'number', 'number'], 'param_names': ['x', 'd', 'k'], 'keyword_only': True},
|
|
|
'ts_arg_min': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_regression': {'min_args': 3, 'max_args': 5, 'arg_types': ['expression', 'expression', 'number', 'number', 'number'], 'param_names': ['y', 'x', 'd', 'lag', 'rettype'],
|
|
|
'keyword_only': True},
|
|
|
'ts_skewness': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_max_diff': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'kth_element': {'min_args': 3, 'max_args': 3, 'arg_types': ['expression', 'number', 'number']},
|
|
|
'hump': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'number'], 'param_names': ['x', 'hump']},
|
|
|
'ts_median': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_delta': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
'ts_poly_regression': {'min_args': 3, 'max_args': 4, 'arg_types': ['expression', 'expression', 'number', 'number'], 'param_names': ['y', 'x', 'd', 'k'], 'keyword_only': True,
|
|
|
'keyword_only_from': 3},
|
|
|
'ts_target_tvr_decay': {'min_args': 1, 'max_args': 4, 'arg_types': ['expression', 'number', 'number', 'number'], 'param_names': ['x', 'lambda_min', 'lambda_max', 'target_tvr'],
|
|
|
'keyword_only': True},
|
|
|
'ts_target_tvr_delta_limit': {'min_args': 2, 'max_args': 5, 'arg_types': ['expression', 'expression', 'number', 'number', 'number'],
|
|
|
'param_names': ['x', 'y', 'lambda_min', 'lambda_max', 'target_tvr'], 'keyword_only': True},
|
|
|
'ts_target_tvr_hump': {'min_args': 1, 'max_args': 4, 'arg_types': ['expression', 'number', 'number', 'number'], 'param_names': ['x', 'lambda_min', 'lambda_max', 'target_tvr'],
|
|
|
'keyword_only': True},
|
|
|
'ts_delta_limit': {'min_args': 2, 'max_args': 3, 'arg_types': ['expression', 'expression', 'number'], 'param_names': ['x', 'y', 'limit_volume'], 'keyword_only': True},
|
|
|
|
|
|
# Special 类别函数
|
|
|
'inst_pnl': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'self_corr': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'in': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'universe_size': {'min_args': 0, 'max_args': 0, 'arg_types': []},
|
|
|
|
|
|
# Missing functions from operators.py
|
|
|
'quantile': {'min_args': 1, 'max_args': 3, 'arg_types': ['expression', 'expression', 'expression'], 'param_names': ['x', 'driver', 'sigma']},
|
|
|
'normalize': {'min_args': 1, 'max_args': 3, 'arg_types': ['expression', 'boolean', 'number']},
|
|
|
'zscore': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
|
|
|
# Logical 类别函数
|
|
|
'or': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'and': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'not': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'is_nan': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'is_not_nan': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'less': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'equal': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'greater': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'is_finite': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'if_else': {'min_args': 3, 'max_args': 3, 'arg_types': ['expression', 'expression', 'expression']},
|
|
|
'not_equal': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'less_equal': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'greater_equal': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
|
|
|
# Vector 类别函数
|
|
|
'vec_kurtosis': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'vec_min': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'vec_count': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'vec_sum': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'vec_skewness': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'vec_max': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'vec_avg': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'vec_range': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'vec_choose': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number'], 'param_names': ['x', 'nth']},
|
|
|
'vec_powersum': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'number'], 'param_names': ['x', 'constant']},
|
|
|
'vec_stddev': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'vec_percentage': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'number'], 'param_names': ['x', 'percentage']},
|
|
|
'vec_ir': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'vec_norm': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'ts_percentage': {'min_args': 2, 'max_args': 3, 'arg_types': ['expression', 'number', 'number'], 'param_names': ['x', 'd', 'percentage']},
|
|
|
'signed_power': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number']},
|
|
|
|
|
|
# Additional functions from test cases
|
|
|
'rank_by_side': {'min_args': 1, 'max_args': 3, 'arg_types': ['expression', 'number', 'number'], 'param_names': ['x', 'rate', 'scale']},
|
|
|
'log_diff': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'nan_mask': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'ts_partial_corr': {'min_args': 4, 'max_args': 4, 'arg_types': ['expression', 'expression', 'expression', 'number']},
|
|
|
'ts_triple_corr': {'min_args': 4, 'max_args': 4, 'arg_types': ['expression', 'expression', 'expression', 'number']},
|
|
|
'clamp': {'min_args': 1, 'max_args': 3, 'arg_types': ['expression', 'expression', 'expression'], 'param_names': ['x', 'lower', 'upper']},
|
|
|
'keep': {'min_args': 2, 'max_args': 3, 'arg_types': ['expression', 'expression', 'number'], 'param_names': ['x', 'condition', 'period']},
|
|
|
'replace': {'min_args': 3, 'max_args': 3, 'arg_types': ['expression', 'expression', 'expression'], 'param_names': ['x', 'target', 'dest']},
|
|
|
'filter': {'min_args': 3, 'max_args': 3, 'arg_types': ['expression', 'expression', 'expression'], 'param_names': ['x', 'h', 't']},
|
|
|
'one_side': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'string'], 'param_names': ['x', 'side']},
|
|
|
'scale_down': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'number'], 'param_names': ['x', 'constant']},
|
|
|
|
|
|
# Arithmetic 类别函数
|
|
|
'add': {'min_args': 2, 'max_args': 101, 'arg_types': ['expression'] * 101},
|
|
|
'multiply': {'min_args': 2, 'max_args': 100, 'arg_types': ['expression'] * 99 + ['boolean'], 'param_names': ['x', 'y', 'filter']},
|
|
|
'sign': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'subtract': {'min_args': 2, 'max_args': 3, 'arg_types': ['expression', 'expression', 'boolean']},
|
|
|
'pasteurize': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'log': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'purify': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'arc_tan': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'max': {'min_args': 2, 'max_args': 100, 'arg_types': ['expression'] * 100},
|
|
|
'to_nan': {'min_args': 1, 'max_args': 3, 'arg_types': ['expression', 'expression', 'boolean']},
|
|
|
'abs': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'sigmoid': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'divide': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'min': {'min_args': 2, 'max_args': 100, 'arg_types': ['expression'] * 100},
|
|
|
'tanh': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'nan_out': {'min_args': 1, 'max_args': 3, 'arg_types': ['expression', 'expression', 'expression'], 'param_names': ['x', 'lower', 'upper']},
|
|
|
'signed_power': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'inverse': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'round': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'sqrt': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
's_log_1p': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'reverse': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'power': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression']},
|
|
|
'densify': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'floor': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression']},
|
|
|
'arc_cos': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression'], 'param_names': ['x']},
|
|
|
'arc_sin': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression'], 'param_names': ['x']},
|
|
|
'ceiling': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression'], 'param_names': ['x']},
|
|
|
'exp': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression'], 'param_names': ['x']},
|
|
|
'fraction': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression'], 'param_names': ['x']},
|
|
|
'round_down': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'expression'], 'param_names': ['x', 'f']},
|
|
|
'is_not_finite': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression'], 'param_names': ['input']},
|
|
|
'negate': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression'], 'param_names': ['input']},
|
|
|
'ts_rank_gmean_amean_diff': {'min_args': 5, 'max_args': 5, 'arg_types': ['expression', 'expression', 'expression', 'expression', 'number'],
|
|
|
'param_names': ['input1', 'input2', 'input3', '...', 'd']},
|
|
|
'ts_vector_neut': {'min_args': 3, 'max_args': 3, 'arg_types': ['expression', 'expression', 'number'], 'param_names': ['x', 'y', 'd']},
|
|
|
'ts_vector_proj': {'min_args': 3, 'max_args': 3, 'arg_types': ['expression', 'expression', 'number'], 'param_names': ['x', 'y', 'd']},
|
|
|
'scale': {'min_args': 1, 'max_args': 4, 'arg_types': ['expression', 'expression', 'expression', 'expression'], 'param_names': ['x', 'scale', 'longscale', 'shortscale']},
|
|
|
'generalized_rank': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'expression'], 'param_names': ['open', 'm']},
|
|
|
'rank_gmean_amean_diff': {'min_args': 4, 'max_args': 4, 'arg_types': ['expression', 'expression', 'expression', 'expression'],
|
|
|
'param_names': ['input1', 'input2', 'input3', '...']},
|
|
|
'truncate': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'expression'], 'param_names': ['x', 'maxPercent']},
|
|
|
'vector_proj': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression'], 'param_names': ['x', 'y']},
|
|
|
'vec_filter': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'expression'], 'param_names': ['vec', 'value']},
|
|
|
'group_coalesce': {'min_args': 4, 'max_args': 4, 'arg_types': ['expression', 'expression', 'expression', 'expression'],
|
|
|
'param_names': ['original_group', 'group2', 'group3', '…']},
|
|
|
'group_percentage': {'min_args': 2, 'max_args': 3, 'arg_types': ['expression', 'category', 'expression'], 'param_names': ['x', 'group', 'percentage']},
|
|
|
'group_vector_neut': {'min_args': 3, 'max_args': 3, 'arg_types': ['expression', 'expression', 'expression'], 'param_names': ['x', 'y', 'g']},
|
|
|
'convert': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'expression'], 'param_names': ['x', 'mode']},
|
|
|
'reduce_avg': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'expression'], 'param_names': ['input', 'threshold']},
|
|
|
'reduce_choose': {'min_args': 2, 'max_args': 3, 'arg_types': ['expression', 'expression', 'expression'], 'param_names': ['input', 'nth', 'ignoreNan']},
|
|
|
'reduce_count': {'min_args': 2, 'max_args': 2, 'arg_types': ['expression', 'expression'], 'param_names': ['input', 'threshold']},
|
|
|
'reduce_ir': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression'], 'param_names': ['input']},
|
|
|
'reduce_kurtosis': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression'], 'param_names': ['input']},
|
|
|
'reduce_max': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression'], 'param_names': ['input']},
|
|
|
'reduce_min': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression'], 'param_names': ['input']},
|
|
|
'reduce_norm': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression'], 'param_names': ['input']},
|
|
|
'reduce_percentage': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'expression'], 'param_names': ['input', 'percentage']},
|
|
|
'reduce_powersum': {'min_args': 1, 'max_args': 3, 'arg_types': ['expression', 'expression', 'expression'], 'param_names': ['input', 'constant', 'precise']},
|
|
|
'reduce_range': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression'], 'param_names': ['input']},
|
|
|
'reduce_skewness': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression'], 'param_names': ['input']},
|
|
|
'reduce_stddev': {'min_args': 1, 'max_args': 2, 'arg_types': ['expression', 'expression'], 'param_names': ['input', 'threshold']},
|
|
|
'reduce_sum': {'min_args': 1, 'max_args': 1, 'arg_types': ['expression'], 'param_names': ['input']},
|
|
|
}
|
|
|
|
|
|
# 2. 定义group类型字段
|
|
|
group_fields = {
|
|
|
'sector', 'subindustry', 'industry', 'exchange', 'country', 'market'
|
|
|
}
|
|
|
|
|
|
# 3. 有效类别集合
|
|
|
valid_categories = group_fields
|
|
|
|
|
|
# 4. 字段命名模式 - 只校验字段是不是数字字母下划线组成
|
|
|
field_patterns = [
|
|
|
re.compile(r'^[a-zA-Z0-9_]+$'),
|
|
|
]
|
|
|
|
|
|
|
|
|
# 4. 抽象语法树节点类型
|
|
|
class ASTNode:
|
|
|
"""抽象语法树节点基类"""
|
|
|
|
|
|
def __init__(self, node_type: str, children: Optional[List['ASTNode']] = None,
|
|
|
value: Optional[Any] = None, line: Optional[int] = None):
|
|
|
self.node_type = node_type
|
|
|
self.children = children or []
|
|
|
self.value = value
|
|
|
self.line = line
|
|
|
|
|
|
def __str__(self) -> str:
|
|
|
return f"ASTNode({self.node_type}, {self.value}, line={self.line})"
|
|
|
|
|
|
def __repr__(self) -> str:
|
|
|
return self.__str__()
|
|
|
|
|
|
|
|
|
class ExpressionValidator:
|
|
|
"""表达式验证器类"""
|
|
|
|
|
|
def __init__(self):
|
|
|
"""初始化词法分析器和语法分析器"""
|
|
|
self.lexer = lex.lex(module=self, debug=False)
|
|
|
self.parser = yacc.yacc(module=self, debug=False)
|
|
|
self.errors = []
|
|
|
self._unit_cache: Dict[int, str] = {}
|
|
|
self._derived_category_cache: Dict[int, bool] = {}
|
|
|
|
|
|
# 词法分析器规则
|
|
|
tokens = ('FUNCTION', 'FIELD', 'NUMBER', 'LPAREN', 'RPAREN',
|
|
|
'PLUS', 'MINUS', 'TIMES', 'DIVIDE', 'COMMA', 'CATEGORY',
|
|
|
'EQUAL', 'ASSIGN', 'IDENTIFIER', 'STRING', 'GREATER', 'LESS', 'GREATEREQUAL', 'LESSEQUAL', 'NOTEQUAL', 'BOOLEAN')
|
|
|
|
|
|
t_ignore = ' \t\n'
|
|
|
|
|
|
t_PLUS = r'\+'
|
|
|
t_MINUS = r'-'
|
|
|
t_TIMES = r'\*'
|
|
|
t_DIVIDE = r'/'
|
|
|
t_LPAREN = r'\('
|
|
|
t_RPAREN = r'\)'
|
|
|
t_COMMA = r','
|
|
|
t_EQUAL = r'=='
|
|
|
t_NOTEQUAL = r'!='
|
|
|
t_GREATEREQUAL = r'>='
|
|
|
t_LESSEQUAL = r'<='
|
|
|
t_GREATER = r'>'
|
|
|
t_LESS = r'<'
|
|
|
t_ASSIGN = r'='
|
|
|
|
|
|
def t_NUMBER(self, t):
|
|
|
r'\d+\.?\d*'
|
|
|
if '.' in t.value:
|
|
|
t.value = float(t.value)
|
|
|
else:
|
|
|
t.value = int(t.value)
|
|
|
return t
|
|
|
|
|
|
def t_STRING(self, t):
|
|
|
r"'[^']*'|\"[^\"]*\""
|
|
|
t.value = t.value[1:-1]
|
|
|
return t
|
|
|
|
|
|
def t_IDENTIFIER(self, t):
|
|
|
r'[a-zA-Z_][a-zA-Z0-9_]*'
|
|
|
if t.value.lower() in {'true', 'false'}:
|
|
|
t.type = 'BOOLEAN'
|
|
|
t.value = t.value.lower()
|
|
|
else:
|
|
|
lexpos = t.lexpos
|
|
|
next_chars = ''
|
|
|
if lexpos + len(t.value) < len(t.lexer.lexdata):
|
|
|
next_pos = lexpos + len(t.value)
|
|
|
while next_pos < len(t.lexer.lexdata) and t.lexer.lexdata[next_pos].isspace():
|
|
|
next_pos += 1
|
|
|
if next_pos < len(t.lexer.lexdata):
|
|
|
next_chars = t.lexer.lexdata[next_pos:next_pos + 1]
|
|
|
|
|
|
if next_chars == '=':
|
|
|
t.type = 'IDENTIFIER'
|
|
|
elif next_chars == '(':
|
|
|
t.type = 'FUNCTION'
|
|
|
t.value = t.value.lower()
|
|
|
elif t.value in {'std', 'k', 'lambda_min', 'lambda_max', 'target_tvr', 'range', 'buckets', 'lag', 'rettype', 'mode', 'nth', 'constant', 'percentage', 'driver', 'sigma',
|
|
|
'rate', 'scale', 'filter', 'lower', 'upper', 'target', 'dest', 'event', 'sensitivity', 'force', 'h', 't', 'period', 'stddev', 'factor', 'k', 'useStd',
|
|
|
'limit', 'gaussian', 'uniform', 'cauchy'}:
|
|
|
t.type = 'IDENTIFIER'
|
|
|
elif t.value.lower() in supported_functions:
|
|
|
t.type = 'FUNCTION'
|
|
|
t.value = t.value.lower()
|
|
|
elif t.value in valid_categories:
|
|
|
t.type = 'CATEGORY'
|
|
|
elif self._is_valid_field(t.value):
|
|
|
t.type = 'FIELD'
|
|
|
else:
|
|
|
t.type = 'IDENTIFIER'
|
|
|
return t
|
|
|
|
|
|
def t_newline(self, t):
|
|
|
r'\n+'
|
|
|
t.lexer.lineno += len(t.value)
|
|
|
|
|
|
def t_error(self, t):
|
|
|
if t:
|
|
|
if not re.match(r'[a-zA-Z0-9_\+\-\*/\(\)\,\s=<>!]', t.value[0]):
|
|
|
self.errors.append(f"非法字符 '{t.value[0]}' (行 {t.lexer.lineno})")
|
|
|
else:
|
|
|
self.errors.append(f"非法标记 '{t.value}' (行 {t.lexer.lineno})")
|
|
|
t.lexer.skip(1)
|
|
|
else:
|
|
|
self.errors.append("词法分析器到达文件末尾")
|
|
|
|
|
|
# 语法分析器规则
|
|
|
def p_expression(self, p):
|
|
|
"""expression : comparison
|
|
|
| expression EQUAL comparison
|
|
|
| expression NOTEQUAL comparison
|
|
|
| expression GREATER comparison
|
|
|
| expression LESS comparison
|
|
|
| expression GREATEREQUAL comparison
|
|
|
| expression LESSEQUAL comparison"""
|
|
|
if len(p) == 2:
|
|
|
p[0] = p[1]
|
|
|
else:
|
|
|
p[0] = ASTNode('binop', [p[1], p[3]], {'op': p[2]})
|
|
|
|
|
|
def p_comparison(self, p):
|
|
|
"""comparison : term
|
|
|
| comparison PLUS term
|
|
|
| comparison MINUS term"""
|
|
|
if len(p) == 2:
|
|
|
p[0] = p[1]
|
|
|
else:
|
|
|
p[0] = ASTNode('binop', [p[1], p[3]], {'op': p[2]})
|
|
|
|
|
|
def p_term(self, p):
|
|
|
"""term : factor
|
|
|
| term TIMES factor
|
|
|
| term DIVIDE factor"""
|
|
|
if len(p) == 2:
|
|
|
p[0] = p[1]
|
|
|
else:
|
|
|
p[0] = ASTNode('binop', [p[1], p[3]], {'op': p[2]})
|
|
|
|
|
|
def p_factor(self, p):
|
|
|
"""factor : NUMBER
|
|
|
| STRING
|
|
|
| FIELD
|
|
|
| CATEGORY
|
|
|
| IDENTIFIER
|
|
|
| BOOLEAN
|
|
|
| MINUS factor
|
|
|
| LPAREN expression RPAREN
|
|
|
| function_call"""
|
|
|
if len(p) == 2:
|
|
|
if p.slice[1].type == 'NUMBER':
|
|
|
p[0] = ASTNode('number', value=p[1])
|
|
|
elif p.slice[1].type == 'STRING':
|
|
|
p[0] = ASTNode('string', value=p[1])
|
|
|
elif p.slice[1].type == 'FIELD':
|
|
|
p[0] = ASTNode('field', value=p[1])
|
|
|
elif p.slice[1].type == 'CATEGORY':
|
|
|
p[0] = ASTNode('category', value=p[1])
|
|
|
elif p.slice[1].type == 'BOOLEAN':
|
|
|
p[0] = ASTNode('boolean', value=p[1])
|
|
|
elif p.slice[1].type == 'IDENTIFIER':
|
|
|
p[0] = ASTNode('identifier', value=p[1])
|
|
|
else:
|
|
|
p[0] = p[1]
|
|
|
elif len(p) == 3:
|
|
|
p[0] = ASTNode('unop', [p[2]], {'op': p[1]})
|
|
|
elif len(p) == 4:
|
|
|
p[0] = p[2]
|
|
|
else:
|
|
|
p[0] = p[1]
|
|
|
|
|
|
def p_function_call(self, p):
|
|
|
'''function_call : FUNCTION LPAREN args RPAREN'''
|
|
|
p[0] = ASTNode('function', p[3], p[1])
|
|
|
|
|
|
def p_args(self, p):
|
|
|
'''args : arg_list
|
|
|
| empty'''
|
|
|
if len(p) == 2 and p[1] is not None:
|
|
|
p[0] = p[1]
|
|
|
else:
|
|
|
p[0] = []
|
|
|
|
|
|
def p_arg_list(self, p):
|
|
|
'''arg_list : arg
|
|
|
| arg_list COMMA arg'''
|
|
|
if len(p) == 2:
|
|
|
p[0] = [p[1]]
|
|
|
else:
|
|
|
p[0] = p[1] + [p[3]]
|
|
|
|
|
|
def p_arg(self, p):
|
|
|
'''arg : expression
|
|
|
| IDENTIFIER ASSIGN expression'''
|
|
|
if len(p) == 2:
|
|
|
p[0] = {'type': 'positional', 'value': p[1]}
|
|
|
else:
|
|
|
p[0] = {'type': 'named', 'name': p[1], 'value': p[3]}
|
|
|
|
|
|
def p_empty(self, p):
|
|
|
'''empty :'''
|
|
|
p[0] = None
|
|
|
|
|
|
def p_error(self, p):
|
|
|
if p:
|
|
|
self.errors.append(f"语法错误在位置 {p.lexpos}: 非法标记 '{p.value}'")
|
|
|
else:
|
|
|
self.errors.append("语法错误: 表达式不完整")
|
|
|
|
|
|
def _is_valid_field(self, field_name: str) -> bool:
|
|
|
"""检查字段名是否符合模式"""
|
|
|
for pattern in field_patterns:
|
|
|
if pattern.match(field_name):
|
|
|
return True
|
|
|
return False
|
|
|
|
|
|
def validate_function(self, node: ASTNode, is_in_group_arg: bool = False) -> List[str]:
|
|
|
"""验证函数调用的参数数量和类型"""
|
|
|
function_name = node.value
|
|
|
args = node.children
|
|
|
function_info = supported_functions.get(function_name)
|
|
|
|
|
|
if not function_info:
|
|
|
return [f"未知函数: {function_name}"]
|
|
|
|
|
|
if function_name == 'add':
|
|
|
return self._validate_add(args, is_in_group_arg)
|
|
|
|
|
|
errors = []
|
|
|
|
|
|
keyword_only_from = function_info.get('keyword_only_from')
|
|
|
if keyword_only_from is None and function_info.get('keyword_only'):
|
|
|
keyword_only_from = function_info.get('min_args', 0)
|
|
|
|
|
|
if len(args) < function_info['min_args']:
|
|
|
errors.append(f"函数 {function_name} 需要至少 {function_info['min_args']} 个参数,但只提供了 {len(args)}")
|
|
|
elif len(args) > function_info['max_args']:
|
|
|
errors.append(f"函数 {function_name} 最多接受 {function_info['max_args']} 个参数,但提供了 {len(args)}")
|
|
|
|
|
|
positional_index = 0
|
|
|
|
|
|
for arg in args:
|
|
|
if isinstance(arg, dict):
|
|
|
if arg['type'] == 'named':
|
|
|
if 'param_names' in function_info and arg['name'] in function_info['param_names']:
|
|
|
param_index = function_info['param_names'].index(arg['name'])
|
|
|
if param_index < len(function_info['arg_types']):
|
|
|
expected_type = function_info['arg_types'][param_index]
|
|
|
arg_errors = self._validate_arg_type(arg['value'], expected_type, param_index, function_name, is_in_group_arg)
|
|
|
errors.extend(arg_errors)
|
|
|
elif function_name == 'winsorize' and arg['name'] in ['std', 'clip']:
|
|
|
arg_errors = self._validate_arg_type(arg['value'], 'number', 0, function_name, is_in_group_arg)
|
|
|
errors.extend(arg_errors)
|
|
|
elif function_name == 'bucket' and arg['name'] in ['range', 'buckets']:
|
|
|
arg_errors = self._validate_arg_type(arg['value'], 'string', 1, function_name, is_in_group_arg)
|
|
|
errors.extend(arg_errors)
|
|
|
else:
|
|
|
errors.append(f"函数 {function_name} 不存在参数 '{arg['name']}'")
|
|
|
elif arg['type'] == 'positional':
|
|
|
if keyword_only_from is not None and positional_index >= keyword_only_from:
|
|
|
param_name = None
|
|
|
if 'param_names' in function_info and positional_index < len(function_info['param_names']):
|
|
|
param_name = function_info['param_names'][positional_index]
|
|
|
if param_name:
|
|
|
errors.append(f"函数 {function_name} 的第{positional_index + 1}个参数必须使用命名参数 '{param_name}='")
|
|
|
else:
|
|
|
errors.append(f"函数 {function_name} 的第{positional_index + 1}个参数必须使用命名参数")
|
|
|
else:
|
|
|
if positional_index < len(function_info['arg_types']):
|
|
|
expected_type = function_info['arg_types'][positional_index]
|
|
|
arg_errors = self._validate_arg_type(arg['value'], expected_type, positional_index, function_name, is_in_group_arg)
|
|
|
errors.extend(arg_errors)
|
|
|
positional_index += 1
|
|
|
else:
|
|
|
errors.append(f"参数 {positional_index + 1} 格式错误")
|
|
|
positional_index += 1
|
|
|
else:
|
|
|
if keyword_only_from is not None and positional_index >= keyword_only_from:
|
|
|
param_name = None
|
|
|
if 'param_names' in function_info and positional_index < len(function_info['param_names']):
|
|
|
param_name = function_info['param_names'][positional_index]
|
|
|
if param_name:
|
|
|
errors.append(f"函数 {function_name} 的第{positional_index + 1}个参数必须使用命名参数 '{param_name}='")
|
|
|
else:
|
|
|
errors.append(f"函数 {function_name} 的第{positional_index + 1}个参数必须使用命名参数")
|
|
|
else:
|
|
|
if positional_index < len(function_info['arg_types']):
|
|
|
expected_type = function_info['arg_types'][positional_index]
|
|
|
arg_errors = self._validate_arg_type(arg, expected_type, positional_index, function_name, is_in_group_arg)
|
|
|
errors.extend(arg_errors)
|
|
|
positional_index += 1
|
|
|
|
|
|
return errors
|
|
|
|
|
|
def _validate_arg_type(self, arg: ASTNode, expected_type: str, arg_index: int, function_name: str, is_in_group_arg: bool = False) -> List[str]:
|
|
|
"""验证参数类型是否符合预期"""
|
|
|
errors = []
|
|
|
|
|
|
def _is_number_like(node: ASTNode) -> bool:
|
|
|
if node is None:
|
|
|
return False
|
|
|
if node.node_type == 'number':
|
|
|
return True
|
|
|
if node.node_type == 'unop' and isinstance(node.value, dict) and node.value.get('op') in {'-', '+'}:
|
|
|
if node.children and hasattr(node.children[0], 'node_type'):
|
|
|
return _is_number_like(node.children[0])
|
|
|
return False
|
|
|
|
|
|
if self._is_derived_category(arg) and expected_type != 'category':
|
|
|
errors.append(
|
|
|
f"Incompatible unit for input of \"{function_name}\" at index {arg_index}, expected \"Unit[]\", found \"Unit[Group:1]\""
|
|
|
)
|
|
|
return errors
|
|
|
|
|
|
if arg.node_type == 'category' and arg.value in group_fields:
|
|
|
if not (function_name.startswith('group_') or is_in_group_arg):
|
|
|
errors.append(f"Group类型字段 '{arg.value}' 只能用于Group类型函数的参数中")
|
|
|
|
|
|
if expected_type == 'expression':
|
|
|
pass
|
|
|
elif expected_type == 'number':
|
|
|
if not _is_number_like(arg):
|
|
|
errors.append(f"参数 {arg_index + 1} 应该是一个数字,但得到 {arg.node_type}")
|
|
|
elif expected_type == 'boolean':
|
|
|
if arg.node_type not in {'boolean', 'number'}:
|
|
|
errors.append(f"参数 {arg_index + 1} 应该是一个布尔值(true/false 或 0/1),但得到 {arg.node_type}")
|
|
|
elif expected_type == 'field':
|
|
|
if arg.node_type != 'field' and arg.node_type != 'category':
|
|
|
errors.append(f"参数 {arg_index + 1} 应该是一个字段,但得到 {arg.node_type}")
|
|
|
elif arg.node_type == 'field' and not self._is_valid_field(arg.value):
|
|
|
errors.append(f"无效的字段名: {arg.value}")
|
|
|
elif expected_type == 'category':
|
|
|
if not function_name.startswith('group_'):
|
|
|
if arg.node_type != 'category':
|
|
|
errors.append(f"参数 {arg_index + 1} 应该是一个类别,但得到 {arg.node_type}")
|
|
|
elif arg.value not in valid_categories:
|
|
|
errors.append(f"无效的类别: {arg.value}")
|
|
|
|
|
|
return errors
|
|
|
|
|
|
def _infer_unit(self, node: ASTNode) -> str:
|
|
|
"""Infer the Unit kind of an AST node."""
|
|
|
if node is None:
|
|
|
return 'unit'
|
|
|
|
|
|
cache_key = id(node)
|
|
|
cached = self._unit_cache.get(cache_key)
|
|
|
if cached is not None:
|
|
|
return cached
|
|
|
|
|
|
unit = 'unit'
|
|
|
|
|
|
if node.node_type in {'number', 'boolean', 'string'}:
|
|
|
unit = 'scalar'
|
|
|
elif node.node_type in {'field', 'identifier'}:
|
|
|
unit = 'unit'
|
|
|
elif node.node_type == 'category':
|
|
|
unit = 'category'
|
|
|
elif node.node_type in {'unop', 'binop'}:
|
|
|
child_units = [self._infer_unit(child) for child in node.children if hasattr(child, 'node_type')]
|
|
|
unit = 'category' if 'category' in child_units else 'unit'
|
|
|
elif node.node_type == 'function':
|
|
|
fname = node.value
|
|
|
if fname in {'bucket', 'group_cartesian_product'}:
|
|
|
unit = 'category'
|
|
|
else:
|
|
|
first_arg = None
|
|
|
for child in node.children:
|
|
|
if isinstance(child, dict):
|
|
|
if child.get('type') == 'positional':
|
|
|
first_arg = child.get('value')
|
|
|
break
|
|
|
else:
|
|
|
first_arg = child
|
|
|
break
|
|
|
|
|
|
if hasattr(first_arg, 'node_type'):
|
|
|
unit = self._infer_unit(first_arg)
|
|
|
else:
|
|
|
unit = 'unit'
|
|
|
|
|
|
self._unit_cache[cache_key] = unit
|
|
|
return unit
|
|
|
|
|
|
def _is_derived_category(self, node: ASTNode) -> bool:
|
|
|
"""Return True if node is a derived category/grouping key."""
|
|
|
if node is None:
|
|
|
return False
|
|
|
|
|
|
cache_key = id(node)
|
|
|
cached = self._derived_category_cache.get(cache_key)
|
|
|
if cached is not None:
|
|
|
return cached
|
|
|
|
|
|
derived = False
|
|
|
if node.node_type == 'function':
|
|
|
if node.value in {'bucket', 'group_cartesian_product'}:
|
|
|
derived = True
|
|
|
else:
|
|
|
function_info = supported_functions.get(node.value, {})
|
|
|
arg_types = function_info.get('arg_types', [])
|
|
|
param_names = function_info.get('param_names', [])
|
|
|
|
|
|
positional_index = 0
|
|
|
for child in node.children:
|
|
|
if isinstance(child, dict):
|
|
|
if child.get('type') == 'named':
|
|
|
name = child.get('name')
|
|
|
value = child.get('value')
|
|
|
|
|
|
expected_type = None
|
|
|
if name in param_names:
|
|
|
param_index = param_names.index(name)
|
|
|
if param_index < len(arg_types):
|
|
|
expected_type = arg_types[param_index]
|
|
|
|
|
|
if expected_type == 'category':
|
|
|
continue
|
|
|
|
|
|
if self._is_derived_category(value):
|
|
|
derived = True
|
|
|
break
|
|
|
elif child.get('type') == 'positional':
|
|
|
value = child.get('value')
|
|
|
expected_type = arg_types[positional_index] if positional_index < len(arg_types) else None
|
|
|
|
|
|
if expected_type != 'category' and self._is_derived_category(value):
|
|
|
derived = True
|
|
|
break
|
|
|
positional_index += 1
|
|
|
else:
|
|
|
expected_type = arg_types[positional_index] if positional_index < len(arg_types) else None
|
|
|
if expected_type != 'category' and self._is_derived_category(child):
|
|
|
derived = True
|
|
|
break
|
|
|
positional_index += 1
|
|
|
elif node.node_type in {'unop', 'binop'}:
|
|
|
derived = any(
|
|
|
self._is_derived_category(child)
|
|
|
for child in node.children
|
|
|
if hasattr(child, 'node_type')
|
|
|
)
|
|
|
|
|
|
self._derived_category_cache[cache_key] = derived
|
|
|
return derived
|
|
|
|
|
|
def _validate_add(self, args: List[Any], is_in_group_arg: bool = False) -> List[str]:
|
|
|
"""Validate add(x, y, ..., filter=false)."""
|
|
|
errors: List[str] = []
|
|
|
|
|
|
if len(args) < 2:
|
|
|
return [f"函数 add 需要至少 2 个参数,但只提供了 {len(args)}"]
|
|
|
|
|
|
named_filter_nodes: List[ASTNode] = []
|
|
|
positional_nodes: List[ASTNode] = []
|
|
|
|
|
|
for arg in args:
|
|
|
if isinstance(arg, dict) and arg.get('type') == 'named':
|
|
|
name = arg.get('name')
|
|
|
value = arg.get('value')
|
|
|
if name != 'filter':
|
|
|
errors.append(f"函数 add 不存在参数 '{name}'")
|
|
|
continue
|
|
|
if not hasattr(value, 'node_type'):
|
|
|
errors.append("函数 add 的参数 filter 格式错误")
|
|
|
continue
|
|
|
named_filter_nodes.append(value)
|
|
|
elif isinstance(arg, dict) and arg.get('type') == 'positional':
|
|
|
value = arg.get('value')
|
|
|
if hasattr(value, 'node_type'):
|
|
|
positional_nodes.append(value)
|
|
|
else:
|
|
|
errors.append("函数 add 的位置参数格式错误")
|
|
|
elif hasattr(arg, 'node_type'):
|
|
|
positional_nodes.append(arg)
|
|
|
else:
|
|
|
errors.append("函数 add 的参数格式错误")
|
|
|
|
|
|
if len(named_filter_nodes) > 1:
|
|
|
errors.append("函数 add 的参数 'filter' 只能出现一次")
|
|
|
|
|
|
positional_filter_node: Optional[ASTNode] = None
|
|
|
if not named_filter_nodes and len(positional_nodes) >= 3:
|
|
|
last = positional_nodes[-1]
|
|
|
if last.node_type == 'boolean' or (last.node_type == 'number' and last.value in {0, 1}):
|
|
|
positional_filter_node = positional_nodes.pop()
|
|
|
|
|
|
if len(positional_nodes) < 2:
|
|
|
errors.append(f"函数 add 需要至少 2 个输入项(不含filter),但只提供了 {len(positional_nodes)}")
|
|
|
|
|
|
for idx, node in enumerate(positional_nodes):
|
|
|
errors.extend(self._validate_arg_type(node, 'expression', idx, 'add', is_in_group_arg))
|
|
|
|
|
|
if positional_filter_node is not None and named_filter_nodes:
|
|
|
errors.append("函数 add 的 filter 不能同时用位置参数和命名参数传递")
|
|
|
if positional_filter_node is not None:
|
|
|
errors.extend(self._validate_arg_type(positional_filter_node, 'boolean', len(positional_nodes), 'add', is_in_group_arg))
|
|
|
if named_filter_nodes:
|
|
|
errors.extend(self._validate_arg_type(named_filter_nodes[0], 'boolean', len(positional_nodes), 'add', is_in_group_arg))
|
|
|
|
|
|
return errors
|
|
|
|
|
|
def validate_ast(self, ast: Optional[ASTNode], is_in_group_arg: bool = False) -> List[str]:
|
|
|
"""递归验证抽象语法树"""
|
|
|
if not ast:
|
|
|
return ["无法解析表达式"]
|
|
|
|
|
|
errors = []
|
|
|
|
|
|
if ast.node_type == 'function':
|
|
|
is_group_function = ast.value.startswith('group_')
|
|
|
current_in_group_arg = is_in_group_arg or is_group_function
|
|
|
function_errors = self.validate_function(ast, current_in_group_arg)
|
|
|
errors.extend(function_errors)
|
|
|
|
|
|
for child in ast.children:
|
|
|
if isinstance(child, dict):
|
|
|
if 'value' in child and hasattr(child['value'], 'node_type'):
|
|
|
child_errors = self.validate_ast(child['value'], current_in_group_arg)
|
|
|
errors.extend(child_errors)
|
|
|
elif hasattr(child, 'node_type'):
|
|
|
child_errors = self.validate_ast(child, current_in_group_arg)
|
|
|
errors.extend(child_errors)
|
|
|
elif ast.node_type in ['unop', 'binop']:
|
|
|
for child in ast.children:
|
|
|
if hasattr(child, 'node_type'):
|
|
|
child_errors = self.validate_ast(child, is_in_group_arg)
|
|
|
errors.extend(child_errors)
|
|
|
elif ast.node_type == 'field':
|
|
|
if not self._is_valid_field(ast.value):
|
|
|
errors.append(f"无效的字段名: {ast.value}")
|
|
|
else:
|
|
|
for child in ast.children:
|
|
|
if isinstance(child, dict):
|
|
|
if 'value' in child and hasattr(child['value'], 'node_type'):
|
|
|
child_errors = self.validate_ast(child['value'], is_in_group_arg)
|
|
|
errors.extend(child_errors)
|
|
|
elif hasattr(child, 'node_type'):
|
|
|
child_errors = self.validate_ast(child, is_in_group_arg)
|
|
|
errors.extend(child_errors)
|
|
|
|
|
|
return errors
|
|
|
|
|
|
def _process_semicolon_expression(self, expression: str) -> Tuple[bool, str]:
|
|
|
"""处理带有分号的表达式,将其转换为不带分号的简化形式"""
|
|
|
|
|
|
def _top_level_equals_positions(stmt: str) -> List[int]:
|
|
|
positions: List[int] = []
|
|
|
paren_depth = 0
|
|
|
bracket_depth = 0
|
|
|
brace_depth = 0
|
|
|
in_single_quote = False
|
|
|
in_double_quote = False
|
|
|
escape = False
|
|
|
|
|
|
for i, ch in enumerate(stmt):
|
|
|
if escape:
|
|
|
escape = False
|
|
|
continue
|
|
|
if ch == '\\':
|
|
|
escape = True
|
|
|
continue
|
|
|
|
|
|
if in_single_quote:
|
|
|
if ch == "'":
|
|
|
in_single_quote = False
|
|
|
continue
|
|
|
if in_double_quote:
|
|
|
if ch == '"':
|
|
|
in_double_quote = False
|
|
|
continue
|
|
|
|
|
|
if ch == "'":
|
|
|
in_single_quote = True
|
|
|
continue
|
|
|
if ch == '"':
|
|
|
in_double_quote = True
|
|
|
continue
|
|
|
|
|
|
if ch == '(':
|
|
|
paren_depth += 1
|
|
|
continue
|
|
|
if ch == ')':
|
|
|
paren_depth = max(0, paren_depth - 1)
|
|
|
continue
|
|
|
if ch == '[':
|
|
|
bracket_depth += 1
|
|
|
continue
|
|
|
if ch == ']':
|
|
|
bracket_depth = max(0, bracket_depth - 1)
|
|
|
continue
|
|
|
if ch == '{':
|
|
|
brace_depth += 1
|
|
|
continue
|
|
|
if ch == '}':
|
|
|
brace_depth = max(0, brace_depth - 1)
|
|
|
continue
|
|
|
|
|
|
if paren_depth or bracket_depth or brace_depth:
|
|
|
continue
|
|
|
|
|
|
if ch != '=':
|
|
|
continue
|
|
|
|
|
|
prev_ch = stmt[i - 1] if i > 0 else ''
|
|
|
next_ch = stmt[i + 1] if i + 1 < len(stmt) else ''
|
|
|
if prev_ch in ['=', '!', '<', '>'] or next_ch == '=':
|
|
|
continue
|
|
|
|
|
|
positions.append(i)
|
|
|
|
|
|
return positions
|
|
|
|
|
|
def _keyword_arg_names(stmt: str):
|
|
|
names = set()
|
|
|
paren_depth = 0
|
|
|
bracket_depth = 0
|
|
|
brace_depth = 0
|
|
|
in_single_quote = False
|
|
|
in_double_quote = False
|
|
|
escape = False
|
|
|
|
|
|
i = 0
|
|
|
while i < len(stmt):
|
|
|
ch = stmt[i]
|
|
|
|
|
|
if escape:
|
|
|
escape = False
|
|
|
i += 1
|
|
|
continue
|
|
|
if ch == '\\':
|
|
|
escape = True
|
|
|
i += 1
|
|
|
continue
|
|
|
|
|
|
if in_single_quote:
|
|
|
if ch == "'":
|
|
|
in_single_quote = False
|
|
|
i += 1
|
|
|
continue
|
|
|
if in_double_quote:
|
|
|
if ch == '"':
|
|
|
in_double_quote = False
|
|
|
i += 1
|
|
|
continue
|
|
|
|
|
|
if ch == "'":
|
|
|
in_single_quote = True
|
|
|
i += 1
|
|
|
continue
|
|
|
if ch == '"':
|
|
|
in_double_quote = True
|
|
|
i += 1
|
|
|
continue
|
|
|
|
|
|
if ch == '(':
|
|
|
paren_depth += 1
|
|
|
i += 1
|
|
|
continue
|
|
|
if ch == ')':
|
|
|
paren_depth = max(0, paren_depth - 1)
|
|
|
i += 1
|
|
|
continue
|
|
|
if ch == '[':
|
|
|
bracket_depth += 1
|
|
|
i += 1
|
|
|
continue
|
|
|
if ch == ']':
|
|
|
bracket_depth = max(0, bracket_depth - 1)
|
|
|
i += 1
|
|
|
continue
|
|
|
if ch == '{':
|
|
|
brace_depth += 1
|
|
|
i += 1
|
|
|
continue
|
|
|
if ch == '}':
|
|
|
brace_depth = max(0, brace_depth - 1)
|
|
|
i += 1
|
|
|
continue
|
|
|
|
|
|
inside_container = bool(paren_depth or bracket_depth or brace_depth)
|
|
|
|
|
|
if inside_container and (ch.isalpha() or ch == '_'):
|
|
|
start = i
|
|
|
i += 1
|
|
|
while i < len(stmt) and (stmt[i].isalnum() or stmt[i] == '_'):
|
|
|
i += 1
|
|
|
name = stmt[start:i]
|
|
|
|
|
|
j = i
|
|
|
while j < len(stmt) and stmt[j].isspace():
|
|
|
j += 1
|
|
|
|
|
|
if j < len(stmt) and stmt[j] == '=':
|
|
|
next_ch = stmt[j + 1] if j + 1 < len(stmt) else ''
|
|
|
if next_ch != '=':
|
|
|
names.add(name.lower())
|
|
|
continue
|
|
|
|
|
|
i += 1
|
|
|
|
|
|
return names
|
|
|
|
|
|
if expression.strip().endswith(';'):
|
|
|
return False, "表达式不能以分号结尾"
|
|
|
|
|
|
statements = [stmt.strip() for stmt in expression.split(';') if stmt.strip()]
|
|
|
if not statements:
|
|
|
return False, "表达式不能为空"
|
|
|
|
|
|
variables = {}
|
|
|
|
|
|
for i, stmt in enumerate(statements[:-1]):
|
|
|
eq_positions = _top_level_equals_positions(stmt)
|
|
|
if not eq_positions:
|
|
|
return False, f"第{i + 1}个语句必须是赋值语句(使用=符号)"
|
|
|
if len(eq_positions) > 1:
|
|
|
return False, f"第{i + 1}个语句只能包含一个赋值符号(=)"
|
|
|
|
|
|
real_equals_pos = eq_positions[0]
|
|
|
|
|
|
var_name = stmt[:real_equals_pos].strip()
|
|
|
var_value = stmt[real_equals_pos + 1:].strip()
|
|
|
|
|
|
if not re.match(r'^[a-zA-Z_][a-zA-Z0-9_]*$', var_name):
|
|
|
return False, f"第{i + 1}个语句的变量名'{var_name}'无效,只能包含字母、数字和下划线,且不能以数字开头"
|
|
|
|
|
|
var_name_lower = var_name.lower()
|
|
|
|
|
|
kw_names = _keyword_arg_names(var_value)
|
|
|
used_vars = re.findall(r'\b[a-zA-Z_][a-zA-Z0-9_]*\b', var_value)
|
|
|
for used_var in used_vars:
|
|
|
used_var_lower = used_var.lower()
|
|
|
if used_var_lower in kw_names:
|
|
|
continue
|
|
|
if used_var_lower not in variables:
|
|
|
if used_var not in supported_functions:
|
|
|
if len(used_var) <= 2:
|
|
|
return False, f"第{i + 1}个语句中使用的变量'{used_var}'未在之前定义"
|
|
|
elif not self._is_valid_field(used_var):
|
|
|
return False, f"第{i + 1}个语句中使用的变量'{used_var}'未在之前定义"
|
|
|
|
|
|
for existing_var, existing_val in variables.items():
|
|
|
var_value = re.sub(rf'\b{existing_var}\b', existing_val, var_value)
|
|
|
|
|
|
variables[var_name_lower] = var_value
|
|
|
|
|
|
final_stmt = statements[-1]
|
|
|
|
|
|
if _top_level_equals_positions(final_stmt):
|
|
|
return False, "最后一个语句不能是赋值语句"
|
|
|
|
|
|
kw_names = _keyword_arg_names(final_stmt)
|
|
|
used_vars = re.findall(r'\b[a-zA-Z_][a-zA-Z0-9_]*\b', final_stmt)
|
|
|
for used_var in used_vars:
|
|
|
used_var_lower = used_var.lower()
|
|
|
if used_var_lower in kw_names:
|
|
|
continue
|
|
|
if used_var_lower not in variables:
|
|
|
if used_var not in supported_functions:
|
|
|
if len(used_var) <= 2:
|
|
|
return False, f"最后一个语句中使用的变量'{used_var}'未在之前定义"
|
|
|
|
|
|
if self._is_valid_field(used_var) or used_var_lower in valid_categories or used_var_lower in group_fields:
|
|
|
continue
|
|
|
|
|
|
return False, f"最后一个语句中使用的变量'{used_var}'未在之前定义"
|
|
|
|
|
|
final_expr = final_stmt
|
|
|
for var_name, var_value in variables.items():
|
|
|
final_expr = re.sub(rf'\b{var_name}\b', var_value, final_expr)
|
|
|
|
|
|
return True, final_expr
|
|
|
|
|
|
def check_expression(self, expression: str) -> Dict[str, Any]:
|
|
|
"""检查表达式格式是否正确"""
|
|
|
self.errors = []
|
|
|
self._unit_cache = {}
|
|
|
self._derived_category_cache = {}
|
|
|
|
|
|
try:
|
|
|
expression = expression.strip()
|
|
|
if not expression:
|
|
|
return {
|
|
|
'valid': False,
|
|
|
'errors': ['表达式不能为空'],
|
|
|
'tokens': [],
|
|
|
'ast': None
|
|
|
}
|
|
|
|
|
|
if ';' in expression:
|
|
|
success, result = self._process_semicolon_expression(expression)
|
|
|
if not success:
|
|
|
return {
|
|
|
'valid': False,
|
|
|
'errors': [result],
|
|
|
'tokens': [],
|
|
|
'ast': None
|
|
|
}
|
|
|
expression = result
|
|
|
|
|
|
self.lexer.lineno = 1
|
|
|
|
|
|
self.lexer.input(expression)
|
|
|
tokens = []
|
|
|
for token in self.lexer:
|
|
|
tokens.append(token)
|
|
|
|
|
|
self.lexer.input(expression)
|
|
|
self.lexer.lineno = 1
|
|
|
|
|
|
ast = self.parser.parse(expression, lexer=self.lexer)
|
|
|
|
|
|
validation_errors = self.validate_ast(ast)
|
|
|
|
|
|
all_errors = self.errors + validation_errors
|
|
|
|
|
|
bracket_count = 0
|
|
|
for char in expression:
|
|
|
if char == '(':
|
|
|
bracket_count += 1
|
|
|
elif char == ')':
|
|
|
bracket_count -= 1
|
|
|
if bracket_count < 0:
|
|
|
all_errors.append("括号不匹配: 右括号过多")
|
|
|
break
|
|
|
if bracket_count > 0:
|
|
|
all_errors.append("括号不匹配: 左括号过多")
|
|
|
|
|
|
return {
|
|
|
'valid': len(all_errors) == 0,
|
|
|
'errors': all_errors,
|
|
|
'tokens': tokens,
|
|
|
'ast': ast
|
|
|
}
|
|
|
except Exception as e:
|
|
|
return {
|
|
|
'valid': False,
|
|
|
'errors': [f"解析错误: {str(e)}"],
|
|
|
'tokens': [],
|
|
|
'ast': None
|
|
|
}
|
|
|
|
|
|
def validate_expressions(self, expressions: List[str]) -> Tuple[List[str], List[Dict]]:
|
|
|
"""批量验证表达式"""
|
|
|
valid_expressions = []
|
|
|
invalid_details = []
|
|
|
|
|
|
for expr in expressions:
|
|
|
if not isinstance(expr, str) or not expr.strip():
|
|
|
invalid_details.append({
|
|
|
'expression': expr,
|
|
|
'error': '空表达式'
|
|
|
})
|
|
|
continue
|
|
|
|
|
|
result = self.check_expression(expr.strip())
|
|
|
if result['valid']:
|
|
|
valid_expressions.append(expr.strip())
|
|
|
else:
|
|
|
invalid_details.append({
|
|
|
'expression': expr,
|
|
|
'errors': result['errors']
|
|
|
})
|
|
|
|
|
|
return valid_expressions, invalid_details
|
|
|
|
|
|
|
|
|
# ==================== 主处理流程 ====================
|
|
|
|
|
|
def process(data_sets_list: List[Dict], llm_template: str) -> Dict:
|
|
|
"""
|
|
|
处理 LLM 生成的模板,生成 Alpha 表达式
|
|
|
|
|
|
参数:
|
|
|
data_sets_list: 数据集字段列表,格式 [{'id': 'field_name'}, ...]
|
|
|
llm_template: LLM 生成的 Markdown 文本
|
|
|
|
|
|
返回:
|
|
|
包含生成结果的字典
|
|
|
"""
|
|
|
# 提取 field_ids
|
|
|
field_ids = [item['id'] for item in data_sets_list]
|
|
|
|
|
|
# 提取 dataset_ids 和 allowed_suffixes
|
|
|
dataset_ids = field_ids
|
|
|
allowed_suffixes = build_allowed_metric_suffixes(field_ids)
|
|
|
dataset_code = detect_dataset_code(dataset_ids)
|
|
|
|
|
|
print(f" 数据集字段数: {len(dataset_ids)}")
|
|
|
print(f" 提取的后缀数: {len(allowed_suffixes)}")
|
|
|
print(f" 数据集代码: {dataset_code}")
|
|
|
|
|
|
# Step 1: 提取 template blocks
|
|
|
print("\n🔍 提取 Template Blocks...")
|
|
|
block_pairs = extract_template_blocks(llm_template)
|
|
|
print(f" 提取到 {len(block_pairs)} 个 Concept blocks")
|
|
|
|
|
|
if not block_pairs:
|
|
|
print("⚠️ 没有找到有效的 Concept blocks")
|
|
|
return {
|
|
|
"success": False,
|
|
|
"error": "没有找到有效的 Concept blocks",
|
|
|
"expressions": [],
|
|
|
"templates": []
|
|
|
}
|
|
|
|
|
|
# Step 2: 规范化 templates
|
|
|
print("\n📝 规范化 Templates...")
|
|
|
normalized_pairs = []
|
|
|
for item in block_pairs:
|
|
|
template = str(item.get("template") or "").strip()
|
|
|
idea_text = str(item.get("idea") or "").strip()
|
|
|
|
|
|
if not template:
|
|
|
continue
|
|
|
|
|
|
normalized_t, ok = normalize_template_placeholders(
|
|
|
template, dataset_ids, allowed_suffixes, dataset_code
|
|
|
)
|
|
|
|
|
|
# 允许没有 {placeholder} 的模板通过
|
|
|
# if ok:
|
|
|
normalized_pairs.append({
|
|
|
"template": normalized_t,
|
|
|
"idea": idea_text,
|
|
|
"original_template": template
|
|
|
})
|
|
|
print(f" ✓ {normalized_t[:60]}...")
|
|
|
# else:
|
|
|
# print(f" ✗ 跳过: {template[:60]}...")
|
|
|
|
|
|
print(f" 规范化成功: {len(normalized_pairs)} 个")
|
|
|
|
|
|
# Step 3: 生成表达式
|
|
|
print("\n🎯 生成 Alpha 表达式...")
|
|
|
all_expressions = []
|
|
|
template_results = []
|
|
|
|
|
|
for idx, pair in enumerate(normalized_pairs, 1):
|
|
|
template = pair["template"]
|
|
|
expressions = match_single_horizon_auto(field_ids, template)
|
|
|
|
|
|
# 如果没有生成表达式(没有{placeholder}),直接把模板当作表达式
|
|
|
if not expressions and template:
|
|
|
expressions = [template]
|
|
|
|
|
|
if expressions:
|
|
|
result_item = {
|
|
|
"template": template,
|
|
|
"original_template": pair.get("original_template", template),
|
|
|
"idea": pair.get("idea", ""),
|
|
|
"expression_count": len(expressions),
|
|
|
"expressions": expressions
|
|
|
}
|
|
|
template_results.append(result_item)
|
|
|
all_expressions.extend(expressions)
|
|
|
|
|
|
print(f" Idea {idx}: {len(expressions):4d} 个表达式 - {template[:50]}...")
|
|
|
|
|
|
print(f"\n 总计生成: {len(all_expressions)} 个表达式")
|
|
|
|
|
|
# Step 4: 验证表达式
|
|
|
print("\n🔍 验证表达式...")
|
|
|
validator = ExpressionValidator()
|
|
|
valid_expressions, invalid_details = validator.validate_expressions(all_expressions)
|
|
|
invalid_count = len(invalid_details)
|
|
|
|
|
|
print(f" 有效: {len(valid_expressions)}")
|
|
|
print(f" 无效: {invalid_count}")
|
|
|
|
|
|
if invalid_details:
|
|
|
print(f" 无效表达式详情:")
|
|
|
for detail in invalid_details[:5]:
|
|
|
print(f" - {detail['expression'][:50]}...: {detail.get('errors', detail.get('error', '未知错误'))}")
|
|
|
if len(invalid_details) > 5:
|
|
|
print(f" ... 还有 {len(invalid_details) - 5} 个无效表达式")
|
|
|
|
|
|
# Step 5: 去重
|
|
|
print("\n🧹 去重...")
|
|
|
unique_expressions = []
|
|
|
seen = set()
|
|
|
for expr in valid_expressions:
|
|
|
if expr not in seen:
|
|
|
unique_expressions.append(expr)
|
|
|
seen.add(expr)
|
|
|
|
|
|
print(f" 去重后: {len(unique_expressions)} 个")
|
|
|
|
|
|
# Step 6: 过滤包含英文逗号的表达式和模板 过滤表达式:只保留包含英文逗号的
|
|
|
print("\n🧹 过滤包含英文逗号的表达式和模板...")
|
|
|
filtered_expressions = [expr for expr in unique_expressions if ',' in expr]
|
|
|
print(f" 过滤后表达式: {len(filtered_expressions)} 个 (原 {len(unique_expressions)} 个)")
|
|
|
|
|
|
# 过滤模板:只保留包含英文逗号的模板,且其表达式也要包含英文逗号
|
|
|
filtered_template_results = []
|
|
|
for template_item in template_results:
|
|
|
template = template_item.get("template", "")
|
|
|
# 检查模板是否包含英文逗号
|
|
|
if ',' not in template:
|
|
|
continue
|
|
|
# 过滤表达式:只保留包含英文逗号的
|
|
|
filtered_exprs = [expr for expr in template_item.get("expressions", []) if ',' in expr]
|
|
|
if filtered_exprs:
|
|
|
filtered_item = template_item.copy()
|
|
|
filtered_item["expressions"] = filtered_exprs
|
|
|
filtered_item["expression_count"] = len(filtered_exprs)
|
|
|
filtered_template_results.append(filtered_item)
|
|
|
|
|
|
print(f" 过滤后模板: {len(filtered_template_results)} 个 (原 {len(template_results)} 个)")
|
|
|
|
|
|
print("\n✅ 处理完成!")
|
|
|
|
|
|
# 打印统计
|
|
|
print("\n" + "=" * 50)
|
|
|
print("处理统计:")
|
|
|
print("=" * 50)
|
|
|
summary = {
|
|
|
"total_templates": len(block_pairs),
|
|
|
"normalized_templates": len(normalized_pairs),
|
|
|
"total_expressions": len(all_expressions),
|
|
|
"valid_expressions": len(valid_expressions),
|
|
|
"unique_expressions": len(unique_expressions),
|
|
|
"invalid_count": invalid_count
|
|
|
}
|
|
|
for key, value in summary.items():
|
|
|
print(f" {key}: {value}")
|
|
|
|
|
|
# 返回结果字典
|
|
|
return {
|
|
|
"success": True,
|
|
|
"summary": summary,
|
|
|
"templates": filtered_template_results,
|
|
|
"expressions": filtered_expressions,
|
|
|
"invalid_details": invalid_details
|
|
|
}
|
|
|
|