You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
 
 
 

432 lines
19 KiB

from flask import Blueprint, render_template, request, jsonify
import requests
import json
import logging
import threading
from concurrent.futures import ThreadPoolExecutor, as_completed
inspiration_house_bp = Blueprint('inspiration_house', __name__)
# Configure logging
logging.basicConfig(level=logging.INFO)
logger = logging.getLogger(__name__)
@inspiration_house_bp.route('/')
def inspiration_house():
"""Main inspiration house page"""
return render_template('inspiration_house.html')
@inspiration_house_bp.route('/api/test-deepseek', methods=['POST'])
def test_deepseek_api():
"""Test API connection for both Deepseek and Kimi"""
try:
api_key = request.headers.get('X-API-Key')
if not api_key:
return jsonify({'success': False, 'error': 'API key is required'}), 400
data = request.get_json()
provider = data.get('provider', 'deepseek')
model_name = data.get('model_name', 'deepseek-chat')
# Test API with a simple request
headers = {
'Authorization': f'Bearer {api_key}',
'Content-Type': 'application/json'
}
test_data = {
'model': model_name,
'messages': [
{'role': 'user', 'content': 'Hello, this is a test message.'}
],
'max_tokens': 10
}
# Choose API endpoint based on provider
if provider == 'kimi':
# Use the correct Kimi API endpoint that was tested and confirmed working
api_url = 'https://api.moonshot.cn/v1/chat/completions'
kimi_data = {
'model': model_name,
'messages': [
{'role': 'user', 'content': 'Hello, this is a test message.'}
],
'max_tokens': 10,
'stream': False
}
logger.info(f"Testing Kimi API with correct endpoint: {api_url}")
response = requests.post(api_url, headers=headers, json=kimi_data, timeout=10)
logger.info(f"Kimi response status: {response.status_code}")
else:
# Default to Deepseek
api_url = 'https://api.deepseek.com/chat/completions'
response = requests.post(api_url, headers=headers, json=test_data, timeout=10)
if response.status_code == 200:
return jsonify({'success': True, 'message': f'{provider.capitalize()} API connection successful'})
else:
error_detail = response.text
return jsonify({'success': False, 'error': f'API returned status {response.status_code}: {error_detail}'}), 400
except requests.exceptions.RequestException as e:
logger.error(f"API test error: {str(e)}")
return jsonify({'success': False, 'error': f'Network error: {str(e)}'}), 500
except Exception as e:
logger.error(f"Unexpected error in API test: {str(e)}")
return jsonify({'success': False, 'error': f'Unexpected error: {str(e)}'}), 500
@inspiration_house_bp.route('/api/evaluate-operator', methods=['POST'])
def evaluate_operator():
"""Evaluate a single operator against the research target"""
try:
api_key = request.headers.get('X-API-Key')
if not api_key:
return jsonify({'success': False, 'error': 'API key is required'}), 400
data = request.get_json()
operator = data.get('operator', {})
research_target = data.get('research_target', '')
current_expression = data.get('current_expression', '')
expression_context = data.get('expression_context', '')
provider = data.get('provider', 'deepseek')
model_name = data.get('model_name', 'deepseek-chat')
if not operator or not research_target:
return jsonify({'success': False, 'error': 'Operator and research target are required'}), 400
# Build the evaluation prompt
system_prompt = get_evaluation_system_prompt()
user_prompt = build_evaluation_prompt(operator, research_target, current_expression, expression_context)
# Make API call based on provider
headers = {
'Authorization': f'Bearer {api_key}',
'Content-Type': 'application/json'
}
if provider == 'kimi':
# Use the correct Kimi API endpoint that was tested and confirmed working
api_url = 'https://api.moonshot.cn/v1/chat/completions'
api_data = {
'model': model_name,
'messages': [
{'role': 'system', 'content': system_prompt},
{'role': 'user', 'content': user_prompt}
],
'max_tokens': 4096,
'temperature': 0.3
}
response = requests.post(api_url, headers=headers, json=api_data, timeout=30)
else:
# Deepseek API structure
api_url = 'https://api.deepseek.com/chat/completions'
api_data = {
'model': model_name,
'messages': [
{'role': 'system', 'content': system_prompt},
{'role': 'user', 'content': user_prompt}
],
'max_tokens': 4096,
'temperature': 0.3
}
response = requests.post(api_url, headers=headers, json=api_data, timeout=30)
if response.status_code == 200:
response_data = response.json()
# Both Kimi and Deepseek use the same OpenAI-compatible response structure
ai_response = response_data['choices'][0]['message']['content']
# Parse the AI response to extract score and reason
score, reason = parse_evaluation_response(ai_response)
return jsonify({
'success': True,
'score': score,
'reason': reason,
'operator': operator.get('name', 'Unknown'),
'research_target': research_target
})
else:
error_detail = response.text
logger.error(f"{provider.capitalize()} API error: {response.status_code} - {error_detail}")
return jsonify({'success': False, 'error': f'API returned status {response.status_code}: {error_detail}'}), 400
except requests.exceptions.RequestException as e:
logger.error(f"API request error: {str(e)}")
return jsonify({'success': False, 'error': f'Network error: {str(e)}'}), 500
except Exception as e:
logger.error(f"Unexpected error in operator evaluation: {str(e)}")
return jsonify({'success': False, 'error': f'Unexpected error: {str(e)}'}), 500
def get_evaluation_system_prompt():
"""Get the system prompt for operator evaluation"""
return """You are an expert quantitative researcher and data scientist specializing in financial data analysis and algorithmic trading. Your task is to evaluate how well a specific BRAIN operator could be ADDED to the current expression to help achieve a given quant research target.
IMPORTANT: You are NOT suggesting replacements for the current expression. You are evaluating how each operator could be COMBINED with or applied AFTER the current expression to move closer to the target.
Your evaluation should be based on:
1. How the operator could be applied AFTER the current expression (e.g., ts_rank(current_expression))
2. The operator's mathematical and statistical properties when combined with the current expression
3. How this combination would better achieve the stated research target
4. The synergistic effects of applying this operator to the current expression
You must provide:
1. A score from 0-10 (where 10 is perfect synergistic addition to achieve the target)
2. A clear, concise reason explaining how this operator could be added to improve the current expression
Scoring Guidelines:
- 8-10: Excellent addition - operator would significantly enhance the current expression for the target
- 6-7: Good addition - operator would meaningfully improve the current expression
- 4-5: Moderate addition - operator could provide some benefit when combined
- 2-3: Weak addition - operator would add little value to the current expression
- 0-1: No value - operator would not help or might even hurt the current expression
Response Format:
Score: [number from 0-10]
Reason: [provide a detailed explanation of how this operator could be added to the current expression to better achieve the target. Include specific examples of how the combination would work and why it would be effective.]
Focus on combination effects and how the operator would enhance the current expression rather than replace it. Be thorough in your explanation."""
def build_evaluation_prompt(operator, research_target, current_expression, expression_context):
"""Build the user prompt for operator evaluation"""
operator_info = f"""
Operator Name: {operator.get('name', 'Unknown')}
Operator Category: {operator.get('category', 'Unknown')}
Operator Description: {operator.get('description', 'No description available')}
"""
context_info = f"""
Research Target: {research_target}
Current Expression: {current_expression if current_expression else 'None provided'}
Expression Context: {expression_context if expression_context else 'None provided'}
"""
return f"""Please evaluate how well this BRAIN operator could be ADDED to the current expression to help achieve the research target.
{operator_info}
{context_info}
IMPORTANT: Consider how this operator could be applied AFTER the current expression (e.g., {operator.get('name', 'operator')}(current_expression)) to enhance the strategy. Do NOT suggest replacing the current expression.
Think about:
- How would applying this operator to the current expression improve the strategy?
- What synergistic effects would this combination create?
- How would this addition move us closer to the research target?
Response Format:
Score: [number from 0-10]
Reason: [provide a detailed explanation of how this operator could be added to the current expression to better achieve the target. Include specific examples of how the combination would work and why it would be effective.]"""
def parse_evaluation_response(response):
"""Parse the AI response to extract score and reason"""
try:
lines = response.strip().split('\n')
score = 0
reason = "No reason provided"
in_reason_section = False
reason_lines = []
for line in lines:
line = line.strip()
if line.lower().startswith('score:'):
try:
score_text = line.split(':', 1)[1].strip()
score = int(score_text)
# Ensure score is within valid range
score = max(0, min(10, score))
except (ValueError, IndexError):
score = 0
elif line.lower().startswith('reason:'):
in_reason_section = True
reason_text = line.split(':', 1)[1].strip()
if reason_text:
reason_lines.append(reason_text)
elif in_reason_section and line:
# Continue collecting reason lines until we hit another section or empty line
if line.lower().startswith(('score:', 'operator:', 'category:')):
break
reason_lines.append(line)
if reason_lines:
reason = ' '.join(reason_lines)
return score, reason
except Exception as e:
logger.error(f"Error parsing evaluation response: {str(e)}")
return 0, f"Error parsing response: {str(e)}"
@inspiration_house_bp.route('/api/batch-evaluate', methods=['POST'])
def batch_evaluate_operators():
"""Evaluate multiple operators in parallel"""
try:
api_key = request.headers.get('X-API-Key')
if not api_key:
return jsonify({'success': False, 'error': 'API key is required'}), 400
data = request.get_json()
operators = data.get('operators', [])
research_target = data.get('research_target', '')
current_expression = data.get('current_expression', '')
expression_context = data.get('expression_context', '')
batch_size = data.get('batch_size', 100) # Get batch size from request
provider = data.get('provider', 'deepseek')
model_name = data.get('model_name', 'deepseek-chat')
if not operators or not research_target:
return jsonify({'success': False, 'error': 'Operators and research target are required'}), 400
# Use ThreadPoolExecutor for parallel processing
# Use full parallelization for all providers
max_workers = len(operators)
logger.info(f"Using {max_workers} workers for API evaluation")
results = []
with ThreadPoolExecutor(max_workers=max_workers) as executor:
# Submit all evaluation tasks
future_to_operator = {
executor.submit(
evaluate_single_operator_async,
api_key,
operator,
research_target,
current_expression,
expression_context,
provider,
model_name
): operator for operator in operators
}
# Collect results as they complete
for future in as_completed(future_to_operator):
operator = future_to_operator[future]
try:
result = future.result()
results.append(result)
except Exception as e:
logger.error(f"Error evaluating operator {operator.get('name', 'Unknown')}: {str(e)}")
results.append({
'operator': operator.get('name', 'Unknown'),
'category': operator.get('category', 'Unknown'),
'score': 0,
'reason': f'Error: {str(e)}',
'timestamp': None
})
return jsonify({
'success': True,
'results': results,
'total_evaluated': len(results),
'workers_used': max_workers,
'total_operators': len(operators)
})
except Exception as e:
logger.error(f"Unexpected error in batch evaluation: {str(e)}")
return jsonify({'success': False, 'error': f'Unexpected error: {str(e)}'}), 500
def evaluate_single_operator_async(api_key, operator, research_target, current_expression, expression_context, provider='deepseek', model_name='deepseek-chat'):
"""Evaluate a single operator asynchronously"""
try:
# Build the evaluation prompt
system_prompt = get_evaluation_system_prompt()
user_prompt = build_evaluation_prompt(operator, research_target, current_expression, expression_context)
# Make API call based on provider
headers = {
'Authorization': f'Bearer {api_key}',
'Content-Type': 'application/json'
}
if provider == 'kimi':
# Use the correct Kimi API endpoint that was tested and confirmed working
api_url = 'https://api.moonshot.cn/v1/chat/completions'
api_data = {
'model': model_name,
'messages': [
{'role': 'system', 'content': system_prompt},
{'role': 'user', 'content': user_prompt}
],
'max_tokens': 4096,
'temperature': 0.3
}
response = requests.post(api_url, headers=headers, json=api_data, timeout=30)
else:
# Deepseek API structure
api_url = 'https://api.deepseek.com/chat/completions'
api_data = {
'model': model_name,
'messages': [
{'role': 'system', 'content': system_prompt},
{'role': 'user', 'content': user_prompt}
],
'max_tokens': 4096,
'temperature': 0.3
}
response = requests.post(api_url, headers=headers, json=api_data, timeout=30)
if response.status_code == 200:
response_data = response.json()
# Both Kimi and Deepseek use the same OpenAI-compatible response structure
ai_response = response_data['choices'][0]['message']['content']
# Parse the AI response to extract score and reason
score, reason = parse_evaluation_response(ai_response)
return {
'operator': operator.get('name', 'Unknown'),
'category': operator.get('category', 'Unknown'),
'score': score,
'reason': reason,
'timestamp': None
}
else:
error_detail = response.text
logger.error(f"{provider.capitalize()} API error for operator {operator.get('name', 'Unknown')}: {response.status_code} - {error_detail}")
return {
'operator': operator.get('name', 'Unknown'),
'category': operator.get('category', 'Unknown'),
'score': 0,
'reason': f'API Error: {response.status_code}',
'timestamp': None
}
except requests.exceptions.Timeout as e:
logger.error(f"Timeout error evaluating operator {operator.get('name', 'Unknown')}: {str(e)}")
return {
'operator': operator.get('name', 'Unknown'),
'category': operator.get('category', 'Unknown'),
'score': 0,
'reason': f'API Timeout: Request took too long to complete',
'timestamp': None
}
except requests.exceptions.RequestException as e:
logger.error(f"Request error evaluating operator {operator.get('name', 'Unknown')}: {str(e)}")
return {
'operator': operator.get('name', 'Unknown'),
'category': operator.get('category', 'Unknown'),
'score': 0,
'reason': f'Network Error: {str(e)}',
'timestamp': None
}
except Exception as e:
logger.error(f"Error evaluating operator {operator.get('name', 'Unknown')}: {str(e)}")
return {
'operator': operator.get('name', 'Unknown'),
'category': operator.get('category', 'Unknown'),
'score': 0,
'reason': f'Error: {str(e)}',
'timestamp': None
}