This commit is contained in:
Dobromir Popov
2025-07-14 17:56:09 +03:00
parent d53a2ba75d
commit 4a55c5ff03
11 changed files with 1509 additions and 118 deletions

View File

@ -737,6 +737,44 @@ class DQNAgent:
return None
def _safe_cnn_forward(self, network, states):
"""Safely call CNN forward method ensuring we always get 5 return values"""
try:
result = network(states)
if isinstance(result, tuple) and len(result) == 5:
return result
elif isinstance(result, tuple) and len(result) == 1:
# Handle case where only q_values are returned (like in empty tensor case)
q_values = result[0]
batch_size = q_values.size(0)
device = q_values.device
default_extrema = torch.zeros(batch_size, 3, device=device)
default_price = torch.zeros(batch_size, 1, device=device)
default_features = torch.zeros(batch_size, 1024, device=device)
default_advanced = torch.zeros(batch_size, 1, device=device)
return q_values, default_extrema, default_price, default_features, default_advanced
else:
# Fallback: create all default tensors
batch_size = states.size(0)
device = states.device
default_q_values = torch.zeros(batch_size, self.n_actions, device=device)
default_extrema = torch.zeros(batch_size, 3, device=device)
default_price = torch.zeros(batch_size, 1, device=device)
default_features = torch.zeros(batch_size, 1024, device=device)
default_advanced = torch.zeros(batch_size, 1, device=device)
return default_q_values, default_extrema, default_price, default_features, default_advanced
except Exception as e:
logger.error(f"Error in CNN forward pass: {e}")
# Fallback: create all default tensors
batch_size = states.size(0)
device = states.device
default_q_values = torch.zeros(batch_size, self.n_actions, device=device)
default_extrema = torch.zeros(batch_size, 3, device=device)
default_price = torch.zeros(batch_size, 1, device=device)
default_features = torch.zeros(batch_size, 1024, device=device)
default_advanced = torch.zeros(batch_size, 1, device=device)
return default_q_values, default_extrema, default_price, default_features, default_advanced
def replay(self, experiences=None):
"""Train the model using experiences from memory"""
@ -995,17 +1033,17 @@ class DQNAgent:
else:
raise ValueError("Invalid arguments to _replay_standard")
# Get current Q values
current_q_values, current_extrema_pred, current_price_pred, hidden_features, current_advanced_pred = self.policy_net(states)
# Get current Q values using safe wrapper
current_q_values, current_extrema_pred, current_price_pred, hidden_features, current_advanced_pred = self._safe_cnn_forward(self.policy_net, states)
current_q_values = current_q_values.gather(1, actions.unsqueeze(1)).squeeze(1)
# Enhanced Double DQN implementation
with torch.no_grad():
if self.use_double_dqn:
# Double DQN: Use policy network to select actions, target network to evaluate
policy_q_values, _, _, _, _ = self.policy_net(next_states)
policy_q_values, _, _, _, _ = self._safe_cnn_forward(self.policy_net, next_states)
next_actions = policy_q_values.argmax(1)
target_q_values_all, _, _, _, _ = self.target_net(next_states)
target_q_values_all, _, _, _, _ = self._safe_cnn_forward(self.target_net, next_states)
next_q_values = target_q_values_all.gather(1, next_actions.unsqueeze(1)).squeeze(1)
else:
# Standard DQN: Use target network for both selection and evaluation

View File

@ -395,16 +395,24 @@ class EnhancedCNN(nn.Module):
# Validate input dimensions to prevent zero-element tensor issues
if x.numel() == 0:
logger.error(f"Forward pass received empty tensor with shape {x.shape}")
# Return default output to prevent crash
default_output = torch.zeros(batch_size, self.n_actions, device=x.device)
return default_output
# Return default outputs for all 5 expected values to prevent crash
default_q_values = torch.zeros(batch_size, self.n_actions, device=x.device)
default_extrema = torch.zeros(batch_size, 3, device=x.device) # bottom/top/neither
default_price_pred = torch.zeros(batch_size, 1, device=x.device)
default_features = torch.zeros(batch_size, 1024, device=x.device)
default_advanced = torch.zeros(batch_size, 1, device=x.device)
return default_q_values, default_extrema, default_price_pred, default_features, default_advanced
# Check for zero feature dimensions
if len(x.shape) > 1 and any(dim == 0 for dim in x.shape[1:]):
logger.error(f"Forward pass received tensor with zero feature dimensions: {x.shape}")
# Return default output to prevent crash
default_output = torch.zeros(batch_size, self.n_actions, device=x.device)
return default_output
# Return default outputs for all 5 expected values to prevent crash
default_q_values = torch.zeros(batch_size, self.n_actions, device=x.device)
default_extrema = torch.zeros(batch_size, 3, device=x.device) # bottom/top/neither
default_price_pred = torch.zeros(batch_size, 1, device=x.device)
default_features = torch.zeros(batch_size, 1024, device=x.device)
default_advanced = torch.zeros(batch_size, 1, device=x.device)
return default_q_values, default_extrema, default_price_pred, default_features, default_advanced
# Process different input shapes
if len(x.shape) > 2:
@ -496,23 +504,15 @@ class EnhancedCNN(nn.Module):
market_regime_pred = self.market_regime_head(features_refined)
risk_pred = self.risk_head(features_refined)
# Package all price predictions
price_predictions = {
'immediate': price_immediate,
'midterm': price_midterm,
'longterm': price_longterm,
'values': price_values
}
# Package all price predictions into a single tensor (use immediate as primary)
# For compatibility with DQN agent, we return price_immediate as the price prediction tensor
price_pred_tensor = price_immediate
# Package additional predictions for enhanced decision making
advanced_predictions = {
'volatility': volatility_pred,
'support_resistance': support_resistance_pred,
'market_regime': market_regime_pred,
'risk_assessment': risk_pred
}
# Package additional predictions into a single tensor (use volatility as primary)
# For compatibility with DQN agent, we return volatility_pred as the advanced prediction tensor
advanced_pred_tensor = volatility_pred
return q_values, extrema_pred, price_predictions, features_refined, advanced_predictions
return q_values, extrema_pred, price_pred_tensor, features_refined, advanced_pred_tensor
def act(self, state, explore=True) -> Tuple[int, float, List[float]]:
"""Enhanced action selection with ultra massive model predictions"""