cleanup, CNN fixes
This commit is contained in:
@ -329,13 +329,13 @@ class EnhancedCNNModel(nn.Module):
|
||||
x = x.unsqueeze(0)
|
||||
elif len(x.shape) > 3:
|
||||
# Input has extra dimensions - flatten to [batch, seq, features]
|
||||
x = x.view(x.shape[0], -1, x.shape[-1])
|
||||
x = x.reshape(x.shape[0], -1, x.shape[-1])
|
||||
|
||||
x = self._memory_barrier(x) # Apply barrier after shape changes
|
||||
batch_size, seq_len, features = x.shape
|
||||
|
||||
# Reshape for processing: [batch, seq, features] -> [batch*seq, features]
|
||||
x_reshaped = x.view(-1, features)
|
||||
x_reshaped = x.reshape(-1, features)
|
||||
x_reshaped = self._memory_barrier(x_reshaped)
|
||||
|
||||
# Input embedding
|
||||
@ -343,7 +343,7 @@ class EnhancedCNNModel(nn.Module):
|
||||
embedded = self._memory_barrier(embedded)
|
||||
|
||||
# Reshape back for conv1d: [batch*seq, channels] -> [batch, channels, seq]
|
||||
embedded = embedded.view(batch_size, seq_len, -1).transpose(1, 2).contiguous()
|
||||
embedded = embedded.reshape(batch_size, seq_len, -1).transpose(1, 2).contiguous()
|
||||
embedded = self._memory_barrier(embedded)
|
||||
|
||||
# Multi-scale feature extraction - ensure each path creates independent tensors
|
||||
@ -380,10 +380,10 @@ class EnhancedCNNModel(nn.Module):
|
||||
|
||||
# Global aggregation - create independent tensors
|
||||
avg_pooled = self.global_pool(attended_features)
|
||||
avg_pooled = self._memory_barrier(avg_pooled.view(avg_pooled.shape[0], -1)) # Flatten instead of squeeze
|
||||
avg_pooled = self._memory_barrier(avg_pooled.reshape(avg_pooled.shape[0], -1)) # Flatten instead of squeeze
|
||||
|
||||
max_pooled = self.global_max_pool(attended_features)
|
||||
max_pooled = self._memory_barrier(max_pooled.view(max_pooled.shape[0], -1)) # Flatten instead of squeeze
|
||||
max_pooled = self._memory_barrier(max_pooled.reshape(max_pooled.shape[0], -1)) # Flatten instead of squeeze
|
||||
|
||||
# Combine global features - create new tensor
|
||||
global_features = torch.cat([avg_pooled, max_pooled], dim=1)
|
||||
@ -399,7 +399,7 @@ class EnhancedCNNModel(nn.Module):
|
||||
|
||||
# Combine all features for final decision (8 regime classes + 1 volatility)
|
||||
# Create completely independent tensors for concatenation
|
||||
vol_pred_flat = self._memory_barrier(volatility_pred.view(volatility_pred.shape[0], -1)) # Flatten instead of squeeze
|
||||
vol_pred_flat = self._memory_barrier(volatility_pred.reshape(volatility_pred.shape[0], -1)) # Flatten instead of squeeze
|
||||
combined_features = torch.cat([processed_features, regime_probs, vol_pred_flat], dim=1)
|
||||
combined_features = self._memory_barrier(combined_features)
|
||||
|
||||
@ -411,15 +411,15 @@ class EnhancedCNNModel(nn.Module):
|
||||
trading_probs = self._memory_barrier(F.softmax(scaled_logits, dim=1))
|
||||
|
||||
# Flatten confidence to ensure consistent shape
|
||||
confidence_flat = self._memory_barrier(confidence.view(confidence.shape[0], -1))
|
||||
volatility_flat = self._memory_barrier(volatility_pred.view(volatility_pred.shape[0], -1))
|
||||
confidence_flat = self._memory_barrier(confidence.reshape(confidence.shape[0], -1))
|
||||
volatility_flat = self._memory_barrier(volatility_pred.reshape(volatility_pred.shape[0], -1))
|
||||
|
||||
return {
|
||||
'logits': self._memory_barrier(trading_logits),
|
||||
'probabilities': self._memory_barrier(trading_probs),
|
||||
'confidence': confidence_flat[:, 0] if confidence_flat.shape[1] > 0 else confidence_flat.view(-1)[0],
|
||||
'confidence': confidence_flat[:, 0] if confidence_flat.shape[1] > 0 else confidence_flat.reshape(-1)[0],
|
||||
'regime': self._memory_barrier(regime_probs),
|
||||
'volatility': volatility_flat[:, 0] if volatility_flat.shape[1] > 0 else volatility_flat.view(-1)[0],
|
||||
'volatility': volatility_flat[:, 0] if volatility_flat.shape[1] > 0 else volatility_flat.reshape(-1)[0],
|
||||
'features': self._memory_barrier(processed_features)
|
||||
}
|
||||
|
||||
|
Reference in New Issue
Block a user