Click here to Skip to main content
15,665,276 members
Please Sign up or sign in to vote.
0.00/5 (No votes)
I am trying to perform a 10-fold cross-validation on a LSTM, the code is the following:
Predict Closing Prices using a 3 day window of previous closing prices.we use window_size = 4 for this.

    from tensorflow.keras.models import Sequential
    from tensorflow.keras.layers import LSTM, Dense, Dropout
    # Build the LSTM model.
    # Define the LSTM RNN model.
    model = Sequential()
    number_units = 9  
    dropout_fraction = 0.5
    # Layer 1
        input_shape=(X_train.shape[1], 1))
    # Layer 2
    # The return_sequences parameter needs to set to True every time we add a new LSTM layer, excluding the final layer.
    model.add(LSTM(units=number_units, return_sequences=True))
    # Layer 3
    # Output layer
    # Compile the model
    model.compile(optimizer="adam", loss="mean_squared_error")
    # Summarize the model
    # Train the model, y_train, epochs=500, shuffle=False, batch_size=5, verbose=1)
    # Evaluate the model for loss
    model.evaluate(X_test, y_test)
    # Make some predictions
    predicted = model.predict(X_test)
    import sklearn.metrics as metrics
    # Evaluating the model
    print('RMSD ( Root Mean Squared Error ) :', np.sqrt(metrics.mean_squared_error(y_test, predicted)))
    print('R-squared :', metrics.r2_score(y_test, predicted))
    # Recover the original prices instead of the scaled version
    predicted_prices = y_test_scaler.inverse_transform(predicted)
    real_prices = y_test_scaler.inverse_transform(y_test.reshape(-1, 1))
    # Create a DataFrame of Real and Predicted values
    stocks = pd.DataFrame({
        "Real": real_prices.ravel(),
        "Predicted": predicted_prices.ravel()
    }, index = dw.index[-len(real_prices): ])

The idea is to perform a 10-fold cross-validation and improve RMSE and R-squared results.

A note about my input:

    X, y = window_data(dw, window_size, feature_col_number1, feature_col_number2, feature_col_number3
                       , feature_col_number4 , feature_col_number5, feature_col_number6, target_col_number)
    # Use 70% of the data for training and the remaineder for testing
    X_split = int(0.7 * len(X))
    y_split = int(0.7 * len(y))
    X_train = X[: X_split]
    X_test = X[X_split:]
    y_train = y[: y_split]
    y_test = y[y_split:]
    #Scaling Data with MinMaxScaler
    from sklearn.preprocessing import MinMaxScaler
    # Use the MinMaxScaler to scale data between 0 and 1.
    x_train_scaler = MinMaxScaler()
    x_test_scaler = MinMaxScaler()
    y_train_scaler = MinMaxScaler()
    y_test_scaler = MinMaxScaler()
    # Fit the scaler for the Training Data
    # Scale the training data
    X_train = x_train_scaler.transform(X_train)
    y_train = y_train_scaler.transform(y_train)
    # Fit the scaler for the Testing Data
    # Scale the y_test data
    X_test = x_test_scaler.transform(X_test)
    y_test = y_test_scaler.transform(y_test)
    # Reshape the features for the model
    X_train = X_train.reshape((X_train.shape[0], X_train.shape[1], 1))
    X_test = X_test.reshape((X_test.shape[0], X_test.shape[1], 1))

How do I enter k_fold cross validation? Its code and where to put it?

What I have tried:

How do I enter k_fold cross validation? Its code and where to put it?
Updated 24-Dec-21 6:47am

This content, along with any associated source code and files, is licensed under The Code Project Open License (CPOL)

CodeProject, 20 Bay Street, 11th Floor Toronto, Ontario, Canada M5J 2N8 +1 (416) 849-8900