memory: weights_csv: -Weights of LSTM layer: w_lstm (for X), u_lstm (for hidden unit H) , b_lstm (biases). -Weights of fully connected layer: w_fc , b_fc. Order of gates in Keras: i, f, c, o (in w_lstm, u_lstm, b_lstm). weights_bin: -.mem files contain binary representation of the weights. luts_bin: -.mem files containt binary representation of the lut contents. Fixed point representation: [sign bit, integer bits, fractional bits]. Sign bit: 0 - positive, 1 - negative.