relu activation
This commit is contained in:
parent
77f8d3bf57
commit
2e22f60c2e
2
tf.js
2
tf.js
@ -12,7 +12,7 @@ var model = tf.sequential();
|
|||||||
|
|
||||||
|
|
||||||
|
|
||||||
model.add(tf.layers.dense({units: hiddenSize, inputShape: [inputShape]}));
|
model.add(tf.layers.dense({units: hiddenSize, inputShape: [inputShape], activation: 'relu'}));
|
||||||
model.add(tf.layers.dense({units: outputShape}));
|
model.add(tf.layers.dense({units: outputShape}));
|
||||||
|
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user