public static ComputationGraphConfiguration getConf() { ComputationGraphConfiguration.GraphBuilder builder = new NeuralNetConfiguration.Builder() .seed(12345) .updater(new Adam(0.01)) .weightInit(WeightInit.RELU) .graphBuilder() .addInputs("in"); String[] poolNames = new String[ngramFilters.length]; int i = 0; for (int ngram : ngramFilters) { String filterName = String.format("ngram%d", ngram); poolNames[i] = String.format("pool%d", ngram); builder = builder.addLayer(filterName, new Convolution1DLayer.Builder() .nOut(numFilters) .kernelSize(ngram) .activation(Activation.RELU) .build(), "in") .addLayer(poolNames[i], new GlobalPoolingLayer.Builder(PoolingType.MAX).build(), filterName); i++; } return builder.addVertex("concat", new MergeVertex(), poolNames) .addLayer("predict", new DenseLayer.Builder().nOut(numClasses).dropOut(dropoutRetain) .activation(Activation.SOFTMAX).build(), "concat") .addLayer("loss", new LossLayer.Builder(LossFunctions.LossFunction.MCXENT).build(), "predict") .setOutputs("loss") .setInputTypes(InputType.recurrent(W2V_VECTOR_SIZE, 1000)) .build(); } }
.nOut(numFactors) .build(), "globalPool") .setOutputs("out") .build();
currentInput = lName; gb.setOutputs(currentInput);
.activation("identity").weightInit(WEIGHT_INIT).learningRateDecayPolicy(learningRatePolicy) .nIn((int) (numHiddenNodes * Math.pow(reduction, 4))).nOut(numOutputsSomaticFrequency).build(), "dense5") .setOutputs(getOutputNames()) .pretrain(false).backprop(true).build();
numLSTMTrueGenotypeHiddenNodes, numLSTMDecoderInputs, numReductionLayers, reductionRate); ComputationGraphConfiguration conf = build .setOutputs(outputNames) .build();
numLSTMTrueGenotypeHiddenNodes, numLSTMDecoderInputs, numReductionLayers, reductionRate); ComputationGraphConfiguration conf = build .setOutputs(outputNames) .build(); System.out.println(conf);
.nOut(numClasses).activation(Activation.SOFTMAX).build(), "avgpool") .setOutputs("output").backprop(true).pretrain(false);
.addLayer("output", new OutputLayer.Builder(LossFunctions.LossFunction.NEGATIVELOGLIKELIHOOD) .nIn(1024).nOut(numLabels).activation(Activation.SOFTMAX).build(), "fc1") .setOutputs("output").backprop(true).pretrain(false);
currentInput = lName; builder.setOutputs(currentInput);
.setOutputs(outputNames) .build();
@Override public ComputationGraph createComputationalGraph(DomainDescriptor domainDescriptor) { LearningRatePolicy learningRatePolicy = LearningRatePolicy.Poly; layerAssembler.setLearningRatePolicy(learningRatePolicy); layerAssembler.initializeBuilder(); int numInputs = domainDescriptor.getNumInputs("input")[0]; int numHiddenNodes = domainDescriptor.getNumHiddenNodes("firstDense"); ComputationGraphConfiguration.GraphBuilder build = layerAssembler.assemble(numInputs, numHiddenNodes, numLayers); int numIn = layerAssembler.getNumOutputs(); WeightInit WEIGHT_INIT = WeightInit.XAVIER; String lastDenseLayerName = layerAssembler.lastLayerName(); build.addLayer("homozygous", new OutputLayer.Builder( domainDescriptor.getOutputLoss("homozygous")) .weightInit(WEIGHT_INIT) .activation("softmax").weightInit(WEIGHT_INIT).learningRateDecayPolicy(learningRatePolicy) .nIn(numIn).nOut(11).build(), lastDenseLayerName); for (int i = 1; i < outputNames.length; i++) { build.addLayer(outputNames[i], new OutputLayer.Builder( domainDescriptor.getOutputLoss(outputNames[i])) .weightInit(WEIGHT_INIT) .activation("softmax").weightInit(WEIGHT_INIT).learningRateDecayPolicy(learningRatePolicy) .nIn(numIn).nOut(2).build(), lastDenseLayerName); } appendMetaDataLayer(domainDescriptor, learningRatePolicy, build, numIn, WEIGHT_INIT, lastDenseLayerName); appendIsVariantLayer(domainDescriptor, learningRatePolicy, build, numIn, WEIGHT_INIT, lastDenseLayerName); ComputationGraphConfiguration conf = build .setOutputs(outputNames) .build(); return new ComputationGraph(conf); }
currentInput = lName; gb.setOutputs(currentInput); gb.setInputTypes(InputType.inferInputType(features));
@Override public GraphConfiguration getValue(double[] values) { //Create ComputationGraphConfiguration... NeuralNetConfiguration.Builder builder = randomGlobalConf(values); ComputationGraphConfiguration.GraphBuilder graphBuilder = builder.graphBuilder(); graphBuilder.addInputs(this.networkInputs); graphBuilder.setOutputs(this.networkOutputs); if (inputTypes != null) graphBuilder.setInputTypes(inputTypes.getValue(values)); //Build/add our layers and vertices: for (LayerConf c : layerSpaces) { org.deeplearning4j.nn.conf.layers.Layer l = c.layerSpace.getValue(values); graphBuilder.addLayer(c.getLayerName(), l, c.getInputs()); } for (VertexConf gv : vertices) { graphBuilder.addVertex(gv.getVertexName(), gv.getGraphVertex(), gv.getInputs()); } if (backprop != null) graphBuilder.backprop(backprop.getValue(values)); if (pretrain != null) graphBuilder.pretrain(pretrain.getValue(values)); if (backpropType != null) graphBuilder.backpropType(backpropType.getValue(values)); if (tbpttFwdLength != null) graphBuilder.tBPTTForwardLength(tbpttFwdLength.getValue(values)); if (tbpttBwdLength != null) graphBuilder.tBPTTBackwardLength(tbpttBwdLength.getValue(values)); ComputationGraphConfiguration configuration = graphBuilder.build(); return new GraphConfiguration(configuration, earlyStoppingConfiguration, numEpochs); }
@Override public ComputationGraph createComputationalGraph(DomainDescriptor domainDescriptor) { LearningRatePolicy learningRatePolicy = LearningRatePolicy.Poly; layerAssembler.setLearningRatePolicy(learningRatePolicy); layerAssembler.initializeBuilder(); int numInputs = domainDescriptor.getNumInputs("input")[0]; int numHiddenNodes = domainDescriptor.getNumHiddenNodes("firstDense"); ComputationGraphConfiguration.GraphBuilder build = layerAssembler.assemble(numInputs, numHiddenNodes, numLayers); int numIn = layerAssembler.getNumOutputs(); WeightInit WEIGHT_INIT = WeightInit.XAVIER; String lastDenseLayerName = layerAssembler.lastLayerName(); String combined = fixRef?"combinedRef":"combined"; build.addLayer(combined, new OutputLayer.Builder( domainDescriptor.getOutputLoss(combined)) .weightInit(WEIGHT_INIT) .activation("softmax").weightInit(WEIGHT_INIT).learningRateDecayPolicy(learningRatePolicy) .nIn(numIn) .nOut(domainDescriptor.getNumOutputs(combined)[0]).build(), lastDenseLayerName); appendMetaDataLayer(domainDescriptor, learningRatePolicy, build, numIn, WEIGHT_INIT, lastDenseLayerName); appendIsVariantLayer(domainDescriptor, learningRatePolicy, build, numIn, WEIGHT_INIT, lastDenseLayerName); ComputationGraphConfiguration conf = build .setOutputs(getOutputNames()) .build(); return new ComputationGraph(conf); }
@Override public ComputationGraph createComputationalGraph(DomainDescriptor domainDescriptor) { LearningRatePolicy learningRatePolicy = LearningRatePolicy.Poly; layerAssembler.setLearningRatePolicy(learningRatePolicy); layerAssembler.initializeBuilder(); int numInputs = domainDescriptor.getNumInputs("input")[0]; int numHiddenNodes = domainDescriptor.getNumHiddenNodes("firstDense"); ComputationGraphConfiguration.GraphBuilder build = layerAssembler.assemble(numInputs, numHiddenNodes, numLayers); int numIn = layerAssembler.getNumOutputs(); WeightInit WEIGHT_INIT = WeightInit.XAVIER; String lastDenseLayerName = layerAssembler.lastLayerName(); build.addLayer("softmaxGenotype", new OutputLayer.Builder( domainDescriptor.getOutputLoss("softmaxGenotype")) .weightInit(WEIGHT_INIT) .activation("softmax").weightInit(WEIGHT_INIT).learningRateDecayPolicy(learningRatePolicy) .nIn(numIn) .nOut(domainDescriptor.getNumOutputs("softmaxGenotype")[0]).build(), lastDenseLayerName).addInputs(); appendMetaDataLayer(domainDescriptor, learningRatePolicy, build, numIn, WEIGHT_INIT, lastDenseLayerName); appendIsVariantLayer(domainDescriptor, learningRatePolicy, build, numIn, WEIGHT_INIT, lastDenseLayerName); ComputationGraphConfiguration conf = build .setOutputs(outputNames) .build(); return new ComputationGraph(conf); }
@Override public ComputationGraph init() { int embeddingSize = 128; ComputationGraphConfiguration.GraphBuilder graph = graphBuilder("input1"); graph.addInputs("input1").setInputTypes(InputType.convolutional(inputShape[2], inputShape[1], inputShape[0])) // Logits .addLayer("bottleneck", new DenseLayer.Builder().nIn(5376).nOut(embeddingSize).build(), "avgpool") // Embeddings .addVertex("embeddings", new L2NormalizeVertex(new int[] {1}, 1e-10), "bottleneck") // Output .addLayer("outputLayer", new CenterLossOutputLayer.Builder() .lossFunction(LossFunctions.LossFunction.NEGATIVELOGLIKELIHOOD) .activation(Activation.SOFTMAX).alpha(0.9).lambda(1e-4) .nIn(embeddingSize).nOut(numClasses).build(), "embeddings") .setOutputs("outputLayer").backprop(true).pretrain(false); ComputationGraphConfiguration conf = graph.build(); ComputationGraph model = new ComputationGraph(conf); model.init(); return model; }
public static void main(String[] args) throws Exception { //Define a simple ComputationGraph: ComputationGraphConfiguration conf = new NeuralNetConfiguration.Builder() .weightInit(WeightInit.XAVIER) .updater(new Nesterovs(0.01, 0.9)) .graphBuilder() .addInputs("in") .addLayer("layer0", new DenseLayer.Builder().nIn(4).nOut(3).activation(Activation.TANH).build(), "in") .addLayer("layer1", new OutputLayer.Builder(LossFunctions.LossFunction.NEGATIVELOGLIKELIHOOD).activation(Activation.SOFTMAX).nIn(3).nOut(3).build(), "layer0") .setOutputs("layer1") .backprop(true).pretrain(false).build(); ComputationGraph net = new ComputationGraph(conf); net.init(); //Save the model File locationToSave = new File("model/MyComputationGraph.zip"); //Where to save the network. Note: the file is in .zip format - can be opened externally boolean saveUpdater = true; //Updater: i.e., the state for Momentum, RMSProp, Adagrad etc. Save this if you want to train your network more in the future ModelSerializer.writeModel(net, locationToSave, saveUpdater); //Load the model ComputationGraph restored = ModelSerializer.restoreComputationGraph(locationToSave); System.out.println("Saved and loaded parameters are equal: " + net.params().equals(restored.params())); System.out.println("Saved and loaded configurations are equal: " + net.getConfiguration().equals(restored.getConfiguration())); }
/** * Default layer setup: Create sequential layer network defined by the order of the layer list * * @param gb GraphBuilder object */ protected void makeDefaultLayerSetup(GraphBuilder gb) { String currentInput = "input"; gb.addInputs(currentInput); // Collect layers for (Layer layer : layers) { String lName = layer.getLayerName(); gb.addLayer(lName, layer.getBackend().clone(), currentInput); currentInput = lName; } gb.setOutputs(currentInput); }
/** * Set outputs to the computation graph, will add to ones that are existing * Also determines the order, like in ComputationGraphConfiguration * @param outputNames * @return */ public GraphBuilder setOutputs(String... outputNames) { initBuilderIfReq(); editedConfigBuilder.setOutputs(outputNames); return this; }
.graphBuilder() .addInputs("input") .setOutputs("output"); .addLayer("globalpooling", new GlobalPoolingLayer.Builder(PoolingType.AVG).build(), "convolution2d_6") .addLayer("loss", new LossLayer.Builder(LossFunctions.LossFunction.NEGATIVELOGLIKELIHOOD).activation(Activation.SOFTMAX).build(), "globalpooling") .setOutputs("loss");