yoavg / theanets

feedforward and recurrent neural networks

Home Page:http://theanets.readthedocs.org

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

theanets

The theanets package provides tools for defining and optimizing several common types of neural network models. It uses Python for rapid development, and under the hood Theano provides graph optimization and fast computations on the GPU.

The package defines models for classification, autoencoding, regression, and prediction. Models can easily be created with any number of feedforward or recurrent layers and combined with different regularizers:

  • L1/L2 weight decay
  • L1/L2 hidden activation penalties (e.g., sparse autoencoders)
  • Dropout/gaussian noise (e.g., denoising autoencoders)

Several optimization algorithms are also included:

Installation

Install the latest published code using pip:

pip install theanets

Or download the current source and run it from there:

git clone http://github.com/lmjohns3/theanets
cd theanets
python setup.py develop

Example

Let's say you wanted to create a classifier and train it on some 100-dimensional data points that you've classified into 10 categories. You can define your model and train it using a few lines of code:

import climate
import theanets
import my_data_set

climate.enable_default_logging()

exp = theanets.Experiment(
    theanets.Classifier,
    layers=(100, 200, 100, 10),
)

exp.train(
    my_data_set.training_data,
    my_data_set.validation_data,
    optimize='sgd',
    learning_rate=0.01,
    momentum=0.5,
    hidden_l1=0.1,
)

exp.network.predict(my_data_set.test_data)

The trained network lets you make predictions on unseen test data.

More Information

Source: https://github.com/lmjohns3/theanets

Documentation: http://theanets.readthedocs.org

Mailing list: https://groups.google.com/forum/#!forum/theanets

About

feedforward and recurrent neural networks

http://theanets.readthedocs.org

License:MIT License


Languages

Language:Python 100.0%