/ruby-dnn

ruby-dnn is a ruby deep learning library.

Primary LanguageRubyMIT LicenseMIT

ruby-dnn

Gem Version Build Status Docs Latest

ruby-dnn is a ruby deep learning library. This library supports full connected neural network and convolution neural network and recurrent neural network. Currently, you can get 99% accuracy with MNIST and 82% with CIFAR 10.

Installation

Add this line to your application's Gemfile:

gem 'ruby-dnn'

And then execute:

$ bundle

Or install it yourself as:

$ gem install ruby-dnn

Usage

MNIST MLP example

model = Sequential.new

model << InputLayer.new(784)

model << Dense.new(256)
model << ReLU.new

model << Dense.new(256)
model << ReLU.new

model << Dense.new(10)

model.setup(Adam.new, SoftmaxCrossEntropy.new)

trainer = DNN::Trainer.new(model)
trainer.fit(x_train, y_train, 10, batch_size: 128, test: [x_test, y_test])
accuracy, loss = trainer.evaluate(x_test, y_test)
puts "accuracy: #{accuracy}"
puts "loss: #{loss}"

When create a model with 'define by run' style:

class MLP < Model
  def initialize
    super
    @d1 = Dense.new(256)
    @d2 = Dense.new(256)
    @d3 = Dense.new(10)
  end

  def forward(x)
    fs = DNN::Functions::FunctionSpace
    x = InputLayer.new(784).(x)
    x = @d1.(x)
    x = fs.relu(x)
    x = @d2.(x)
    x = fs.relu(x)
    x = @d3.(x)
    x
  end
end

model = MLP.new

model.setup(Adam.new, SoftmaxCrossEntropy.new)

trainer = DNN::Trainer.new(model)
trainer.fit(x_train, y_train, 10, batch_size: 128, test: [x_test, y_test])
accuracy, loss = trainer.evaluate(x_test, y_test)
puts "accuracy: #{accuracy}"
puts "loss: #{loss}"

Please refer to examples for basic usage.
If you want to know more detailed information, please refer to the source code.

Sample

Implemented

Implemented classes
Connections Dense, Conv2D, Conv2DTranspose, Embedding, SimpleRNN, LSTM, GRU
Activations Sigmoid, Tanh, Softsign, Softplus, Swish, ReLU, LeakyReLU, ELU, Mish
Basic Flatten, Reshape, Dropout, BatchNormalization
Pooling MaxPool2D, AvgPool2D, GlobalAvgPool2D, UnPool2D
Optimizers SGD, Nesterov, AdaGrad, RMSProp, AdaDelta, RMSPropGraves, Adam, AdaBound
Losses MeanSquaredError, MeanAbsoluteError, Hinge, HuberLoss, SoftmaxCrossEntropy, SigmoidCrossEntropy

Datasets

By setting the environment variable RUBY_DNN_DOWNLOADS_PATH, you can specify the path to read dataset.

  • Iris
  • MNIST
  • Fashion-MNIST
  • CIFAR-10
  • CIFAR-100
  • STL-10

Use GPU

If you do require "cumo/narray" before require "dnn", you can run it on GPU. Or, set the environment variable RUBY_DNN_USE_CUMO to ENABLE to force the GPU to be used.

Use Numo Linalg

When running on a CPU, you can speed it up by using Numo Linalg. In this case, Numo Linalg is automatically loaded by setting the environment variable RUNY_DNN_USE_NUMO_LINALG to ENABLE.

TODO

  • Write a test.
  • Write a document.
  • Improve performance when using GPU.

Development

After checking out the repo, run bin/setup to install dependencies. Then, run rake "test" to run the tests. You can also run bin/console for an interactive prompt that will allow you to experiment.

To install this gem onto your local machine, run bundle exec rake install. To release a new version, update the version number in version.rb, and then run bundle exec rake release, which will create a git tag for the version, push git commits and tags, and push the .gem file to rubygems.org.

Contributing

Bug reports and pull requests are welcome on GitHub at https://github.com/unagiootoro/ruby-dnn. This project is intended to be a safe, welcoming space for collaboration, and contributors are expected to adhere to the Contributor Covenant code of conduct.

License

The gem is available as open source under the terms of the MIT License.

Code of Conduct

Everyone interacting in the ruby-dnn project’s codebases, issue trackers, chat rooms and mailing lists is expected to follow the code of conduct.