Popularity
0.9
Growing
Activity
9.2
Growing
25
3
2

Description

ruby-dnn is a ruby deep learning library. This library supports full connected neural network and convolution neural network. Currently, you can get 99% accuracy with MNIST and 74% with CIFAR 10.

Monthly Downloads: 2,982
Programming language: Ruby
License: MIT License
Latest version: v1.1.5

ruby-dnn alternatives and similar gems

Based on the "Machine Learning" category

Do you think we are missing an alternative of ruby-dnn or a related project?

Add another 'Machine Learning' Gem

README

ruby-dnn

Gem Version
Build Status

ruby-dnn is a ruby deep learning library. This library supports full connected neural network and convolution neural network and recurrent neural network. Currently, you can get 99% accuracy with MNIST and 82% with CIFAR 10.

Installation

Add this line to your application's Gemfile:

gem 'ruby-dnn'

And then execute:

$ bundle

Or install it yourself as:

$ gem install ruby-dnn

Usage

MNIST MLP example

model = Sequential.new

model << InputLayer.new(784)

model << Dense.new(256)
model << ReLU.new

model << Dense.new(256)
model << ReLU.new

model << Dense.new(10)

model.setup(Adam.new, SoftmaxCrossEntropy.new)

model.train(x_train, y_train, 10, batch_size: 128, test: [x_test, y_test])

accuracy, loss = model.evaluate(x_test, y_test)
puts "accuracy: #{accuracy}"
puts "loss: #{loss}"

When create a model with 'define by run' style:

class MLP < Model
  def initialize
    super
    @d1 = Dense.new(256)
    @d2 = Dense.new(256)
    @d3 = Dense.new(10)
  end

  def forward(x)
    x = InputLayer.new(784).(x)
    x = @d1.(x)
    x = ReLU.(x)
    x = @d2.(x)
    x = ReLU.(x)
    x = @d3.(x)
    x
  end
end

model = MLP.new

model.setup(Adam.new, SoftmaxCrossEntropy.new)

model.train(x_train, y_train, 10, batch_size: 128, test: [x_test, y_test])

accuracy, loss = model.evaluate(x_test, y_test)
puts "accuracy: #{accuracy}"
puts "loss: #{loss}"

Please refer to examples for basic usage.
If you want to know more detailed information, please refer to the source code.

Implemented

Implemented classes
Connections Dense, Conv2D, Conv2DTranspose, Embedding, SimpleRNN, LSTM, GRU
Activations Sigmoid, Tanh, Softsign, Softplus, Swish, ReLU, LeakyReLU, ELU, Mish
Basic Flatten, Reshape, Dropout, BatchNormalization
Pooling MaxPool2D, AvgPool2D, GlobalAvgPool2D, UnPool2D
Optimizers SGD, Nesterov, AdaGrad, RMSProp, AdaDelta, RMSPropGraves, Adam, AdaBound
Losses MeanSquaredError, MeanAbsoluteError, Hinge, HuberLoss, SoftmaxCrossEntropy, SigmoidCrossEntropy

Datasets

By setting the environment variable "RUBY_DNN_DOWNLOADS_PATH", you can specify the path to read dataset.

● Iris
● MNIST
● Fashion-MNIST
● CIFAR-10
● CIFAR-100
● STL-10

Examples

● VAE
● DCGAN
● Pix2pix

TODO

● Write a test.
● Write a document.
● Support to GPU.

Development

After checking out the repo, run bin/setup to install dependencies. Then, run rake "spec" to run the tests. You can also run bin/console for an interactive prompt that will allow you to experiment.

To install this gem onto your local machine, run bundle exec rake install. To release a new version, update the version number in version.rb, and then run bundle exec rake release, which will create a git tag for the version, push git commits and tags, and push the .gem file to rubygems.org.

Contributing

Bug reports and pull requests are welcome on GitHub at https://github.com/unagiootoro/ruby-dnn. This project is intended to be a safe, welcoming space for collaboration, and contributors are expected to adhere to the Contributor Covenant code of conduct.

License

The gem is available as open source under the terms of the MIT License.

Code of Conduct

Everyone interacting in the ruby-dnn project’s codebases, issue trackers, chat rooms and mailing lists is expected to follow the code of conduct.


*Note that all licence references and agreements mentioned in the ruby-dnn README section above are relevant to that project's source code only.