Questions tagged [mxnet]

Apache MXNet is a fast, scalable, open source deep learning framework that supports Python, R, Scala, C++, Julia, Go, and Javascript. MXNet can run on multiple CPUs and multiple GPUs distributed across multiple hosts, with almost linear scalability.

Apache MXNet: Flexible and Efficient Library for Deep Learning

For more details see http://mxnet.io/

Flexible

Supports both imperative and symbolic programming

Portable

Runs on CPUs or GPUs, on clusters, servers, desktops, or mobile phones

Multiple Languages

Supports over 7 programming languages, including C++, Python, R, Scala, Julia, Matlab, Javascript and Perl.

Auto-Differentiation

Calculates the gradient automatically for training a model

Distributed on Cloud

Supports distributed training on multiple CPU/GPU machines, including AWS, GCE, Azure, and Yarn clusters

Performance

Optimized C++ backend engine parallelizes both I/O and computation

701 questions
2
votes
1 answer

Reset GPU memory using Keras 1.2.2 with MXnet backend

I'm using AWS p2.x8large and trying to evaluate my model using k-fold cross validation. After the first repetition my GPUs memory is full and when I try to train once again I receive a cuda memory problem. My question is how to reset the GPU memory…
Jenia Golbstein
  • 374
  • 2
  • 12
2
votes
1 answer

Mxnet - Understanding weight shape for Dense Layer

I am understanding Mxnet framework. While following this Linear Regression article I see the following piece of code which is creating a Dense Layer: net = gluon.nn.Dense(1, in_units=2) # input dimension= 2, output dimension = 1 but why does…
AgentX
  • 1,402
  • 3
  • 23
  • 38
2
votes
2 answers

Retrain mxnet model in R

I have created a neural network with mxnet. Now I want to train this model iteratively on new data points. After I simulated a new data point I want to make a new gradient descent update on this model. I do not want to save the model to an external…
needRhelp
  • 2,948
  • 2
  • 24
  • 48
2
votes
1 answer

Mxnet Gluon custom data iterator

I have written a custom data iterator using mx.io.DataIter class. What's the easiest way to use this data iterator with Gluon interface? I went through the documentation and couldn't find an easy way to do so. One of my idea was to use it as…
krishnakamathk
  • 143
  • 2
  • 9
2
votes
1 answer

How does MXNET C++ code bind with its python API?

1.When defining custom operator in MXNet with C++, how does the custom operator defined in C++ generate its python API automatically? 2.How can I finde the corresponding codes? 3. What's the difference between defining custom operator with C++ and…
2
votes
1 answer

How to use mx.sym.Correlation?

Suppose I have two feature maps F1 and F2 output by a network. I want to compute convolution of F1 and F2. Assume that F1 has shape (1, C, 10, 10) and F2 has shape (1, C, 3, 3) and the wanted result should have shape (1, 1, 8, 8) if pad = 0, stride…
Squids Zyd
  • 137
  • 1
  • 8
2
votes
1 answer

"Compile with USE_CUDA=1 to enable GPU usage" Message with MXNet on AWS Deep Learning AMI

I would like to train a neural network whilst utilising all 4 GPU's on my g2.8xarge EC2 instance using MXNet. I am using the following AWS Deep Learning Linux community AMI: Deep Learning AMI Amazon Linux - 3.3_Oct2017 - ami-999844e0) As per these…
2
votes
1 answer

Keras with mxnet-backend and 8 GPUs trains slow when using ImageDataGenerator

Following this great post: Scaling Keras Model Training to Multiple GPUs I tried to upgrade my model to run in parallel on my multiple GPUs instance. At first I ran the MNIST example as proposed here: MNIST in Keras with the additional syntax in the…
Jenia Golbstein
  • 374
  • 2
  • 12
2
votes
1 answer

R mxnet 1 pass returns NAN as loss value

Is this expected behaviour? library(mxnet) hidden_u_1 <- 100 activ_hidden_1 <- 'tanh' hidden_u_2 <- 1 learn_rate <- 0.001 initializer <- mx.init.uniform(1) optimizer <- 'rmsprop' #sgd loss <- mx.metric.mse device.cpu <- mx.cpu() mini_batch…
Alexey Burnakov
  • 259
  • 2
  • 14
2
votes
1 answer

Have MXNet used Nvidia's NCCL library for multi-GPU communication?

In Nvidia website, they claimed MXNet uses NCCL (https://developer.nvidia.com/nccl). However, I haven't found any reference from MXNet's github repository that they actually use NCCL library. In the chainer blog, they also claimed that chainer…
Saiful
  • 61
  • 5
2
votes
1 answer

Using Index Arrays on Columns of an MXNet NDArray

Given an index array index and, say, a matrix A I want a matrix B with the corresponding permutation of the columns of A. In Numpy I would do the following, >>> A = np.arange(6).reshape(2,3); A array([[0, 1, 2], [3, 4, 5]]) >>> index =…
2
votes
1 answer

How to build MXNET custom operators into separate library/package with C++?

I want to try out some new activation functions and I have successfully created custom operators in C++ on top of the latest mxnet source code. But what I really like to have is to build a separate library (.so file in ubuntu) just for my custom…
ck.liao
  • 21
  • 1
  • 3
2
votes
1 answer

MXNET multi-iterators (combine a .rec iterator with an NDArray iterator)

How do I create a combined iterator in MXNET? For example, given a record (.rec) iterator if I want to change the labels corresponding to each image then there are two options: a) Create a new rec iterator with the same data(images) and new…
Ashish Khetan
  • 121
  • 1
  • 6
2
votes
1 answer

How to print the processing steps/report from model.fit MXNet Python

I am trying to train my 20x20 images dataset using MXNet deep learning library, you can see the code below: the question is when I run it, although it shows no errors it returns nothing, I mean it does not show any processing like : epoch 0 :…
esraa
  • 41
  • 5
2
votes
1 answer

Mxnet element wise multiply

In MXNet, if I wanted to create a vector of weights that multiplied each input, i.e. to have w*x_i and then backprop over the weights w how would I do this? I tried: y_hat = input w1 = mx.sym.Variable("w1") y_hat = mx.symbol.broadcast_mul(w1,…
Drew
  • 65
  • 5