Questions tagged [mxnet]

Apache MXNet is a fast, scalable, open source deep learning framework that supports Python, R, Scala, C++, Julia, Go, and Javascript. MXNet can run on multiple CPUs and multiple GPUs distributed across multiple hosts, with almost linear scalability.

Apache MXNet: Flexible and Efficient Library for Deep Learning

For more details see http://mxnet.io/

Flexible

Supports both imperative and symbolic programming

Portable

Runs on CPUs or GPUs, on clusters, servers, desktops, or mobile phones

Multiple Languages

Supports over 7 programming languages, including C++, Python, R, Scala, Julia, Matlab, Javascript and Perl.

Auto-Differentiation

Calculates the gradient automatically for training a model

Distributed on Cloud

Supports distributed training on multiple CPU/GPU machines, including AWS, GCE, Azure, and Yarn clusters

Performance

Optimized C++ backend engine parallelizes both I/O and computation

701 questions
1
vote
1 answer

Issue when Re-implement Matrix Factorization in Pytorch

I try to implement matrix factorization in Pytorch as the data extractor and model. The original model is written in mxnet. Here I try to use the same idea in Pytorch. Here is my code, it can be runned directly in codelab import torch import…
jason
  • 1,998
  • 3
  • 22
  • 42
1
vote
1 answer

How to extract typical rows in MXNet?

These are data(Batch size 2) and batch index import mxnet as mx data=mx.nd.array(range(24)).reshape(2,3,4) index=mx.nd.array([[0,1],[1,2]]) How to get the selected data? I tried the Pick and take functions, but don't know how to do it.
partida
  • 501
  • 4
  • 20
1
vote
0 answers

Is there a way to figure out if an NDArray is trainable?

Is there a way to figure out if an NDArray variable has a gradient/requires backprop, similar to the requires_grad attribute in pytorch? I've tried checking whether x.grad is None, but that doesn't work for intermediate variables: import mxnet as…
Raphael Koh
  • 111
  • 9
1
vote
1 answer

Mxnet MMS - Access from external server

I am trying to use Mxnet MMS (multi-model-server) to make my Mxnet model accessible by my website. The Mxnet MMS can run without problem and I can use the served model on the same machine. However, I want to access the served model from another…
Can
  • 145
  • 1
  • 11
1
vote
1 answer

Load Amazon Sagemaker NTM model locally for inference

I have trained a Sagemaker NTM model which is a neural topic model, directly on the AWS sagemaker platform. Once training is complete you are able to download the mxnet model files. Once unpacked the files contain: params symbol.json meta.json I…
dendog
  • 2,976
  • 5
  • 26
  • 63
1
vote
1 answer

Installing autogluon in Amazon EMR

I want to be able to use both Pyspark and AutoGluon libraries in a notebook backed by an EMR cluster. I have tried to install AutoGluon using the bootstrap script for the EMR cluster (emr-5.30.1) with the following sudo python3 -m pip install…
1
vote
0 answers

MXNET Gluon Error when calling two consecutive backward - Check failed: type_ != nullptr

I need to call backward function for two different models, separately. # Loss_1 #----------------------
 

Losses_1 = [] with autograd.record():
 
 for batch_per_gpu in batch
es: pred = model_1(batch_per_gpu)
 …
Batool
  • 957
  • 1
  • 12
  • 20
1
vote
2 answers

mxnet cause gpu memory leak inference with dynamic shape input, while using flask

Sorry, I don't put on any code to prove The problem is: when I use a simple for loop like this: for _ in range(2000): rnum = random.randint(1, 5) img = np.random.rand(rnum, 3, 112, 112) mxnet_model.inference(img) it will work…
RayChang
  • 13
  • 3
1
vote
0 answers

how to free memory from mx.ndarray

I write a custom op in Python layer. implemented with operators of mx.nd.op_name, It work normal when the shape of input is in the same. But it tells out of memory when its' shape is different.the custom op as below. It seems like the memory of…
cxt123
  • 21
  • 1
  • 3
1
vote
1 answer

mxnet-cu102mkl==1.6.0 disappered from pypi today, why?

Looks like that version disappeared recently: ERROR: Could not find a version that satisfies the requirement mxnet-cu102mkl==1.6.0 (from versions: 1.6.0.post0) 11:39:11 ERROR: No matching distribution found for mxnet-cu102mkl==1.6.0 The top hit…
Marcin Owsiany
  • 486
  • 5
  • 8
1
vote
0 answers

Save/Load MXNet model parameters using NumPy

How can I save the parameters for an MXNet model into a NumPy file (.npy)? After doing so, how can I load these parameters from the .npy file back into my model? Here is a minimal example to save the parameters for MXNet model using MXNet…
Masquerade
  • 3,580
  • 5
  • 20
  • 37
1
vote
0 answers

Docker MXNet/Python Error: executable file not found in $PATH": unknown

I have built a docker image on my ubuntu machine (Ubuntu 20.04) and I am now trying to get it running. However, I seem to get the same error every time no matter what. My base image for the docker container is python:3.7.6-stretch This is my docker…
Luca01
  • 23
  • 4
1
vote
2 answers

Why is my CPU doing matrix operations faster than GPU instead?

When I tried to verify that the GPU does matrix operations over the CPU, I got unexpected results.CPU performs better than GPU according to my experience result, it makes me confused. I used cpu and gpu to do matrix multiplication…
qianfeng
  • 21
  • 3
1
vote
0 answers

How long does compiling mxnet take?

I was following the repository for RoITransformer_DOTA. There I first cloned the mxnet repository and then used the following command to compile it make -j $(nproc) USE_OPENCV=1 USE_BLAS=openblas USE_CUDA=1 USE_CUDA_PATH=/usr/local/cuda…
1
vote
1 answer

What is the least total batch size for SyncBatchNorm

For the normal BatchNorm, the least batch size per GPU is 2. I wonder if I use the SyncBatchNorm, can I use batch_size=1 for every GPU with more than a single GPU? I.e, the total_batch_size is more than 1 but batch_size_per_gpu is 1. I would…
one
  • 2,205
  • 1
  • 15
  • 37