Replies: 1 comment
-
The algorithm supports JSON format for sparse data as described in https://docs.aws.amazon.com/sagemaker/latest/dg/cdf-inference.html#common-in-formats. Sparse data can also be provided in protobuf format. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
ISSUE
I trained Factorization machine model on sagemaker with MXNET. Now for training I followed this notebook . My train data is sparse(300 M rows, 4M columns) I used
smac.write_spmatrix_to_sparse_tensor
to write my sparse training data to s3 for training. I was able to train and deploy an endpoint for inference. But problem starts when I try to inference, my feature dimension is 4 Million. So when I try to pass a dense vector of this size topredict
for inference, I getRequest Entity Too Large
. On checking I found that sagemaker has 5mb size limit. only option I can think of is to somehow pass the sparse vector butpredict
is not accepting that. Can you guys help?I tried sending sparse tensor, with content-type as
protobuf
, didn't work.I am using
RealTimePredictor
class aspredictor_cls
, is there anything I can do with this class so that it acceptsparse vector
and maybe convert it on server side if needed. Any Suggestions?Beta Was this translation helpful? Give feedback.
All reactions