blob: 8a725a718ccd40e3b5a7224d2ba2086754602df6 [file] [log] [blame]
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
"""MovieLens data handling: download, parse, and expose as DataIter
"""
import os
import mxnet as mx
from mxnet import gluon
def load_mldataset(filename):
"""Not particularly fast code to parse the text file and load it into three NDArray's
and product an NDArrayIter
"""
user = []
item = []
score = []
with open(filename) as f:
for line in f:
tks = line.strip().split('\t')
if len(tks) != 4:
continue
user.append(int(tks[0]))
item.append(int(tks[1]))
score.append(float(tks[2]))
user = mx.np.array(user)
item = mx.np.array(item)
score = mx.np.array(score)
return gluon.data.ArrayDataset(user, item, score)
def ensure_local_data(prefix):
if not os.path.exists(f"{prefix}.zip"):
print(f"Downloading MovieLens data: {prefix}")
# MovieLens 100k dataset from https://grouplens.org/datasets/movielens/
# This dataset is copy right to GroupLens Research Group at the University of Minnesota,
# and licensed under their usage license.
# For full text of the usage license, see http://files.grouplens.org/datasets/movielens/ml-100k-README.txt
os.system(f"wget http://files.grouplens.org/datasets/movielens/{prefix}.zip")
os.system(f"unzip {prefix}.zip")
def get_dataset(prefix='ml-100k'):
"""Returns a pair of NDArrayDataIter, one for train, one for test.
"""
ensure_local_data(prefix)
return (load_mldataset(f'./{prefix}/u1.base'),
load_mldataset(f'./{prefix}/u1.test'))
def max_id(fname):
mu = 0
mi = 0
for line in open(fname):
tks = line.strip().split('\t')
if len(tks) != 4:
continue
mu = max(mu, int(tks[0]))
mi = max(mi, int(tks[1]))
return mu + 1, mi + 1