Skip to content

Commit

Permalink
Remove Mahout support
Browse files Browse the repository at this point in the history
  • Loading branch information
chu11 committed Jan 19, 2022
1 parent b10c39d commit 5311c92
Show file tree
Hide file tree
Showing 50 changed files with 19 additions and 6,312 deletions.
1 change: 1 addition & 0 deletions NEWS
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@ Magpie 3.0 includes some new software support, but primarily removes a
lot of older support.

Legacy support Removed
- Mahout

Backwards Compatibility Notes

Expand Down
14 changes: 7 additions & 7 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,11 +2,11 @@ Magpie
------

Magpie contains a number of scripts for running Big Data software in
HPC environments. Thus far, Hadoop, Spark, Hbase, Storm, Pig, Mahout,
Phoenix, Kafka, Zeppelin, Zookeeper, and Alluxio are supported. It currently
supports running over the parallel file system Lustre and running over
any generic network filesytem. There is scheduler/resource manager
support for Slurm, Moab, Torque, and LSF.
HPC environments. Thus far, Hadoop, Spark, Hbase, Storm, Pig,
Phoenix, Kafka, Zeppelin, Zookeeper, and Alluxio are supported. It
currently supports running over the parallel file system Lustre and
running over any generic network filesytem. There is
scheduler/resource manager support for Slurm, Moab, Torque, and LSF.

Some of the features presently supported:

Expand Down Expand Up @@ -80,8 +80,6 @@ Zookeeper - 3.4.X

Storm - 0.9.X, 0.10.X, 1.0.X, 1.1.X, 1.2.X

Mahout - 0.11.X, 0.12.X, 0.13.0

Phoenix - 4.5.X, 4.6.0, 4.7.0, 4.8.X, 4.9.0, 4.10.1, 4.11.0, 4.12.0,
4.13.X, 4.14.0

Expand Down Expand Up @@ -117,6 +115,8 @@ Removed in Magpie 2.0

Removed in Magpie 3.0

- Mahout

Documentation
-------------

Expand Down
10 changes: 2 additions & 8 deletions doc/README
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@ Magpie

Magpie contains a number of scripts for running Big Data software in
HPC environments. Thus far, Hadoop, Spark, Hbase, Hive, Storm, Pig,
Mahout, Phoenix, Kafka, Zeppelin, and Zookeeper are supported. It
Phoenix, Kafka, Zeppelin, and Zookeeper are supported. It
currently supports running over the parallel file system Lustre and
running over any generic network filesytem. There is
scheduler/resource manager support for Slurm, Moab, Torque, and LSF.
Expand Down Expand Up @@ -252,8 +252,6 @@ Storm - 0.9.3, 0.9.4, 0.9.5, 0.9.6, 0.9.7, 0.10.0, 0.10.1, 0.10.2,
1.0.0, 1.0.1, 1.0.2, 1.0.3, 1.0.4, 1.1.0, 1.1.1, 1.1.2, 1.2.0,
1.2.1, 1.2.2, 1.2.3

Mahout - 0.11.0+, 0.11.1+, 0.11.2+, 0.12.0+, 0.12.1+, 0.12.2, 0.13.0

Phoenix - 4.5.0-Hbase-1.0+, 4.5.0-Hbase-1.1+, 4.5.1-Hbase-1.0+,
4.5.1-Hbase-1.1+, 4.5.2-HBase-1.0+, 4.5.2-HBase-1.1+,
4.6.0-Hbase-1.0+, 4.6.0-Hbase-1.1, 4.7.0-Hbase-1.0+,
Expand Down Expand Up @@ -288,9 +286,6 @@ to be a good starting point to use in running jobs.
Pig 0.13.X, 0.14.X w/ Hadoop 2.6.X
Pig 0.15.X -> 0.17.X w/ Hadoop 2.7.X

Mahout 0.11.X w/ Hadoop 2.7.X
Mahout 0.12.X w/ Hadoop 2.7.X

Hbase 0.98.X w/ Hadoop 2.2.X, Zookeeper 3.4.X
Hbase 0.99.X -> 1.6.X w/ Hadoop 2.7.X, Zookeeper 3.4.X

Expand Down Expand Up @@ -376,7 +371,7 @@ considered on the side of experimental.

- Experimental

Packages: Kafka, Zeppelin, Mahout, Hive, TensorFlow w/ & w/o
Packages: Kafka, Zeppelin, Hive, TensorFlow w/ & w/o
Horovod, Ray

Documentation
Expand All @@ -388,7 +383,6 @@ files.

Hadoop - See README.hadoop
Pig - See README.pig
Mahout - See README.mahout
Hbase - See README.hbase
Hive - See README.hive
Spark - See README.spark
Expand Down
126 changes: 0 additions & 126 deletions doc/README.mahout

This file was deleted.

48 changes: 5 additions & 43 deletions magpie-check-inputs
Original file line number Diff line number Diff line change
Expand Up @@ -189,7 +189,7 @@ __Magpie_check_deprecated_configs () {
# Flag deprecated settings for user
#

oldmodes="HADOOP_MODE PIG_MODE MAHOUT_MODE HBASE_MODE PHOENIX_MODE SPARK_MODE KAFKA_MODE ZEPPELIN_MODE STORM_MODE ZOOKEEPER_MODE"
oldmodes="HADOOP_MODE PIG_MODE HBASE_MODE PHOENIX_MODE SPARK_MODE KAFKA_MODE ZEPPELIN_MODE STORM_MODE ZOOKEEPER_MODE"
oldprojects="HADOOP_UDA_SETUP TACHYON_SETUP"
oldfeatures="HDFS_FEDERATION_NAMENODE_COUNT HADOOP_PER_JOB_HDFS_PATH ZOOKEEPER_PER_JOB_DATA_DIR HADOOP_RAWNETWORKFS_BLOCKSIZE"
oldvars="SPARK_USE_YARN MAGPIE_SCRIPT_PATH MAGPIE_SCRIPT_ARGS"
Expand Down Expand Up @@ -257,7 +257,6 @@ if [ "${MAGPIE_JOB_TYPE}" != "hadoop" ] \
&& [ "${MAGPIE_JOB_TYPE}" != "hbase" ] \
&& [ "${MAGPIE_JOB_TYPE}" != "phoenix" ] \
&& [ "${MAGPIE_JOB_TYPE}" != "pig" ] \
&& [ "${MAGPIE_JOB_TYPE}" != "mahout" ] \
&& [ "${MAGPIE_JOB_TYPE}" != "spark" ] \
&& [ "${MAGPIE_JOB_TYPE}" != "kafka" ] \
&& [ "${MAGPIE_JOB_TYPE}" != "zeppelin" ] \
Expand All @@ -273,7 +272,7 @@ if [ "${MAGPIE_JOB_TYPE}" != "hadoop" ] \
&& [ "${MAGPIE_JOB_TYPE}" != "interactive" ] \
&& [ "${MAGPIE_JOB_TYPE}" != "setuponly" ]
then
echo "MAGPIE_JOB_TYPE must be set to hadoop, hbase, pig, mahout, phoenix, spark, \
echo "MAGPIE_JOB_TYPE must be set to hadoop, hbase, pig, phoenix, spark, \
kafka, zeppelin, storm, hive, zookeeper, tensorflow, tensorflow-horovod, ray, \
alluxio, testall, script, interactive, or setuponly"
exit 1
Expand Down Expand Up @@ -368,7 +367,7 @@ nodecount=${MAGPIE_NODE_COUNT}
# nodecountmaster is a counter to count the master only once
nodecountmaster=1

magpieprojects="HADOOP PIG MAHOUT HBASE PHOENIX SPARK KAFKA ZEPPELIN STORM HIVE ZOOKEEPER TENSORFLOW TENSORFLOW_HOROVOD RAY ALLUXIO"
magpieprojects="HADOOP PIG HBASE PHOENIX SPARK KAFKA ZEPPELIN STORM HIVE ZOOKEEPER TENSORFLOW TENSORFLOW_HOROVOD RAY ALLUXIO"

for project in ${magpieprojects}
do
Expand All @@ -379,7 +378,6 @@ done
# Did user turn on SOMETHING to run
#
# Pig is not "something", b/c it runs on top of hadoop
# Mahout is not "something", b/c it runs on top of hadoop

if [ "${HADOOP_SETUP}" != "yes" ] \
&& [ "${HBASE_SETUP}" != "yes" ] \
Expand All @@ -402,7 +400,7 @@ fi

# If java required, was it set to something reasonable

magpieprojects_java="HADOOP PIG MAHOUT HBASE PHOENIX SPARK KAFKA ZEPPELIN STORM HIVE ZOOKEEPER ALLUXIO"
magpieprojects_java="HADOOP PIG HBASE PHOENIX SPARK KAFKA ZEPPELIN STORM HIVE ZOOKEEPER ALLUXIO"
for project in ${magpieprojects_java}
do
setupvar="${project}_SETUP"
Expand All @@ -415,7 +413,7 @@ done

# Did user turn on something matching job run type

magpieprojects="HADOOP PIG MAHOUT HBASE PHOENIX SPARK KAFKA ZEPPELIN STORM HIVE ZOOKEEPER TENSORFLOW TENSORFLOW_HOROVOD RAY ALLUXIO"
magpieprojects="HADOOP PIG HBASE PHOENIX SPARK KAFKA ZEPPELIN STORM HIVE ZOOKEEPER TENSORFLOW TENSORFLOW_HOROVOD RAY ALLUXIO"

for project in ${magpieprojects}
do
Expand All @@ -433,7 +431,6 @@ done
if [ "${MAGPIE_JOB_TYPE}" == "testall" ] \
&& [ "${HADOOP_SETUP}" != "yes" ] \
&& [ "${PIG_SETUP}" != "yes" ] \
&& [ "${MAHOUT_SETUP}" != "yes" ] \
&& [ "${HBASE_SETUP}" != "yes" ] \
&& [ "${SPARK_SETUP}" != "yes" ] \
&& [ "${KAFKA_SETUP}" != "yes" ] \
Expand Down Expand Up @@ -681,41 +678,6 @@ then
fi
fi

#
# Check Mahout Inputs
#

if [ "${MAHOUT_SETUP}" == "yes" ]
then
__Magpie_check_must_be_set "JAVA_HOME" "for Mahout"

__Magpie_check_must_be_set "MAHOUT_VERSION" "to run Mahout"

__Magpie_check_if_version_format_correct "MAHOUT_VERSION"

__Magpie_check_must_be_set_and_is_directory "MAHOUT_HOME" "to run Mahout"

__Magpie_check_must_be_set "MAHOUT_LOCAL_DIR" "to run Mahout"

__Magpie_check_is_enabled "Hadoop" "Mahout"

if [ "${MAHOUT_JOB}" != "clustersyntheticcontrol" ]
then
echo "MAHOUT_JOB must be set to clustersyntheticcontrol"
exit 1
fi

if (! Magpie_hadoop_setup_type_enables_yarn \
|| ! Magpie_hadoop_setup_type_enables_hdfs \
|| ! Magpie_hadoop_filesystem_mode_is_hdfs_type) \
&& [ "${MAGPIE_JOB_TYPE}" == "mahout" ] \
&& [ "${MAHOUT_JOB}" == "clustersyntheticcontrol" ]
then
echo "HADOOP_SETUP_TYPE must be set to MR for MAHOUT_JOB=${MAHOUT_JOB}"
exit 1
fi
fi

#
# Check Hbase Inputs
#
Expand Down
6 changes: 0 additions & 6 deletions magpie-run
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,6 @@ source ${MAGPIE_SCRIPTS_HOME}/magpie/run/magpie-run-project-alluxio
source ${MAGPIE_SCRIPTS_HOME}/magpie/run/magpie-run-project-hadoop
source ${MAGPIE_SCRIPTS_HOME}/magpie/run/magpie-run-project-hbase
source ${MAGPIE_SCRIPTS_HOME}/magpie/run/magpie-run-project-kafka
source ${MAGPIE_SCRIPTS_HOME}/magpie/run/magpie-run-project-mahout
source ${MAGPIE_SCRIPTS_HOME}/magpie/run/magpie-run-project-pig
source ${MAGPIE_SCRIPTS_HOME}/magpie/run/magpie-run-project-phoenix
source ${MAGPIE_SCRIPTS_HOME}/magpie/run/magpie-run-project-hive
Expand Down Expand Up @@ -123,8 +122,6 @@ Magpie_run_start_alluxio

# After Hadoop setup, requires Hadoop
Magpie_run_start_pig
# After Hadoop setup, requires Hadoop
Magpie_run_start_mahout

# After Zookeeper setup, requires Zookeeper
# Will set magpie_run_hbase_should_be_torndown & magpie_run_hbase_setup_successful appropriately
Expand Down Expand Up @@ -200,9 +197,6 @@ then
elif [ "${MAGPIE_JOB_TYPE}" == "pig" ]
then
Magpie_run_pig
elif [ "${MAGPIE_JOB_TYPE}" == "mahout" ]
then
Magpie_run_mahout
elif [ "${MAGPIE_JOB_TYPE}" == "hadoop" ]
then
Magpie_run_hadoop
Expand Down
2 changes: 1 addition & 1 deletion magpie-setup-core
Original file line number Diff line number Diff line change
Expand Up @@ -66,7 +66,7 @@ fi
# Setup primary conf, log, and local scratchspace directories for projects
#

magpieprojects="HADOOP PIG MAHOUT HBASE HIVE PHOENIX SPARK KAFKA ZEPPELIN STORM ZOOKEEPER RAY ALLUXIO"
magpieprojects="HADOOP PIG HBASE HIVE PHOENIX SPARK KAFKA ZEPPELIN STORM ZOOKEEPER RAY ALLUXIO"

for project in ${magpieprojects}
do
Expand Down
8 changes: 0 additions & 8 deletions magpie-setup-projects
Original file line number Diff line number Diff line change
Expand Up @@ -42,14 +42,6 @@ then
fi
fi

if [ "${MAHOUT_SETUP}" == "yes" ]
then
(${MAGPIE_SCRIPTS_HOME}/magpie/setup/magpie-setup-project-mahout)
if [ $? -ne 0 ] ; then
exit 1
fi
fi

if [ "${HBASE_SETUP}" == "yes" ]
then
(${MAGPIE_SCRIPTS_HOME}/magpie/setup/magpie-setup-project-hbase)
Expand Down
Loading

0 comments on commit 5311c92

Please sign in to comment.