Mahout命令使用参数介绍

版本:0.5

输入命令: mahout

预期结果:

Running on hadoop, using HADOOP_HOME=/home/username/local/hadoop

HADOOP_CONF_DIR=/home/username/local/hadoop/conf

An example program must be given as the first argument.

Valid program names are:

  arff.vector: : Generate Vectors from an ARFF file or directory

  canopy: : Canopy clustering

  cat: : Print a file or resource as the logistic regression models would see it

  cleansvd: : Cleanup and verification of SVD output

  clusterdump: : Dump cluster output to text

  dirichlet: : Dirichlet Clustering

  eigencuts: : Eigencuts spectral clustering

  evaluateFactorization: : compute RMSE of a rating matrix factorization against probes in memory

  evaluateFactorizationParallel: : compute RMSE of a rating matrix factorization against probes

  fkmeans: : Fuzzy K-means clustering

  fpg: : Frequent Pattern Growth

  itemsimilarity: : Compute the item-item-similarities for item-based collaborative filtering

  kmeans: : K-means clustering

  lda: : Latent Dirchlet Allocation

  ldatopics: : LDA Print Topics

  lucene.vector: : Generate Vectors from a Lucene index

  matrixmult: : Take the product of two matrices

  meanshift: : Mean Shift clustering

  parallelALS: : ALS-WR factorization of a rating matrix

  predictFromFactorization: : predict preferences from a factorization of a rating matrix

  prepare20newsgroups: : Reformat 20 newsgroups data

  recommenditembased: : Compute recommendations using item-based collaborative filtering

  rowid: : Map SequenceFile<Text,VectorWritable> to {SequenceFile<IntWritable,VectorWritable>, SequenceFile<IntWritable,Text>}

  rowsimilarity: : Compute the pairwise similarities of the rows of a matrix

  runlogistic: : Run a logistic regression model against CSV data

  seq2sparse: : Sparse Vector generation from Text sequence files (从文本文件生成稀疏向量序列

  seqdirectory: : Generate sequence files (of Text) from a directory(从一个目录生成序列文件)

  seqdumper: : Generic Sequence File dumper (将SequenceFile文件转换成可读的文本形式)

  seqwiki: : Wikipedia xml dump to sequence file

  spectralkmeans: : Spectral k-means clustering

  splitDataset: : split a rating dataset into training and probe parts

  ssvd: : Stochastic SVD

  svd: : Lanczos Singular Value Decomposition

  testclassifier: : Test Bayes Classifier

  trainclassifier: : Train Bayes Classifier

  trainlogistic: : Train a logistic regression using stochastic gradient descent

  transpose: : Take the transpose of a matrix

  vectordump: : Dump vectors from a sequence file to text

  wikipediaDataSetCreator: : Splits data set of wikipedia wrt feature like country

  wikipediaXMLSplitter: : Reads wikipedia data and creates ch


  • 1
    点赞
  • 1
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值