spark中的遍历dataframe

  def main(args: Array[String]) = {
    val spark = SparkSession.builder().appName("p2") //.master("local")
      .enableHiveSupport().getOrCreate()
    import spark.implicits._
    val nowdate = LocalDate.now()
    val statis_day = nowdate.minusDays(1).toString.split("-").mkString("")//当天时间
    val pre01_day = date_before(statis_day, "1")
    //前一天日期
    //val pre01_day=args(0)
    val pre31_day = date_before(pre01_day, "7")
    //前三十一天日期

    var df= spark.sql(
      s"""
         |select idea_third_categ,
         |case when percentile_approx(qvalue, 0.2)>0 then percentile_approx(qvalue, 0.2)
         | else "1.0" end as x1,
         |case when percentile_approx(qvalue, 0.4)>0 then percentile_approx(qvalue, 0.4)
         | else "1.0" end as  x2,
         |case when percentile_approx(qvalue, 0.6)>0 then percentile_approx(qvalue, 0.6)
         | else "1.0" end as x3,
         |case when percentile_approx(qvalue, 0.8)>0 then percentile_approx(qvalue, 0.8)
         | else "1.0" end as x4,
         | "1.0" as x5
         |from aps.cpc_pv_log
         |where statis_date<=${pre01_day} and
         |statis_date>${pre31_day}
         |group by idea_third_categ
      """.stripMargin).rdd
    //df.createOrReplaceTempView("df")

    val resultDF = spark.createDataFrame(df.map(item=>{
      val s1= item.getAs[String]("idea_third_categ")
      val ab = new ArrayBuffer[String]()
      var List = new Array[String](5)
      List(0) = item.getAs[String]("x1")
      List(1) = item.getAs[String]("x2")
      List(2) = item.getAs[String]("x3")
      List(3) = item.getAs[String]("x4")
      List(4) = item.getAs[String]("x5")
      for (i <- 0 to 3) {
        breakable {
          if (List(i) == "1.0" | List(i + 1) == List(i)) {
            break()
          }
          //buffer1+=data
          ab += List(i)
        }
      }
      var s2=ab.mkString("#")
      (s1,s2)
    })).toDF("Card3_Id", "quantile")

    resultDF.show(20)
    resultDF.createOrReplaceTempView("dframe")
    val pv_cart3 = spark.sql(
      s"""insert overwrite table predict.pctr_quantile partition (statis_date='${statis_day}')
         |select Card3_Id,quantile
         |from dframe
            """.stripMargin).repartition(10)
    println("dataframe导入成功")
  }

   var df= spark.sql(
      s"""
         |select idea_third_categ,
         |case when percentile_approx(qvalue, 0.2)>0 then percentile_approx(qvalue, 0.2)
         | else "1.0" end as x1,
         |case when percentile_approx(qvalue, 0.4)>0 then percentile_approx(qvalue, 0.4)
         | else "1.0" end as  x2,
         |case when percentile_approx(qvalue, 0.6)>0 then percentile_approx(qvalue, 0.6)
         | else "1.0" end as x3,
         |case when percentile_approx(qvalue, 0.8)>0 then percentile_approx(qvalue, 0.8)
         | else "1.0" end as x4,
         | "1.0" as x5
         |from aps.cpc_pv_log
         |where statis_date<=${pre01_day} and
         |statis_date>${pre31_day}
         |group by idea_third_categ
      """.stripMargin).rdd.map(x=>{
      val s1= x(0).toString
      val ab = new ArrayBuffer[String]()
      var List = new Array[String](5)
      List(0) = x(1).toString
      List(1) = x(2).toString
      List(2) = x(3).toString
      List(3) = x(4).toString
      List(4) = x(5).toString
      for (i <- 0 to 3) {
        breakable {
          if (List(i) == "1.0" | List(i + 1) == List(i)) {
            break()
          }
          //buffer1+=data
          ab += List(i)
        }
      }
      var s2=ab.mkString("#")
      (s1,s2)
    }).toDF("Card3_Id", "quantile")

  • 0
    点赞
  • 5
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值