import org.apache.spark.{SparkConf, SparkContext}
/**
* Created by songcl on 2016/7/15.
*/
object day_uv {
def main(args: Array[String]): Unit = {
/**
* Created by songcl on 2016/7/5.
*/
val conf = new SparkConf().setAppName("Simple Application").setMaster("spark://10.0.58.21:7077")
val sc = new SparkContext(conf)
val url2 = "jdbc:mysql://rds3dabp9v2v7v596tai.mysql.rds.aliyuncs.com/r2d2?user=r2d2_admin&password=Vj0kHdve3"
val format = new java.text.SimpleDateFormat("yyyyMMdd")
val dat01 = format.format(new java.util.Date().getTime() - 1 * 24 * 60 * 60 * 1000)
val dat02 = format.format(new java.util.Date().getTime() - 0 * 24 * 60 * 60 * 1000)
val dat03 = format.format(new java.util.Date().getTime() - 2 * 24 * 60 * 60 * 1000)
val format2 = new java.text.SimpleDateFormat("yyyy-MM-dd")
val dat = format2.format(new java.util.Date().getTime() - 1 * 24 * 60 * 60 * 1000)
val log01 = sc.textFile("hdfs://10.0.58.21:9000/data/logstash."+dat01+".log")
val log02 = sc.textFile("hdfs://10.0.58.21:9000/data/logstash."+dat02+".log")
val log03 = sc.textFile("hdfs://10.0.58.21:9000/data/logstash."+dat03+".log")
val log2=log01 union log02 union log03
val log=log2.filter(line=>line.contains(dat))
val rowRDD = log2.map(line => (line.split("\"message\":\"").last.split(" ").head.trim(), line.split("account: ").last.split(", args:").head)).filter({ case (k, v) => k==dat &&v.length()==8 && !k.contains("TypeError:") }).distinct()
val sqlContext = new org.apache.spark.sql.SQLContext(sc)
import sqlContext.implicits._
val df = rowRDD.toDF("date", "No")
//val dfed = df.filter(df("created") < dat1 && df("created") > dat2)
// val dfed = df.filter(df("date") >= dat)
val ed = df.registerTempTable("kv")
val q = sqlContext.sql("select kv.date,count(distinct kv.No) from kv group by kv.date")
import sqlContext.implicits._
val df2 = q.toDF("stat_time", "uv")
// k.insertIntoJDBC(url2, "utest", false)
// df2.insertIntoJDBC(url2, "userlog", false)
import sqlContext.implicits._
/// val df2 = q.toDF("stat_time", "uv")
df2.insertIntoJDBC(url2, "day_uv", false)
}
}